diff --git a/epochX/cudacpp/CODEGEN/PLUGIN/CUDACPP_SA_OUTPUT/CHANGELOG.md b/epochX/cudacpp/CODEGEN/PLUGIN/CUDACPP_SA_OUTPUT/CHANGELOG.md index 652e68d055..1109bb0fd9 100644 --- a/epochX/cudacpp/CODEGEN/PLUGIN/CUDACPP_SA_OUTPUT/CHANGELOG.md +++ b/epochX/cudacpp/CODEGEN/PLUGIN/CUDACPP_SA_OUTPUT/CHANGELOG.md @@ -6,12 +6,23 @@ The format is loosely based on [Keep a Changelog](https://keepachangelog.com). -------------------------------------------------------------------------------- -## [Unreleased] - 2024-10-06 +## [Unreleased] - 2024-10-08 + +### Added + +- Infrastructure issues + - AV ([#945]) Add cudacpp_bldall runcard to produce multi-backend gridpacks. + - AV ([#700]) Add cudacpp_helinl and cudacpp_hrdcod runcards to support HELINL=1 and HRDCOD=1 builds. + - AV ([#957]) In internal "tlau" tests, instrument python code in gridpacks to provide timing profiles for event generation. + - AV Enhance internal "tlau" tests and add a few test logs and various scripts to analyse them. ### Changed - Updated cudacpp version to 1.00.01. +- Infrastructure issues + - AV ([#700]) Renamed the floating_type runcard as cudacpp_fptype for consistency with other cudacpp runcards. + ### Fixed - Platform-specific issues @@ -56,9 +67,12 @@ The format is loosely based on [Keep a Changelog](https://keepachangelog.com). [Unreleased]: https://github.com/madgraph5/madgraph4gpu/releases/compare/cudacpp_for3.6.0_v1.00.00...HEAD [#601]: https://github.com/madgraph5/madgraph4gpu/issues/601 +[#700]: https://github.com/madgraph5/madgraph4gpu/issues/700 [#846]: https://github.com/madgraph5/madgraph4gpu/issues/846 [#942]: https://github.com/madgraph5/madgraph4gpu/issues/942 [#944]: https://github.com/madgraph5/madgraph4gpu/issues/944 +[#945]: https://github.com/madgraph5/madgraph4gpu/issues/945 +[#957]: https://github.com/madgraph5/madgraph4gpu/issues/957 [#959]: https://github.com/madgraph5/madgraph4gpu/issues/959 [#993]: https://github.com/madgraph5/madgraph4gpu/issues/993 [#995]: https://github.com/madgraph5/madgraph4gpu/issues/995 diff --git a/epochX/cudacpp/CODEGEN/PLUGIN/CUDACPP_SA_OUTPUT/MG5aMC_patches/PROD/patch.common b/epochX/cudacpp/CODEGEN/PLUGIN/CUDACPP_SA_OUTPUT/MG5aMC_patches/PROD/patch.common index ce1c49dc2a..bfdf1c751f 100644 --- a/epochX/cudacpp/CODEGEN/PLUGIN/CUDACPP_SA_OUTPUT/MG5aMC_patches/PROD/patch.common +++ b/epochX/cudacpp/CODEGEN/PLUGIN/CUDACPP_SA_OUTPUT/MG5aMC_patches/PROD/patch.common @@ -1,12 +1,17 @@ diff --git b/epochX/cudacpp/gg_tt.mad/SubProcesses/makefile a/epochX/cudacpp/gg_tt.mad/SubProcesses/makefile -index 348c283be..49e6800ff 100644 +index 348c283be..2e5bff046 100644 --- b/epochX/cudacpp/gg_tt.mad/SubProcesses/makefile +++ a/epochX/cudacpp/gg_tt.mad/SubProcesses/makefile -@@ -1,6 +1,37 @@ +@@ -1,6 +1,42 @@ +SHELL := /bin/bash + include ../../Source/make_opts + ++# Export the relevant settings from make_opts to builds using the cudacpp makefile (#700) ++export FPTYPE ++export HELINL ++export HRDCOD ++ +# Determine CUDACPP_BUILDDIR based on the user-defined choices of BACKEND, FPTYPE, HELINL, HRDCOD and USEBUILDDIR (#829) +# Stop with an error if BACKEND=cuda and nvcc is missing or if BACKEND=hip and hipcc is missing +include ../../src/cudacpp_config.mk @@ -40,7 +45,7 @@ index 348c283be..49e6800ff 100644 # Load additional dependencies of the bias module, if present ifeq (,$(wildcard ../bias_dependencies)) BIASDEPENDENCIES = -@@ -24,7 +55,20 @@ else +@@ -24,7 +60,20 @@ else MADLOOP_LIB = endif @@ -62,7 +67,7 @@ index 348c283be..49e6800ff 100644 LIBS = $(LIBDIR)libbias.$(libext) $(LIBDIR)libdhelas.$(libext) $(LIBDIR)libdsample.$(libext) $(LIBDIR)libgeneric.$(libext) $(LIBDIR)libpdf.$(libext) $(LIBDIR)libgammaUPC.$(libext) $(LIBDIR)libmodel.$(libext) $(LIBDIR)libcernlib.$(libext) $(MADLOOP_LIB) $(LOOP_LIBS) -@@ -43,41 +87,148 @@ ifeq ($(strip $(MATRIX_HEL)),) +@@ -43,41 +92,150 @@ ifeq ($(strip $(MATRIX_HEL)),) endif @@ -153,32 +158,34 @@ index 348c283be..49e6800ff 100644 + rm -f $(PROG) + ln -s $(PROG)_fortran $(PROG) + -+madevent_cuda_link: -+ $(MAKE) USEGTEST=0 BACKEND=cuda $(CUDACPP_BUILDDIR)/$(PROG)_cuda ++madevent_cudacpp_link: all + rm -f $(PROG) ++ifeq ($(BACKEND),cuda) + ln -s $(CUDACPP_BUILDDIR)/$(PROG)_cuda $(PROG) -+ -+madevent_hip_link: -+ $(MAKE) USEGTEST=0 BACKEND=hip $(CUDACPP_BUILDDIR)/$(PROG)_hip -+ rm -f $(PROG) ++else ifeq ($(BACKEND),hip) + ln -s $(CUDACPP_BUILDDIR)/$(PROG)_hip $(PROG) -+ -+madevent_cpp_link: -+ $(MAKE) USEGTEST=0 BACKEND=cppauto $(CUDACPP_BUILDDIR)/$(PROG)_cpp -+ rm -f $(PROG) ++else + ln -s $(CUDACPP_BUILDDIR)/$(PROG)_cpp $(PROG) ++endif ++ ++madevent_cuda_link: ++ $(MAKE) USEGTEST=0 USEBUILDDIR=1 BACKEND=cuda madevent_cudacpp_link -$(LIBDIR)libpdf.$(libext): - cd ../../Source/PDF; make -+override SUPPORTED_AVXS = cppnone cppsse4 cppavx2 cpp512y cpp512z cppauto -+madevent_%_link: -+ @if [ '$(words $(filter $*, $(SUPPORTED_AVXS)))' != '1' ]; then echo "ERROR! Invalid target '$@' (supported madevent_cpp*_link targets are: $(foreach avx,$(SUPPORTED_AVXS),'madevent_cpp$(avx)_link'))"; exit 1; fi -+ $(MAKE) USEGTEST=0 BACKEND=$* $(CUDACPP_BUILDDIR)/$(PROG)_cpp -+ rm -f $(PROG) -+ ln -s $(CUDACPP_BUILDDIR)/$(PROG)_cpp $(PROG) ++madevent_hip_link: ++ $(MAKE) USEGTEST=0 USEBUILDDIR=1 BACKEND=hip madevent_cudacpp_link -$(LIBDIR)libgammaUPC.$(libext): - cd ../../Source/PDF/gammaUPC; make ++madevent_cpp_link: ++ $(MAKE) USEGTEST=0 USEBUILDDIR=1 BACKEND=cppauto madevent_cudacpp_link ++ ++override SUPPORTED_AVXS = cppnone cppsse4 cppavx2 cpp512y cpp512z cppauto ++madevent_%_link: ++ @if [ '$(words $(filter $*, $(SUPPORTED_AVXS)))' != '1' ]; then echo "ERROR! Invalid target '$@' (supported madevent_cpp*_link targets are: $(foreach avx,$(SUPPORTED_AVXS),'madevent_$(avx)_link'))"; exit 1; fi ++ $(MAKE) USEGTEST=0 USEBUILDDIR=1 BACKEND=$* madevent_cudacpp_link ++ +# Building $(PROG)_cpp no longer builds $(PROG)_cuda if CUDACPP_BACKENDLIB for cuda exists (this was the case in the past to allow cpp-only builds #503) +$(CUDACPP_BUILDDIR)/$(PROG)_cpp: $(PROCESS) $(DSIG_cudacpp) auto_dsig.o $(LIBS) $(MATRIX) counters.o ompnumthreads.o $(CUDACPP_BUILDDIR)/.cudacpplibs + $(FC) -o $@ $(PROCESS) $(DSIG_cudacpp) auto_dsig.o $(MATRIX) $(LINKLIBS) $(BIASDEPENDENCIES) $(OMPFLAGS) counters.o ompnumthreads.o -L$(LIBDIR)/$(CUDACPP_BUILDDIR) -l$(CUDACPP_COMMONLIB) -l$(CUDACPP_BACKENDLIB) $(LIBFLAGSRPATH) $(LDFLAGS) @@ -227,7 +234,7 @@ index 348c283be..49e6800ff 100644 # Dependencies -@@ -97,5 +248,80 @@ unwgt.o: genps.inc nexternal.inc symswap.inc cluster.inc run.inc message.inc \ +@@ -97,5 +255,80 @@ unwgt.o: genps.inc nexternal.inc symswap.inc cluster.inc run.inc message.inc \ run_config.inc initcluster.o: message.inc diff --git a/epochX/cudacpp/CODEGEN/PLUGIN/CUDACPP_SA_OUTPUT/launch_plugin.py b/epochX/cudacpp/CODEGEN/PLUGIN/CUDACPP_SA_OUTPUT/launch_plugin.py index 0924927785..19d65fa195 100644 --- a/epochX/cudacpp/CODEGEN/PLUGIN/CUDACPP_SA_OUTPUT/launch_plugin.py +++ b/epochX/cudacpp/CODEGEN/PLUGIN/CUDACPP_SA_OUTPUT/launch_plugin.py @@ -33,10 +33,16 @@ def compile(self, *args, **opts): if 'cwd' in opts and os.path.basename(opts['cwd']) == 'Source': path = pjoin(opts['cwd'], 'make_opts') common_run_interface.CommonRunCmd.update_make_opts_full(path, - {'FPTYPE': self.run_card['floating_type'] }) - misc.sprint('FPTYPE checked') + {'FPTYPE': self.run_card['cudacpp_fptype'], + 'HELINL': self.run_card['cudacpp_helinl'], + 'HRDCOD': self.run_card['cudacpp_hrdcod'] }) + misc.sprint('FPTYPE, HELINL, HRDCOD checked') cudacpp_supported_backends = [ 'fortran', 'cuda', 'hip', 'cpp', 'cppnone', 'cppsse4', 'cppavx2', 'cpp512y', 'cpp512z', 'cppauto' ] if args and args[0][0] == 'madevent' and hasattr(self, 'run_card'): + if self.run_card['cudacpp_bldall'] == True: # pre-build all backends #945 + logger.info("Pre-building madevent in madevent_interface.py with ALL matrix elements") + args[0][0] = 'bldall' + misc.compile(nb_core=self.options['nb_core'], *args, **opts) cudacpp_backend = self.run_card['cudacpp_backend'].lower() # the default value is defined in launch_plugin.py logger.info("Building madevent in madevent_interface.py with '%s' matrix elements"%cudacpp_backend) if cudacpp_backend in cudacpp_supported_backends : @@ -47,14 +53,18 @@ def compile(self, *args, **opts): else: return misc.compile(nb_core=self.options['nb_core'], *args, **opts) -# Phase-Space Optimization ------------------------------------------------------------------------------------ +# CUDACPP runcard block ------------------------------------------------------------------------------------ template_on = \ """#*********************************************************************** # SIMD/GPU configuration for the CUDACPP plugin #************************************************************************ - %(cudacpp_backend)s = cudacpp_backend ! CUDACPP backend: fortran, cuda, hip, cpp, cppnone, cppsse4, cppavx2, cpp512y, cpp512z, cppauto + %(cudacpp_backend)s = cudacpp_backend ! CUDACPP backend: fortran, cuda, hip, cpp (DEFAULT), cppnone, cppsse4, cppavx2, cpp512y, cpp512z, cppauto +#*** WARNING! The following cudacpp runcards are experimental! Users should normally change only the cudacpp_backend card *** + %(cudacpp_fptype)s = cudacpp_fptype ! CUDACPP floating point precision: f (single), d (double), m (mixed, DEFAULT: double for amplitudes, single for colors) + %(cudacpp_hrdcod)s = cudacpp_hrdcod ! CUDACPP parameter hardcoding: 0 (DEFAULT, parameters not hardcoded: read param_card.dat at runtime), 1 (hardcoded parameters) + %(cudacpp_helinl)s = cudacpp_helinl ! CUDACPP helicity amplitude inlining: 0 (DEFAULT, ordinary inlining of templates), 1 (aggressive inlining with 'always inline') + %(cudacpp_bldall)s = cudacpp_bldall ! CUDACPP build all available backends in separate build directories: False, True """ - template_off = '' plugin_block = banner_mod.RunBlock('simd', template_on=template_on, template_off=template_off) @@ -75,8 +85,14 @@ def reset_simd(self, old_value, new_value, name): def reset_makeopts(self, old_value, new_value, name): if not hasattr(self, 'path'): raise Exception - if name == 'floating_type': + if name == 'cudacpp_fptype': common_run_interface.CommonRunCmd.update_make_opts_full({'FPTYPE': new_value}) + elif name == 'cudacpp_hrdcod': + raise Exception('Cannot change cudacpp_hrdcod') + elif name == 'cudacpp_helinl': + raise Exception('Cannot change cudacpp_helinl') + elif name == 'cudacpp_bldall': + raise Exception('Cannot change cudacpp_bldall') else: raise Exception Sourcedir = pjoin(os.path.dirname(os.path.dirname(self.path)), 'Source') @@ -84,14 +100,34 @@ def reset_makeopts(self, old_value, new_value, name): def default_setup(self): super().default_setup() - self.add_param('floating_type', 'm', include=False, hidden=True, - fct_mod=(self.reset_makeopts,(),{}), - allowed=['m','d','f'], - comment='floating point precision: f (single), d (double), m (mixed: double for amplitudes, single for colors)' - ) cudacpp_supported_backends = [ 'fortran', 'cuda', 'hip', 'cpp', 'cppnone', 'cppsse4', 'cppavx2', 'cpp512y', 'cpp512z', 'cppauto' ] - self.add_param('cudacpp_backend', 'cpp', include=False, hidden=False, + self.add_param('cudacpp_backend', 'cpp', + include=False, # AV: 'include=True' would add "CUDACPP_BACKEND = 'cpp'" to run_card.inc + hidden=False, # AV: keep cudacpp_backend in runcard template and keep 'hidden='False' allowed=cudacpp_supported_backends) + self.add_param('cudacpp_fptype', 'm', + include=False, # AV: 'include=True' would add "CUDACPP_FPTYPE = 'm'" to run_card.inc (if fct_mod is removed, else codegen fails) + hidden=False, # AV: add cudacpp_backend to runcard template and keep 'hidden='False' + fct_mod=(self.reset_makeopts,(),{}), # AV: I assume this forces a 'make cleanavx' if FPTYPE changes? + allowed=['m','d','f'] + ) + self.add_param('cudacpp_helinl', '0', + include=False, # AV: no need to add this parameter to run_card.inc + hidden=False, # AV: add cudacpp_helinl to runcard template and keep 'hidden='False' + fct_mod=(self.reset_makeopts,(),{}), # AV: I assume this raises an exception if cudacpp_helinl changes? + allowed=['0','1'] + ) + self.add_param('cudacpp_hrdcod', '0', + include=False, # AV: no need to add this parameter to run_card.inc + hidden=False, # AV: add cudacpp_hrdcod to runcard template and keep 'hidden='False' + fct_mod=(self.reset_makeopts,(),{}), # AV: I assume this raises an exception if cudacpp_hrdcod changes? + allowed=['0','1'] + ) + self.add_param('cudacpp_bldall', False, + include=False, # AV: no need to add this parameter to run_card.inc + hidden=False, # AV: add cudacpp_bldall to runcard template and keep 'hidden='False' + fct_mod=(self.reset_makeopts,(),{}), # AV: I assume this raises an exception if cudacpp_bldall changes? + ) self['vector_size'] = 16 # already setup in default class (just change value) self['aloha_flag'] = '--fast-math' self['matrix_flag'] = '-O3' diff --git a/epochX/cudacpp/CODEGEN/generateAndCompare.sh b/epochX/cudacpp/CODEGEN/generateAndCompare.sh index 6221b1cfee..372ac0676d 100755 --- a/epochX/cudacpp/CODEGEN/generateAndCompare.sh +++ b/epochX/cudacpp/CODEGEN/generateAndCompare.sh @@ -199,6 +199,38 @@ function codeGenAndDiff() susy_gg_ulul) cmd="import model MSSM_SLHA2; generate g g > ul ul~" ;; + pp_dy012j) # from Jin #942 + cmd="import model sm-no_b_mass + define p = u d c s b u~ d~ c~ s~ b~ g + define j = p + define ell+ = e+ mu+ ta+ + define ell- = e- mu- ta- + define nu = ve vm vt + define nubar = ve~ vm~ vt~ + generate p p > ell+ ell- @0 + add process p p > ell+ ell- j @1 + add process p p > ell+ ell- j j @2" + ;; + pp_dy3j) # from Jin #942 + cmd="import model sm-no_b_mass + define p = u d c s b u~ d~ c~ s~ b~ g + define j = p + define ell+ = e+ mu+ ta+ + define ell- = e- mu- ta- + define nu = ve vm vt + define nubar = ve~ vm~ vt~ + generate p p > ell+ ell- j j j @0" + ;; + pp_dy4j) # extend dy3j from Jin #942 + cmd="import model sm-no_b_mass + define p = u d c s b u~ d~ c~ s~ b~ g + define j = p + define ell+ = e+ mu+ ta+ + define ell- = e- mu- ta- + define nu = ve vm vt + define nubar = ve~ vm~ vt~ + generate p p > ell+ ell- j j j j @0" + ;; atlas) cmd="import model sm-no_b_mass define p = g u c d s b u~ c~ d~ s~ b~ diff --git a/epochX/cudacpp/bldall.sh b/epochX/cudacpp/bldall.sh new file mode 100755 index 0000000000..02b212babc --- /dev/null +++ b/epochX/cudacpp/bldall.sh @@ -0,0 +1,32 @@ +#!/bin/bash + +if [ "$1" == "" ] || [ "$2" != "" ]; then + echo "Usage: $0 " + exit 1 +fi +spdir=${1}/SubProcesses + +if [ ! -d ${spdir} ]; then + echo "ERROR! Directory not found: ${spdir}" + exit 1 +fi +cd ${spdir} + +START=$(date +%s) + +TMP=${START} +for pdir in P*; do + echo "--------------------------------------------------------------------------------" + cd ${pdir} + pwd + sleep 1 + make -j bldall + cd - > /dev/null + END=$(date +%s) + echo "ELAPSED: $((END-TMP)) seconds" + TMP=${END} +done +echo "================================================================================" + +END=$(date +%s) +echo "ELAPSED: $((END-START)) seconds" diff --git a/epochX/cudacpp/ee_mumu.mad/Cards/run_card.dat b/epochX/cudacpp/ee_mumu.mad/Cards/run_card.dat index 74f70b567b..fe540b1793 100644 --- a/epochX/cudacpp/ee_mumu.mad/Cards/run_card.dat +++ b/epochX/cudacpp/ee_mumu.mad/Cards/run_card.dat @@ -203,5 +203,10 @@ systematics = systematics_program ! none, systematics [python], SysCalc [deprece #*********************************************************************** # SIMD/GPU configuration for the CUDACPP plugin #************************************************************************ - cpp = cudacpp_backend ! CUDACPP backend: fortran, cuda, hip, cpp, cppnone, cppsse4, cppavx2, cpp512y, cpp512z, cppauto + cpp = cudacpp_backend ! CUDACPP backend: fortran, cuda, hip, cpp (DEFAULT), cppnone, cppsse4, cppavx2, cpp512y, cpp512z, cppauto +#*** WARNING! The following cudacpp runcards are experimental! Users should normally change only the cudacpp_backend card *** + m = cudacpp_fptype ! CUDACPP floating point precision: f (single), d (double), m (mixed, DEFAULT: double for amplitudes, single for colors) + 0 = cudacpp_hrdcod ! CUDACPP parameter hardcoding: 0 (DEFAULT, parameters not hardcoded: read param_card.dat at runtime), 1 (hardcoded parameters) + 0 = cudacpp_helinl ! CUDACPP helicity amplitude inlining: 0 (DEFAULT, ordinary inlining of templates), 1 (aggressive inlining with 'always inline') + False = cudacpp_bldall ! CUDACPP build all available backends in separate build directories: False, True diff --git a/epochX/cudacpp/ee_mumu.mad/Cards/run_card_default.dat b/epochX/cudacpp/ee_mumu.mad/Cards/run_card_default.dat index 68ee164d00..9bc9dc5361 100644 --- a/epochX/cudacpp/ee_mumu.mad/Cards/run_card_default.dat +++ b/epochX/cudacpp/ee_mumu.mad/Cards/run_card_default.dat @@ -203,5 +203,10 @@ systematics = systematics_program ! none, systematics [python], SysCalc [deprece #*********************************************************************** # SIMD/GPU configuration for the CUDACPP plugin #************************************************************************ - cpp = cudacpp_backend ! CUDACPP backend: fortran, cuda, hip, cpp, cppnone, cppsse4, cppavx2, cpp512y, cpp512z, cppauto + cpp = cudacpp_backend ! CUDACPP backend: fortran, cuda, hip, cpp (DEFAULT), cppnone, cppsse4, cppavx2, cpp512y, cpp512z, cppauto +#*** WARNING! The following cudacpp runcards are experimental! Users should normally change only the cudacpp_backend card *** + m = cudacpp_fptype ! CUDACPP floating point precision: f (single), d (double), m (mixed, DEFAULT: double for amplitudes, single for colors) + 0 = cudacpp_hrdcod ! CUDACPP parameter hardcoding: 0 (DEFAULT, parameters not hardcoded: read param_card.dat at runtime), 1 (hardcoded parameters) + 0 = cudacpp_helinl ! CUDACPP helicity amplitude inlining: 0 (DEFAULT, ordinary inlining of templates), 1 (aggressive inlining with 'always inline') + False = cudacpp_bldall ! CUDACPP build all available backends in separate build directories: False, True diff --git a/epochX/cudacpp/ee_mumu.mad/SubProcesses/makefile b/epochX/cudacpp/ee_mumu.mad/SubProcesses/makefile index 49e6800fff..2e5bff046b 100644 --- a/epochX/cudacpp/ee_mumu.mad/SubProcesses/makefile +++ b/epochX/cudacpp/ee_mumu.mad/SubProcesses/makefile @@ -2,6 +2,11 @@ SHELL := /bin/bash include ../../Source/make_opts +# Export the relevant settings from make_opts to builds using the cudacpp makefile (#700) +export FPTYPE +export HELINL +export HRDCOD + # Determine CUDACPP_BUILDDIR based on the user-defined choices of BACKEND, FPTYPE, HELINL, HRDCOD and USEBUILDDIR (#829) # Stop with an error if BACKEND=cuda and nvcc is missing or if BACKEND=hip and hipcc is missing include ../../src/cudacpp_config.mk @@ -162,27 +167,29 @@ madevent_fortran_link: $(PROG)_fortran rm -f $(PROG) ln -s $(PROG)_fortran $(PROG) -madevent_cuda_link: - $(MAKE) USEGTEST=0 BACKEND=cuda $(CUDACPP_BUILDDIR)/$(PROG)_cuda +madevent_cudacpp_link: all rm -f $(PROG) +ifeq ($(BACKEND),cuda) ln -s $(CUDACPP_BUILDDIR)/$(PROG)_cuda $(PROG) +else ifeq ($(BACKEND),hip) + ln -s $(CUDACPP_BUILDDIR)/$(PROG)_hip $(PROG) +else + ln -s $(CUDACPP_BUILDDIR)/$(PROG)_cpp $(PROG) +endif + +madevent_cuda_link: + $(MAKE) USEGTEST=0 USEBUILDDIR=1 BACKEND=cuda madevent_cudacpp_link madevent_hip_link: - $(MAKE) USEGTEST=0 BACKEND=hip $(CUDACPP_BUILDDIR)/$(PROG)_hip - rm -f $(PROG) - ln -s $(CUDACPP_BUILDDIR)/$(PROG)_hip $(PROG) + $(MAKE) USEGTEST=0 USEBUILDDIR=1 BACKEND=hip madevent_cudacpp_link madevent_cpp_link: - $(MAKE) USEGTEST=0 BACKEND=cppauto $(CUDACPP_BUILDDIR)/$(PROG)_cpp - rm -f $(PROG) - ln -s $(CUDACPP_BUILDDIR)/$(PROG)_cpp $(PROG) + $(MAKE) USEGTEST=0 USEBUILDDIR=1 BACKEND=cppauto madevent_cudacpp_link override SUPPORTED_AVXS = cppnone cppsse4 cppavx2 cpp512y cpp512z cppauto madevent_%_link: - @if [ '$(words $(filter $*, $(SUPPORTED_AVXS)))' != '1' ]; then echo "ERROR! Invalid target '$@' (supported madevent_cpp*_link targets are: $(foreach avx,$(SUPPORTED_AVXS),'madevent_cpp$(avx)_link'))"; exit 1; fi - $(MAKE) USEGTEST=0 BACKEND=$* $(CUDACPP_BUILDDIR)/$(PROG)_cpp - rm -f $(PROG) - ln -s $(CUDACPP_BUILDDIR)/$(PROG)_cpp $(PROG) + @if [ '$(words $(filter $*, $(SUPPORTED_AVXS)))' != '1' ]; then echo "ERROR! Invalid target '$@' (supported madevent_cpp*_link targets are: $(foreach avx,$(SUPPORTED_AVXS),'madevent_$(avx)_link'))"; exit 1; fi + $(MAKE) USEGTEST=0 USEBUILDDIR=1 BACKEND=$* madevent_cudacpp_link # Building $(PROG)_cpp no longer builds $(PROG)_cuda if CUDACPP_BACKENDLIB for cuda exists (this was the case in the past to allow cpp-only builds #503) $(CUDACPP_BUILDDIR)/$(PROG)_cpp: $(PROCESS) $(DSIG_cudacpp) auto_dsig.o $(LIBS) $(MATRIX) counters.o ompnumthreads.o $(CUDACPP_BUILDDIR)/.cudacpplibs diff --git a/epochX/cudacpp/ee_mumu.mad/bin/internal/launch_plugin.py b/epochX/cudacpp/ee_mumu.mad/bin/internal/launch_plugin.py index 0924927785..19d65fa195 100644 --- a/epochX/cudacpp/ee_mumu.mad/bin/internal/launch_plugin.py +++ b/epochX/cudacpp/ee_mumu.mad/bin/internal/launch_plugin.py @@ -33,10 +33,16 @@ def compile(self, *args, **opts): if 'cwd' in opts and os.path.basename(opts['cwd']) == 'Source': path = pjoin(opts['cwd'], 'make_opts') common_run_interface.CommonRunCmd.update_make_opts_full(path, - {'FPTYPE': self.run_card['floating_type'] }) - misc.sprint('FPTYPE checked') + {'FPTYPE': self.run_card['cudacpp_fptype'], + 'HELINL': self.run_card['cudacpp_helinl'], + 'HRDCOD': self.run_card['cudacpp_hrdcod'] }) + misc.sprint('FPTYPE, HELINL, HRDCOD checked') cudacpp_supported_backends = [ 'fortran', 'cuda', 'hip', 'cpp', 'cppnone', 'cppsse4', 'cppavx2', 'cpp512y', 'cpp512z', 'cppauto' ] if args and args[0][0] == 'madevent' and hasattr(self, 'run_card'): + if self.run_card['cudacpp_bldall'] == True: # pre-build all backends #945 + logger.info("Pre-building madevent in madevent_interface.py with ALL matrix elements") + args[0][0] = 'bldall' + misc.compile(nb_core=self.options['nb_core'], *args, **opts) cudacpp_backend = self.run_card['cudacpp_backend'].lower() # the default value is defined in launch_plugin.py logger.info("Building madevent in madevent_interface.py with '%s' matrix elements"%cudacpp_backend) if cudacpp_backend in cudacpp_supported_backends : @@ -47,14 +53,18 @@ def compile(self, *args, **opts): else: return misc.compile(nb_core=self.options['nb_core'], *args, **opts) -# Phase-Space Optimization ------------------------------------------------------------------------------------ +# CUDACPP runcard block ------------------------------------------------------------------------------------ template_on = \ """#*********************************************************************** # SIMD/GPU configuration for the CUDACPP plugin #************************************************************************ - %(cudacpp_backend)s = cudacpp_backend ! CUDACPP backend: fortran, cuda, hip, cpp, cppnone, cppsse4, cppavx2, cpp512y, cpp512z, cppauto + %(cudacpp_backend)s = cudacpp_backend ! CUDACPP backend: fortran, cuda, hip, cpp (DEFAULT), cppnone, cppsse4, cppavx2, cpp512y, cpp512z, cppauto +#*** WARNING! The following cudacpp runcards are experimental! Users should normally change only the cudacpp_backend card *** + %(cudacpp_fptype)s = cudacpp_fptype ! CUDACPP floating point precision: f (single), d (double), m (mixed, DEFAULT: double for amplitudes, single for colors) + %(cudacpp_hrdcod)s = cudacpp_hrdcod ! CUDACPP parameter hardcoding: 0 (DEFAULT, parameters not hardcoded: read param_card.dat at runtime), 1 (hardcoded parameters) + %(cudacpp_helinl)s = cudacpp_helinl ! CUDACPP helicity amplitude inlining: 0 (DEFAULT, ordinary inlining of templates), 1 (aggressive inlining with 'always inline') + %(cudacpp_bldall)s = cudacpp_bldall ! CUDACPP build all available backends in separate build directories: False, True """ - template_off = '' plugin_block = banner_mod.RunBlock('simd', template_on=template_on, template_off=template_off) @@ -75,8 +85,14 @@ def reset_simd(self, old_value, new_value, name): def reset_makeopts(self, old_value, new_value, name): if not hasattr(self, 'path'): raise Exception - if name == 'floating_type': + if name == 'cudacpp_fptype': common_run_interface.CommonRunCmd.update_make_opts_full({'FPTYPE': new_value}) + elif name == 'cudacpp_hrdcod': + raise Exception('Cannot change cudacpp_hrdcod') + elif name == 'cudacpp_helinl': + raise Exception('Cannot change cudacpp_helinl') + elif name == 'cudacpp_bldall': + raise Exception('Cannot change cudacpp_bldall') else: raise Exception Sourcedir = pjoin(os.path.dirname(os.path.dirname(self.path)), 'Source') @@ -84,14 +100,34 @@ def reset_makeopts(self, old_value, new_value, name): def default_setup(self): super().default_setup() - self.add_param('floating_type', 'm', include=False, hidden=True, - fct_mod=(self.reset_makeopts,(),{}), - allowed=['m','d','f'], - comment='floating point precision: f (single), d (double), m (mixed: double for amplitudes, single for colors)' - ) cudacpp_supported_backends = [ 'fortran', 'cuda', 'hip', 'cpp', 'cppnone', 'cppsse4', 'cppavx2', 'cpp512y', 'cpp512z', 'cppauto' ] - self.add_param('cudacpp_backend', 'cpp', include=False, hidden=False, + self.add_param('cudacpp_backend', 'cpp', + include=False, # AV: 'include=True' would add "CUDACPP_BACKEND = 'cpp'" to run_card.inc + hidden=False, # AV: keep cudacpp_backend in runcard template and keep 'hidden='False' allowed=cudacpp_supported_backends) + self.add_param('cudacpp_fptype', 'm', + include=False, # AV: 'include=True' would add "CUDACPP_FPTYPE = 'm'" to run_card.inc (if fct_mod is removed, else codegen fails) + hidden=False, # AV: add cudacpp_backend to runcard template and keep 'hidden='False' + fct_mod=(self.reset_makeopts,(),{}), # AV: I assume this forces a 'make cleanavx' if FPTYPE changes? + allowed=['m','d','f'] + ) + self.add_param('cudacpp_helinl', '0', + include=False, # AV: no need to add this parameter to run_card.inc + hidden=False, # AV: add cudacpp_helinl to runcard template and keep 'hidden='False' + fct_mod=(self.reset_makeopts,(),{}), # AV: I assume this raises an exception if cudacpp_helinl changes? + allowed=['0','1'] + ) + self.add_param('cudacpp_hrdcod', '0', + include=False, # AV: no need to add this parameter to run_card.inc + hidden=False, # AV: add cudacpp_hrdcod to runcard template and keep 'hidden='False' + fct_mod=(self.reset_makeopts,(),{}), # AV: I assume this raises an exception if cudacpp_hrdcod changes? + allowed=['0','1'] + ) + self.add_param('cudacpp_bldall', False, + include=False, # AV: no need to add this parameter to run_card.inc + hidden=False, # AV: add cudacpp_bldall to runcard template and keep 'hidden='False' + fct_mod=(self.reset_makeopts,(),{}), # AV: I assume this raises an exception if cudacpp_bldall changes? + ) self['vector_size'] = 16 # already setup in default class (just change value) self['aloha_flag'] = '--fast-math' self['matrix_flag'] = '-O3' diff --git a/epochX/cudacpp/gg_tt.mad/Cards/run_card.dat b/epochX/cudacpp/gg_tt.mad/Cards/run_card.dat index 6b82577032..e63cab27e3 100644 --- a/epochX/cudacpp/gg_tt.mad/Cards/run_card.dat +++ b/epochX/cudacpp/gg_tt.mad/Cards/run_card.dat @@ -161,5 +161,10 @@ systematics = systematics_program ! none, systematics [python], SysCalc [deprece #*********************************************************************** # SIMD/GPU configuration for the CUDACPP plugin #************************************************************************ - cpp = cudacpp_backend ! CUDACPP backend: fortran, cuda, hip, cpp, cppnone, cppsse4, cppavx2, cpp512y, cpp512z, cppauto + cpp = cudacpp_backend ! CUDACPP backend: fortran, cuda, hip, cpp (DEFAULT), cppnone, cppsse4, cppavx2, cpp512y, cpp512z, cppauto +#*** WARNING! The following cudacpp runcards are experimental! Users should normally change only the cudacpp_backend card *** + m = cudacpp_fptype ! CUDACPP floating point precision: f (single), d (double), m (mixed, DEFAULT: double for amplitudes, single for colors) + 0 = cudacpp_hrdcod ! CUDACPP parameter hardcoding: 0 (DEFAULT, parameters not hardcoded: read param_card.dat at runtime), 1 (hardcoded parameters) + 0 = cudacpp_helinl ! CUDACPP helicity amplitude inlining: 0 (DEFAULT, ordinary inlining of templates), 1 (aggressive inlining with 'always inline') + False = cudacpp_bldall ! CUDACPP build all available backends in separate build directories: False, True diff --git a/epochX/cudacpp/gg_tt.mad/Cards/run_card_default.dat b/epochX/cudacpp/gg_tt.mad/Cards/run_card_default.dat index b8db871c35..a32f0a05ee 100644 --- a/epochX/cudacpp/gg_tt.mad/Cards/run_card_default.dat +++ b/epochX/cudacpp/gg_tt.mad/Cards/run_card_default.dat @@ -161,5 +161,10 @@ systematics = systematics_program ! none, systematics [python], SysCalc [deprece #*********************************************************************** # SIMD/GPU configuration for the CUDACPP plugin #************************************************************************ - cpp = cudacpp_backend ! CUDACPP backend: fortran, cuda, hip, cpp, cppnone, cppsse4, cppavx2, cpp512y, cpp512z, cppauto + cpp = cudacpp_backend ! CUDACPP backend: fortran, cuda, hip, cpp (DEFAULT), cppnone, cppsse4, cppavx2, cpp512y, cpp512z, cppauto +#*** WARNING! The following cudacpp runcards are experimental! Users should normally change only the cudacpp_backend card *** + m = cudacpp_fptype ! CUDACPP floating point precision: f (single), d (double), m (mixed, DEFAULT: double for amplitudes, single for colors) + 0 = cudacpp_hrdcod ! CUDACPP parameter hardcoding: 0 (DEFAULT, parameters not hardcoded: read param_card.dat at runtime), 1 (hardcoded parameters) + 0 = cudacpp_helinl ! CUDACPP helicity amplitude inlining: 0 (DEFAULT, ordinary inlining of templates), 1 (aggressive inlining with 'always inline') + False = cudacpp_bldall ! CUDACPP build all available backends in separate build directories: False, True diff --git a/epochX/cudacpp/gg_tt.mad/SubProcesses/makefile b/epochX/cudacpp/gg_tt.mad/SubProcesses/makefile index 49e6800fff..2e5bff046b 100644 --- a/epochX/cudacpp/gg_tt.mad/SubProcesses/makefile +++ b/epochX/cudacpp/gg_tt.mad/SubProcesses/makefile @@ -2,6 +2,11 @@ SHELL := /bin/bash include ../../Source/make_opts +# Export the relevant settings from make_opts to builds using the cudacpp makefile (#700) +export FPTYPE +export HELINL +export HRDCOD + # Determine CUDACPP_BUILDDIR based on the user-defined choices of BACKEND, FPTYPE, HELINL, HRDCOD and USEBUILDDIR (#829) # Stop with an error if BACKEND=cuda and nvcc is missing or if BACKEND=hip and hipcc is missing include ../../src/cudacpp_config.mk @@ -162,27 +167,29 @@ madevent_fortran_link: $(PROG)_fortran rm -f $(PROG) ln -s $(PROG)_fortran $(PROG) -madevent_cuda_link: - $(MAKE) USEGTEST=0 BACKEND=cuda $(CUDACPP_BUILDDIR)/$(PROG)_cuda +madevent_cudacpp_link: all rm -f $(PROG) +ifeq ($(BACKEND),cuda) ln -s $(CUDACPP_BUILDDIR)/$(PROG)_cuda $(PROG) +else ifeq ($(BACKEND),hip) + ln -s $(CUDACPP_BUILDDIR)/$(PROG)_hip $(PROG) +else + ln -s $(CUDACPP_BUILDDIR)/$(PROG)_cpp $(PROG) +endif + +madevent_cuda_link: + $(MAKE) USEGTEST=0 USEBUILDDIR=1 BACKEND=cuda madevent_cudacpp_link madevent_hip_link: - $(MAKE) USEGTEST=0 BACKEND=hip $(CUDACPP_BUILDDIR)/$(PROG)_hip - rm -f $(PROG) - ln -s $(CUDACPP_BUILDDIR)/$(PROG)_hip $(PROG) + $(MAKE) USEGTEST=0 USEBUILDDIR=1 BACKEND=hip madevent_cudacpp_link madevent_cpp_link: - $(MAKE) USEGTEST=0 BACKEND=cppauto $(CUDACPP_BUILDDIR)/$(PROG)_cpp - rm -f $(PROG) - ln -s $(CUDACPP_BUILDDIR)/$(PROG)_cpp $(PROG) + $(MAKE) USEGTEST=0 USEBUILDDIR=1 BACKEND=cppauto madevent_cudacpp_link override SUPPORTED_AVXS = cppnone cppsse4 cppavx2 cpp512y cpp512z cppauto madevent_%_link: - @if [ '$(words $(filter $*, $(SUPPORTED_AVXS)))' != '1' ]; then echo "ERROR! Invalid target '$@' (supported madevent_cpp*_link targets are: $(foreach avx,$(SUPPORTED_AVXS),'madevent_cpp$(avx)_link'))"; exit 1; fi - $(MAKE) USEGTEST=0 BACKEND=$* $(CUDACPP_BUILDDIR)/$(PROG)_cpp - rm -f $(PROG) - ln -s $(CUDACPP_BUILDDIR)/$(PROG)_cpp $(PROG) + @if [ '$(words $(filter $*, $(SUPPORTED_AVXS)))' != '1' ]; then echo "ERROR! Invalid target '$@' (supported madevent_cpp*_link targets are: $(foreach avx,$(SUPPORTED_AVXS),'madevent_$(avx)_link'))"; exit 1; fi + $(MAKE) USEGTEST=0 USEBUILDDIR=1 BACKEND=$* madevent_cudacpp_link # Building $(PROG)_cpp no longer builds $(PROG)_cuda if CUDACPP_BACKENDLIB for cuda exists (this was the case in the past to allow cpp-only builds #503) $(CUDACPP_BUILDDIR)/$(PROG)_cpp: $(PROCESS) $(DSIG_cudacpp) auto_dsig.o $(LIBS) $(MATRIX) counters.o ompnumthreads.o $(CUDACPP_BUILDDIR)/.cudacpplibs diff --git a/epochX/cudacpp/gg_tt.mad/bin/internal/launch_plugin.py b/epochX/cudacpp/gg_tt.mad/bin/internal/launch_plugin.py index 0924927785..19d65fa195 100644 --- a/epochX/cudacpp/gg_tt.mad/bin/internal/launch_plugin.py +++ b/epochX/cudacpp/gg_tt.mad/bin/internal/launch_plugin.py @@ -33,10 +33,16 @@ def compile(self, *args, **opts): if 'cwd' in opts and os.path.basename(opts['cwd']) == 'Source': path = pjoin(opts['cwd'], 'make_opts') common_run_interface.CommonRunCmd.update_make_opts_full(path, - {'FPTYPE': self.run_card['floating_type'] }) - misc.sprint('FPTYPE checked') + {'FPTYPE': self.run_card['cudacpp_fptype'], + 'HELINL': self.run_card['cudacpp_helinl'], + 'HRDCOD': self.run_card['cudacpp_hrdcod'] }) + misc.sprint('FPTYPE, HELINL, HRDCOD checked') cudacpp_supported_backends = [ 'fortran', 'cuda', 'hip', 'cpp', 'cppnone', 'cppsse4', 'cppavx2', 'cpp512y', 'cpp512z', 'cppauto' ] if args and args[0][0] == 'madevent' and hasattr(self, 'run_card'): + if self.run_card['cudacpp_bldall'] == True: # pre-build all backends #945 + logger.info("Pre-building madevent in madevent_interface.py with ALL matrix elements") + args[0][0] = 'bldall' + misc.compile(nb_core=self.options['nb_core'], *args, **opts) cudacpp_backend = self.run_card['cudacpp_backend'].lower() # the default value is defined in launch_plugin.py logger.info("Building madevent in madevent_interface.py with '%s' matrix elements"%cudacpp_backend) if cudacpp_backend in cudacpp_supported_backends : @@ -47,14 +53,18 @@ def compile(self, *args, **opts): else: return misc.compile(nb_core=self.options['nb_core'], *args, **opts) -# Phase-Space Optimization ------------------------------------------------------------------------------------ +# CUDACPP runcard block ------------------------------------------------------------------------------------ template_on = \ """#*********************************************************************** # SIMD/GPU configuration for the CUDACPP plugin #************************************************************************ - %(cudacpp_backend)s = cudacpp_backend ! CUDACPP backend: fortran, cuda, hip, cpp, cppnone, cppsse4, cppavx2, cpp512y, cpp512z, cppauto + %(cudacpp_backend)s = cudacpp_backend ! CUDACPP backend: fortran, cuda, hip, cpp (DEFAULT), cppnone, cppsse4, cppavx2, cpp512y, cpp512z, cppauto +#*** WARNING! The following cudacpp runcards are experimental! Users should normally change only the cudacpp_backend card *** + %(cudacpp_fptype)s = cudacpp_fptype ! CUDACPP floating point precision: f (single), d (double), m (mixed, DEFAULT: double for amplitudes, single for colors) + %(cudacpp_hrdcod)s = cudacpp_hrdcod ! CUDACPP parameter hardcoding: 0 (DEFAULT, parameters not hardcoded: read param_card.dat at runtime), 1 (hardcoded parameters) + %(cudacpp_helinl)s = cudacpp_helinl ! CUDACPP helicity amplitude inlining: 0 (DEFAULT, ordinary inlining of templates), 1 (aggressive inlining with 'always inline') + %(cudacpp_bldall)s = cudacpp_bldall ! CUDACPP build all available backends in separate build directories: False, True """ - template_off = '' plugin_block = banner_mod.RunBlock('simd', template_on=template_on, template_off=template_off) @@ -75,8 +85,14 @@ def reset_simd(self, old_value, new_value, name): def reset_makeopts(self, old_value, new_value, name): if not hasattr(self, 'path'): raise Exception - if name == 'floating_type': + if name == 'cudacpp_fptype': common_run_interface.CommonRunCmd.update_make_opts_full({'FPTYPE': new_value}) + elif name == 'cudacpp_hrdcod': + raise Exception('Cannot change cudacpp_hrdcod') + elif name == 'cudacpp_helinl': + raise Exception('Cannot change cudacpp_helinl') + elif name == 'cudacpp_bldall': + raise Exception('Cannot change cudacpp_bldall') else: raise Exception Sourcedir = pjoin(os.path.dirname(os.path.dirname(self.path)), 'Source') @@ -84,14 +100,34 @@ def reset_makeopts(self, old_value, new_value, name): def default_setup(self): super().default_setup() - self.add_param('floating_type', 'm', include=False, hidden=True, - fct_mod=(self.reset_makeopts,(),{}), - allowed=['m','d','f'], - comment='floating point precision: f (single), d (double), m (mixed: double for amplitudes, single for colors)' - ) cudacpp_supported_backends = [ 'fortran', 'cuda', 'hip', 'cpp', 'cppnone', 'cppsse4', 'cppavx2', 'cpp512y', 'cpp512z', 'cppauto' ] - self.add_param('cudacpp_backend', 'cpp', include=False, hidden=False, + self.add_param('cudacpp_backend', 'cpp', + include=False, # AV: 'include=True' would add "CUDACPP_BACKEND = 'cpp'" to run_card.inc + hidden=False, # AV: keep cudacpp_backend in runcard template and keep 'hidden='False' allowed=cudacpp_supported_backends) + self.add_param('cudacpp_fptype', 'm', + include=False, # AV: 'include=True' would add "CUDACPP_FPTYPE = 'm'" to run_card.inc (if fct_mod is removed, else codegen fails) + hidden=False, # AV: add cudacpp_backend to runcard template and keep 'hidden='False' + fct_mod=(self.reset_makeopts,(),{}), # AV: I assume this forces a 'make cleanavx' if FPTYPE changes? + allowed=['m','d','f'] + ) + self.add_param('cudacpp_helinl', '0', + include=False, # AV: no need to add this parameter to run_card.inc + hidden=False, # AV: add cudacpp_helinl to runcard template and keep 'hidden='False' + fct_mod=(self.reset_makeopts,(),{}), # AV: I assume this raises an exception if cudacpp_helinl changes? + allowed=['0','1'] + ) + self.add_param('cudacpp_hrdcod', '0', + include=False, # AV: no need to add this parameter to run_card.inc + hidden=False, # AV: add cudacpp_hrdcod to runcard template and keep 'hidden='False' + fct_mod=(self.reset_makeopts,(),{}), # AV: I assume this raises an exception if cudacpp_hrdcod changes? + allowed=['0','1'] + ) + self.add_param('cudacpp_bldall', False, + include=False, # AV: no need to add this parameter to run_card.inc + hidden=False, # AV: add cudacpp_bldall to runcard template and keep 'hidden='False' + fct_mod=(self.reset_makeopts,(),{}), # AV: I assume this raises an exception if cudacpp_bldall changes? + ) self['vector_size'] = 16 # already setup in default class (just change value) self['aloha_flag'] = '--fast-math' self['matrix_flag'] = '-O3' diff --git a/epochX/cudacpp/gg_tt01g.mad/Cards/run_card.dat b/epochX/cudacpp/gg_tt01g.mad/Cards/run_card.dat index 1711d30371..4d793781ba 100644 --- a/epochX/cudacpp/gg_tt01g.mad/Cards/run_card.dat +++ b/epochX/cudacpp/gg_tt01g.mad/Cards/run_card.dat @@ -182,5 +182,10 @@ systematics = systematics_program ! none, systematics [python], SysCalc [deprece #*********************************************************************** # SIMD/GPU configuration for the CUDACPP plugin #************************************************************************ - cpp = cudacpp_backend ! CUDACPP backend: fortran, cuda, hip, cpp, cppnone, cppsse4, cppavx2, cpp512y, cpp512z, cppauto + cpp = cudacpp_backend ! CUDACPP backend: fortran, cuda, hip, cpp (DEFAULT), cppnone, cppsse4, cppavx2, cpp512y, cpp512z, cppauto +#*** WARNING! The following cudacpp runcards are experimental! Users should normally change only the cudacpp_backend card *** + m = cudacpp_fptype ! CUDACPP floating point precision: f (single), d (double), m (mixed, DEFAULT: double for amplitudes, single for colors) + 0 = cudacpp_hrdcod ! CUDACPP parameter hardcoding: 0 (DEFAULT, parameters not hardcoded: read param_card.dat at runtime), 1 (hardcoded parameters) + 0 = cudacpp_helinl ! CUDACPP helicity amplitude inlining: 0 (DEFAULT, ordinary inlining of templates), 1 (aggressive inlining with 'always inline') + False = cudacpp_bldall ! CUDACPP build all available backends in separate build directories: False, True diff --git a/epochX/cudacpp/gg_tt01g.mad/Cards/run_card_default.dat b/epochX/cudacpp/gg_tt01g.mad/Cards/run_card_default.dat index 364dbd21b0..7e3105aeb0 100644 --- a/epochX/cudacpp/gg_tt01g.mad/Cards/run_card_default.dat +++ b/epochX/cudacpp/gg_tt01g.mad/Cards/run_card_default.dat @@ -182,5 +182,10 @@ systematics = systematics_program ! none, systematics [python], SysCalc [deprece #*********************************************************************** # SIMD/GPU configuration for the CUDACPP plugin #************************************************************************ - cpp = cudacpp_backend ! CUDACPP backend: fortran, cuda, hip, cpp, cppnone, cppsse4, cppavx2, cpp512y, cpp512z, cppauto + cpp = cudacpp_backend ! CUDACPP backend: fortran, cuda, hip, cpp (DEFAULT), cppnone, cppsse4, cppavx2, cpp512y, cpp512z, cppauto +#*** WARNING! The following cudacpp runcards are experimental! Users should normally change only the cudacpp_backend card *** + m = cudacpp_fptype ! CUDACPP floating point precision: f (single), d (double), m (mixed, DEFAULT: double for amplitudes, single for colors) + 0 = cudacpp_hrdcod ! CUDACPP parameter hardcoding: 0 (DEFAULT, parameters not hardcoded: read param_card.dat at runtime), 1 (hardcoded parameters) + 0 = cudacpp_helinl ! CUDACPP helicity amplitude inlining: 0 (DEFAULT, ordinary inlining of templates), 1 (aggressive inlining with 'always inline') + False = cudacpp_bldall ! CUDACPP build all available backends in separate build directories: False, True diff --git a/epochX/cudacpp/gg_tt01g.mad/SubProcesses/makefile b/epochX/cudacpp/gg_tt01g.mad/SubProcesses/makefile index 49e6800fff..2e5bff046b 100644 --- a/epochX/cudacpp/gg_tt01g.mad/SubProcesses/makefile +++ b/epochX/cudacpp/gg_tt01g.mad/SubProcesses/makefile @@ -2,6 +2,11 @@ SHELL := /bin/bash include ../../Source/make_opts +# Export the relevant settings from make_opts to builds using the cudacpp makefile (#700) +export FPTYPE +export HELINL +export HRDCOD + # Determine CUDACPP_BUILDDIR based on the user-defined choices of BACKEND, FPTYPE, HELINL, HRDCOD and USEBUILDDIR (#829) # Stop with an error if BACKEND=cuda and nvcc is missing or if BACKEND=hip and hipcc is missing include ../../src/cudacpp_config.mk @@ -162,27 +167,29 @@ madevent_fortran_link: $(PROG)_fortran rm -f $(PROG) ln -s $(PROG)_fortran $(PROG) -madevent_cuda_link: - $(MAKE) USEGTEST=0 BACKEND=cuda $(CUDACPP_BUILDDIR)/$(PROG)_cuda +madevent_cudacpp_link: all rm -f $(PROG) +ifeq ($(BACKEND),cuda) ln -s $(CUDACPP_BUILDDIR)/$(PROG)_cuda $(PROG) +else ifeq ($(BACKEND),hip) + ln -s $(CUDACPP_BUILDDIR)/$(PROG)_hip $(PROG) +else + ln -s $(CUDACPP_BUILDDIR)/$(PROG)_cpp $(PROG) +endif + +madevent_cuda_link: + $(MAKE) USEGTEST=0 USEBUILDDIR=1 BACKEND=cuda madevent_cudacpp_link madevent_hip_link: - $(MAKE) USEGTEST=0 BACKEND=hip $(CUDACPP_BUILDDIR)/$(PROG)_hip - rm -f $(PROG) - ln -s $(CUDACPP_BUILDDIR)/$(PROG)_hip $(PROG) + $(MAKE) USEGTEST=0 USEBUILDDIR=1 BACKEND=hip madevent_cudacpp_link madevent_cpp_link: - $(MAKE) USEGTEST=0 BACKEND=cppauto $(CUDACPP_BUILDDIR)/$(PROG)_cpp - rm -f $(PROG) - ln -s $(CUDACPP_BUILDDIR)/$(PROG)_cpp $(PROG) + $(MAKE) USEGTEST=0 USEBUILDDIR=1 BACKEND=cppauto madevent_cudacpp_link override SUPPORTED_AVXS = cppnone cppsse4 cppavx2 cpp512y cpp512z cppauto madevent_%_link: - @if [ '$(words $(filter $*, $(SUPPORTED_AVXS)))' != '1' ]; then echo "ERROR! Invalid target '$@' (supported madevent_cpp*_link targets are: $(foreach avx,$(SUPPORTED_AVXS),'madevent_cpp$(avx)_link'))"; exit 1; fi - $(MAKE) USEGTEST=0 BACKEND=$* $(CUDACPP_BUILDDIR)/$(PROG)_cpp - rm -f $(PROG) - ln -s $(CUDACPP_BUILDDIR)/$(PROG)_cpp $(PROG) + @if [ '$(words $(filter $*, $(SUPPORTED_AVXS)))' != '1' ]; then echo "ERROR! Invalid target '$@' (supported madevent_cpp*_link targets are: $(foreach avx,$(SUPPORTED_AVXS),'madevent_$(avx)_link'))"; exit 1; fi + $(MAKE) USEGTEST=0 USEBUILDDIR=1 BACKEND=$* madevent_cudacpp_link # Building $(PROG)_cpp no longer builds $(PROG)_cuda if CUDACPP_BACKENDLIB for cuda exists (this was the case in the past to allow cpp-only builds #503) $(CUDACPP_BUILDDIR)/$(PROG)_cpp: $(PROCESS) $(DSIG_cudacpp) auto_dsig.o $(LIBS) $(MATRIX) counters.o ompnumthreads.o $(CUDACPP_BUILDDIR)/.cudacpplibs diff --git a/epochX/cudacpp/gg_tt01g.mad/bin/internal/launch_plugin.py b/epochX/cudacpp/gg_tt01g.mad/bin/internal/launch_plugin.py index 0924927785..19d65fa195 100644 --- a/epochX/cudacpp/gg_tt01g.mad/bin/internal/launch_plugin.py +++ b/epochX/cudacpp/gg_tt01g.mad/bin/internal/launch_plugin.py @@ -33,10 +33,16 @@ def compile(self, *args, **opts): if 'cwd' in opts and os.path.basename(opts['cwd']) == 'Source': path = pjoin(opts['cwd'], 'make_opts') common_run_interface.CommonRunCmd.update_make_opts_full(path, - {'FPTYPE': self.run_card['floating_type'] }) - misc.sprint('FPTYPE checked') + {'FPTYPE': self.run_card['cudacpp_fptype'], + 'HELINL': self.run_card['cudacpp_helinl'], + 'HRDCOD': self.run_card['cudacpp_hrdcod'] }) + misc.sprint('FPTYPE, HELINL, HRDCOD checked') cudacpp_supported_backends = [ 'fortran', 'cuda', 'hip', 'cpp', 'cppnone', 'cppsse4', 'cppavx2', 'cpp512y', 'cpp512z', 'cppauto' ] if args and args[0][0] == 'madevent' and hasattr(self, 'run_card'): + if self.run_card['cudacpp_bldall'] == True: # pre-build all backends #945 + logger.info("Pre-building madevent in madevent_interface.py with ALL matrix elements") + args[0][0] = 'bldall' + misc.compile(nb_core=self.options['nb_core'], *args, **opts) cudacpp_backend = self.run_card['cudacpp_backend'].lower() # the default value is defined in launch_plugin.py logger.info("Building madevent in madevent_interface.py with '%s' matrix elements"%cudacpp_backend) if cudacpp_backend in cudacpp_supported_backends : @@ -47,14 +53,18 @@ def compile(self, *args, **opts): else: return misc.compile(nb_core=self.options['nb_core'], *args, **opts) -# Phase-Space Optimization ------------------------------------------------------------------------------------ +# CUDACPP runcard block ------------------------------------------------------------------------------------ template_on = \ """#*********************************************************************** # SIMD/GPU configuration for the CUDACPP plugin #************************************************************************ - %(cudacpp_backend)s = cudacpp_backend ! CUDACPP backend: fortran, cuda, hip, cpp, cppnone, cppsse4, cppavx2, cpp512y, cpp512z, cppauto + %(cudacpp_backend)s = cudacpp_backend ! CUDACPP backend: fortran, cuda, hip, cpp (DEFAULT), cppnone, cppsse4, cppavx2, cpp512y, cpp512z, cppauto +#*** WARNING! The following cudacpp runcards are experimental! Users should normally change only the cudacpp_backend card *** + %(cudacpp_fptype)s = cudacpp_fptype ! CUDACPP floating point precision: f (single), d (double), m (mixed, DEFAULT: double for amplitudes, single for colors) + %(cudacpp_hrdcod)s = cudacpp_hrdcod ! CUDACPP parameter hardcoding: 0 (DEFAULT, parameters not hardcoded: read param_card.dat at runtime), 1 (hardcoded parameters) + %(cudacpp_helinl)s = cudacpp_helinl ! CUDACPP helicity amplitude inlining: 0 (DEFAULT, ordinary inlining of templates), 1 (aggressive inlining with 'always inline') + %(cudacpp_bldall)s = cudacpp_bldall ! CUDACPP build all available backends in separate build directories: False, True """ - template_off = '' plugin_block = banner_mod.RunBlock('simd', template_on=template_on, template_off=template_off) @@ -75,8 +85,14 @@ def reset_simd(self, old_value, new_value, name): def reset_makeopts(self, old_value, new_value, name): if not hasattr(self, 'path'): raise Exception - if name == 'floating_type': + if name == 'cudacpp_fptype': common_run_interface.CommonRunCmd.update_make_opts_full({'FPTYPE': new_value}) + elif name == 'cudacpp_hrdcod': + raise Exception('Cannot change cudacpp_hrdcod') + elif name == 'cudacpp_helinl': + raise Exception('Cannot change cudacpp_helinl') + elif name == 'cudacpp_bldall': + raise Exception('Cannot change cudacpp_bldall') else: raise Exception Sourcedir = pjoin(os.path.dirname(os.path.dirname(self.path)), 'Source') @@ -84,14 +100,34 @@ def reset_makeopts(self, old_value, new_value, name): def default_setup(self): super().default_setup() - self.add_param('floating_type', 'm', include=False, hidden=True, - fct_mod=(self.reset_makeopts,(),{}), - allowed=['m','d','f'], - comment='floating point precision: f (single), d (double), m (mixed: double for amplitudes, single for colors)' - ) cudacpp_supported_backends = [ 'fortran', 'cuda', 'hip', 'cpp', 'cppnone', 'cppsse4', 'cppavx2', 'cpp512y', 'cpp512z', 'cppauto' ] - self.add_param('cudacpp_backend', 'cpp', include=False, hidden=False, + self.add_param('cudacpp_backend', 'cpp', + include=False, # AV: 'include=True' would add "CUDACPP_BACKEND = 'cpp'" to run_card.inc + hidden=False, # AV: keep cudacpp_backend in runcard template and keep 'hidden='False' allowed=cudacpp_supported_backends) + self.add_param('cudacpp_fptype', 'm', + include=False, # AV: 'include=True' would add "CUDACPP_FPTYPE = 'm'" to run_card.inc (if fct_mod is removed, else codegen fails) + hidden=False, # AV: add cudacpp_backend to runcard template and keep 'hidden='False' + fct_mod=(self.reset_makeopts,(),{}), # AV: I assume this forces a 'make cleanavx' if FPTYPE changes? + allowed=['m','d','f'] + ) + self.add_param('cudacpp_helinl', '0', + include=False, # AV: no need to add this parameter to run_card.inc + hidden=False, # AV: add cudacpp_helinl to runcard template and keep 'hidden='False' + fct_mod=(self.reset_makeopts,(),{}), # AV: I assume this raises an exception if cudacpp_helinl changes? + allowed=['0','1'] + ) + self.add_param('cudacpp_hrdcod', '0', + include=False, # AV: no need to add this parameter to run_card.inc + hidden=False, # AV: add cudacpp_hrdcod to runcard template and keep 'hidden='False' + fct_mod=(self.reset_makeopts,(),{}), # AV: I assume this raises an exception if cudacpp_hrdcod changes? + allowed=['0','1'] + ) + self.add_param('cudacpp_bldall', False, + include=False, # AV: no need to add this parameter to run_card.inc + hidden=False, # AV: add cudacpp_bldall to runcard template and keep 'hidden='False' + fct_mod=(self.reset_makeopts,(),{}), # AV: I assume this raises an exception if cudacpp_bldall changes? + ) self['vector_size'] = 16 # already setup in default class (just change value) self['aloha_flag'] = '--fast-math' self['matrix_flag'] = '-O3' diff --git a/epochX/cudacpp/gg_ttg.mad/Cards/run_card.dat b/epochX/cudacpp/gg_ttg.mad/Cards/run_card.dat index d087670827..22884b593b 100644 --- a/epochX/cudacpp/gg_ttg.mad/Cards/run_card.dat +++ b/epochX/cudacpp/gg_ttg.mad/Cards/run_card.dat @@ -164,5 +164,10 @@ systematics = systematics_program ! none, systematics [python], SysCalc [deprece #*********************************************************************** # SIMD/GPU configuration for the CUDACPP plugin #************************************************************************ - cpp = cudacpp_backend ! CUDACPP backend: fortran, cuda, hip, cpp, cppnone, cppsse4, cppavx2, cpp512y, cpp512z, cppauto + cpp = cudacpp_backend ! CUDACPP backend: fortran, cuda, hip, cpp (DEFAULT), cppnone, cppsse4, cppavx2, cpp512y, cpp512z, cppauto +#*** WARNING! The following cudacpp runcards are experimental! Users should normally change only the cudacpp_backend card *** + m = cudacpp_fptype ! CUDACPP floating point precision: f (single), d (double), m (mixed, DEFAULT: double for amplitudes, single for colors) + 0 = cudacpp_hrdcod ! CUDACPP parameter hardcoding: 0 (DEFAULT, parameters not hardcoded: read param_card.dat at runtime), 1 (hardcoded parameters) + 0 = cudacpp_helinl ! CUDACPP helicity amplitude inlining: 0 (DEFAULT, ordinary inlining of templates), 1 (aggressive inlining with 'always inline') + False = cudacpp_bldall ! CUDACPP build all available backends in separate build directories: False, True diff --git a/epochX/cudacpp/gg_ttg.mad/Cards/run_card_default.dat b/epochX/cudacpp/gg_ttg.mad/Cards/run_card_default.dat index 43e93cbf40..c9521d5fcf 100644 --- a/epochX/cudacpp/gg_ttg.mad/Cards/run_card_default.dat +++ b/epochX/cudacpp/gg_ttg.mad/Cards/run_card_default.dat @@ -164,5 +164,10 @@ systematics = systematics_program ! none, systematics [python], SysCalc [deprece #*********************************************************************** # SIMD/GPU configuration for the CUDACPP plugin #************************************************************************ - cpp = cudacpp_backend ! CUDACPP backend: fortran, cuda, hip, cpp, cppnone, cppsse4, cppavx2, cpp512y, cpp512z, cppauto + cpp = cudacpp_backend ! CUDACPP backend: fortran, cuda, hip, cpp (DEFAULT), cppnone, cppsse4, cppavx2, cpp512y, cpp512z, cppauto +#*** WARNING! The following cudacpp runcards are experimental! Users should normally change only the cudacpp_backend card *** + m = cudacpp_fptype ! CUDACPP floating point precision: f (single), d (double), m (mixed, DEFAULT: double for amplitudes, single for colors) + 0 = cudacpp_hrdcod ! CUDACPP parameter hardcoding: 0 (DEFAULT, parameters not hardcoded: read param_card.dat at runtime), 1 (hardcoded parameters) + 0 = cudacpp_helinl ! CUDACPP helicity amplitude inlining: 0 (DEFAULT, ordinary inlining of templates), 1 (aggressive inlining with 'always inline') + False = cudacpp_bldall ! CUDACPP build all available backends in separate build directories: False, True diff --git a/epochX/cudacpp/gg_ttg.mad/SubProcesses/makefile b/epochX/cudacpp/gg_ttg.mad/SubProcesses/makefile index 49e6800fff..2e5bff046b 100644 --- a/epochX/cudacpp/gg_ttg.mad/SubProcesses/makefile +++ b/epochX/cudacpp/gg_ttg.mad/SubProcesses/makefile @@ -2,6 +2,11 @@ SHELL := /bin/bash include ../../Source/make_opts +# Export the relevant settings from make_opts to builds using the cudacpp makefile (#700) +export FPTYPE +export HELINL +export HRDCOD + # Determine CUDACPP_BUILDDIR based on the user-defined choices of BACKEND, FPTYPE, HELINL, HRDCOD and USEBUILDDIR (#829) # Stop with an error if BACKEND=cuda and nvcc is missing or if BACKEND=hip and hipcc is missing include ../../src/cudacpp_config.mk @@ -162,27 +167,29 @@ madevent_fortran_link: $(PROG)_fortran rm -f $(PROG) ln -s $(PROG)_fortran $(PROG) -madevent_cuda_link: - $(MAKE) USEGTEST=0 BACKEND=cuda $(CUDACPP_BUILDDIR)/$(PROG)_cuda +madevent_cudacpp_link: all rm -f $(PROG) +ifeq ($(BACKEND),cuda) ln -s $(CUDACPP_BUILDDIR)/$(PROG)_cuda $(PROG) +else ifeq ($(BACKEND),hip) + ln -s $(CUDACPP_BUILDDIR)/$(PROG)_hip $(PROG) +else + ln -s $(CUDACPP_BUILDDIR)/$(PROG)_cpp $(PROG) +endif + +madevent_cuda_link: + $(MAKE) USEGTEST=0 USEBUILDDIR=1 BACKEND=cuda madevent_cudacpp_link madevent_hip_link: - $(MAKE) USEGTEST=0 BACKEND=hip $(CUDACPP_BUILDDIR)/$(PROG)_hip - rm -f $(PROG) - ln -s $(CUDACPP_BUILDDIR)/$(PROG)_hip $(PROG) + $(MAKE) USEGTEST=0 USEBUILDDIR=1 BACKEND=hip madevent_cudacpp_link madevent_cpp_link: - $(MAKE) USEGTEST=0 BACKEND=cppauto $(CUDACPP_BUILDDIR)/$(PROG)_cpp - rm -f $(PROG) - ln -s $(CUDACPP_BUILDDIR)/$(PROG)_cpp $(PROG) + $(MAKE) USEGTEST=0 USEBUILDDIR=1 BACKEND=cppauto madevent_cudacpp_link override SUPPORTED_AVXS = cppnone cppsse4 cppavx2 cpp512y cpp512z cppauto madevent_%_link: - @if [ '$(words $(filter $*, $(SUPPORTED_AVXS)))' != '1' ]; then echo "ERROR! Invalid target '$@' (supported madevent_cpp*_link targets are: $(foreach avx,$(SUPPORTED_AVXS),'madevent_cpp$(avx)_link'))"; exit 1; fi - $(MAKE) USEGTEST=0 BACKEND=$* $(CUDACPP_BUILDDIR)/$(PROG)_cpp - rm -f $(PROG) - ln -s $(CUDACPP_BUILDDIR)/$(PROG)_cpp $(PROG) + @if [ '$(words $(filter $*, $(SUPPORTED_AVXS)))' != '1' ]; then echo "ERROR! Invalid target '$@' (supported madevent_cpp*_link targets are: $(foreach avx,$(SUPPORTED_AVXS),'madevent_$(avx)_link'))"; exit 1; fi + $(MAKE) USEGTEST=0 USEBUILDDIR=1 BACKEND=$* madevent_cudacpp_link # Building $(PROG)_cpp no longer builds $(PROG)_cuda if CUDACPP_BACKENDLIB for cuda exists (this was the case in the past to allow cpp-only builds #503) $(CUDACPP_BUILDDIR)/$(PROG)_cpp: $(PROCESS) $(DSIG_cudacpp) auto_dsig.o $(LIBS) $(MATRIX) counters.o ompnumthreads.o $(CUDACPP_BUILDDIR)/.cudacpplibs diff --git a/epochX/cudacpp/gg_ttg.mad/bin/internal/launch_plugin.py b/epochX/cudacpp/gg_ttg.mad/bin/internal/launch_plugin.py index 0924927785..19d65fa195 100644 --- a/epochX/cudacpp/gg_ttg.mad/bin/internal/launch_plugin.py +++ b/epochX/cudacpp/gg_ttg.mad/bin/internal/launch_plugin.py @@ -33,10 +33,16 @@ def compile(self, *args, **opts): if 'cwd' in opts and os.path.basename(opts['cwd']) == 'Source': path = pjoin(opts['cwd'], 'make_opts') common_run_interface.CommonRunCmd.update_make_opts_full(path, - {'FPTYPE': self.run_card['floating_type'] }) - misc.sprint('FPTYPE checked') + {'FPTYPE': self.run_card['cudacpp_fptype'], + 'HELINL': self.run_card['cudacpp_helinl'], + 'HRDCOD': self.run_card['cudacpp_hrdcod'] }) + misc.sprint('FPTYPE, HELINL, HRDCOD checked') cudacpp_supported_backends = [ 'fortran', 'cuda', 'hip', 'cpp', 'cppnone', 'cppsse4', 'cppavx2', 'cpp512y', 'cpp512z', 'cppauto' ] if args and args[0][0] == 'madevent' and hasattr(self, 'run_card'): + if self.run_card['cudacpp_bldall'] == True: # pre-build all backends #945 + logger.info("Pre-building madevent in madevent_interface.py with ALL matrix elements") + args[0][0] = 'bldall' + misc.compile(nb_core=self.options['nb_core'], *args, **opts) cudacpp_backend = self.run_card['cudacpp_backend'].lower() # the default value is defined in launch_plugin.py logger.info("Building madevent in madevent_interface.py with '%s' matrix elements"%cudacpp_backend) if cudacpp_backend in cudacpp_supported_backends : @@ -47,14 +53,18 @@ def compile(self, *args, **opts): else: return misc.compile(nb_core=self.options['nb_core'], *args, **opts) -# Phase-Space Optimization ------------------------------------------------------------------------------------ +# CUDACPP runcard block ------------------------------------------------------------------------------------ template_on = \ """#*********************************************************************** # SIMD/GPU configuration for the CUDACPP plugin #************************************************************************ - %(cudacpp_backend)s = cudacpp_backend ! CUDACPP backend: fortran, cuda, hip, cpp, cppnone, cppsse4, cppavx2, cpp512y, cpp512z, cppauto + %(cudacpp_backend)s = cudacpp_backend ! CUDACPP backend: fortran, cuda, hip, cpp (DEFAULT), cppnone, cppsse4, cppavx2, cpp512y, cpp512z, cppauto +#*** WARNING! The following cudacpp runcards are experimental! Users should normally change only the cudacpp_backend card *** + %(cudacpp_fptype)s = cudacpp_fptype ! CUDACPP floating point precision: f (single), d (double), m (mixed, DEFAULT: double for amplitudes, single for colors) + %(cudacpp_hrdcod)s = cudacpp_hrdcod ! CUDACPP parameter hardcoding: 0 (DEFAULT, parameters not hardcoded: read param_card.dat at runtime), 1 (hardcoded parameters) + %(cudacpp_helinl)s = cudacpp_helinl ! CUDACPP helicity amplitude inlining: 0 (DEFAULT, ordinary inlining of templates), 1 (aggressive inlining with 'always inline') + %(cudacpp_bldall)s = cudacpp_bldall ! CUDACPP build all available backends in separate build directories: False, True """ - template_off = '' plugin_block = banner_mod.RunBlock('simd', template_on=template_on, template_off=template_off) @@ -75,8 +85,14 @@ def reset_simd(self, old_value, new_value, name): def reset_makeopts(self, old_value, new_value, name): if not hasattr(self, 'path'): raise Exception - if name == 'floating_type': + if name == 'cudacpp_fptype': common_run_interface.CommonRunCmd.update_make_opts_full({'FPTYPE': new_value}) + elif name == 'cudacpp_hrdcod': + raise Exception('Cannot change cudacpp_hrdcod') + elif name == 'cudacpp_helinl': + raise Exception('Cannot change cudacpp_helinl') + elif name == 'cudacpp_bldall': + raise Exception('Cannot change cudacpp_bldall') else: raise Exception Sourcedir = pjoin(os.path.dirname(os.path.dirname(self.path)), 'Source') @@ -84,14 +100,34 @@ def reset_makeopts(self, old_value, new_value, name): def default_setup(self): super().default_setup() - self.add_param('floating_type', 'm', include=False, hidden=True, - fct_mod=(self.reset_makeopts,(),{}), - allowed=['m','d','f'], - comment='floating point precision: f (single), d (double), m (mixed: double for amplitudes, single for colors)' - ) cudacpp_supported_backends = [ 'fortran', 'cuda', 'hip', 'cpp', 'cppnone', 'cppsse4', 'cppavx2', 'cpp512y', 'cpp512z', 'cppauto' ] - self.add_param('cudacpp_backend', 'cpp', include=False, hidden=False, + self.add_param('cudacpp_backend', 'cpp', + include=False, # AV: 'include=True' would add "CUDACPP_BACKEND = 'cpp'" to run_card.inc + hidden=False, # AV: keep cudacpp_backend in runcard template and keep 'hidden='False' allowed=cudacpp_supported_backends) + self.add_param('cudacpp_fptype', 'm', + include=False, # AV: 'include=True' would add "CUDACPP_FPTYPE = 'm'" to run_card.inc (if fct_mod is removed, else codegen fails) + hidden=False, # AV: add cudacpp_backend to runcard template and keep 'hidden='False' + fct_mod=(self.reset_makeopts,(),{}), # AV: I assume this forces a 'make cleanavx' if FPTYPE changes? + allowed=['m','d','f'] + ) + self.add_param('cudacpp_helinl', '0', + include=False, # AV: no need to add this parameter to run_card.inc + hidden=False, # AV: add cudacpp_helinl to runcard template and keep 'hidden='False' + fct_mod=(self.reset_makeopts,(),{}), # AV: I assume this raises an exception if cudacpp_helinl changes? + allowed=['0','1'] + ) + self.add_param('cudacpp_hrdcod', '0', + include=False, # AV: no need to add this parameter to run_card.inc + hidden=False, # AV: add cudacpp_hrdcod to runcard template and keep 'hidden='False' + fct_mod=(self.reset_makeopts,(),{}), # AV: I assume this raises an exception if cudacpp_hrdcod changes? + allowed=['0','1'] + ) + self.add_param('cudacpp_bldall', False, + include=False, # AV: no need to add this parameter to run_card.inc + hidden=False, # AV: add cudacpp_bldall to runcard template and keep 'hidden='False' + fct_mod=(self.reset_makeopts,(),{}), # AV: I assume this raises an exception if cudacpp_bldall changes? + ) self['vector_size'] = 16 # already setup in default class (just change value) self['aloha_flag'] = '--fast-math' self['matrix_flag'] = '-O3' diff --git a/epochX/cudacpp/gg_ttgg.mad/Cards/run_card.dat b/epochX/cudacpp/gg_ttgg.mad/Cards/run_card.dat index ecdc7fd25c..92d33b0e6c 100644 --- a/epochX/cudacpp/gg_ttgg.mad/Cards/run_card.dat +++ b/epochX/cudacpp/gg_ttgg.mad/Cards/run_card.dat @@ -189,5 +189,10 @@ systematics = systematics_program ! none, systematics [python], SysCalc [deprece #*********************************************************************** # SIMD/GPU configuration for the CUDACPP plugin #************************************************************************ - cpp = cudacpp_backend ! CUDACPP backend: fortran, cuda, hip, cpp, cppnone, cppsse4, cppavx2, cpp512y, cpp512z, cppauto + cpp = cudacpp_backend ! CUDACPP backend: fortran, cuda, hip, cpp (DEFAULT), cppnone, cppsse4, cppavx2, cpp512y, cpp512z, cppauto +#*** WARNING! The following cudacpp runcards are experimental! Users should normally change only the cudacpp_backend card *** + m = cudacpp_fptype ! CUDACPP floating point precision: f (single), d (double), m (mixed, DEFAULT: double for amplitudes, single for colors) + 0 = cudacpp_hrdcod ! CUDACPP parameter hardcoding: 0 (DEFAULT, parameters not hardcoded: read param_card.dat at runtime), 1 (hardcoded parameters) + 0 = cudacpp_helinl ! CUDACPP helicity amplitude inlining: 0 (DEFAULT, ordinary inlining of templates), 1 (aggressive inlining with 'always inline') + False = cudacpp_bldall ! CUDACPP build all available backends in separate build directories: False, True diff --git a/epochX/cudacpp/gg_ttgg.mad/Cards/run_card_default.dat b/epochX/cudacpp/gg_ttgg.mad/Cards/run_card_default.dat index 7ec841d6c2..fb59ffc9fb 100644 --- a/epochX/cudacpp/gg_ttgg.mad/Cards/run_card_default.dat +++ b/epochX/cudacpp/gg_ttgg.mad/Cards/run_card_default.dat @@ -189,5 +189,10 @@ systematics = systematics_program ! none, systematics [python], SysCalc [deprece #*********************************************************************** # SIMD/GPU configuration for the CUDACPP plugin #************************************************************************ - cpp = cudacpp_backend ! CUDACPP backend: fortran, cuda, hip, cpp, cppnone, cppsse4, cppavx2, cpp512y, cpp512z, cppauto + cpp = cudacpp_backend ! CUDACPP backend: fortran, cuda, hip, cpp (DEFAULT), cppnone, cppsse4, cppavx2, cpp512y, cpp512z, cppauto +#*** WARNING! The following cudacpp runcards are experimental! Users should normally change only the cudacpp_backend card *** + m = cudacpp_fptype ! CUDACPP floating point precision: f (single), d (double), m (mixed, DEFAULT: double for amplitudes, single for colors) + 0 = cudacpp_hrdcod ! CUDACPP parameter hardcoding: 0 (DEFAULT, parameters not hardcoded: read param_card.dat at runtime), 1 (hardcoded parameters) + 0 = cudacpp_helinl ! CUDACPP helicity amplitude inlining: 0 (DEFAULT, ordinary inlining of templates), 1 (aggressive inlining with 'always inline') + False = cudacpp_bldall ! CUDACPP build all available backends in separate build directories: False, True diff --git a/epochX/cudacpp/gg_ttgg.mad/SubProcesses/makefile b/epochX/cudacpp/gg_ttgg.mad/SubProcesses/makefile index 49e6800fff..2e5bff046b 100644 --- a/epochX/cudacpp/gg_ttgg.mad/SubProcesses/makefile +++ b/epochX/cudacpp/gg_ttgg.mad/SubProcesses/makefile @@ -2,6 +2,11 @@ SHELL := /bin/bash include ../../Source/make_opts +# Export the relevant settings from make_opts to builds using the cudacpp makefile (#700) +export FPTYPE +export HELINL +export HRDCOD + # Determine CUDACPP_BUILDDIR based on the user-defined choices of BACKEND, FPTYPE, HELINL, HRDCOD and USEBUILDDIR (#829) # Stop with an error if BACKEND=cuda and nvcc is missing or if BACKEND=hip and hipcc is missing include ../../src/cudacpp_config.mk @@ -162,27 +167,29 @@ madevent_fortran_link: $(PROG)_fortran rm -f $(PROG) ln -s $(PROG)_fortran $(PROG) -madevent_cuda_link: - $(MAKE) USEGTEST=0 BACKEND=cuda $(CUDACPP_BUILDDIR)/$(PROG)_cuda +madevent_cudacpp_link: all rm -f $(PROG) +ifeq ($(BACKEND),cuda) ln -s $(CUDACPP_BUILDDIR)/$(PROG)_cuda $(PROG) +else ifeq ($(BACKEND),hip) + ln -s $(CUDACPP_BUILDDIR)/$(PROG)_hip $(PROG) +else + ln -s $(CUDACPP_BUILDDIR)/$(PROG)_cpp $(PROG) +endif + +madevent_cuda_link: + $(MAKE) USEGTEST=0 USEBUILDDIR=1 BACKEND=cuda madevent_cudacpp_link madevent_hip_link: - $(MAKE) USEGTEST=0 BACKEND=hip $(CUDACPP_BUILDDIR)/$(PROG)_hip - rm -f $(PROG) - ln -s $(CUDACPP_BUILDDIR)/$(PROG)_hip $(PROG) + $(MAKE) USEGTEST=0 USEBUILDDIR=1 BACKEND=hip madevent_cudacpp_link madevent_cpp_link: - $(MAKE) USEGTEST=0 BACKEND=cppauto $(CUDACPP_BUILDDIR)/$(PROG)_cpp - rm -f $(PROG) - ln -s $(CUDACPP_BUILDDIR)/$(PROG)_cpp $(PROG) + $(MAKE) USEGTEST=0 USEBUILDDIR=1 BACKEND=cppauto madevent_cudacpp_link override SUPPORTED_AVXS = cppnone cppsse4 cppavx2 cpp512y cpp512z cppauto madevent_%_link: - @if [ '$(words $(filter $*, $(SUPPORTED_AVXS)))' != '1' ]; then echo "ERROR! Invalid target '$@' (supported madevent_cpp*_link targets are: $(foreach avx,$(SUPPORTED_AVXS),'madevent_cpp$(avx)_link'))"; exit 1; fi - $(MAKE) USEGTEST=0 BACKEND=$* $(CUDACPP_BUILDDIR)/$(PROG)_cpp - rm -f $(PROG) - ln -s $(CUDACPP_BUILDDIR)/$(PROG)_cpp $(PROG) + @if [ '$(words $(filter $*, $(SUPPORTED_AVXS)))' != '1' ]; then echo "ERROR! Invalid target '$@' (supported madevent_cpp*_link targets are: $(foreach avx,$(SUPPORTED_AVXS),'madevent_$(avx)_link'))"; exit 1; fi + $(MAKE) USEGTEST=0 USEBUILDDIR=1 BACKEND=$* madevent_cudacpp_link # Building $(PROG)_cpp no longer builds $(PROG)_cuda if CUDACPP_BACKENDLIB for cuda exists (this was the case in the past to allow cpp-only builds #503) $(CUDACPP_BUILDDIR)/$(PROG)_cpp: $(PROCESS) $(DSIG_cudacpp) auto_dsig.o $(LIBS) $(MATRIX) counters.o ompnumthreads.o $(CUDACPP_BUILDDIR)/.cudacpplibs diff --git a/epochX/cudacpp/gg_ttgg.mad/bin/internal/launch_plugin.py b/epochX/cudacpp/gg_ttgg.mad/bin/internal/launch_plugin.py index 0924927785..19d65fa195 100644 --- a/epochX/cudacpp/gg_ttgg.mad/bin/internal/launch_plugin.py +++ b/epochX/cudacpp/gg_ttgg.mad/bin/internal/launch_plugin.py @@ -33,10 +33,16 @@ def compile(self, *args, **opts): if 'cwd' in opts and os.path.basename(opts['cwd']) == 'Source': path = pjoin(opts['cwd'], 'make_opts') common_run_interface.CommonRunCmd.update_make_opts_full(path, - {'FPTYPE': self.run_card['floating_type'] }) - misc.sprint('FPTYPE checked') + {'FPTYPE': self.run_card['cudacpp_fptype'], + 'HELINL': self.run_card['cudacpp_helinl'], + 'HRDCOD': self.run_card['cudacpp_hrdcod'] }) + misc.sprint('FPTYPE, HELINL, HRDCOD checked') cudacpp_supported_backends = [ 'fortran', 'cuda', 'hip', 'cpp', 'cppnone', 'cppsse4', 'cppavx2', 'cpp512y', 'cpp512z', 'cppauto' ] if args and args[0][0] == 'madevent' and hasattr(self, 'run_card'): + if self.run_card['cudacpp_bldall'] == True: # pre-build all backends #945 + logger.info("Pre-building madevent in madevent_interface.py with ALL matrix elements") + args[0][0] = 'bldall' + misc.compile(nb_core=self.options['nb_core'], *args, **opts) cudacpp_backend = self.run_card['cudacpp_backend'].lower() # the default value is defined in launch_plugin.py logger.info("Building madevent in madevent_interface.py with '%s' matrix elements"%cudacpp_backend) if cudacpp_backend in cudacpp_supported_backends : @@ -47,14 +53,18 @@ def compile(self, *args, **opts): else: return misc.compile(nb_core=self.options['nb_core'], *args, **opts) -# Phase-Space Optimization ------------------------------------------------------------------------------------ +# CUDACPP runcard block ------------------------------------------------------------------------------------ template_on = \ """#*********************************************************************** # SIMD/GPU configuration for the CUDACPP plugin #************************************************************************ - %(cudacpp_backend)s = cudacpp_backend ! CUDACPP backend: fortran, cuda, hip, cpp, cppnone, cppsse4, cppavx2, cpp512y, cpp512z, cppauto + %(cudacpp_backend)s = cudacpp_backend ! CUDACPP backend: fortran, cuda, hip, cpp (DEFAULT), cppnone, cppsse4, cppavx2, cpp512y, cpp512z, cppauto +#*** WARNING! The following cudacpp runcards are experimental! Users should normally change only the cudacpp_backend card *** + %(cudacpp_fptype)s = cudacpp_fptype ! CUDACPP floating point precision: f (single), d (double), m (mixed, DEFAULT: double for amplitudes, single for colors) + %(cudacpp_hrdcod)s = cudacpp_hrdcod ! CUDACPP parameter hardcoding: 0 (DEFAULT, parameters not hardcoded: read param_card.dat at runtime), 1 (hardcoded parameters) + %(cudacpp_helinl)s = cudacpp_helinl ! CUDACPP helicity amplitude inlining: 0 (DEFAULT, ordinary inlining of templates), 1 (aggressive inlining with 'always inline') + %(cudacpp_bldall)s = cudacpp_bldall ! CUDACPP build all available backends in separate build directories: False, True """ - template_off = '' plugin_block = banner_mod.RunBlock('simd', template_on=template_on, template_off=template_off) @@ -75,8 +85,14 @@ def reset_simd(self, old_value, new_value, name): def reset_makeopts(self, old_value, new_value, name): if not hasattr(self, 'path'): raise Exception - if name == 'floating_type': + if name == 'cudacpp_fptype': common_run_interface.CommonRunCmd.update_make_opts_full({'FPTYPE': new_value}) + elif name == 'cudacpp_hrdcod': + raise Exception('Cannot change cudacpp_hrdcod') + elif name == 'cudacpp_helinl': + raise Exception('Cannot change cudacpp_helinl') + elif name == 'cudacpp_bldall': + raise Exception('Cannot change cudacpp_bldall') else: raise Exception Sourcedir = pjoin(os.path.dirname(os.path.dirname(self.path)), 'Source') @@ -84,14 +100,34 @@ def reset_makeopts(self, old_value, new_value, name): def default_setup(self): super().default_setup() - self.add_param('floating_type', 'm', include=False, hidden=True, - fct_mod=(self.reset_makeopts,(),{}), - allowed=['m','d','f'], - comment='floating point precision: f (single), d (double), m (mixed: double for amplitudes, single for colors)' - ) cudacpp_supported_backends = [ 'fortran', 'cuda', 'hip', 'cpp', 'cppnone', 'cppsse4', 'cppavx2', 'cpp512y', 'cpp512z', 'cppauto' ] - self.add_param('cudacpp_backend', 'cpp', include=False, hidden=False, + self.add_param('cudacpp_backend', 'cpp', + include=False, # AV: 'include=True' would add "CUDACPP_BACKEND = 'cpp'" to run_card.inc + hidden=False, # AV: keep cudacpp_backend in runcard template and keep 'hidden='False' allowed=cudacpp_supported_backends) + self.add_param('cudacpp_fptype', 'm', + include=False, # AV: 'include=True' would add "CUDACPP_FPTYPE = 'm'" to run_card.inc (if fct_mod is removed, else codegen fails) + hidden=False, # AV: add cudacpp_backend to runcard template and keep 'hidden='False' + fct_mod=(self.reset_makeopts,(),{}), # AV: I assume this forces a 'make cleanavx' if FPTYPE changes? + allowed=['m','d','f'] + ) + self.add_param('cudacpp_helinl', '0', + include=False, # AV: no need to add this parameter to run_card.inc + hidden=False, # AV: add cudacpp_helinl to runcard template and keep 'hidden='False' + fct_mod=(self.reset_makeopts,(),{}), # AV: I assume this raises an exception if cudacpp_helinl changes? + allowed=['0','1'] + ) + self.add_param('cudacpp_hrdcod', '0', + include=False, # AV: no need to add this parameter to run_card.inc + hidden=False, # AV: add cudacpp_hrdcod to runcard template and keep 'hidden='False' + fct_mod=(self.reset_makeopts,(),{}), # AV: I assume this raises an exception if cudacpp_hrdcod changes? + allowed=['0','1'] + ) + self.add_param('cudacpp_bldall', False, + include=False, # AV: no need to add this parameter to run_card.inc + hidden=False, # AV: add cudacpp_bldall to runcard template and keep 'hidden='False' + fct_mod=(self.reset_makeopts,(),{}), # AV: I assume this raises an exception if cudacpp_bldall changes? + ) self['vector_size'] = 16 # already setup in default class (just change value) self['aloha_flag'] = '--fast-math' self['matrix_flag'] = '-O3' diff --git a/epochX/cudacpp/gg_ttggg.mad/Cards/run_card.dat b/epochX/cudacpp/gg_ttggg.mad/Cards/run_card.dat index a08f93d92b..4cee4c235a 100644 --- a/epochX/cudacpp/gg_ttggg.mad/Cards/run_card.dat +++ b/epochX/cudacpp/gg_ttggg.mad/Cards/run_card.dat @@ -195,5 +195,10 @@ systematics = systematics_program ! none, systematics [python], SysCalc [deprece #*********************************************************************** # SIMD/GPU configuration for the CUDACPP plugin #************************************************************************ - cpp = cudacpp_backend ! CUDACPP backend: fortran, cuda, hip, cpp, cppnone, cppsse4, cppavx2, cpp512y, cpp512z, cppauto + cpp = cudacpp_backend ! CUDACPP backend: fortran, cuda, hip, cpp (DEFAULT), cppnone, cppsse4, cppavx2, cpp512y, cpp512z, cppauto +#*** WARNING! The following cudacpp runcards are experimental! Users should normally change only the cudacpp_backend card *** + m = cudacpp_fptype ! CUDACPP floating point precision: f (single), d (double), m (mixed, DEFAULT: double for amplitudes, single for colors) + 0 = cudacpp_hrdcod ! CUDACPP parameter hardcoding: 0 (DEFAULT, parameters not hardcoded: read param_card.dat at runtime), 1 (hardcoded parameters) + 0 = cudacpp_helinl ! CUDACPP helicity amplitude inlining: 0 (DEFAULT, ordinary inlining of templates), 1 (aggressive inlining with 'always inline') + False = cudacpp_bldall ! CUDACPP build all available backends in separate build directories: False, True diff --git a/epochX/cudacpp/gg_ttggg.mad/Cards/run_card_default.dat b/epochX/cudacpp/gg_ttggg.mad/Cards/run_card_default.dat index 48050a5fd7..a11218fa6e 100644 --- a/epochX/cudacpp/gg_ttggg.mad/Cards/run_card_default.dat +++ b/epochX/cudacpp/gg_ttggg.mad/Cards/run_card_default.dat @@ -195,5 +195,10 @@ systematics = systematics_program ! none, systematics [python], SysCalc [deprece #*********************************************************************** # SIMD/GPU configuration for the CUDACPP plugin #************************************************************************ - cpp = cudacpp_backend ! CUDACPP backend: fortran, cuda, hip, cpp, cppnone, cppsse4, cppavx2, cpp512y, cpp512z, cppauto + cpp = cudacpp_backend ! CUDACPP backend: fortran, cuda, hip, cpp (DEFAULT), cppnone, cppsse4, cppavx2, cpp512y, cpp512z, cppauto +#*** WARNING! The following cudacpp runcards are experimental! Users should normally change only the cudacpp_backend card *** + m = cudacpp_fptype ! CUDACPP floating point precision: f (single), d (double), m (mixed, DEFAULT: double for amplitudes, single for colors) + 0 = cudacpp_hrdcod ! CUDACPP parameter hardcoding: 0 (DEFAULT, parameters not hardcoded: read param_card.dat at runtime), 1 (hardcoded parameters) + 0 = cudacpp_helinl ! CUDACPP helicity amplitude inlining: 0 (DEFAULT, ordinary inlining of templates), 1 (aggressive inlining with 'always inline') + False = cudacpp_bldall ! CUDACPP build all available backends in separate build directories: False, True diff --git a/epochX/cudacpp/gg_ttggg.mad/SubProcesses/makefile b/epochX/cudacpp/gg_ttggg.mad/SubProcesses/makefile index 49e6800fff..2e5bff046b 100644 --- a/epochX/cudacpp/gg_ttggg.mad/SubProcesses/makefile +++ b/epochX/cudacpp/gg_ttggg.mad/SubProcesses/makefile @@ -2,6 +2,11 @@ SHELL := /bin/bash include ../../Source/make_opts +# Export the relevant settings from make_opts to builds using the cudacpp makefile (#700) +export FPTYPE +export HELINL +export HRDCOD + # Determine CUDACPP_BUILDDIR based on the user-defined choices of BACKEND, FPTYPE, HELINL, HRDCOD and USEBUILDDIR (#829) # Stop with an error if BACKEND=cuda and nvcc is missing or if BACKEND=hip and hipcc is missing include ../../src/cudacpp_config.mk @@ -162,27 +167,29 @@ madevent_fortran_link: $(PROG)_fortran rm -f $(PROG) ln -s $(PROG)_fortran $(PROG) -madevent_cuda_link: - $(MAKE) USEGTEST=0 BACKEND=cuda $(CUDACPP_BUILDDIR)/$(PROG)_cuda +madevent_cudacpp_link: all rm -f $(PROG) +ifeq ($(BACKEND),cuda) ln -s $(CUDACPP_BUILDDIR)/$(PROG)_cuda $(PROG) +else ifeq ($(BACKEND),hip) + ln -s $(CUDACPP_BUILDDIR)/$(PROG)_hip $(PROG) +else + ln -s $(CUDACPP_BUILDDIR)/$(PROG)_cpp $(PROG) +endif + +madevent_cuda_link: + $(MAKE) USEGTEST=0 USEBUILDDIR=1 BACKEND=cuda madevent_cudacpp_link madevent_hip_link: - $(MAKE) USEGTEST=0 BACKEND=hip $(CUDACPP_BUILDDIR)/$(PROG)_hip - rm -f $(PROG) - ln -s $(CUDACPP_BUILDDIR)/$(PROG)_hip $(PROG) + $(MAKE) USEGTEST=0 USEBUILDDIR=1 BACKEND=hip madevent_cudacpp_link madevent_cpp_link: - $(MAKE) USEGTEST=0 BACKEND=cppauto $(CUDACPP_BUILDDIR)/$(PROG)_cpp - rm -f $(PROG) - ln -s $(CUDACPP_BUILDDIR)/$(PROG)_cpp $(PROG) + $(MAKE) USEGTEST=0 USEBUILDDIR=1 BACKEND=cppauto madevent_cudacpp_link override SUPPORTED_AVXS = cppnone cppsse4 cppavx2 cpp512y cpp512z cppauto madevent_%_link: - @if [ '$(words $(filter $*, $(SUPPORTED_AVXS)))' != '1' ]; then echo "ERROR! Invalid target '$@' (supported madevent_cpp*_link targets are: $(foreach avx,$(SUPPORTED_AVXS),'madevent_cpp$(avx)_link'))"; exit 1; fi - $(MAKE) USEGTEST=0 BACKEND=$* $(CUDACPP_BUILDDIR)/$(PROG)_cpp - rm -f $(PROG) - ln -s $(CUDACPP_BUILDDIR)/$(PROG)_cpp $(PROG) + @if [ '$(words $(filter $*, $(SUPPORTED_AVXS)))' != '1' ]; then echo "ERROR! Invalid target '$@' (supported madevent_cpp*_link targets are: $(foreach avx,$(SUPPORTED_AVXS),'madevent_$(avx)_link'))"; exit 1; fi + $(MAKE) USEGTEST=0 USEBUILDDIR=1 BACKEND=$* madevent_cudacpp_link # Building $(PROG)_cpp no longer builds $(PROG)_cuda if CUDACPP_BACKENDLIB for cuda exists (this was the case in the past to allow cpp-only builds #503) $(CUDACPP_BUILDDIR)/$(PROG)_cpp: $(PROCESS) $(DSIG_cudacpp) auto_dsig.o $(LIBS) $(MATRIX) counters.o ompnumthreads.o $(CUDACPP_BUILDDIR)/.cudacpplibs diff --git a/epochX/cudacpp/gg_ttggg.mad/bin/internal/launch_plugin.py b/epochX/cudacpp/gg_ttggg.mad/bin/internal/launch_plugin.py index 0924927785..19d65fa195 100644 --- a/epochX/cudacpp/gg_ttggg.mad/bin/internal/launch_plugin.py +++ b/epochX/cudacpp/gg_ttggg.mad/bin/internal/launch_plugin.py @@ -33,10 +33,16 @@ def compile(self, *args, **opts): if 'cwd' in opts and os.path.basename(opts['cwd']) == 'Source': path = pjoin(opts['cwd'], 'make_opts') common_run_interface.CommonRunCmd.update_make_opts_full(path, - {'FPTYPE': self.run_card['floating_type'] }) - misc.sprint('FPTYPE checked') + {'FPTYPE': self.run_card['cudacpp_fptype'], + 'HELINL': self.run_card['cudacpp_helinl'], + 'HRDCOD': self.run_card['cudacpp_hrdcod'] }) + misc.sprint('FPTYPE, HELINL, HRDCOD checked') cudacpp_supported_backends = [ 'fortran', 'cuda', 'hip', 'cpp', 'cppnone', 'cppsse4', 'cppavx2', 'cpp512y', 'cpp512z', 'cppauto' ] if args and args[0][0] == 'madevent' and hasattr(self, 'run_card'): + if self.run_card['cudacpp_bldall'] == True: # pre-build all backends #945 + logger.info("Pre-building madevent in madevent_interface.py with ALL matrix elements") + args[0][0] = 'bldall' + misc.compile(nb_core=self.options['nb_core'], *args, **opts) cudacpp_backend = self.run_card['cudacpp_backend'].lower() # the default value is defined in launch_plugin.py logger.info("Building madevent in madevent_interface.py with '%s' matrix elements"%cudacpp_backend) if cudacpp_backend in cudacpp_supported_backends : @@ -47,14 +53,18 @@ def compile(self, *args, **opts): else: return misc.compile(nb_core=self.options['nb_core'], *args, **opts) -# Phase-Space Optimization ------------------------------------------------------------------------------------ +# CUDACPP runcard block ------------------------------------------------------------------------------------ template_on = \ """#*********************************************************************** # SIMD/GPU configuration for the CUDACPP plugin #************************************************************************ - %(cudacpp_backend)s = cudacpp_backend ! CUDACPP backend: fortran, cuda, hip, cpp, cppnone, cppsse4, cppavx2, cpp512y, cpp512z, cppauto + %(cudacpp_backend)s = cudacpp_backend ! CUDACPP backend: fortran, cuda, hip, cpp (DEFAULT), cppnone, cppsse4, cppavx2, cpp512y, cpp512z, cppauto +#*** WARNING! The following cudacpp runcards are experimental! Users should normally change only the cudacpp_backend card *** + %(cudacpp_fptype)s = cudacpp_fptype ! CUDACPP floating point precision: f (single), d (double), m (mixed, DEFAULT: double for amplitudes, single for colors) + %(cudacpp_hrdcod)s = cudacpp_hrdcod ! CUDACPP parameter hardcoding: 0 (DEFAULT, parameters not hardcoded: read param_card.dat at runtime), 1 (hardcoded parameters) + %(cudacpp_helinl)s = cudacpp_helinl ! CUDACPP helicity amplitude inlining: 0 (DEFAULT, ordinary inlining of templates), 1 (aggressive inlining with 'always inline') + %(cudacpp_bldall)s = cudacpp_bldall ! CUDACPP build all available backends in separate build directories: False, True """ - template_off = '' plugin_block = banner_mod.RunBlock('simd', template_on=template_on, template_off=template_off) @@ -75,8 +85,14 @@ def reset_simd(self, old_value, new_value, name): def reset_makeopts(self, old_value, new_value, name): if not hasattr(self, 'path'): raise Exception - if name == 'floating_type': + if name == 'cudacpp_fptype': common_run_interface.CommonRunCmd.update_make_opts_full({'FPTYPE': new_value}) + elif name == 'cudacpp_hrdcod': + raise Exception('Cannot change cudacpp_hrdcod') + elif name == 'cudacpp_helinl': + raise Exception('Cannot change cudacpp_helinl') + elif name == 'cudacpp_bldall': + raise Exception('Cannot change cudacpp_bldall') else: raise Exception Sourcedir = pjoin(os.path.dirname(os.path.dirname(self.path)), 'Source') @@ -84,14 +100,34 @@ def reset_makeopts(self, old_value, new_value, name): def default_setup(self): super().default_setup() - self.add_param('floating_type', 'm', include=False, hidden=True, - fct_mod=(self.reset_makeopts,(),{}), - allowed=['m','d','f'], - comment='floating point precision: f (single), d (double), m (mixed: double for amplitudes, single for colors)' - ) cudacpp_supported_backends = [ 'fortran', 'cuda', 'hip', 'cpp', 'cppnone', 'cppsse4', 'cppavx2', 'cpp512y', 'cpp512z', 'cppauto' ] - self.add_param('cudacpp_backend', 'cpp', include=False, hidden=False, + self.add_param('cudacpp_backend', 'cpp', + include=False, # AV: 'include=True' would add "CUDACPP_BACKEND = 'cpp'" to run_card.inc + hidden=False, # AV: keep cudacpp_backend in runcard template and keep 'hidden='False' allowed=cudacpp_supported_backends) + self.add_param('cudacpp_fptype', 'm', + include=False, # AV: 'include=True' would add "CUDACPP_FPTYPE = 'm'" to run_card.inc (if fct_mod is removed, else codegen fails) + hidden=False, # AV: add cudacpp_backend to runcard template and keep 'hidden='False' + fct_mod=(self.reset_makeopts,(),{}), # AV: I assume this forces a 'make cleanavx' if FPTYPE changes? + allowed=['m','d','f'] + ) + self.add_param('cudacpp_helinl', '0', + include=False, # AV: no need to add this parameter to run_card.inc + hidden=False, # AV: add cudacpp_helinl to runcard template and keep 'hidden='False' + fct_mod=(self.reset_makeopts,(),{}), # AV: I assume this raises an exception if cudacpp_helinl changes? + allowed=['0','1'] + ) + self.add_param('cudacpp_hrdcod', '0', + include=False, # AV: no need to add this parameter to run_card.inc + hidden=False, # AV: add cudacpp_hrdcod to runcard template and keep 'hidden='False' + fct_mod=(self.reset_makeopts,(),{}), # AV: I assume this raises an exception if cudacpp_hrdcod changes? + allowed=['0','1'] + ) + self.add_param('cudacpp_bldall', False, + include=False, # AV: no need to add this parameter to run_card.inc + hidden=False, # AV: add cudacpp_bldall to runcard template and keep 'hidden='False' + fct_mod=(self.reset_makeopts,(),{}), # AV: I assume this raises an exception if cudacpp_bldall changes? + ) self['vector_size'] = 16 # already setup in default class (just change value) self['aloha_flag'] = '--fast-math' self['matrix_flag'] = '-O3' diff --git a/epochX/cudacpp/gq_ttq.mad/Cards/run_card.dat b/epochX/cudacpp/gq_ttq.mad/Cards/run_card.dat index 66a805e521..97e585f71e 100644 --- a/epochX/cudacpp/gq_ttq.mad/Cards/run_card.dat +++ b/epochX/cudacpp/gq_ttq.mad/Cards/run_card.dat @@ -166,5 +166,10 @@ systematics = systematics_program ! none, systematics [python], SysCalc [deprece #*********************************************************************** # SIMD/GPU configuration for the CUDACPP plugin #************************************************************************ - cpp = cudacpp_backend ! CUDACPP backend: fortran, cuda, hip, cpp, cppnone, cppsse4, cppavx2, cpp512y, cpp512z, cppauto + cpp = cudacpp_backend ! CUDACPP backend: fortran, cuda, hip, cpp (DEFAULT), cppnone, cppsse4, cppavx2, cpp512y, cpp512z, cppauto +#*** WARNING! The following cudacpp runcards are experimental! Users should normally change only the cudacpp_backend card *** + m = cudacpp_fptype ! CUDACPP floating point precision: f (single), d (double), m (mixed, DEFAULT: double for amplitudes, single for colors) + 0 = cudacpp_hrdcod ! CUDACPP parameter hardcoding: 0 (DEFAULT, parameters not hardcoded: read param_card.dat at runtime), 1 (hardcoded parameters) + 0 = cudacpp_helinl ! CUDACPP helicity amplitude inlining: 0 (DEFAULT, ordinary inlining of templates), 1 (aggressive inlining with 'always inline') + False = cudacpp_bldall ! CUDACPP build all available backends in separate build directories: False, True diff --git a/epochX/cudacpp/gq_ttq.mad/Cards/run_card_default.dat b/epochX/cudacpp/gq_ttq.mad/Cards/run_card_default.dat index 8c0f1e2199..214c16aefe 100644 --- a/epochX/cudacpp/gq_ttq.mad/Cards/run_card_default.dat +++ b/epochX/cudacpp/gq_ttq.mad/Cards/run_card_default.dat @@ -166,5 +166,10 @@ systematics = systematics_program ! none, systematics [python], SysCalc [deprece #*********************************************************************** # SIMD/GPU configuration for the CUDACPP plugin #************************************************************************ - cpp = cudacpp_backend ! CUDACPP backend: fortran, cuda, hip, cpp, cppnone, cppsse4, cppavx2, cpp512y, cpp512z, cppauto + cpp = cudacpp_backend ! CUDACPP backend: fortran, cuda, hip, cpp (DEFAULT), cppnone, cppsse4, cppavx2, cpp512y, cpp512z, cppauto +#*** WARNING! The following cudacpp runcards are experimental! Users should normally change only the cudacpp_backend card *** + m = cudacpp_fptype ! CUDACPP floating point precision: f (single), d (double), m (mixed, DEFAULT: double for amplitudes, single for colors) + 0 = cudacpp_hrdcod ! CUDACPP parameter hardcoding: 0 (DEFAULT, parameters not hardcoded: read param_card.dat at runtime), 1 (hardcoded parameters) + 0 = cudacpp_helinl ! CUDACPP helicity amplitude inlining: 0 (DEFAULT, ordinary inlining of templates), 1 (aggressive inlining with 'always inline') + False = cudacpp_bldall ! CUDACPP build all available backends in separate build directories: False, True diff --git a/epochX/cudacpp/gq_ttq.mad/SubProcesses/makefile b/epochX/cudacpp/gq_ttq.mad/SubProcesses/makefile index 49e6800fff..2e5bff046b 100644 --- a/epochX/cudacpp/gq_ttq.mad/SubProcesses/makefile +++ b/epochX/cudacpp/gq_ttq.mad/SubProcesses/makefile @@ -2,6 +2,11 @@ SHELL := /bin/bash include ../../Source/make_opts +# Export the relevant settings from make_opts to builds using the cudacpp makefile (#700) +export FPTYPE +export HELINL +export HRDCOD + # Determine CUDACPP_BUILDDIR based on the user-defined choices of BACKEND, FPTYPE, HELINL, HRDCOD and USEBUILDDIR (#829) # Stop with an error if BACKEND=cuda and nvcc is missing or if BACKEND=hip and hipcc is missing include ../../src/cudacpp_config.mk @@ -162,27 +167,29 @@ madevent_fortran_link: $(PROG)_fortran rm -f $(PROG) ln -s $(PROG)_fortran $(PROG) -madevent_cuda_link: - $(MAKE) USEGTEST=0 BACKEND=cuda $(CUDACPP_BUILDDIR)/$(PROG)_cuda +madevent_cudacpp_link: all rm -f $(PROG) +ifeq ($(BACKEND),cuda) ln -s $(CUDACPP_BUILDDIR)/$(PROG)_cuda $(PROG) +else ifeq ($(BACKEND),hip) + ln -s $(CUDACPP_BUILDDIR)/$(PROG)_hip $(PROG) +else + ln -s $(CUDACPP_BUILDDIR)/$(PROG)_cpp $(PROG) +endif + +madevent_cuda_link: + $(MAKE) USEGTEST=0 USEBUILDDIR=1 BACKEND=cuda madevent_cudacpp_link madevent_hip_link: - $(MAKE) USEGTEST=0 BACKEND=hip $(CUDACPP_BUILDDIR)/$(PROG)_hip - rm -f $(PROG) - ln -s $(CUDACPP_BUILDDIR)/$(PROG)_hip $(PROG) + $(MAKE) USEGTEST=0 USEBUILDDIR=1 BACKEND=hip madevent_cudacpp_link madevent_cpp_link: - $(MAKE) USEGTEST=0 BACKEND=cppauto $(CUDACPP_BUILDDIR)/$(PROG)_cpp - rm -f $(PROG) - ln -s $(CUDACPP_BUILDDIR)/$(PROG)_cpp $(PROG) + $(MAKE) USEGTEST=0 USEBUILDDIR=1 BACKEND=cppauto madevent_cudacpp_link override SUPPORTED_AVXS = cppnone cppsse4 cppavx2 cpp512y cpp512z cppauto madevent_%_link: - @if [ '$(words $(filter $*, $(SUPPORTED_AVXS)))' != '1' ]; then echo "ERROR! Invalid target '$@' (supported madevent_cpp*_link targets are: $(foreach avx,$(SUPPORTED_AVXS),'madevent_cpp$(avx)_link'))"; exit 1; fi - $(MAKE) USEGTEST=0 BACKEND=$* $(CUDACPP_BUILDDIR)/$(PROG)_cpp - rm -f $(PROG) - ln -s $(CUDACPP_BUILDDIR)/$(PROG)_cpp $(PROG) + @if [ '$(words $(filter $*, $(SUPPORTED_AVXS)))' != '1' ]; then echo "ERROR! Invalid target '$@' (supported madevent_cpp*_link targets are: $(foreach avx,$(SUPPORTED_AVXS),'madevent_$(avx)_link'))"; exit 1; fi + $(MAKE) USEGTEST=0 USEBUILDDIR=1 BACKEND=$* madevent_cudacpp_link # Building $(PROG)_cpp no longer builds $(PROG)_cuda if CUDACPP_BACKENDLIB for cuda exists (this was the case in the past to allow cpp-only builds #503) $(CUDACPP_BUILDDIR)/$(PROG)_cpp: $(PROCESS) $(DSIG_cudacpp) auto_dsig.o $(LIBS) $(MATRIX) counters.o ompnumthreads.o $(CUDACPP_BUILDDIR)/.cudacpplibs diff --git a/epochX/cudacpp/gq_ttq.mad/bin/internal/launch_plugin.py b/epochX/cudacpp/gq_ttq.mad/bin/internal/launch_plugin.py index 0924927785..19d65fa195 100644 --- a/epochX/cudacpp/gq_ttq.mad/bin/internal/launch_plugin.py +++ b/epochX/cudacpp/gq_ttq.mad/bin/internal/launch_plugin.py @@ -33,10 +33,16 @@ def compile(self, *args, **opts): if 'cwd' in opts and os.path.basename(opts['cwd']) == 'Source': path = pjoin(opts['cwd'], 'make_opts') common_run_interface.CommonRunCmd.update_make_opts_full(path, - {'FPTYPE': self.run_card['floating_type'] }) - misc.sprint('FPTYPE checked') + {'FPTYPE': self.run_card['cudacpp_fptype'], + 'HELINL': self.run_card['cudacpp_helinl'], + 'HRDCOD': self.run_card['cudacpp_hrdcod'] }) + misc.sprint('FPTYPE, HELINL, HRDCOD checked') cudacpp_supported_backends = [ 'fortran', 'cuda', 'hip', 'cpp', 'cppnone', 'cppsse4', 'cppavx2', 'cpp512y', 'cpp512z', 'cppauto' ] if args and args[0][0] == 'madevent' and hasattr(self, 'run_card'): + if self.run_card['cudacpp_bldall'] == True: # pre-build all backends #945 + logger.info("Pre-building madevent in madevent_interface.py with ALL matrix elements") + args[0][0] = 'bldall' + misc.compile(nb_core=self.options['nb_core'], *args, **opts) cudacpp_backend = self.run_card['cudacpp_backend'].lower() # the default value is defined in launch_plugin.py logger.info("Building madevent in madevent_interface.py with '%s' matrix elements"%cudacpp_backend) if cudacpp_backend in cudacpp_supported_backends : @@ -47,14 +53,18 @@ def compile(self, *args, **opts): else: return misc.compile(nb_core=self.options['nb_core'], *args, **opts) -# Phase-Space Optimization ------------------------------------------------------------------------------------ +# CUDACPP runcard block ------------------------------------------------------------------------------------ template_on = \ """#*********************************************************************** # SIMD/GPU configuration for the CUDACPP plugin #************************************************************************ - %(cudacpp_backend)s = cudacpp_backend ! CUDACPP backend: fortran, cuda, hip, cpp, cppnone, cppsse4, cppavx2, cpp512y, cpp512z, cppauto + %(cudacpp_backend)s = cudacpp_backend ! CUDACPP backend: fortran, cuda, hip, cpp (DEFAULT), cppnone, cppsse4, cppavx2, cpp512y, cpp512z, cppauto +#*** WARNING! The following cudacpp runcards are experimental! Users should normally change only the cudacpp_backend card *** + %(cudacpp_fptype)s = cudacpp_fptype ! CUDACPP floating point precision: f (single), d (double), m (mixed, DEFAULT: double for amplitudes, single for colors) + %(cudacpp_hrdcod)s = cudacpp_hrdcod ! CUDACPP parameter hardcoding: 0 (DEFAULT, parameters not hardcoded: read param_card.dat at runtime), 1 (hardcoded parameters) + %(cudacpp_helinl)s = cudacpp_helinl ! CUDACPP helicity amplitude inlining: 0 (DEFAULT, ordinary inlining of templates), 1 (aggressive inlining with 'always inline') + %(cudacpp_bldall)s = cudacpp_bldall ! CUDACPP build all available backends in separate build directories: False, True """ - template_off = '' plugin_block = banner_mod.RunBlock('simd', template_on=template_on, template_off=template_off) @@ -75,8 +85,14 @@ def reset_simd(self, old_value, new_value, name): def reset_makeopts(self, old_value, new_value, name): if not hasattr(self, 'path'): raise Exception - if name == 'floating_type': + if name == 'cudacpp_fptype': common_run_interface.CommonRunCmd.update_make_opts_full({'FPTYPE': new_value}) + elif name == 'cudacpp_hrdcod': + raise Exception('Cannot change cudacpp_hrdcod') + elif name == 'cudacpp_helinl': + raise Exception('Cannot change cudacpp_helinl') + elif name == 'cudacpp_bldall': + raise Exception('Cannot change cudacpp_bldall') else: raise Exception Sourcedir = pjoin(os.path.dirname(os.path.dirname(self.path)), 'Source') @@ -84,14 +100,34 @@ def reset_makeopts(self, old_value, new_value, name): def default_setup(self): super().default_setup() - self.add_param('floating_type', 'm', include=False, hidden=True, - fct_mod=(self.reset_makeopts,(),{}), - allowed=['m','d','f'], - comment='floating point precision: f (single), d (double), m (mixed: double for amplitudes, single for colors)' - ) cudacpp_supported_backends = [ 'fortran', 'cuda', 'hip', 'cpp', 'cppnone', 'cppsse4', 'cppavx2', 'cpp512y', 'cpp512z', 'cppauto' ] - self.add_param('cudacpp_backend', 'cpp', include=False, hidden=False, + self.add_param('cudacpp_backend', 'cpp', + include=False, # AV: 'include=True' would add "CUDACPP_BACKEND = 'cpp'" to run_card.inc + hidden=False, # AV: keep cudacpp_backend in runcard template and keep 'hidden='False' allowed=cudacpp_supported_backends) + self.add_param('cudacpp_fptype', 'm', + include=False, # AV: 'include=True' would add "CUDACPP_FPTYPE = 'm'" to run_card.inc (if fct_mod is removed, else codegen fails) + hidden=False, # AV: add cudacpp_backend to runcard template and keep 'hidden='False' + fct_mod=(self.reset_makeopts,(),{}), # AV: I assume this forces a 'make cleanavx' if FPTYPE changes? + allowed=['m','d','f'] + ) + self.add_param('cudacpp_helinl', '0', + include=False, # AV: no need to add this parameter to run_card.inc + hidden=False, # AV: add cudacpp_helinl to runcard template and keep 'hidden='False' + fct_mod=(self.reset_makeopts,(),{}), # AV: I assume this raises an exception if cudacpp_helinl changes? + allowed=['0','1'] + ) + self.add_param('cudacpp_hrdcod', '0', + include=False, # AV: no need to add this parameter to run_card.inc + hidden=False, # AV: add cudacpp_hrdcod to runcard template and keep 'hidden='False' + fct_mod=(self.reset_makeopts,(),{}), # AV: I assume this raises an exception if cudacpp_hrdcod changes? + allowed=['0','1'] + ) + self.add_param('cudacpp_bldall', False, + include=False, # AV: no need to add this parameter to run_card.inc + hidden=False, # AV: add cudacpp_bldall to runcard template and keep 'hidden='False' + fct_mod=(self.reset_makeopts,(),{}), # AV: I assume this raises an exception if cudacpp_bldall changes? + ) self['vector_size'] = 16 # already setup in default class (just change value) self['aloha_flag'] = '--fast-math' self['matrix_flag'] = '-O3' diff --git a/epochX/cudacpp/heft_gg_bb.mad/Cards/run_card.dat b/epochX/cudacpp/heft_gg_bb.mad/Cards/run_card.dat index 8af20dc4e4..0845611cf8 100644 --- a/epochX/cudacpp/heft_gg_bb.mad/Cards/run_card.dat +++ b/epochX/cudacpp/heft_gg_bb.mad/Cards/run_card.dat @@ -174,5 +174,10 @@ systematics = systematics_program ! none, systematics [python], SysCalc [deprece #*********************************************************************** # SIMD/GPU configuration for the CUDACPP plugin #************************************************************************ - cpp = cudacpp_backend ! CUDACPP backend: fortran, cuda, hip, cpp, cppnone, cppsse4, cppavx2, cpp512y, cpp512z, cppauto + cpp = cudacpp_backend ! CUDACPP backend: fortran, cuda, hip, cpp (DEFAULT), cppnone, cppsse4, cppavx2, cpp512y, cpp512z, cppauto +#*** WARNING! The following cudacpp runcards are experimental! Users should normally change only the cudacpp_backend card *** + m = cudacpp_fptype ! CUDACPP floating point precision: f (single), d (double), m (mixed, DEFAULT: double for amplitudes, single for colors) + 0 = cudacpp_hrdcod ! CUDACPP parameter hardcoding: 0 (DEFAULT, parameters not hardcoded: read param_card.dat at runtime), 1 (hardcoded parameters) + 0 = cudacpp_helinl ! CUDACPP helicity amplitude inlining: 0 (DEFAULT, ordinary inlining of templates), 1 (aggressive inlining with 'always inline') + False = cudacpp_bldall ! CUDACPP build all available backends in separate build directories: False, True diff --git a/epochX/cudacpp/heft_gg_bb.mad/Cards/run_card_default.dat b/epochX/cudacpp/heft_gg_bb.mad/Cards/run_card_default.dat index 0815703ee4..16e591b1ad 100644 --- a/epochX/cudacpp/heft_gg_bb.mad/Cards/run_card_default.dat +++ b/epochX/cudacpp/heft_gg_bb.mad/Cards/run_card_default.dat @@ -174,5 +174,10 @@ systematics = systematics_program ! none, systematics [python], SysCalc [deprece #*********************************************************************** # SIMD/GPU configuration for the CUDACPP plugin #************************************************************************ - cpp = cudacpp_backend ! CUDACPP backend: fortran, cuda, hip, cpp, cppnone, cppsse4, cppavx2, cpp512y, cpp512z, cppauto + cpp = cudacpp_backend ! CUDACPP backend: fortran, cuda, hip, cpp (DEFAULT), cppnone, cppsse4, cppavx2, cpp512y, cpp512z, cppauto +#*** WARNING! The following cudacpp runcards are experimental! Users should normally change only the cudacpp_backend card *** + m = cudacpp_fptype ! CUDACPP floating point precision: f (single), d (double), m (mixed, DEFAULT: double for amplitudes, single for colors) + 0 = cudacpp_hrdcod ! CUDACPP parameter hardcoding: 0 (DEFAULT, parameters not hardcoded: read param_card.dat at runtime), 1 (hardcoded parameters) + 0 = cudacpp_helinl ! CUDACPP helicity amplitude inlining: 0 (DEFAULT, ordinary inlining of templates), 1 (aggressive inlining with 'always inline') + False = cudacpp_bldall ! CUDACPP build all available backends in separate build directories: False, True diff --git a/epochX/cudacpp/heft_gg_bb.mad/SubProcesses/makefile b/epochX/cudacpp/heft_gg_bb.mad/SubProcesses/makefile index 49e6800fff..2e5bff046b 100644 --- a/epochX/cudacpp/heft_gg_bb.mad/SubProcesses/makefile +++ b/epochX/cudacpp/heft_gg_bb.mad/SubProcesses/makefile @@ -2,6 +2,11 @@ SHELL := /bin/bash include ../../Source/make_opts +# Export the relevant settings from make_opts to builds using the cudacpp makefile (#700) +export FPTYPE +export HELINL +export HRDCOD + # Determine CUDACPP_BUILDDIR based on the user-defined choices of BACKEND, FPTYPE, HELINL, HRDCOD and USEBUILDDIR (#829) # Stop with an error if BACKEND=cuda and nvcc is missing or if BACKEND=hip and hipcc is missing include ../../src/cudacpp_config.mk @@ -162,27 +167,29 @@ madevent_fortran_link: $(PROG)_fortran rm -f $(PROG) ln -s $(PROG)_fortran $(PROG) -madevent_cuda_link: - $(MAKE) USEGTEST=0 BACKEND=cuda $(CUDACPP_BUILDDIR)/$(PROG)_cuda +madevent_cudacpp_link: all rm -f $(PROG) +ifeq ($(BACKEND),cuda) ln -s $(CUDACPP_BUILDDIR)/$(PROG)_cuda $(PROG) +else ifeq ($(BACKEND),hip) + ln -s $(CUDACPP_BUILDDIR)/$(PROG)_hip $(PROG) +else + ln -s $(CUDACPP_BUILDDIR)/$(PROG)_cpp $(PROG) +endif + +madevent_cuda_link: + $(MAKE) USEGTEST=0 USEBUILDDIR=1 BACKEND=cuda madevent_cudacpp_link madevent_hip_link: - $(MAKE) USEGTEST=0 BACKEND=hip $(CUDACPP_BUILDDIR)/$(PROG)_hip - rm -f $(PROG) - ln -s $(CUDACPP_BUILDDIR)/$(PROG)_hip $(PROG) + $(MAKE) USEGTEST=0 USEBUILDDIR=1 BACKEND=hip madevent_cudacpp_link madevent_cpp_link: - $(MAKE) USEGTEST=0 BACKEND=cppauto $(CUDACPP_BUILDDIR)/$(PROG)_cpp - rm -f $(PROG) - ln -s $(CUDACPP_BUILDDIR)/$(PROG)_cpp $(PROG) + $(MAKE) USEGTEST=0 USEBUILDDIR=1 BACKEND=cppauto madevent_cudacpp_link override SUPPORTED_AVXS = cppnone cppsse4 cppavx2 cpp512y cpp512z cppauto madevent_%_link: - @if [ '$(words $(filter $*, $(SUPPORTED_AVXS)))' != '1' ]; then echo "ERROR! Invalid target '$@' (supported madevent_cpp*_link targets are: $(foreach avx,$(SUPPORTED_AVXS),'madevent_cpp$(avx)_link'))"; exit 1; fi - $(MAKE) USEGTEST=0 BACKEND=$* $(CUDACPP_BUILDDIR)/$(PROG)_cpp - rm -f $(PROG) - ln -s $(CUDACPP_BUILDDIR)/$(PROG)_cpp $(PROG) + @if [ '$(words $(filter $*, $(SUPPORTED_AVXS)))' != '1' ]; then echo "ERROR! Invalid target '$@' (supported madevent_cpp*_link targets are: $(foreach avx,$(SUPPORTED_AVXS),'madevent_$(avx)_link'))"; exit 1; fi + $(MAKE) USEGTEST=0 USEBUILDDIR=1 BACKEND=$* madevent_cudacpp_link # Building $(PROG)_cpp no longer builds $(PROG)_cuda if CUDACPP_BACKENDLIB for cuda exists (this was the case in the past to allow cpp-only builds #503) $(CUDACPP_BUILDDIR)/$(PROG)_cpp: $(PROCESS) $(DSIG_cudacpp) auto_dsig.o $(LIBS) $(MATRIX) counters.o ompnumthreads.o $(CUDACPP_BUILDDIR)/.cudacpplibs diff --git a/epochX/cudacpp/heft_gg_bb.mad/bin/internal/launch_plugin.py b/epochX/cudacpp/heft_gg_bb.mad/bin/internal/launch_plugin.py index 0924927785..19d65fa195 100644 --- a/epochX/cudacpp/heft_gg_bb.mad/bin/internal/launch_plugin.py +++ b/epochX/cudacpp/heft_gg_bb.mad/bin/internal/launch_plugin.py @@ -33,10 +33,16 @@ def compile(self, *args, **opts): if 'cwd' in opts and os.path.basename(opts['cwd']) == 'Source': path = pjoin(opts['cwd'], 'make_opts') common_run_interface.CommonRunCmd.update_make_opts_full(path, - {'FPTYPE': self.run_card['floating_type'] }) - misc.sprint('FPTYPE checked') + {'FPTYPE': self.run_card['cudacpp_fptype'], + 'HELINL': self.run_card['cudacpp_helinl'], + 'HRDCOD': self.run_card['cudacpp_hrdcod'] }) + misc.sprint('FPTYPE, HELINL, HRDCOD checked') cudacpp_supported_backends = [ 'fortran', 'cuda', 'hip', 'cpp', 'cppnone', 'cppsse4', 'cppavx2', 'cpp512y', 'cpp512z', 'cppauto' ] if args and args[0][0] == 'madevent' and hasattr(self, 'run_card'): + if self.run_card['cudacpp_bldall'] == True: # pre-build all backends #945 + logger.info("Pre-building madevent in madevent_interface.py with ALL matrix elements") + args[0][0] = 'bldall' + misc.compile(nb_core=self.options['nb_core'], *args, **opts) cudacpp_backend = self.run_card['cudacpp_backend'].lower() # the default value is defined in launch_plugin.py logger.info("Building madevent in madevent_interface.py with '%s' matrix elements"%cudacpp_backend) if cudacpp_backend in cudacpp_supported_backends : @@ -47,14 +53,18 @@ def compile(self, *args, **opts): else: return misc.compile(nb_core=self.options['nb_core'], *args, **opts) -# Phase-Space Optimization ------------------------------------------------------------------------------------ +# CUDACPP runcard block ------------------------------------------------------------------------------------ template_on = \ """#*********************************************************************** # SIMD/GPU configuration for the CUDACPP plugin #************************************************************************ - %(cudacpp_backend)s = cudacpp_backend ! CUDACPP backend: fortran, cuda, hip, cpp, cppnone, cppsse4, cppavx2, cpp512y, cpp512z, cppauto + %(cudacpp_backend)s = cudacpp_backend ! CUDACPP backend: fortran, cuda, hip, cpp (DEFAULT), cppnone, cppsse4, cppavx2, cpp512y, cpp512z, cppauto +#*** WARNING! The following cudacpp runcards are experimental! Users should normally change only the cudacpp_backend card *** + %(cudacpp_fptype)s = cudacpp_fptype ! CUDACPP floating point precision: f (single), d (double), m (mixed, DEFAULT: double for amplitudes, single for colors) + %(cudacpp_hrdcod)s = cudacpp_hrdcod ! CUDACPP parameter hardcoding: 0 (DEFAULT, parameters not hardcoded: read param_card.dat at runtime), 1 (hardcoded parameters) + %(cudacpp_helinl)s = cudacpp_helinl ! CUDACPP helicity amplitude inlining: 0 (DEFAULT, ordinary inlining of templates), 1 (aggressive inlining with 'always inline') + %(cudacpp_bldall)s = cudacpp_bldall ! CUDACPP build all available backends in separate build directories: False, True """ - template_off = '' plugin_block = banner_mod.RunBlock('simd', template_on=template_on, template_off=template_off) @@ -75,8 +85,14 @@ def reset_simd(self, old_value, new_value, name): def reset_makeopts(self, old_value, new_value, name): if not hasattr(self, 'path'): raise Exception - if name == 'floating_type': + if name == 'cudacpp_fptype': common_run_interface.CommonRunCmd.update_make_opts_full({'FPTYPE': new_value}) + elif name == 'cudacpp_hrdcod': + raise Exception('Cannot change cudacpp_hrdcod') + elif name == 'cudacpp_helinl': + raise Exception('Cannot change cudacpp_helinl') + elif name == 'cudacpp_bldall': + raise Exception('Cannot change cudacpp_bldall') else: raise Exception Sourcedir = pjoin(os.path.dirname(os.path.dirname(self.path)), 'Source') @@ -84,14 +100,34 @@ def reset_makeopts(self, old_value, new_value, name): def default_setup(self): super().default_setup() - self.add_param('floating_type', 'm', include=False, hidden=True, - fct_mod=(self.reset_makeopts,(),{}), - allowed=['m','d','f'], - comment='floating point precision: f (single), d (double), m (mixed: double for amplitudes, single for colors)' - ) cudacpp_supported_backends = [ 'fortran', 'cuda', 'hip', 'cpp', 'cppnone', 'cppsse4', 'cppavx2', 'cpp512y', 'cpp512z', 'cppauto' ] - self.add_param('cudacpp_backend', 'cpp', include=False, hidden=False, + self.add_param('cudacpp_backend', 'cpp', + include=False, # AV: 'include=True' would add "CUDACPP_BACKEND = 'cpp'" to run_card.inc + hidden=False, # AV: keep cudacpp_backend in runcard template and keep 'hidden='False' allowed=cudacpp_supported_backends) + self.add_param('cudacpp_fptype', 'm', + include=False, # AV: 'include=True' would add "CUDACPP_FPTYPE = 'm'" to run_card.inc (if fct_mod is removed, else codegen fails) + hidden=False, # AV: add cudacpp_backend to runcard template and keep 'hidden='False' + fct_mod=(self.reset_makeopts,(),{}), # AV: I assume this forces a 'make cleanavx' if FPTYPE changes? + allowed=['m','d','f'] + ) + self.add_param('cudacpp_helinl', '0', + include=False, # AV: no need to add this parameter to run_card.inc + hidden=False, # AV: add cudacpp_helinl to runcard template and keep 'hidden='False' + fct_mod=(self.reset_makeopts,(),{}), # AV: I assume this raises an exception if cudacpp_helinl changes? + allowed=['0','1'] + ) + self.add_param('cudacpp_hrdcod', '0', + include=False, # AV: no need to add this parameter to run_card.inc + hidden=False, # AV: add cudacpp_hrdcod to runcard template and keep 'hidden='False' + fct_mod=(self.reset_makeopts,(),{}), # AV: I assume this raises an exception if cudacpp_hrdcod changes? + allowed=['0','1'] + ) + self.add_param('cudacpp_bldall', False, + include=False, # AV: no need to add this parameter to run_card.inc + hidden=False, # AV: add cudacpp_bldall to runcard template and keep 'hidden='False' + fct_mod=(self.reset_makeopts,(),{}), # AV: I assume this raises an exception if cudacpp_bldall changes? + ) self['vector_size'] = 16 # already setup in default class (just change value) self['aloha_flag'] = '--fast-math' self['matrix_flag'] = '-O3' diff --git a/epochX/cudacpp/nobm_pp_ttW.mad/Cards/run_card.dat b/epochX/cudacpp/nobm_pp_ttW.mad/Cards/run_card.dat index 5eca3e3f2b..45bb1a0aa2 100644 --- a/epochX/cudacpp/nobm_pp_ttW.mad/Cards/run_card.dat +++ b/epochX/cudacpp/nobm_pp_ttW.mad/Cards/run_card.dat @@ -184,5 +184,10 @@ systematics = systematics_program ! none, systematics [python], SysCalc [deprece #*********************************************************************** # SIMD/GPU configuration for the CUDACPP plugin #************************************************************************ - cpp = cudacpp_backend ! CUDACPP backend: fortran, cuda, hip, cpp, cppnone, cppsse4, cppavx2, cpp512y, cpp512z, cppauto + cpp = cudacpp_backend ! CUDACPP backend: fortran, cuda, hip, cpp (DEFAULT), cppnone, cppsse4, cppavx2, cpp512y, cpp512z, cppauto +#*** WARNING! The following cudacpp runcards are experimental! Users should normally change only the cudacpp_backend card *** + m = cudacpp_fptype ! CUDACPP floating point precision: f (single), d (double), m (mixed, DEFAULT: double for amplitudes, single for colors) + 0 = cudacpp_hrdcod ! CUDACPP parameter hardcoding: 0 (DEFAULT, parameters not hardcoded: read param_card.dat at runtime), 1 (hardcoded parameters) + 0 = cudacpp_helinl ! CUDACPP helicity amplitude inlining: 0 (DEFAULT, ordinary inlining of templates), 1 (aggressive inlining with 'always inline') + False = cudacpp_bldall ! CUDACPP build all available backends in separate build directories: False, True diff --git a/epochX/cudacpp/nobm_pp_ttW.mad/Cards/run_card_default.dat b/epochX/cudacpp/nobm_pp_ttW.mad/Cards/run_card_default.dat index 3b445d02a0..07d7ead34f 100644 --- a/epochX/cudacpp/nobm_pp_ttW.mad/Cards/run_card_default.dat +++ b/epochX/cudacpp/nobm_pp_ttW.mad/Cards/run_card_default.dat @@ -184,5 +184,10 @@ systematics = systematics_program ! none, systematics [python], SysCalc [deprece #*********************************************************************** # SIMD/GPU configuration for the CUDACPP plugin #************************************************************************ - cpp = cudacpp_backend ! CUDACPP backend: fortran, cuda, hip, cpp, cppnone, cppsse4, cppavx2, cpp512y, cpp512z, cppauto + cpp = cudacpp_backend ! CUDACPP backend: fortran, cuda, hip, cpp (DEFAULT), cppnone, cppsse4, cppavx2, cpp512y, cpp512z, cppauto +#*** WARNING! The following cudacpp runcards are experimental! Users should normally change only the cudacpp_backend card *** + m = cudacpp_fptype ! CUDACPP floating point precision: f (single), d (double), m (mixed, DEFAULT: double for amplitudes, single for colors) + 0 = cudacpp_hrdcod ! CUDACPP parameter hardcoding: 0 (DEFAULT, parameters not hardcoded: read param_card.dat at runtime), 1 (hardcoded parameters) + 0 = cudacpp_helinl ! CUDACPP helicity amplitude inlining: 0 (DEFAULT, ordinary inlining of templates), 1 (aggressive inlining with 'always inline') + False = cudacpp_bldall ! CUDACPP build all available backends in separate build directories: False, True diff --git a/epochX/cudacpp/nobm_pp_ttW.mad/SubProcesses/makefile b/epochX/cudacpp/nobm_pp_ttW.mad/SubProcesses/makefile index 49e6800fff..2e5bff046b 100644 --- a/epochX/cudacpp/nobm_pp_ttW.mad/SubProcesses/makefile +++ b/epochX/cudacpp/nobm_pp_ttW.mad/SubProcesses/makefile @@ -2,6 +2,11 @@ SHELL := /bin/bash include ../../Source/make_opts +# Export the relevant settings from make_opts to builds using the cudacpp makefile (#700) +export FPTYPE +export HELINL +export HRDCOD + # Determine CUDACPP_BUILDDIR based on the user-defined choices of BACKEND, FPTYPE, HELINL, HRDCOD and USEBUILDDIR (#829) # Stop with an error if BACKEND=cuda and nvcc is missing or if BACKEND=hip and hipcc is missing include ../../src/cudacpp_config.mk @@ -162,27 +167,29 @@ madevent_fortran_link: $(PROG)_fortran rm -f $(PROG) ln -s $(PROG)_fortran $(PROG) -madevent_cuda_link: - $(MAKE) USEGTEST=0 BACKEND=cuda $(CUDACPP_BUILDDIR)/$(PROG)_cuda +madevent_cudacpp_link: all rm -f $(PROG) +ifeq ($(BACKEND),cuda) ln -s $(CUDACPP_BUILDDIR)/$(PROG)_cuda $(PROG) +else ifeq ($(BACKEND),hip) + ln -s $(CUDACPP_BUILDDIR)/$(PROG)_hip $(PROG) +else + ln -s $(CUDACPP_BUILDDIR)/$(PROG)_cpp $(PROG) +endif + +madevent_cuda_link: + $(MAKE) USEGTEST=0 USEBUILDDIR=1 BACKEND=cuda madevent_cudacpp_link madevent_hip_link: - $(MAKE) USEGTEST=0 BACKEND=hip $(CUDACPP_BUILDDIR)/$(PROG)_hip - rm -f $(PROG) - ln -s $(CUDACPP_BUILDDIR)/$(PROG)_hip $(PROG) + $(MAKE) USEGTEST=0 USEBUILDDIR=1 BACKEND=hip madevent_cudacpp_link madevent_cpp_link: - $(MAKE) USEGTEST=0 BACKEND=cppauto $(CUDACPP_BUILDDIR)/$(PROG)_cpp - rm -f $(PROG) - ln -s $(CUDACPP_BUILDDIR)/$(PROG)_cpp $(PROG) + $(MAKE) USEGTEST=0 USEBUILDDIR=1 BACKEND=cppauto madevent_cudacpp_link override SUPPORTED_AVXS = cppnone cppsse4 cppavx2 cpp512y cpp512z cppauto madevent_%_link: - @if [ '$(words $(filter $*, $(SUPPORTED_AVXS)))' != '1' ]; then echo "ERROR! Invalid target '$@' (supported madevent_cpp*_link targets are: $(foreach avx,$(SUPPORTED_AVXS),'madevent_cpp$(avx)_link'))"; exit 1; fi - $(MAKE) USEGTEST=0 BACKEND=$* $(CUDACPP_BUILDDIR)/$(PROG)_cpp - rm -f $(PROG) - ln -s $(CUDACPP_BUILDDIR)/$(PROG)_cpp $(PROG) + @if [ '$(words $(filter $*, $(SUPPORTED_AVXS)))' != '1' ]; then echo "ERROR! Invalid target '$@' (supported madevent_cpp*_link targets are: $(foreach avx,$(SUPPORTED_AVXS),'madevent_$(avx)_link'))"; exit 1; fi + $(MAKE) USEGTEST=0 USEBUILDDIR=1 BACKEND=$* madevent_cudacpp_link # Building $(PROG)_cpp no longer builds $(PROG)_cuda if CUDACPP_BACKENDLIB for cuda exists (this was the case in the past to allow cpp-only builds #503) $(CUDACPP_BUILDDIR)/$(PROG)_cpp: $(PROCESS) $(DSIG_cudacpp) auto_dsig.o $(LIBS) $(MATRIX) counters.o ompnumthreads.o $(CUDACPP_BUILDDIR)/.cudacpplibs diff --git a/epochX/cudacpp/nobm_pp_ttW.mad/bin/internal/launch_plugin.py b/epochX/cudacpp/nobm_pp_ttW.mad/bin/internal/launch_plugin.py index 0924927785..19d65fa195 100644 --- a/epochX/cudacpp/nobm_pp_ttW.mad/bin/internal/launch_plugin.py +++ b/epochX/cudacpp/nobm_pp_ttW.mad/bin/internal/launch_plugin.py @@ -33,10 +33,16 @@ def compile(self, *args, **opts): if 'cwd' in opts and os.path.basename(opts['cwd']) == 'Source': path = pjoin(opts['cwd'], 'make_opts') common_run_interface.CommonRunCmd.update_make_opts_full(path, - {'FPTYPE': self.run_card['floating_type'] }) - misc.sprint('FPTYPE checked') + {'FPTYPE': self.run_card['cudacpp_fptype'], + 'HELINL': self.run_card['cudacpp_helinl'], + 'HRDCOD': self.run_card['cudacpp_hrdcod'] }) + misc.sprint('FPTYPE, HELINL, HRDCOD checked') cudacpp_supported_backends = [ 'fortran', 'cuda', 'hip', 'cpp', 'cppnone', 'cppsse4', 'cppavx2', 'cpp512y', 'cpp512z', 'cppauto' ] if args and args[0][0] == 'madevent' and hasattr(self, 'run_card'): + if self.run_card['cudacpp_bldall'] == True: # pre-build all backends #945 + logger.info("Pre-building madevent in madevent_interface.py with ALL matrix elements") + args[0][0] = 'bldall' + misc.compile(nb_core=self.options['nb_core'], *args, **opts) cudacpp_backend = self.run_card['cudacpp_backend'].lower() # the default value is defined in launch_plugin.py logger.info("Building madevent in madevent_interface.py with '%s' matrix elements"%cudacpp_backend) if cudacpp_backend in cudacpp_supported_backends : @@ -47,14 +53,18 @@ def compile(self, *args, **opts): else: return misc.compile(nb_core=self.options['nb_core'], *args, **opts) -# Phase-Space Optimization ------------------------------------------------------------------------------------ +# CUDACPP runcard block ------------------------------------------------------------------------------------ template_on = \ """#*********************************************************************** # SIMD/GPU configuration for the CUDACPP plugin #************************************************************************ - %(cudacpp_backend)s = cudacpp_backend ! CUDACPP backend: fortran, cuda, hip, cpp, cppnone, cppsse4, cppavx2, cpp512y, cpp512z, cppauto + %(cudacpp_backend)s = cudacpp_backend ! CUDACPP backend: fortran, cuda, hip, cpp (DEFAULT), cppnone, cppsse4, cppavx2, cpp512y, cpp512z, cppauto +#*** WARNING! The following cudacpp runcards are experimental! Users should normally change only the cudacpp_backend card *** + %(cudacpp_fptype)s = cudacpp_fptype ! CUDACPP floating point precision: f (single), d (double), m (mixed, DEFAULT: double for amplitudes, single for colors) + %(cudacpp_hrdcod)s = cudacpp_hrdcod ! CUDACPP parameter hardcoding: 0 (DEFAULT, parameters not hardcoded: read param_card.dat at runtime), 1 (hardcoded parameters) + %(cudacpp_helinl)s = cudacpp_helinl ! CUDACPP helicity amplitude inlining: 0 (DEFAULT, ordinary inlining of templates), 1 (aggressive inlining with 'always inline') + %(cudacpp_bldall)s = cudacpp_bldall ! CUDACPP build all available backends in separate build directories: False, True """ - template_off = '' plugin_block = banner_mod.RunBlock('simd', template_on=template_on, template_off=template_off) @@ -75,8 +85,14 @@ def reset_simd(self, old_value, new_value, name): def reset_makeopts(self, old_value, new_value, name): if not hasattr(self, 'path'): raise Exception - if name == 'floating_type': + if name == 'cudacpp_fptype': common_run_interface.CommonRunCmd.update_make_opts_full({'FPTYPE': new_value}) + elif name == 'cudacpp_hrdcod': + raise Exception('Cannot change cudacpp_hrdcod') + elif name == 'cudacpp_helinl': + raise Exception('Cannot change cudacpp_helinl') + elif name == 'cudacpp_bldall': + raise Exception('Cannot change cudacpp_bldall') else: raise Exception Sourcedir = pjoin(os.path.dirname(os.path.dirname(self.path)), 'Source') @@ -84,14 +100,34 @@ def reset_makeopts(self, old_value, new_value, name): def default_setup(self): super().default_setup() - self.add_param('floating_type', 'm', include=False, hidden=True, - fct_mod=(self.reset_makeopts,(),{}), - allowed=['m','d','f'], - comment='floating point precision: f (single), d (double), m (mixed: double for amplitudes, single for colors)' - ) cudacpp_supported_backends = [ 'fortran', 'cuda', 'hip', 'cpp', 'cppnone', 'cppsse4', 'cppavx2', 'cpp512y', 'cpp512z', 'cppauto' ] - self.add_param('cudacpp_backend', 'cpp', include=False, hidden=False, + self.add_param('cudacpp_backend', 'cpp', + include=False, # AV: 'include=True' would add "CUDACPP_BACKEND = 'cpp'" to run_card.inc + hidden=False, # AV: keep cudacpp_backend in runcard template and keep 'hidden='False' allowed=cudacpp_supported_backends) + self.add_param('cudacpp_fptype', 'm', + include=False, # AV: 'include=True' would add "CUDACPP_FPTYPE = 'm'" to run_card.inc (if fct_mod is removed, else codegen fails) + hidden=False, # AV: add cudacpp_backend to runcard template and keep 'hidden='False' + fct_mod=(self.reset_makeopts,(),{}), # AV: I assume this forces a 'make cleanavx' if FPTYPE changes? + allowed=['m','d','f'] + ) + self.add_param('cudacpp_helinl', '0', + include=False, # AV: no need to add this parameter to run_card.inc + hidden=False, # AV: add cudacpp_helinl to runcard template and keep 'hidden='False' + fct_mod=(self.reset_makeopts,(),{}), # AV: I assume this raises an exception if cudacpp_helinl changes? + allowed=['0','1'] + ) + self.add_param('cudacpp_hrdcod', '0', + include=False, # AV: no need to add this parameter to run_card.inc + hidden=False, # AV: add cudacpp_hrdcod to runcard template and keep 'hidden='False' + fct_mod=(self.reset_makeopts,(),{}), # AV: I assume this raises an exception if cudacpp_hrdcod changes? + allowed=['0','1'] + ) + self.add_param('cudacpp_bldall', False, + include=False, # AV: no need to add this parameter to run_card.inc + hidden=False, # AV: add cudacpp_bldall to runcard template and keep 'hidden='False' + fct_mod=(self.reset_makeopts,(),{}), # AV: I assume this raises an exception if cudacpp_bldall changes? + ) self['vector_size'] = 16 # already setup in default class (just change value) self['aloha_flag'] = '--fast-math' self['matrix_flag'] = '-O3' diff --git a/epochX/cudacpp/pp_tt012j.mad/Cards/run_card.dat b/epochX/cudacpp/pp_tt012j.mad/Cards/run_card.dat index 5eb60f35df..5f467b9ac3 100644 --- a/epochX/cudacpp/pp_tt012j.mad/Cards/run_card.dat +++ b/epochX/cudacpp/pp_tt012j.mad/Cards/run_card.dat @@ -207,5 +207,10 @@ systematics = systematics_program ! none, systematics [python], SysCalc [deprece #*********************************************************************** # SIMD/GPU configuration for the CUDACPP plugin #************************************************************************ - cpp = cudacpp_backend ! CUDACPP backend: fortran, cuda, hip, cpp, cppnone, cppsse4, cppavx2, cpp512y, cpp512z, cppauto + cpp = cudacpp_backend ! CUDACPP backend: fortran, cuda, hip, cpp (DEFAULT), cppnone, cppsse4, cppavx2, cpp512y, cpp512z, cppauto +#*** WARNING! The following cudacpp runcards are experimental! Users should normally change only the cudacpp_backend card *** + m = cudacpp_fptype ! CUDACPP floating point precision: f (single), d (double), m (mixed, DEFAULT: double for amplitudes, single for colors) + 0 = cudacpp_hrdcod ! CUDACPP parameter hardcoding: 0 (DEFAULT, parameters not hardcoded: read param_card.dat at runtime), 1 (hardcoded parameters) + 0 = cudacpp_helinl ! CUDACPP helicity amplitude inlining: 0 (DEFAULT, ordinary inlining of templates), 1 (aggressive inlining with 'always inline') + False = cudacpp_bldall ! CUDACPP build all available backends in separate build directories: False, True diff --git a/epochX/cudacpp/pp_tt012j.mad/Cards/run_card_default.dat b/epochX/cudacpp/pp_tt012j.mad/Cards/run_card_default.dat index 38810a6b83..ac7ad1f0b1 100644 --- a/epochX/cudacpp/pp_tt012j.mad/Cards/run_card_default.dat +++ b/epochX/cudacpp/pp_tt012j.mad/Cards/run_card_default.dat @@ -207,5 +207,10 @@ systematics = systematics_program ! none, systematics [python], SysCalc [deprece #*********************************************************************** # SIMD/GPU configuration for the CUDACPP plugin #************************************************************************ - cpp = cudacpp_backend ! CUDACPP backend: fortran, cuda, hip, cpp, cppnone, cppsse4, cppavx2, cpp512y, cpp512z, cppauto + cpp = cudacpp_backend ! CUDACPP backend: fortran, cuda, hip, cpp (DEFAULT), cppnone, cppsse4, cppavx2, cpp512y, cpp512z, cppauto +#*** WARNING! The following cudacpp runcards are experimental! Users should normally change only the cudacpp_backend card *** + m = cudacpp_fptype ! CUDACPP floating point precision: f (single), d (double), m (mixed, DEFAULT: double for amplitudes, single for colors) + 0 = cudacpp_hrdcod ! CUDACPP parameter hardcoding: 0 (DEFAULT, parameters not hardcoded: read param_card.dat at runtime), 1 (hardcoded parameters) + 0 = cudacpp_helinl ! CUDACPP helicity amplitude inlining: 0 (DEFAULT, ordinary inlining of templates), 1 (aggressive inlining with 'always inline') + False = cudacpp_bldall ! CUDACPP build all available backends in separate build directories: False, True diff --git a/epochX/cudacpp/pp_tt012j.mad/SubProcesses/makefile b/epochX/cudacpp/pp_tt012j.mad/SubProcesses/makefile index 49e6800fff..2e5bff046b 100644 --- a/epochX/cudacpp/pp_tt012j.mad/SubProcesses/makefile +++ b/epochX/cudacpp/pp_tt012j.mad/SubProcesses/makefile @@ -2,6 +2,11 @@ SHELL := /bin/bash include ../../Source/make_opts +# Export the relevant settings from make_opts to builds using the cudacpp makefile (#700) +export FPTYPE +export HELINL +export HRDCOD + # Determine CUDACPP_BUILDDIR based on the user-defined choices of BACKEND, FPTYPE, HELINL, HRDCOD and USEBUILDDIR (#829) # Stop with an error if BACKEND=cuda and nvcc is missing or if BACKEND=hip and hipcc is missing include ../../src/cudacpp_config.mk @@ -162,27 +167,29 @@ madevent_fortran_link: $(PROG)_fortran rm -f $(PROG) ln -s $(PROG)_fortran $(PROG) -madevent_cuda_link: - $(MAKE) USEGTEST=0 BACKEND=cuda $(CUDACPP_BUILDDIR)/$(PROG)_cuda +madevent_cudacpp_link: all rm -f $(PROG) +ifeq ($(BACKEND),cuda) ln -s $(CUDACPP_BUILDDIR)/$(PROG)_cuda $(PROG) +else ifeq ($(BACKEND),hip) + ln -s $(CUDACPP_BUILDDIR)/$(PROG)_hip $(PROG) +else + ln -s $(CUDACPP_BUILDDIR)/$(PROG)_cpp $(PROG) +endif + +madevent_cuda_link: + $(MAKE) USEGTEST=0 USEBUILDDIR=1 BACKEND=cuda madevent_cudacpp_link madevent_hip_link: - $(MAKE) USEGTEST=0 BACKEND=hip $(CUDACPP_BUILDDIR)/$(PROG)_hip - rm -f $(PROG) - ln -s $(CUDACPP_BUILDDIR)/$(PROG)_hip $(PROG) + $(MAKE) USEGTEST=0 USEBUILDDIR=1 BACKEND=hip madevent_cudacpp_link madevent_cpp_link: - $(MAKE) USEGTEST=0 BACKEND=cppauto $(CUDACPP_BUILDDIR)/$(PROG)_cpp - rm -f $(PROG) - ln -s $(CUDACPP_BUILDDIR)/$(PROG)_cpp $(PROG) + $(MAKE) USEGTEST=0 USEBUILDDIR=1 BACKEND=cppauto madevent_cudacpp_link override SUPPORTED_AVXS = cppnone cppsse4 cppavx2 cpp512y cpp512z cppauto madevent_%_link: - @if [ '$(words $(filter $*, $(SUPPORTED_AVXS)))' != '1' ]; then echo "ERROR! Invalid target '$@' (supported madevent_cpp*_link targets are: $(foreach avx,$(SUPPORTED_AVXS),'madevent_cpp$(avx)_link'))"; exit 1; fi - $(MAKE) USEGTEST=0 BACKEND=$* $(CUDACPP_BUILDDIR)/$(PROG)_cpp - rm -f $(PROG) - ln -s $(CUDACPP_BUILDDIR)/$(PROG)_cpp $(PROG) + @if [ '$(words $(filter $*, $(SUPPORTED_AVXS)))' != '1' ]; then echo "ERROR! Invalid target '$@' (supported madevent_cpp*_link targets are: $(foreach avx,$(SUPPORTED_AVXS),'madevent_$(avx)_link'))"; exit 1; fi + $(MAKE) USEGTEST=0 USEBUILDDIR=1 BACKEND=$* madevent_cudacpp_link # Building $(PROG)_cpp no longer builds $(PROG)_cuda if CUDACPP_BACKENDLIB for cuda exists (this was the case in the past to allow cpp-only builds #503) $(CUDACPP_BUILDDIR)/$(PROG)_cpp: $(PROCESS) $(DSIG_cudacpp) auto_dsig.o $(LIBS) $(MATRIX) counters.o ompnumthreads.o $(CUDACPP_BUILDDIR)/.cudacpplibs diff --git a/epochX/cudacpp/pp_tt012j.mad/bin/internal/launch_plugin.py b/epochX/cudacpp/pp_tt012j.mad/bin/internal/launch_plugin.py index 0924927785..19d65fa195 100644 --- a/epochX/cudacpp/pp_tt012j.mad/bin/internal/launch_plugin.py +++ b/epochX/cudacpp/pp_tt012j.mad/bin/internal/launch_plugin.py @@ -33,10 +33,16 @@ def compile(self, *args, **opts): if 'cwd' in opts and os.path.basename(opts['cwd']) == 'Source': path = pjoin(opts['cwd'], 'make_opts') common_run_interface.CommonRunCmd.update_make_opts_full(path, - {'FPTYPE': self.run_card['floating_type'] }) - misc.sprint('FPTYPE checked') + {'FPTYPE': self.run_card['cudacpp_fptype'], + 'HELINL': self.run_card['cudacpp_helinl'], + 'HRDCOD': self.run_card['cudacpp_hrdcod'] }) + misc.sprint('FPTYPE, HELINL, HRDCOD checked') cudacpp_supported_backends = [ 'fortran', 'cuda', 'hip', 'cpp', 'cppnone', 'cppsse4', 'cppavx2', 'cpp512y', 'cpp512z', 'cppauto' ] if args and args[0][0] == 'madevent' and hasattr(self, 'run_card'): + if self.run_card['cudacpp_bldall'] == True: # pre-build all backends #945 + logger.info("Pre-building madevent in madevent_interface.py with ALL matrix elements") + args[0][0] = 'bldall' + misc.compile(nb_core=self.options['nb_core'], *args, **opts) cudacpp_backend = self.run_card['cudacpp_backend'].lower() # the default value is defined in launch_plugin.py logger.info("Building madevent in madevent_interface.py with '%s' matrix elements"%cudacpp_backend) if cudacpp_backend in cudacpp_supported_backends : @@ -47,14 +53,18 @@ def compile(self, *args, **opts): else: return misc.compile(nb_core=self.options['nb_core'], *args, **opts) -# Phase-Space Optimization ------------------------------------------------------------------------------------ +# CUDACPP runcard block ------------------------------------------------------------------------------------ template_on = \ """#*********************************************************************** # SIMD/GPU configuration for the CUDACPP plugin #************************************************************************ - %(cudacpp_backend)s = cudacpp_backend ! CUDACPP backend: fortran, cuda, hip, cpp, cppnone, cppsse4, cppavx2, cpp512y, cpp512z, cppauto + %(cudacpp_backend)s = cudacpp_backend ! CUDACPP backend: fortran, cuda, hip, cpp (DEFAULT), cppnone, cppsse4, cppavx2, cpp512y, cpp512z, cppauto +#*** WARNING! The following cudacpp runcards are experimental! Users should normally change only the cudacpp_backend card *** + %(cudacpp_fptype)s = cudacpp_fptype ! CUDACPP floating point precision: f (single), d (double), m (mixed, DEFAULT: double for amplitudes, single for colors) + %(cudacpp_hrdcod)s = cudacpp_hrdcod ! CUDACPP parameter hardcoding: 0 (DEFAULT, parameters not hardcoded: read param_card.dat at runtime), 1 (hardcoded parameters) + %(cudacpp_helinl)s = cudacpp_helinl ! CUDACPP helicity amplitude inlining: 0 (DEFAULT, ordinary inlining of templates), 1 (aggressive inlining with 'always inline') + %(cudacpp_bldall)s = cudacpp_bldall ! CUDACPP build all available backends in separate build directories: False, True """ - template_off = '' plugin_block = banner_mod.RunBlock('simd', template_on=template_on, template_off=template_off) @@ -75,8 +85,14 @@ def reset_simd(self, old_value, new_value, name): def reset_makeopts(self, old_value, new_value, name): if not hasattr(self, 'path'): raise Exception - if name == 'floating_type': + if name == 'cudacpp_fptype': common_run_interface.CommonRunCmd.update_make_opts_full({'FPTYPE': new_value}) + elif name == 'cudacpp_hrdcod': + raise Exception('Cannot change cudacpp_hrdcod') + elif name == 'cudacpp_helinl': + raise Exception('Cannot change cudacpp_helinl') + elif name == 'cudacpp_bldall': + raise Exception('Cannot change cudacpp_bldall') else: raise Exception Sourcedir = pjoin(os.path.dirname(os.path.dirname(self.path)), 'Source') @@ -84,14 +100,34 @@ def reset_makeopts(self, old_value, new_value, name): def default_setup(self): super().default_setup() - self.add_param('floating_type', 'm', include=False, hidden=True, - fct_mod=(self.reset_makeopts,(),{}), - allowed=['m','d','f'], - comment='floating point precision: f (single), d (double), m (mixed: double for amplitudes, single for colors)' - ) cudacpp_supported_backends = [ 'fortran', 'cuda', 'hip', 'cpp', 'cppnone', 'cppsse4', 'cppavx2', 'cpp512y', 'cpp512z', 'cppauto' ] - self.add_param('cudacpp_backend', 'cpp', include=False, hidden=False, + self.add_param('cudacpp_backend', 'cpp', + include=False, # AV: 'include=True' would add "CUDACPP_BACKEND = 'cpp'" to run_card.inc + hidden=False, # AV: keep cudacpp_backend in runcard template and keep 'hidden='False' allowed=cudacpp_supported_backends) + self.add_param('cudacpp_fptype', 'm', + include=False, # AV: 'include=True' would add "CUDACPP_FPTYPE = 'm'" to run_card.inc (if fct_mod is removed, else codegen fails) + hidden=False, # AV: add cudacpp_backend to runcard template and keep 'hidden='False' + fct_mod=(self.reset_makeopts,(),{}), # AV: I assume this forces a 'make cleanavx' if FPTYPE changes? + allowed=['m','d','f'] + ) + self.add_param('cudacpp_helinl', '0', + include=False, # AV: no need to add this parameter to run_card.inc + hidden=False, # AV: add cudacpp_helinl to runcard template and keep 'hidden='False' + fct_mod=(self.reset_makeopts,(),{}), # AV: I assume this raises an exception if cudacpp_helinl changes? + allowed=['0','1'] + ) + self.add_param('cudacpp_hrdcod', '0', + include=False, # AV: no need to add this parameter to run_card.inc + hidden=False, # AV: add cudacpp_hrdcod to runcard template and keep 'hidden='False' + fct_mod=(self.reset_makeopts,(),{}), # AV: I assume this raises an exception if cudacpp_hrdcod changes? + allowed=['0','1'] + ) + self.add_param('cudacpp_bldall', False, + include=False, # AV: no need to add this parameter to run_card.inc + hidden=False, # AV: add cudacpp_bldall to runcard template and keep 'hidden='False' + fct_mod=(self.reset_makeopts,(),{}), # AV: I assume this raises an exception if cudacpp_bldall changes? + ) self['vector_size'] = 16 # already setup in default class (just change value) self['aloha_flag'] = '--fast-math' self['matrix_flag'] = '-O3' diff --git a/epochX/cudacpp/smeft_gg_tttt.mad/Cards/run_card.dat b/epochX/cudacpp/smeft_gg_tttt.mad/Cards/run_card.dat index 6b82577032..e63cab27e3 100644 --- a/epochX/cudacpp/smeft_gg_tttt.mad/Cards/run_card.dat +++ b/epochX/cudacpp/smeft_gg_tttt.mad/Cards/run_card.dat @@ -161,5 +161,10 @@ systematics = systematics_program ! none, systematics [python], SysCalc [deprece #*********************************************************************** # SIMD/GPU configuration for the CUDACPP plugin #************************************************************************ - cpp = cudacpp_backend ! CUDACPP backend: fortran, cuda, hip, cpp, cppnone, cppsse4, cppavx2, cpp512y, cpp512z, cppauto + cpp = cudacpp_backend ! CUDACPP backend: fortran, cuda, hip, cpp (DEFAULT), cppnone, cppsse4, cppavx2, cpp512y, cpp512z, cppauto +#*** WARNING! The following cudacpp runcards are experimental! Users should normally change only the cudacpp_backend card *** + m = cudacpp_fptype ! CUDACPP floating point precision: f (single), d (double), m (mixed, DEFAULT: double for amplitudes, single for colors) + 0 = cudacpp_hrdcod ! CUDACPP parameter hardcoding: 0 (DEFAULT, parameters not hardcoded: read param_card.dat at runtime), 1 (hardcoded parameters) + 0 = cudacpp_helinl ! CUDACPP helicity amplitude inlining: 0 (DEFAULT, ordinary inlining of templates), 1 (aggressive inlining with 'always inline') + False = cudacpp_bldall ! CUDACPP build all available backends in separate build directories: False, True diff --git a/epochX/cudacpp/smeft_gg_tttt.mad/Cards/run_card_default.dat b/epochX/cudacpp/smeft_gg_tttt.mad/Cards/run_card_default.dat index b8db871c35..a32f0a05ee 100644 --- a/epochX/cudacpp/smeft_gg_tttt.mad/Cards/run_card_default.dat +++ b/epochX/cudacpp/smeft_gg_tttt.mad/Cards/run_card_default.dat @@ -161,5 +161,10 @@ systematics = systematics_program ! none, systematics [python], SysCalc [deprece #*********************************************************************** # SIMD/GPU configuration for the CUDACPP plugin #************************************************************************ - cpp = cudacpp_backend ! CUDACPP backend: fortran, cuda, hip, cpp, cppnone, cppsse4, cppavx2, cpp512y, cpp512z, cppauto + cpp = cudacpp_backend ! CUDACPP backend: fortran, cuda, hip, cpp (DEFAULT), cppnone, cppsse4, cppavx2, cpp512y, cpp512z, cppauto +#*** WARNING! The following cudacpp runcards are experimental! Users should normally change only the cudacpp_backend card *** + m = cudacpp_fptype ! CUDACPP floating point precision: f (single), d (double), m (mixed, DEFAULT: double for amplitudes, single for colors) + 0 = cudacpp_hrdcod ! CUDACPP parameter hardcoding: 0 (DEFAULT, parameters not hardcoded: read param_card.dat at runtime), 1 (hardcoded parameters) + 0 = cudacpp_helinl ! CUDACPP helicity amplitude inlining: 0 (DEFAULT, ordinary inlining of templates), 1 (aggressive inlining with 'always inline') + False = cudacpp_bldall ! CUDACPP build all available backends in separate build directories: False, True diff --git a/epochX/cudacpp/smeft_gg_tttt.mad/SubProcesses/makefile b/epochX/cudacpp/smeft_gg_tttt.mad/SubProcesses/makefile index 49e6800fff..2e5bff046b 100644 --- a/epochX/cudacpp/smeft_gg_tttt.mad/SubProcesses/makefile +++ b/epochX/cudacpp/smeft_gg_tttt.mad/SubProcesses/makefile @@ -2,6 +2,11 @@ SHELL := /bin/bash include ../../Source/make_opts +# Export the relevant settings from make_opts to builds using the cudacpp makefile (#700) +export FPTYPE +export HELINL +export HRDCOD + # Determine CUDACPP_BUILDDIR based on the user-defined choices of BACKEND, FPTYPE, HELINL, HRDCOD and USEBUILDDIR (#829) # Stop with an error if BACKEND=cuda and nvcc is missing or if BACKEND=hip and hipcc is missing include ../../src/cudacpp_config.mk @@ -162,27 +167,29 @@ madevent_fortran_link: $(PROG)_fortran rm -f $(PROG) ln -s $(PROG)_fortran $(PROG) -madevent_cuda_link: - $(MAKE) USEGTEST=0 BACKEND=cuda $(CUDACPP_BUILDDIR)/$(PROG)_cuda +madevent_cudacpp_link: all rm -f $(PROG) +ifeq ($(BACKEND),cuda) ln -s $(CUDACPP_BUILDDIR)/$(PROG)_cuda $(PROG) +else ifeq ($(BACKEND),hip) + ln -s $(CUDACPP_BUILDDIR)/$(PROG)_hip $(PROG) +else + ln -s $(CUDACPP_BUILDDIR)/$(PROG)_cpp $(PROG) +endif + +madevent_cuda_link: + $(MAKE) USEGTEST=0 USEBUILDDIR=1 BACKEND=cuda madevent_cudacpp_link madevent_hip_link: - $(MAKE) USEGTEST=0 BACKEND=hip $(CUDACPP_BUILDDIR)/$(PROG)_hip - rm -f $(PROG) - ln -s $(CUDACPP_BUILDDIR)/$(PROG)_hip $(PROG) + $(MAKE) USEGTEST=0 USEBUILDDIR=1 BACKEND=hip madevent_cudacpp_link madevent_cpp_link: - $(MAKE) USEGTEST=0 BACKEND=cppauto $(CUDACPP_BUILDDIR)/$(PROG)_cpp - rm -f $(PROG) - ln -s $(CUDACPP_BUILDDIR)/$(PROG)_cpp $(PROG) + $(MAKE) USEGTEST=0 USEBUILDDIR=1 BACKEND=cppauto madevent_cudacpp_link override SUPPORTED_AVXS = cppnone cppsse4 cppavx2 cpp512y cpp512z cppauto madevent_%_link: - @if [ '$(words $(filter $*, $(SUPPORTED_AVXS)))' != '1' ]; then echo "ERROR! Invalid target '$@' (supported madevent_cpp*_link targets are: $(foreach avx,$(SUPPORTED_AVXS),'madevent_cpp$(avx)_link'))"; exit 1; fi - $(MAKE) USEGTEST=0 BACKEND=$* $(CUDACPP_BUILDDIR)/$(PROG)_cpp - rm -f $(PROG) - ln -s $(CUDACPP_BUILDDIR)/$(PROG)_cpp $(PROG) + @if [ '$(words $(filter $*, $(SUPPORTED_AVXS)))' != '1' ]; then echo "ERROR! Invalid target '$@' (supported madevent_cpp*_link targets are: $(foreach avx,$(SUPPORTED_AVXS),'madevent_$(avx)_link'))"; exit 1; fi + $(MAKE) USEGTEST=0 USEBUILDDIR=1 BACKEND=$* madevent_cudacpp_link # Building $(PROG)_cpp no longer builds $(PROG)_cuda if CUDACPP_BACKENDLIB for cuda exists (this was the case in the past to allow cpp-only builds #503) $(CUDACPP_BUILDDIR)/$(PROG)_cpp: $(PROCESS) $(DSIG_cudacpp) auto_dsig.o $(LIBS) $(MATRIX) counters.o ompnumthreads.o $(CUDACPP_BUILDDIR)/.cudacpplibs diff --git a/epochX/cudacpp/smeft_gg_tttt.mad/bin/internal/launch_plugin.py b/epochX/cudacpp/smeft_gg_tttt.mad/bin/internal/launch_plugin.py index 0924927785..19d65fa195 100644 --- a/epochX/cudacpp/smeft_gg_tttt.mad/bin/internal/launch_plugin.py +++ b/epochX/cudacpp/smeft_gg_tttt.mad/bin/internal/launch_plugin.py @@ -33,10 +33,16 @@ def compile(self, *args, **opts): if 'cwd' in opts and os.path.basename(opts['cwd']) == 'Source': path = pjoin(opts['cwd'], 'make_opts') common_run_interface.CommonRunCmd.update_make_opts_full(path, - {'FPTYPE': self.run_card['floating_type'] }) - misc.sprint('FPTYPE checked') + {'FPTYPE': self.run_card['cudacpp_fptype'], + 'HELINL': self.run_card['cudacpp_helinl'], + 'HRDCOD': self.run_card['cudacpp_hrdcod'] }) + misc.sprint('FPTYPE, HELINL, HRDCOD checked') cudacpp_supported_backends = [ 'fortran', 'cuda', 'hip', 'cpp', 'cppnone', 'cppsse4', 'cppavx2', 'cpp512y', 'cpp512z', 'cppauto' ] if args and args[0][0] == 'madevent' and hasattr(self, 'run_card'): + if self.run_card['cudacpp_bldall'] == True: # pre-build all backends #945 + logger.info("Pre-building madevent in madevent_interface.py with ALL matrix elements") + args[0][0] = 'bldall' + misc.compile(nb_core=self.options['nb_core'], *args, **opts) cudacpp_backend = self.run_card['cudacpp_backend'].lower() # the default value is defined in launch_plugin.py logger.info("Building madevent in madevent_interface.py with '%s' matrix elements"%cudacpp_backend) if cudacpp_backend in cudacpp_supported_backends : @@ -47,14 +53,18 @@ def compile(self, *args, **opts): else: return misc.compile(nb_core=self.options['nb_core'], *args, **opts) -# Phase-Space Optimization ------------------------------------------------------------------------------------ +# CUDACPP runcard block ------------------------------------------------------------------------------------ template_on = \ """#*********************************************************************** # SIMD/GPU configuration for the CUDACPP plugin #************************************************************************ - %(cudacpp_backend)s = cudacpp_backend ! CUDACPP backend: fortran, cuda, hip, cpp, cppnone, cppsse4, cppavx2, cpp512y, cpp512z, cppauto + %(cudacpp_backend)s = cudacpp_backend ! CUDACPP backend: fortran, cuda, hip, cpp (DEFAULT), cppnone, cppsse4, cppavx2, cpp512y, cpp512z, cppauto +#*** WARNING! The following cudacpp runcards are experimental! Users should normally change only the cudacpp_backend card *** + %(cudacpp_fptype)s = cudacpp_fptype ! CUDACPP floating point precision: f (single), d (double), m (mixed, DEFAULT: double for amplitudes, single for colors) + %(cudacpp_hrdcod)s = cudacpp_hrdcod ! CUDACPP parameter hardcoding: 0 (DEFAULT, parameters not hardcoded: read param_card.dat at runtime), 1 (hardcoded parameters) + %(cudacpp_helinl)s = cudacpp_helinl ! CUDACPP helicity amplitude inlining: 0 (DEFAULT, ordinary inlining of templates), 1 (aggressive inlining with 'always inline') + %(cudacpp_bldall)s = cudacpp_bldall ! CUDACPP build all available backends in separate build directories: False, True """ - template_off = '' plugin_block = banner_mod.RunBlock('simd', template_on=template_on, template_off=template_off) @@ -75,8 +85,14 @@ def reset_simd(self, old_value, new_value, name): def reset_makeopts(self, old_value, new_value, name): if not hasattr(self, 'path'): raise Exception - if name == 'floating_type': + if name == 'cudacpp_fptype': common_run_interface.CommonRunCmd.update_make_opts_full({'FPTYPE': new_value}) + elif name == 'cudacpp_hrdcod': + raise Exception('Cannot change cudacpp_hrdcod') + elif name == 'cudacpp_helinl': + raise Exception('Cannot change cudacpp_helinl') + elif name == 'cudacpp_bldall': + raise Exception('Cannot change cudacpp_bldall') else: raise Exception Sourcedir = pjoin(os.path.dirname(os.path.dirname(self.path)), 'Source') @@ -84,14 +100,34 @@ def reset_makeopts(self, old_value, new_value, name): def default_setup(self): super().default_setup() - self.add_param('floating_type', 'm', include=False, hidden=True, - fct_mod=(self.reset_makeopts,(),{}), - allowed=['m','d','f'], - comment='floating point precision: f (single), d (double), m (mixed: double for amplitudes, single for colors)' - ) cudacpp_supported_backends = [ 'fortran', 'cuda', 'hip', 'cpp', 'cppnone', 'cppsse4', 'cppavx2', 'cpp512y', 'cpp512z', 'cppauto' ] - self.add_param('cudacpp_backend', 'cpp', include=False, hidden=False, + self.add_param('cudacpp_backend', 'cpp', + include=False, # AV: 'include=True' would add "CUDACPP_BACKEND = 'cpp'" to run_card.inc + hidden=False, # AV: keep cudacpp_backend in runcard template and keep 'hidden='False' allowed=cudacpp_supported_backends) + self.add_param('cudacpp_fptype', 'm', + include=False, # AV: 'include=True' would add "CUDACPP_FPTYPE = 'm'" to run_card.inc (if fct_mod is removed, else codegen fails) + hidden=False, # AV: add cudacpp_backend to runcard template and keep 'hidden='False' + fct_mod=(self.reset_makeopts,(),{}), # AV: I assume this forces a 'make cleanavx' if FPTYPE changes? + allowed=['m','d','f'] + ) + self.add_param('cudacpp_helinl', '0', + include=False, # AV: no need to add this parameter to run_card.inc + hidden=False, # AV: add cudacpp_helinl to runcard template and keep 'hidden='False' + fct_mod=(self.reset_makeopts,(),{}), # AV: I assume this raises an exception if cudacpp_helinl changes? + allowed=['0','1'] + ) + self.add_param('cudacpp_hrdcod', '0', + include=False, # AV: no need to add this parameter to run_card.inc + hidden=False, # AV: add cudacpp_hrdcod to runcard template and keep 'hidden='False' + fct_mod=(self.reset_makeopts,(),{}), # AV: I assume this raises an exception if cudacpp_hrdcod changes? + allowed=['0','1'] + ) + self.add_param('cudacpp_bldall', False, + include=False, # AV: no need to add this parameter to run_card.inc + hidden=False, # AV: add cudacpp_bldall to runcard template and keep 'hidden='False' + fct_mod=(self.reset_makeopts,(),{}), # AV: I assume this raises an exception if cudacpp_bldall changes? + ) self['vector_size'] = 16 # already setup in default class (just change value) self['aloha_flag'] = '--fast-math' self['matrix_flag'] = '-O3' diff --git a/epochX/cudacpp/susy_gg_t1t1.mad/Cards/run_card.dat b/epochX/cudacpp/susy_gg_t1t1.mad/Cards/run_card.dat index 6b82577032..e63cab27e3 100644 --- a/epochX/cudacpp/susy_gg_t1t1.mad/Cards/run_card.dat +++ b/epochX/cudacpp/susy_gg_t1t1.mad/Cards/run_card.dat @@ -161,5 +161,10 @@ systematics = systematics_program ! none, systematics [python], SysCalc [deprece #*********************************************************************** # SIMD/GPU configuration for the CUDACPP plugin #************************************************************************ - cpp = cudacpp_backend ! CUDACPP backend: fortran, cuda, hip, cpp, cppnone, cppsse4, cppavx2, cpp512y, cpp512z, cppauto + cpp = cudacpp_backend ! CUDACPP backend: fortran, cuda, hip, cpp (DEFAULT), cppnone, cppsse4, cppavx2, cpp512y, cpp512z, cppauto +#*** WARNING! The following cudacpp runcards are experimental! Users should normally change only the cudacpp_backend card *** + m = cudacpp_fptype ! CUDACPP floating point precision: f (single), d (double), m (mixed, DEFAULT: double for amplitudes, single for colors) + 0 = cudacpp_hrdcod ! CUDACPP parameter hardcoding: 0 (DEFAULT, parameters not hardcoded: read param_card.dat at runtime), 1 (hardcoded parameters) + 0 = cudacpp_helinl ! CUDACPP helicity amplitude inlining: 0 (DEFAULT, ordinary inlining of templates), 1 (aggressive inlining with 'always inline') + False = cudacpp_bldall ! CUDACPP build all available backends in separate build directories: False, True diff --git a/epochX/cudacpp/susy_gg_t1t1.mad/Cards/run_card_default.dat b/epochX/cudacpp/susy_gg_t1t1.mad/Cards/run_card_default.dat index b8db871c35..a32f0a05ee 100644 --- a/epochX/cudacpp/susy_gg_t1t1.mad/Cards/run_card_default.dat +++ b/epochX/cudacpp/susy_gg_t1t1.mad/Cards/run_card_default.dat @@ -161,5 +161,10 @@ systematics = systematics_program ! none, systematics [python], SysCalc [deprece #*********************************************************************** # SIMD/GPU configuration for the CUDACPP plugin #************************************************************************ - cpp = cudacpp_backend ! CUDACPP backend: fortran, cuda, hip, cpp, cppnone, cppsse4, cppavx2, cpp512y, cpp512z, cppauto + cpp = cudacpp_backend ! CUDACPP backend: fortran, cuda, hip, cpp (DEFAULT), cppnone, cppsse4, cppavx2, cpp512y, cpp512z, cppauto +#*** WARNING! The following cudacpp runcards are experimental! Users should normally change only the cudacpp_backend card *** + m = cudacpp_fptype ! CUDACPP floating point precision: f (single), d (double), m (mixed, DEFAULT: double for amplitudes, single for colors) + 0 = cudacpp_hrdcod ! CUDACPP parameter hardcoding: 0 (DEFAULT, parameters not hardcoded: read param_card.dat at runtime), 1 (hardcoded parameters) + 0 = cudacpp_helinl ! CUDACPP helicity amplitude inlining: 0 (DEFAULT, ordinary inlining of templates), 1 (aggressive inlining with 'always inline') + False = cudacpp_bldall ! CUDACPP build all available backends in separate build directories: False, True diff --git a/epochX/cudacpp/susy_gg_t1t1.mad/SubProcesses/makefile b/epochX/cudacpp/susy_gg_t1t1.mad/SubProcesses/makefile index 49e6800fff..2e5bff046b 100644 --- a/epochX/cudacpp/susy_gg_t1t1.mad/SubProcesses/makefile +++ b/epochX/cudacpp/susy_gg_t1t1.mad/SubProcesses/makefile @@ -2,6 +2,11 @@ SHELL := /bin/bash include ../../Source/make_opts +# Export the relevant settings from make_opts to builds using the cudacpp makefile (#700) +export FPTYPE +export HELINL +export HRDCOD + # Determine CUDACPP_BUILDDIR based on the user-defined choices of BACKEND, FPTYPE, HELINL, HRDCOD and USEBUILDDIR (#829) # Stop with an error if BACKEND=cuda and nvcc is missing or if BACKEND=hip and hipcc is missing include ../../src/cudacpp_config.mk @@ -162,27 +167,29 @@ madevent_fortran_link: $(PROG)_fortran rm -f $(PROG) ln -s $(PROG)_fortran $(PROG) -madevent_cuda_link: - $(MAKE) USEGTEST=0 BACKEND=cuda $(CUDACPP_BUILDDIR)/$(PROG)_cuda +madevent_cudacpp_link: all rm -f $(PROG) +ifeq ($(BACKEND),cuda) ln -s $(CUDACPP_BUILDDIR)/$(PROG)_cuda $(PROG) +else ifeq ($(BACKEND),hip) + ln -s $(CUDACPP_BUILDDIR)/$(PROG)_hip $(PROG) +else + ln -s $(CUDACPP_BUILDDIR)/$(PROG)_cpp $(PROG) +endif + +madevent_cuda_link: + $(MAKE) USEGTEST=0 USEBUILDDIR=1 BACKEND=cuda madevent_cudacpp_link madevent_hip_link: - $(MAKE) USEGTEST=0 BACKEND=hip $(CUDACPP_BUILDDIR)/$(PROG)_hip - rm -f $(PROG) - ln -s $(CUDACPP_BUILDDIR)/$(PROG)_hip $(PROG) + $(MAKE) USEGTEST=0 USEBUILDDIR=1 BACKEND=hip madevent_cudacpp_link madevent_cpp_link: - $(MAKE) USEGTEST=0 BACKEND=cppauto $(CUDACPP_BUILDDIR)/$(PROG)_cpp - rm -f $(PROG) - ln -s $(CUDACPP_BUILDDIR)/$(PROG)_cpp $(PROG) + $(MAKE) USEGTEST=0 USEBUILDDIR=1 BACKEND=cppauto madevent_cudacpp_link override SUPPORTED_AVXS = cppnone cppsse4 cppavx2 cpp512y cpp512z cppauto madevent_%_link: - @if [ '$(words $(filter $*, $(SUPPORTED_AVXS)))' != '1' ]; then echo "ERROR! Invalid target '$@' (supported madevent_cpp*_link targets are: $(foreach avx,$(SUPPORTED_AVXS),'madevent_cpp$(avx)_link'))"; exit 1; fi - $(MAKE) USEGTEST=0 BACKEND=$* $(CUDACPP_BUILDDIR)/$(PROG)_cpp - rm -f $(PROG) - ln -s $(CUDACPP_BUILDDIR)/$(PROG)_cpp $(PROG) + @if [ '$(words $(filter $*, $(SUPPORTED_AVXS)))' != '1' ]; then echo "ERROR! Invalid target '$@' (supported madevent_cpp*_link targets are: $(foreach avx,$(SUPPORTED_AVXS),'madevent_$(avx)_link'))"; exit 1; fi + $(MAKE) USEGTEST=0 USEBUILDDIR=1 BACKEND=$* madevent_cudacpp_link # Building $(PROG)_cpp no longer builds $(PROG)_cuda if CUDACPP_BACKENDLIB for cuda exists (this was the case in the past to allow cpp-only builds #503) $(CUDACPP_BUILDDIR)/$(PROG)_cpp: $(PROCESS) $(DSIG_cudacpp) auto_dsig.o $(LIBS) $(MATRIX) counters.o ompnumthreads.o $(CUDACPP_BUILDDIR)/.cudacpplibs diff --git a/epochX/cudacpp/susy_gg_t1t1.mad/bin/internal/launch_plugin.py b/epochX/cudacpp/susy_gg_t1t1.mad/bin/internal/launch_plugin.py index 0924927785..19d65fa195 100644 --- a/epochX/cudacpp/susy_gg_t1t1.mad/bin/internal/launch_plugin.py +++ b/epochX/cudacpp/susy_gg_t1t1.mad/bin/internal/launch_plugin.py @@ -33,10 +33,16 @@ def compile(self, *args, **opts): if 'cwd' in opts and os.path.basename(opts['cwd']) == 'Source': path = pjoin(opts['cwd'], 'make_opts') common_run_interface.CommonRunCmd.update_make_opts_full(path, - {'FPTYPE': self.run_card['floating_type'] }) - misc.sprint('FPTYPE checked') + {'FPTYPE': self.run_card['cudacpp_fptype'], + 'HELINL': self.run_card['cudacpp_helinl'], + 'HRDCOD': self.run_card['cudacpp_hrdcod'] }) + misc.sprint('FPTYPE, HELINL, HRDCOD checked') cudacpp_supported_backends = [ 'fortran', 'cuda', 'hip', 'cpp', 'cppnone', 'cppsse4', 'cppavx2', 'cpp512y', 'cpp512z', 'cppauto' ] if args and args[0][0] == 'madevent' and hasattr(self, 'run_card'): + if self.run_card['cudacpp_bldall'] == True: # pre-build all backends #945 + logger.info("Pre-building madevent in madevent_interface.py with ALL matrix elements") + args[0][0] = 'bldall' + misc.compile(nb_core=self.options['nb_core'], *args, **opts) cudacpp_backend = self.run_card['cudacpp_backend'].lower() # the default value is defined in launch_plugin.py logger.info("Building madevent in madevent_interface.py with '%s' matrix elements"%cudacpp_backend) if cudacpp_backend in cudacpp_supported_backends : @@ -47,14 +53,18 @@ def compile(self, *args, **opts): else: return misc.compile(nb_core=self.options['nb_core'], *args, **opts) -# Phase-Space Optimization ------------------------------------------------------------------------------------ +# CUDACPP runcard block ------------------------------------------------------------------------------------ template_on = \ """#*********************************************************************** # SIMD/GPU configuration for the CUDACPP plugin #************************************************************************ - %(cudacpp_backend)s = cudacpp_backend ! CUDACPP backend: fortran, cuda, hip, cpp, cppnone, cppsse4, cppavx2, cpp512y, cpp512z, cppauto + %(cudacpp_backend)s = cudacpp_backend ! CUDACPP backend: fortran, cuda, hip, cpp (DEFAULT), cppnone, cppsse4, cppavx2, cpp512y, cpp512z, cppauto +#*** WARNING! The following cudacpp runcards are experimental! Users should normally change only the cudacpp_backend card *** + %(cudacpp_fptype)s = cudacpp_fptype ! CUDACPP floating point precision: f (single), d (double), m (mixed, DEFAULT: double for amplitudes, single for colors) + %(cudacpp_hrdcod)s = cudacpp_hrdcod ! CUDACPP parameter hardcoding: 0 (DEFAULT, parameters not hardcoded: read param_card.dat at runtime), 1 (hardcoded parameters) + %(cudacpp_helinl)s = cudacpp_helinl ! CUDACPP helicity amplitude inlining: 0 (DEFAULT, ordinary inlining of templates), 1 (aggressive inlining with 'always inline') + %(cudacpp_bldall)s = cudacpp_bldall ! CUDACPP build all available backends in separate build directories: False, True """ - template_off = '' plugin_block = banner_mod.RunBlock('simd', template_on=template_on, template_off=template_off) @@ -75,8 +85,14 @@ def reset_simd(self, old_value, new_value, name): def reset_makeopts(self, old_value, new_value, name): if not hasattr(self, 'path'): raise Exception - if name == 'floating_type': + if name == 'cudacpp_fptype': common_run_interface.CommonRunCmd.update_make_opts_full({'FPTYPE': new_value}) + elif name == 'cudacpp_hrdcod': + raise Exception('Cannot change cudacpp_hrdcod') + elif name == 'cudacpp_helinl': + raise Exception('Cannot change cudacpp_helinl') + elif name == 'cudacpp_bldall': + raise Exception('Cannot change cudacpp_bldall') else: raise Exception Sourcedir = pjoin(os.path.dirname(os.path.dirname(self.path)), 'Source') @@ -84,14 +100,34 @@ def reset_makeopts(self, old_value, new_value, name): def default_setup(self): super().default_setup() - self.add_param('floating_type', 'm', include=False, hidden=True, - fct_mod=(self.reset_makeopts,(),{}), - allowed=['m','d','f'], - comment='floating point precision: f (single), d (double), m (mixed: double for amplitudes, single for colors)' - ) cudacpp_supported_backends = [ 'fortran', 'cuda', 'hip', 'cpp', 'cppnone', 'cppsse4', 'cppavx2', 'cpp512y', 'cpp512z', 'cppauto' ] - self.add_param('cudacpp_backend', 'cpp', include=False, hidden=False, + self.add_param('cudacpp_backend', 'cpp', + include=False, # AV: 'include=True' would add "CUDACPP_BACKEND = 'cpp'" to run_card.inc + hidden=False, # AV: keep cudacpp_backend in runcard template and keep 'hidden='False' allowed=cudacpp_supported_backends) + self.add_param('cudacpp_fptype', 'm', + include=False, # AV: 'include=True' would add "CUDACPP_FPTYPE = 'm'" to run_card.inc (if fct_mod is removed, else codegen fails) + hidden=False, # AV: add cudacpp_backend to runcard template and keep 'hidden='False' + fct_mod=(self.reset_makeopts,(),{}), # AV: I assume this forces a 'make cleanavx' if FPTYPE changes? + allowed=['m','d','f'] + ) + self.add_param('cudacpp_helinl', '0', + include=False, # AV: no need to add this parameter to run_card.inc + hidden=False, # AV: add cudacpp_helinl to runcard template and keep 'hidden='False' + fct_mod=(self.reset_makeopts,(),{}), # AV: I assume this raises an exception if cudacpp_helinl changes? + allowed=['0','1'] + ) + self.add_param('cudacpp_hrdcod', '0', + include=False, # AV: no need to add this parameter to run_card.inc + hidden=False, # AV: add cudacpp_hrdcod to runcard template and keep 'hidden='False' + fct_mod=(self.reset_makeopts,(),{}), # AV: I assume this raises an exception if cudacpp_hrdcod changes? + allowed=['0','1'] + ) + self.add_param('cudacpp_bldall', False, + include=False, # AV: no need to add this parameter to run_card.inc + hidden=False, # AV: add cudacpp_bldall to runcard template and keep 'hidden='False' + fct_mod=(self.reset_makeopts,(),{}), # AV: I assume this raises an exception if cudacpp_bldall changes? + ) self['vector_size'] = 16 # already setup in default class (just change value) self['aloha_flag'] = '--fast-math' self['matrix_flag'] = '-O3' diff --git a/epochX/cudacpp/susy_gg_tt.mad/Cards/run_card.dat b/epochX/cudacpp/susy_gg_tt.mad/Cards/run_card.dat index 6b82577032..e63cab27e3 100644 --- a/epochX/cudacpp/susy_gg_tt.mad/Cards/run_card.dat +++ b/epochX/cudacpp/susy_gg_tt.mad/Cards/run_card.dat @@ -161,5 +161,10 @@ systematics = systematics_program ! none, systematics [python], SysCalc [deprece #*********************************************************************** # SIMD/GPU configuration for the CUDACPP plugin #************************************************************************ - cpp = cudacpp_backend ! CUDACPP backend: fortran, cuda, hip, cpp, cppnone, cppsse4, cppavx2, cpp512y, cpp512z, cppauto + cpp = cudacpp_backend ! CUDACPP backend: fortran, cuda, hip, cpp (DEFAULT), cppnone, cppsse4, cppavx2, cpp512y, cpp512z, cppauto +#*** WARNING! The following cudacpp runcards are experimental! Users should normally change only the cudacpp_backend card *** + m = cudacpp_fptype ! CUDACPP floating point precision: f (single), d (double), m (mixed, DEFAULT: double for amplitudes, single for colors) + 0 = cudacpp_hrdcod ! CUDACPP parameter hardcoding: 0 (DEFAULT, parameters not hardcoded: read param_card.dat at runtime), 1 (hardcoded parameters) + 0 = cudacpp_helinl ! CUDACPP helicity amplitude inlining: 0 (DEFAULT, ordinary inlining of templates), 1 (aggressive inlining with 'always inline') + False = cudacpp_bldall ! CUDACPP build all available backends in separate build directories: False, True diff --git a/epochX/cudacpp/susy_gg_tt.mad/Cards/run_card_default.dat b/epochX/cudacpp/susy_gg_tt.mad/Cards/run_card_default.dat index b8db871c35..a32f0a05ee 100644 --- a/epochX/cudacpp/susy_gg_tt.mad/Cards/run_card_default.dat +++ b/epochX/cudacpp/susy_gg_tt.mad/Cards/run_card_default.dat @@ -161,5 +161,10 @@ systematics = systematics_program ! none, systematics [python], SysCalc [deprece #*********************************************************************** # SIMD/GPU configuration for the CUDACPP plugin #************************************************************************ - cpp = cudacpp_backend ! CUDACPP backend: fortran, cuda, hip, cpp, cppnone, cppsse4, cppavx2, cpp512y, cpp512z, cppauto + cpp = cudacpp_backend ! CUDACPP backend: fortran, cuda, hip, cpp (DEFAULT), cppnone, cppsse4, cppavx2, cpp512y, cpp512z, cppauto +#*** WARNING! The following cudacpp runcards are experimental! Users should normally change only the cudacpp_backend card *** + m = cudacpp_fptype ! CUDACPP floating point precision: f (single), d (double), m (mixed, DEFAULT: double for amplitudes, single for colors) + 0 = cudacpp_hrdcod ! CUDACPP parameter hardcoding: 0 (DEFAULT, parameters not hardcoded: read param_card.dat at runtime), 1 (hardcoded parameters) + 0 = cudacpp_helinl ! CUDACPP helicity amplitude inlining: 0 (DEFAULT, ordinary inlining of templates), 1 (aggressive inlining with 'always inline') + False = cudacpp_bldall ! CUDACPP build all available backends in separate build directories: False, True diff --git a/epochX/cudacpp/susy_gg_tt.mad/SubProcesses/makefile b/epochX/cudacpp/susy_gg_tt.mad/SubProcesses/makefile index 49e6800fff..2e5bff046b 100644 --- a/epochX/cudacpp/susy_gg_tt.mad/SubProcesses/makefile +++ b/epochX/cudacpp/susy_gg_tt.mad/SubProcesses/makefile @@ -2,6 +2,11 @@ SHELL := /bin/bash include ../../Source/make_opts +# Export the relevant settings from make_opts to builds using the cudacpp makefile (#700) +export FPTYPE +export HELINL +export HRDCOD + # Determine CUDACPP_BUILDDIR based on the user-defined choices of BACKEND, FPTYPE, HELINL, HRDCOD and USEBUILDDIR (#829) # Stop with an error if BACKEND=cuda and nvcc is missing or if BACKEND=hip and hipcc is missing include ../../src/cudacpp_config.mk @@ -162,27 +167,29 @@ madevent_fortran_link: $(PROG)_fortran rm -f $(PROG) ln -s $(PROG)_fortran $(PROG) -madevent_cuda_link: - $(MAKE) USEGTEST=0 BACKEND=cuda $(CUDACPP_BUILDDIR)/$(PROG)_cuda +madevent_cudacpp_link: all rm -f $(PROG) +ifeq ($(BACKEND),cuda) ln -s $(CUDACPP_BUILDDIR)/$(PROG)_cuda $(PROG) +else ifeq ($(BACKEND),hip) + ln -s $(CUDACPP_BUILDDIR)/$(PROG)_hip $(PROG) +else + ln -s $(CUDACPP_BUILDDIR)/$(PROG)_cpp $(PROG) +endif + +madevent_cuda_link: + $(MAKE) USEGTEST=0 USEBUILDDIR=1 BACKEND=cuda madevent_cudacpp_link madevent_hip_link: - $(MAKE) USEGTEST=0 BACKEND=hip $(CUDACPP_BUILDDIR)/$(PROG)_hip - rm -f $(PROG) - ln -s $(CUDACPP_BUILDDIR)/$(PROG)_hip $(PROG) + $(MAKE) USEGTEST=0 USEBUILDDIR=1 BACKEND=hip madevent_cudacpp_link madevent_cpp_link: - $(MAKE) USEGTEST=0 BACKEND=cppauto $(CUDACPP_BUILDDIR)/$(PROG)_cpp - rm -f $(PROG) - ln -s $(CUDACPP_BUILDDIR)/$(PROG)_cpp $(PROG) + $(MAKE) USEGTEST=0 USEBUILDDIR=1 BACKEND=cppauto madevent_cudacpp_link override SUPPORTED_AVXS = cppnone cppsse4 cppavx2 cpp512y cpp512z cppauto madevent_%_link: - @if [ '$(words $(filter $*, $(SUPPORTED_AVXS)))' != '1' ]; then echo "ERROR! Invalid target '$@' (supported madevent_cpp*_link targets are: $(foreach avx,$(SUPPORTED_AVXS),'madevent_cpp$(avx)_link'))"; exit 1; fi - $(MAKE) USEGTEST=0 BACKEND=$* $(CUDACPP_BUILDDIR)/$(PROG)_cpp - rm -f $(PROG) - ln -s $(CUDACPP_BUILDDIR)/$(PROG)_cpp $(PROG) + @if [ '$(words $(filter $*, $(SUPPORTED_AVXS)))' != '1' ]; then echo "ERROR! Invalid target '$@' (supported madevent_cpp*_link targets are: $(foreach avx,$(SUPPORTED_AVXS),'madevent_$(avx)_link'))"; exit 1; fi + $(MAKE) USEGTEST=0 USEBUILDDIR=1 BACKEND=$* madevent_cudacpp_link # Building $(PROG)_cpp no longer builds $(PROG)_cuda if CUDACPP_BACKENDLIB for cuda exists (this was the case in the past to allow cpp-only builds #503) $(CUDACPP_BUILDDIR)/$(PROG)_cpp: $(PROCESS) $(DSIG_cudacpp) auto_dsig.o $(LIBS) $(MATRIX) counters.o ompnumthreads.o $(CUDACPP_BUILDDIR)/.cudacpplibs diff --git a/epochX/cudacpp/susy_gg_tt.mad/bin/internal/launch_plugin.py b/epochX/cudacpp/susy_gg_tt.mad/bin/internal/launch_plugin.py index 0924927785..19d65fa195 100644 --- a/epochX/cudacpp/susy_gg_tt.mad/bin/internal/launch_plugin.py +++ b/epochX/cudacpp/susy_gg_tt.mad/bin/internal/launch_plugin.py @@ -33,10 +33,16 @@ def compile(self, *args, **opts): if 'cwd' in opts and os.path.basename(opts['cwd']) == 'Source': path = pjoin(opts['cwd'], 'make_opts') common_run_interface.CommonRunCmd.update_make_opts_full(path, - {'FPTYPE': self.run_card['floating_type'] }) - misc.sprint('FPTYPE checked') + {'FPTYPE': self.run_card['cudacpp_fptype'], + 'HELINL': self.run_card['cudacpp_helinl'], + 'HRDCOD': self.run_card['cudacpp_hrdcod'] }) + misc.sprint('FPTYPE, HELINL, HRDCOD checked') cudacpp_supported_backends = [ 'fortran', 'cuda', 'hip', 'cpp', 'cppnone', 'cppsse4', 'cppavx2', 'cpp512y', 'cpp512z', 'cppauto' ] if args and args[0][0] == 'madevent' and hasattr(self, 'run_card'): + if self.run_card['cudacpp_bldall'] == True: # pre-build all backends #945 + logger.info("Pre-building madevent in madevent_interface.py with ALL matrix elements") + args[0][0] = 'bldall' + misc.compile(nb_core=self.options['nb_core'], *args, **opts) cudacpp_backend = self.run_card['cudacpp_backend'].lower() # the default value is defined in launch_plugin.py logger.info("Building madevent in madevent_interface.py with '%s' matrix elements"%cudacpp_backend) if cudacpp_backend in cudacpp_supported_backends : @@ -47,14 +53,18 @@ def compile(self, *args, **opts): else: return misc.compile(nb_core=self.options['nb_core'], *args, **opts) -# Phase-Space Optimization ------------------------------------------------------------------------------------ +# CUDACPP runcard block ------------------------------------------------------------------------------------ template_on = \ """#*********************************************************************** # SIMD/GPU configuration for the CUDACPP plugin #************************************************************************ - %(cudacpp_backend)s = cudacpp_backend ! CUDACPP backend: fortran, cuda, hip, cpp, cppnone, cppsse4, cppavx2, cpp512y, cpp512z, cppauto + %(cudacpp_backend)s = cudacpp_backend ! CUDACPP backend: fortran, cuda, hip, cpp (DEFAULT), cppnone, cppsse4, cppavx2, cpp512y, cpp512z, cppauto +#*** WARNING! The following cudacpp runcards are experimental! Users should normally change only the cudacpp_backend card *** + %(cudacpp_fptype)s = cudacpp_fptype ! CUDACPP floating point precision: f (single), d (double), m (mixed, DEFAULT: double for amplitudes, single for colors) + %(cudacpp_hrdcod)s = cudacpp_hrdcod ! CUDACPP parameter hardcoding: 0 (DEFAULT, parameters not hardcoded: read param_card.dat at runtime), 1 (hardcoded parameters) + %(cudacpp_helinl)s = cudacpp_helinl ! CUDACPP helicity amplitude inlining: 0 (DEFAULT, ordinary inlining of templates), 1 (aggressive inlining with 'always inline') + %(cudacpp_bldall)s = cudacpp_bldall ! CUDACPP build all available backends in separate build directories: False, True """ - template_off = '' plugin_block = banner_mod.RunBlock('simd', template_on=template_on, template_off=template_off) @@ -75,8 +85,14 @@ def reset_simd(self, old_value, new_value, name): def reset_makeopts(self, old_value, new_value, name): if not hasattr(self, 'path'): raise Exception - if name == 'floating_type': + if name == 'cudacpp_fptype': common_run_interface.CommonRunCmd.update_make_opts_full({'FPTYPE': new_value}) + elif name == 'cudacpp_hrdcod': + raise Exception('Cannot change cudacpp_hrdcod') + elif name == 'cudacpp_helinl': + raise Exception('Cannot change cudacpp_helinl') + elif name == 'cudacpp_bldall': + raise Exception('Cannot change cudacpp_bldall') else: raise Exception Sourcedir = pjoin(os.path.dirname(os.path.dirname(self.path)), 'Source') @@ -84,14 +100,34 @@ def reset_makeopts(self, old_value, new_value, name): def default_setup(self): super().default_setup() - self.add_param('floating_type', 'm', include=False, hidden=True, - fct_mod=(self.reset_makeopts,(),{}), - allowed=['m','d','f'], - comment='floating point precision: f (single), d (double), m (mixed: double for amplitudes, single for colors)' - ) cudacpp_supported_backends = [ 'fortran', 'cuda', 'hip', 'cpp', 'cppnone', 'cppsse4', 'cppavx2', 'cpp512y', 'cpp512z', 'cppauto' ] - self.add_param('cudacpp_backend', 'cpp', include=False, hidden=False, + self.add_param('cudacpp_backend', 'cpp', + include=False, # AV: 'include=True' would add "CUDACPP_BACKEND = 'cpp'" to run_card.inc + hidden=False, # AV: keep cudacpp_backend in runcard template and keep 'hidden='False' allowed=cudacpp_supported_backends) + self.add_param('cudacpp_fptype', 'm', + include=False, # AV: 'include=True' would add "CUDACPP_FPTYPE = 'm'" to run_card.inc (if fct_mod is removed, else codegen fails) + hidden=False, # AV: add cudacpp_backend to runcard template and keep 'hidden='False' + fct_mod=(self.reset_makeopts,(),{}), # AV: I assume this forces a 'make cleanavx' if FPTYPE changes? + allowed=['m','d','f'] + ) + self.add_param('cudacpp_helinl', '0', + include=False, # AV: no need to add this parameter to run_card.inc + hidden=False, # AV: add cudacpp_helinl to runcard template and keep 'hidden='False' + fct_mod=(self.reset_makeopts,(),{}), # AV: I assume this raises an exception if cudacpp_helinl changes? + allowed=['0','1'] + ) + self.add_param('cudacpp_hrdcod', '0', + include=False, # AV: no need to add this parameter to run_card.inc + hidden=False, # AV: add cudacpp_hrdcod to runcard template and keep 'hidden='False' + fct_mod=(self.reset_makeopts,(),{}), # AV: I assume this raises an exception if cudacpp_hrdcod changes? + allowed=['0','1'] + ) + self.add_param('cudacpp_bldall', False, + include=False, # AV: no need to add this parameter to run_card.inc + hidden=False, # AV: add cudacpp_bldall to runcard template and keep 'hidden='False' + fct_mod=(self.reset_makeopts,(),{}), # AV: I assume this raises an exception if cudacpp_bldall changes? + ) self['vector_size'] = 16 # already setup in default class (just change value) self['aloha_flag'] = '--fast-math' self['matrix_flag'] = '-O3' diff --git a/epochX/cudacpp/tlau/fromgridpacks/gg_tt.mad/cpp512y/output.txt b/epochX/cudacpp/tlau/fromgridpacks/gg_tt.mad/cpp512y/output.txt new file mode 100644 index 0000000000..fc51da0a3b --- /dev/null +++ b/epochX/cudacpp/tlau/fromgridpacks/gg_tt.mad/cpp512y/output.txt @@ -0,0 +1,49 @@ +START: Wed Aug 7 02:04:39 PM CEST 2024 +CUDACPP_RUNTIME_DISABLEFPE is not set +lrwxrwxrwx. 1 avalassi zg 35 Aug 7 14:04 madevent/SubProcesses/P1_gg_ttx/madevent -> build.512y_m_inl0_hrd0/madevent_cpp +Now generating 10000 events with random seed 21 and granularity 1 +************************************************************ +* * +* W E L C O M E to * +* M A D G R A P H 5 _ a M C @ N L O * +* M A D E V E N T * +* * +* * * * +* * * * * * +* * * * * 5 * * * * * +* * * * * * +* * * * +* * +* VERSION 3.5.3_lo_vect * +* * +* The MadGraph5_aMC@NLO Development Team - Find us at * +* https://server06.fynu.ucl.ac.be/projects/madgraph * +* * +* Type 'help' for in-line help. * +* * +************************************************************ +INFO: load configuration from /data/avalassi/GPU2024/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gg_tt.mad/madevent/Cards/me5_configuration.txt +INFO: load configuration from /data/avalassi/GPU2024/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gg_tt.mad/madevent/Cards/me5_configuration.txt +Using default text editor "vi". Set another one in ./input/mg5_configuration.txt +No valid eps viewer found. Please set in ./input/mg5_configuration.txt +Using default web browser "firefox". Set another one in ./input/mg5_configuration.txt +WRITE GRIDCARD /data/avalassi/GPU2024/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gg_tt.mad/madevent +generate 10000 events +P1_gg_ttx +Using random number seed offset = 21 +INFO: Generating 10000.0 unweighted events. +sum of cpu time of last step: 2 seconds +INFO: need to improve 2 channels +combine_events +store_events + === Results Summary for run: GridRun_21 tag: tag_1 === + + Cross-section : 0 +- 0 pb + Nb of events : 10000 + +INFO: No version of lhapdf. Can not run systematics computation +decay_events -from_cards +DONE +write ./events.lhe.gz +END: Wed Aug 7 02:04:46 PM CEST 2024 +ELAPSED: 7 seconds diff --git a/epochX/cudacpp/tlau/fromgridpacks/gg_tt.mad/cpp512z/output.txt b/epochX/cudacpp/tlau/fromgridpacks/gg_tt.mad/cpp512z/output.txt new file mode 100644 index 0000000000..70c0ed00a0 --- /dev/null +++ b/epochX/cudacpp/tlau/fromgridpacks/gg_tt.mad/cpp512z/output.txt @@ -0,0 +1,49 @@ +START: Wed Aug 7 02:04:47 PM CEST 2024 +CUDACPP_RUNTIME_DISABLEFPE is not set +lrwxrwxrwx. 1 avalassi zg 35 Aug 7 14:04 madevent/SubProcesses/P1_gg_ttx/madevent -> build.512z_m_inl0_hrd0/madevent_cpp +Now generating 10000 events with random seed 21 and granularity 1 +************************************************************ +* * +* W E L C O M E to * +* M A D G R A P H 5 _ a M C @ N L O * +* M A D E V E N T * +* * +* * * * +* * * * * * +* * * * * 5 * * * * * +* * * * * * +* * * * +* * +* VERSION 3.5.3_lo_vect * +* * +* The MadGraph5_aMC@NLO Development Team - Find us at * +* https://server06.fynu.ucl.ac.be/projects/madgraph * +* * +* Type 'help' for in-line help. * +* * +************************************************************ +INFO: load configuration from /data/avalassi/GPU2024/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gg_tt.mad/madevent/Cards/me5_configuration.txt +INFO: load configuration from /data/avalassi/GPU2024/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gg_tt.mad/madevent/Cards/me5_configuration.txt +Using default text editor "vi". Set another one in ./input/mg5_configuration.txt +No valid eps viewer found. Please set in ./input/mg5_configuration.txt +Using default web browser "firefox". Set another one in ./input/mg5_configuration.txt +WRITE GRIDCARD /data/avalassi/GPU2024/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gg_tt.mad/madevent +generate 10000 events +P1_gg_ttx +Using random number seed offset = 21 +INFO: Generating 10000.0 unweighted events. +sum of cpu time of last step: 2 seconds +INFO: need to improve 2 channels +combine_events +store_events + === Results Summary for run: GridRun_21 tag: tag_1 === + + Cross-section : 0 +- 0 pb + Nb of events : 10000 + +INFO: No version of lhapdf. Can not run systematics computation +decay_events -from_cards +DONE +write ./events.lhe.gz +END: Wed Aug 7 02:04:54 PM CEST 2024 +ELAPSED: 7 seconds diff --git a/epochX/cudacpp/tlau/fromgridpacks/gg_tt.mad/cppavx2/output.txt b/epochX/cudacpp/tlau/fromgridpacks/gg_tt.mad/cppavx2/output.txt new file mode 100644 index 0000000000..d6e61c44c4 --- /dev/null +++ b/epochX/cudacpp/tlau/fromgridpacks/gg_tt.mad/cppavx2/output.txt @@ -0,0 +1,49 @@ +START: Wed Aug 7 02:04:31 PM CEST 2024 +CUDACPP_RUNTIME_DISABLEFPE is not set +lrwxrwxrwx. 1 avalassi zg 35 Aug 7 14:04 madevent/SubProcesses/P1_gg_ttx/madevent -> build.avx2_m_inl0_hrd0/madevent_cpp +Now generating 10000 events with random seed 21 and granularity 1 +************************************************************ +* * +* W E L C O M E to * +* M A D G R A P H 5 _ a M C @ N L O * +* M A D E V E N T * +* * +* * * * +* * * * * * +* * * * * 5 * * * * * +* * * * * * +* * * * +* * +* VERSION 3.5.3_lo_vect * +* * +* The MadGraph5_aMC@NLO Development Team - Find us at * +* https://server06.fynu.ucl.ac.be/projects/madgraph * +* * +* Type 'help' for in-line help. * +* * +************************************************************ +INFO: load configuration from /data/avalassi/GPU2024/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gg_tt.mad/madevent/Cards/me5_configuration.txt +INFO: load configuration from /data/avalassi/GPU2024/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gg_tt.mad/madevent/Cards/me5_configuration.txt +Using default text editor "vi". Set another one in ./input/mg5_configuration.txt +No valid eps viewer found. Please set in ./input/mg5_configuration.txt +Using default web browser "firefox". Set another one in ./input/mg5_configuration.txt +WRITE GRIDCARD /data/avalassi/GPU2024/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gg_tt.mad/madevent +generate 10000 events +P1_gg_ttx +Using random number seed offset = 21 +INFO: Generating 10000.0 unweighted events. +sum of cpu time of last step: 2 seconds +INFO: need to improve 2 channels +combine_events +store_events + === Results Summary for run: GridRun_21 tag: tag_1 === + + Cross-section : 0 +- 0 pb + Nb of events : 10000 + +INFO: No version of lhapdf. Can not run systematics computation +decay_events -from_cards +DONE +write ./events.lhe.gz +END: Wed Aug 7 02:04:38 PM CEST 2024 +ELAPSED: 7 seconds diff --git a/epochX/cudacpp/tlau/fromgridpacks/gg_tt.mad/cppnone/output.txt b/epochX/cudacpp/tlau/fromgridpacks/gg_tt.mad/cppnone/output.txt new file mode 100644 index 0000000000..698e0c92ea --- /dev/null +++ b/epochX/cudacpp/tlau/fromgridpacks/gg_tt.mad/cppnone/output.txt @@ -0,0 +1,49 @@ +START: Wed Aug 7 02:04:14 PM CEST 2024 +CUDACPP_RUNTIME_DISABLEFPE is not set +lrwxrwxrwx. 1 avalassi zg 35 Aug 7 14:04 madevent/SubProcesses/P1_gg_ttx/madevent -> build.none_m_inl0_hrd0/madevent_cpp +Now generating 10000 events with random seed 21 and granularity 1 +************************************************************ +* * +* W E L C O M E to * +* M A D G R A P H 5 _ a M C @ N L O * +* M A D E V E N T * +* * +* * * * +* * * * * * +* * * * * 5 * * * * * +* * * * * * +* * * * +* * +* VERSION 3.5.3_lo_vect * +* * +* The MadGraph5_aMC@NLO Development Team - Find us at * +* https://server06.fynu.ucl.ac.be/projects/madgraph * +* * +* Type 'help' for in-line help. * +* * +************************************************************ +INFO: load configuration from /data/avalassi/GPU2024/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gg_tt.mad/madevent/Cards/me5_configuration.txt +INFO: load configuration from /data/avalassi/GPU2024/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gg_tt.mad/madevent/Cards/me5_configuration.txt +Using default text editor "vi". Set another one in ./input/mg5_configuration.txt +No valid eps viewer found. Please set in ./input/mg5_configuration.txt +Using default web browser "firefox". Set another one in ./input/mg5_configuration.txt +WRITE GRIDCARD /data/avalassi/GPU2024/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gg_tt.mad/madevent +generate 10000 events +P1_gg_ttx +Using random number seed offset = 21 +INFO: Generating 10000.0 unweighted events. +sum of cpu time of last step: 2 seconds +INFO: need to improve 2 channels +combine_events +store_events + === Results Summary for run: GridRun_21 tag: tag_1 === + + Cross-section : 0 +- 0 pb + Nb of events : 10000 + +INFO: No version of lhapdf. Can not run systematics computation +decay_events -from_cards +DONE +write ./events.lhe.gz +END: Wed Aug 7 02:04:22 PM CEST 2024 +ELAPSED: 8 seconds diff --git a/epochX/cudacpp/tlau/fromgridpacks/gg_tt.mad/cppsse4/output.txt b/epochX/cudacpp/tlau/fromgridpacks/gg_tt.mad/cppsse4/output.txt new file mode 100644 index 0000000000..a11b1ce562 --- /dev/null +++ b/epochX/cudacpp/tlau/fromgridpacks/gg_tt.mad/cppsse4/output.txt @@ -0,0 +1,49 @@ +START: Wed Aug 7 02:04:23 PM CEST 2024 +CUDACPP_RUNTIME_DISABLEFPE is not set +lrwxrwxrwx. 1 avalassi zg 35 Aug 7 14:04 madevent/SubProcesses/P1_gg_ttx/madevent -> build.sse4_m_inl0_hrd0/madevent_cpp +Now generating 10000 events with random seed 21 and granularity 1 +************************************************************ +* * +* W E L C O M E to * +* M A D G R A P H 5 _ a M C @ N L O * +* M A D E V E N T * +* * +* * * * +* * * * * * +* * * * * 5 * * * * * +* * * * * * +* * * * +* * +* VERSION 3.5.3_lo_vect * +* * +* The MadGraph5_aMC@NLO Development Team - Find us at * +* https://server06.fynu.ucl.ac.be/projects/madgraph * +* * +* Type 'help' for in-line help. * +* * +************************************************************ +INFO: load configuration from /data/avalassi/GPU2024/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gg_tt.mad/madevent/Cards/me5_configuration.txt +INFO: load configuration from /data/avalassi/GPU2024/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gg_tt.mad/madevent/Cards/me5_configuration.txt +Using default text editor "vi". Set another one in ./input/mg5_configuration.txt +No valid eps viewer found. Please set in ./input/mg5_configuration.txt +Using default web browser "firefox". Set another one in ./input/mg5_configuration.txt +WRITE GRIDCARD /data/avalassi/GPU2024/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gg_tt.mad/madevent +generate 10000 events +P1_gg_ttx +Using random number seed offset = 21 +INFO: Generating 10000.0 unweighted events. +sum of cpu time of last step: 2 seconds +INFO: need to improve 2 channels +combine_events +store_events + === Results Summary for run: GridRun_21 tag: tag_1 === + + Cross-section : 0 +- 0 pb + Nb of events : 10000 + +INFO: No version of lhapdf. Can not run systematics computation +decay_events -from_cards +DONE +write ./events.lhe.gz +END: Wed Aug 7 02:04:30 PM CEST 2024 +ELAPSED: 7 seconds diff --git a/epochX/cudacpp/tlau/fromgridpacks/gg_tt.mad/fortran/output.txt b/epochX/cudacpp/tlau/fromgridpacks/gg_tt.mad/fortran/output.txt new file mode 100644 index 0000000000..72e3f12dc2 --- /dev/null +++ b/epochX/cudacpp/tlau/fromgridpacks/gg_tt.mad/fortran/output.txt @@ -0,0 +1,49 @@ +START: Wed Aug 7 02:04:05 PM CEST 2024 +CUDACPP_RUNTIME_DISABLEFPE is not set +lrwxrwxrwx. 1 avalassi zg 16 Aug 7 14:04 madevent/SubProcesses/P1_gg_ttx/madevent -> madevent_fortran +Now generating 10000 events with random seed 21 and granularity 1 +************************************************************ +* * +* W E L C O M E to * +* M A D G R A P H 5 _ a M C @ N L O * +* M A D E V E N T * +* * +* * * * +* * * * * * +* * * * * 5 * * * * * +* * * * * * +* * * * +* * +* VERSION 3.5.3_lo_vect * +* * +* The MadGraph5_aMC@NLO Development Team - Find us at * +* https://server06.fynu.ucl.ac.be/projects/madgraph * +* * +* Type 'help' for in-line help. * +* * +************************************************************ +INFO: load configuration from /data/avalassi/GPU2024/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gg_tt.mad/madevent/Cards/me5_configuration.txt +INFO: load configuration from /data/avalassi/GPU2024/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gg_tt.mad/madevent/Cards/me5_configuration.txt +Using default text editor "vi". Set another one in ./input/mg5_configuration.txt +No valid eps viewer found. Please set in ./input/mg5_configuration.txt +Using default web browser "firefox". Set another one in ./input/mg5_configuration.txt +WRITE GRIDCARD /data/avalassi/GPU2024/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gg_tt.mad/madevent +generate 10000 events +P1_gg_ttx +Using random number seed offset = 21 +INFO: Generating 10000.0 unweighted events. +sum of cpu time of last step: 2 seconds +INFO: need to improve 2 channels +combine_events +store_events + === Results Summary for run: GridRun_21 tag: tag_1 === + + Cross-section : 0 +- 0 pb + Nb of events : 10000 + +INFO: No version of lhapdf. Can not run systematics computation +decay_events -from_cards +DONE +write ./events.lhe.gz +END: Wed Aug 7 02:04:13 PM CEST 2024 +ELAPSED: 8 seconds diff --git a/epochX/cudacpp/tlau/fromgridpacks/gq_ttq.mad/cpp512y/output.txt b/epochX/cudacpp/tlau/fromgridpacks/gq_ttq.mad/cpp512y/output.txt new file mode 100644 index 0000000000..93fd0d9bd3 --- /dev/null +++ b/epochX/cudacpp/tlau/fromgridpacks/gq_ttq.mad/cpp512y/output.txt @@ -0,0 +1,111 @@ +START: Tue Sep 10 06:50:18 PM CEST 2024 +On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU (nproc=4)] [GPU: 1x Tesla V100S-PCIE-32GB]: +CUDACPP_RUNTIME_DISABLEFPE is not set +lrwxrwxrwx. 1 avalassi zg 35 Sep 10 18:50 madevent/SubProcesses/P1_gu_ttxu/madevent -> build.512y_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 10 18:50 madevent/SubProcesses/P1_gux_ttxux/madevent -> build.512y_m_inl0_hrd0/madevent_cpp +Now generating 100 events with random seed 21 and granularity 1 +************************************************************ +* * +* W E L C O M E to * +* M A D G R A P H 5 _ a M C @ N L O * +* M A D E V E N T * +* * +* * * * +* * * * * * +* * * * * 5 * * * * * +* * * * * * +* * * * +* * +* VERSION 3.5.3_lo_vect * +* * +* The MadGraph5_aMC@NLO Development Team - Find us at * +* https://server06.fynu.ucl.ac.be/projects/madgraph * +* * +* Type 'help' for in-line help. * +* * +************************************************************ +INFO: load configuration from /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/Cards/me5_configuration.txt +INFO: load configuration from /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/Cards/me5_configuration.txt +Using default text editor "vi". Set another one in ./input/mg5_configuration.txt +Using default eps viewer "evince". Set another one in ./input/mg5_configuration.txt +Using default web browser "firefox". Set another one in ./input/mg5_configuration.txt +WRITE GRIDCARD /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent +__CUDACPP_DEBUG: GridPackCmd.__init__ will call self.launch +__CUDACPP_DEBUG: GridPackCmd.launch starting +__CUDACPP_DEBUG: GridPackCmd.launch (1) restore_data +generate 100 events +P1_gu_ttxu +P1_gux_ttxux +__CUDACPP_DEBUG: GridPackCmd.launch (2) refine4grid +__CUDACPP_DEBUG: GridPackCmd.refine4grid starting +Using random number seed offset = 21 +INFO: Generating 100.0 unweighted events. +__CUDACPP_DEBUG: gen_ximprove.launch starting +sum of cpu time of last step: 11 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event starting +INFO: need to improve 7 channels +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gu_ttxu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gu_ttxu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gu_ttxu/ajob1' in 3.1240 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gu_ttxu/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gu_ttxu/G1/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 0.8520s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8077s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0435s for 16384 events => throughput is 3.76E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0008s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gu_ttxu/G2/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 0.5064s + [COUNTERS] Fortran Overhead ( 0 ) : 0.4615s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0440s for 16384 events => throughput is 3.72E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0008s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gu_ttxu/G4/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 0.8798s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8355s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0435s for 16384 events => throughput is 3.77E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0008s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gu_ttxu/G5/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 0.8299s + [COUNTERS] Fortran Overhead ( 0 ) : 0.7845s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0446s for 16384 events => throughput is 3.67E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0008s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gu_ttxu/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gux_ttxux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gux_ttxux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gux_ttxux/ajob1' in 2.4848 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gux_ttxux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gux_ttxux/G1/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 0.8803s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8349s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0446s for 16384 events => throughput is 3.67E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0009s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gux_ttxux/G2/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 0.6988s + [COUNTERS] Fortran Overhead ( 0 ) : 0.6538s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0441s for 16384 events => throughput is 3.72E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0009s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gux_ttxux/G5/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 0.8624s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8175s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0440s for 16384 events => throughput is 3.72E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0008s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gux_ttxux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished in 5.6128 seconds +__CUDACPP_DEBUG: gen_ximprove.launch finished in 5.6162 seconds +__CUDACPP_DEBUG: GridPackCmd.refine4grid finished in 5.6169 seconds +__CUDACPP_DEBUG: GridPackCmd.launch (3a) combine_events +combine_events +__CUDACPP_DEBUG: GridPackCmd.launch (3b) store_events +store_events + === Results Summary for run: GridRun_21 tag: tag_1 === + + Cross-section : 0 +- 0 pb + Nb of events : 100 + +INFO: No version of lhapdf. Can not run systematics computation +decay_events -from_cards +__CUDACPP_DEBUG: GridPackCmd.launch finished in 6.3790 seconds +__CUDACPP_DEBUG: GridPackCmd.__init__ back from self.launch +DONE +write ./events.lhe.gz +END: Tue Sep 10 06:50:26 PM CEST 2024 +ELAPSED: 8 seconds diff --git a/epochX/cudacpp/tlau/fromgridpacks/gq_ttq.mad/cpp512z/output.txt b/epochX/cudacpp/tlau/fromgridpacks/gq_ttq.mad/cpp512z/output.txt new file mode 100644 index 0000000000..e80e0c135b --- /dev/null +++ b/epochX/cudacpp/tlau/fromgridpacks/gq_ttq.mad/cpp512z/output.txt @@ -0,0 +1,111 @@ +START: Tue Sep 10 06:50:27 PM CEST 2024 +On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU (nproc=4)] [GPU: 1x Tesla V100S-PCIE-32GB]: +CUDACPP_RUNTIME_DISABLEFPE is not set +lrwxrwxrwx. 1 avalassi zg 35 Sep 10 18:50 madevent/SubProcesses/P1_gu_ttxu/madevent -> build.512z_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 10 18:50 madevent/SubProcesses/P1_gux_ttxux/madevent -> build.512z_m_inl0_hrd0/madevent_cpp +Now generating 100 events with random seed 21 and granularity 1 +************************************************************ +* * +* W E L C O M E to * +* M A D G R A P H 5 _ a M C @ N L O * +* M A D E V E N T * +* * +* * * * +* * * * * * +* * * * * 5 * * * * * +* * * * * * +* * * * +* * +* VERSION 3.5.3_lo_vect * +* * +* The MadGraph5_aMC@NLO Development Team - Find us at * +* https://server06.fynu.ucl.ac.be/projects/madgraph * +* * +* Type 'help' for in-line help. * +* * +************************************************************ +INFO: load configuration from /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/Cards/me5_configuration.txt +INFO: load configuration from /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/Cards/me5_configuration.txt +Using default text editor "vi". Set another one in ./input/mg5_configuration.txt +Using default eps viewer "evince". Set another one in ./input/mg5_configuration.txt +Using default web browser "firefox". Set another one in ./input/mg5_configuration.txt +WRITE GRIDCARD /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent +__CUDACPP_DEBUG: GridPackCmd.__init__ will call self.launch +__CUDACPP_DEBUG: GridPackCmd.launch starting +__CUDACPP_DEBUG: GridPackCmd.launch (1) restore_data +generate 100 events +P1_gu_ttxu +P1_gux_ttxux +__CUDACPP_DEBUG: GridPackCmd.launch (2) refine4grid +__CUDACPP_DEBUG: GridPackCmd.refine4grid starting +Using random number seed offset = 21 +INFO: Generating 100.0 unweighted events. +__CUDACPP_DEBUG: gen_ximprove.launch starting +sum of cpu time of last step: 11 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event starting +INFO: need to improve 7 channels +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gu_ttxu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gu_ttxu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gu_ttxu/ajob1' in 3.2303 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gu_ttxu/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gu_ttxu/G1/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 0.8830s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8117s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0702s for 16384 events => throughput is 2.33E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0010s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gu_ttxu/G2/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 0.5345s + [COUNTERS] Fortran Overhead ( 0 ) : 0.4632s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0703s for 16384 events => throughput is 2.33E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0010s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gu_ttxu/G4/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 0.9012s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8304s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0699s for 16384 events => throughput is 2.34E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0010s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gu_ttxu/G5/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 0.8560s + [COUNTERS] Fortran Overhead ( 0 ) : 0.7846s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0703s for 16384 events => throughput is 2.33E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0010s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gu_ttxu/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gux_ttxux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gux_ttxux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gux_ttxux/ajob1' in 2.5720 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gux_ttxux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gux_ttxux/G1/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 0.9186s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8466s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0710s for 16384 events => throughput is 2.31E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0010s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gux_ttxux/G2/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 0.7194s + [COUNTERS] Fortran Overhead ( 0 ) : 0.6484s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0700s for 16384 events => throughput is 2.34E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0009s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gux_ttxux/G5/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 0.8905s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8178s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0717s for 16384 events => throughput is 2.28E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0010s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gux_ttxux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished in 5.8065 seconds +__CUDACPP_DEBUG: gen_ximprove.launch finished in 5.8100 seconds +__CUDACPP_DEBUG: GridPackCmd.refine4grid finished in 5.8108 seconds +__CUDACPP_DEBUG: GridPackCmd.launch (3a) combine_events +combine_events +__CUDACPP_DEBUG: GridPackCmd.launch (3b) store_events +store_events + === Results Summary for run: GridRun_21 tag: tag_1 === + + Cross-section : 0 +- 0 pb + Nb of events : 100 + +INFO: No version of lhapdf. Can not run systematics computation +decay_events -from_cards +__CUDACPP_DEBUG: GridPackCmd.launch finished in 6.5685 seconds +__CUDACPP_DEBUG: GridPackCmd.__init__ back from self.launch +DONE +write ./events.lhe.gz +END: Tue Sep 10 06:50:36 PM CEST 2024 +ELAPSED: 9 seconds diff --git a/epochX/cudacpp/tlau/fromgridpacks/gq_ttq.mad/cppavx2/output.txt b/epochX/cudacpp/tlau/fromgridpacks/gq_ttq.mad/cppavx2/output.txt new file mode 100644 index 0000000000..1a939ad250 --- /dev/null +++ b/epochX/cudacpp/tlau/fromgridpacks/gq_ttq.mad/cppavx2/output.txt @@ -0,0 +1,111 @@ +START: Tue Sep 10 06:50:08 PM CEST 2024 +On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU (nproc=4)] [GPU: 1x Tesla V100S-PCIE-32GB]: +CUDACPP_RUNTIME_DISABLEFPE is not set +lrwxrwxrwx. 1 avalassi zg 35 Sep 10 18:50 madevent/SubProcesses/P1_gu_ttxu/madevent -> build.avx2_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 10 18:50 madevent/SubProcesses/P1_gux_ttxux/madevent -> build.avx2_m_inl0_hrd0/madevent_cpp +Now generating 100 events with random seed 21 and granularity 1 +************************************************************ +* * +* W E L C O M E to * +* M A D G R A P H 5 _ a M C @ N L O * +* M A D E V E N T * +* * +* * * * +* * * * * * +* * * * * 5 * * * * * +* * * * * * +* * * * +* * +* VERSION 3.5.3_lo_vect * +* * +* The MadGraph5_aMC@NLO Development Team - Find us at * +* https://server06.fynu.ucl.ac.be/projects/madgraph * +* * +* Type 'help' for in-line help. * +* * +************************************************************ +INFO: load configuration from /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/Cards/me5_configuration.txt +INFO: load configuration from /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/Cards/me5_configuration.txt +Using default text editor "vi". Set another one in ./input/mg5_configuration.txt +Using default eps viewer "evince". Set another one in ./input/mg5_configuration.txt +Using default web browser "firefox". Set another one in ./input/mg5_configuration.txt +WRITE GRIDCARD /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent +__CUDACPP_DEBUG: GridPackCmd.__init__ will call self.launch +__CUDACPP_DEBUG: GridPackCmd.launch starting +__CUDACPP_DEBUG: GridPackCmd.launch (1) restore_data +generate 100 events +P1_gu_ttxu +P1_gux_ttxux +__CUDACPP_DEBUG: GridPackCmd.launch (2) refine4grid +__CUDACPP_DEBUG: GridPackCmd.refine4grid starting +Using random number seed offset = 21 +INFO: Generating 100.0 unweighted events. +__CUDACPP_DEBUG: gen_ximprove.launch starting +sum of cpu time of last step: 11 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event starting +INFO: need to improve 7 channels +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gu_ttxu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gu_ttxu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gu_ttxu/ajob1' in 3.1393 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gu_ttxu/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gu_ttxu/G1/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 0.8550s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8058s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0483s for 16384 events => throughput is 3.39E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0009s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gu_ttxu/G2/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 0.5118s + [COUNTERS] Fortran Overhead ( 0 ) : 0.4618s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0492s for 16384 events => throughput is 3.33E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0008s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gu_ttxu/G4/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 0.8867s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8364s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0495s for 16384 events => throughput is 3.31E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0009s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gu_ttxu/G5/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 0.8300s + [COUNTERS] Fortran Overhead ( 0 ) : 0.7794s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0496s for 16384 events => throughput is 3.30E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0009s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gu_ttxu/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gux_ttxux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gux_ttxux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gux_ttxux/ajob1' in 2.4929 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gux_ttxux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gux_ttxux/G1/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 0.8829s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8328s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0493s for 16384 events => throughput is 3.33E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0008s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gux_ttxux/G2/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 0.6995s + [COUNTERS] Fortran Overhead ( 0 ) : 0.6495s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0492s for 16384 events => throughput is 3.33E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0008s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gux_ttxux/G5/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 0.8674s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8166s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0500s for 16384 events => throughput is 3.28E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0008s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gux_ttxux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished in 5.6362 seconds +__CUDACPP_DEBUG: gen_ximprove.launch finished in 5.6395 seconds +__CUDACPP_DEBUG: GridPackCmd.refine4grid finished in 5.6401 seconds +__CUDACPP_DEBUG: GridPackCmd.launch (3a) combine_events +combine_events +__CUDACPP_DEBUG: GridPackCmd.launch (3b) store_events +store_events + === Results Summary for run: GridRun_21 tag: tag_1 === + + Cross-section : 0 +- 0 pb + Nb of events : 100 + +INFO: No version of lhapdf. Can not run systematics computation +decay_events -from_cards +__CUDACPP_DEBUG: GridPackCmd.launch finished in 6.3991 seconds +__CUDACPP_DEBUG: GridPackCmd.__init__ back from self.launch +DONE +write ./events.lhe.gz +END: Tue Sep 10 06:50:17 PM CEST 2024 +ELAPSED: 9 seconds diff --git a/epochX/cudacpp/tlau/fromgridpacks/gq_ttq.mad/cppnone/output.txt b/epochX/cudacpp/tlau/fromgridpacks/gq_ttq.mad/cppnone/output.txt new file mode 100644 index 0000000000..ff522f5e82 --- /dev/null +++ b/epochX/cudacpp/tlau/fromgridpacks/gq_ttq.mad/cppnone/output.txt @@ -0,0 +1,111 @@ +START: Tue Sep 10 06:49:49 PM CEST 2024 +On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU (nproc=4)] [GPU: 1x Tesla V100S-PCIE-32GB]: +CUDACPP_RUNTIME_DISABLEFPE is not set +lrwxrwxrwx. 1 avalassi zg 35 Sep 10 18:49 madevent/SubProcesses/P1_gu_ttxu/madevent -> build.none_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 10 18:49 madevent/SubProcesses/P1_gux_ttxux/madevent -> build.none_m_inl0_hrd0/madevent_cpp +Now generating 100 events with random seed 21 and granularity 1 +************************************************************ +* * +* W E L C O M E to * +* M A D G R A P H 5 _ a M C @ N L O * +* M A D E V E N T * +* * +* * * * +* * * * * * +* * * * * 5 * * * * * +* * * * * * +* * * * +* * +* VERSION 3.5.3_lo_vect * +* * +* The MadGraph5_aMC@NLO Development Team - Find us at * +* https://server06.fynu.ucl.ac.be/projects/madgraph * +* * +* Type 'help' for in-line help. * +* * +************************************************************ +INFO: load configuration from /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/Cards/me5_configuration.txt +INFO: load configuration from /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/Cards/me5_configuration.txt +Using default text editor "vi". Set another one in ./input/mg5_configuration.txt +Using default eps viewer "evince". Set another one in ./input/mg5_configuration.txt +Using default web browser "firefox". Set another one in ./input/mg5_configuration.txt +WRITE GRIDCARD /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent +__CUDACPP_DEBUG: GridPackCmd.__init__ will call self.launch +__CUDACPP_DEBUG: GridPackCmd.launch starting +__CUDACPP_DEBUG: GridPackCmd.launch (1) restore_data +generate 100 events +P1_gu_ttxu +P1_gux_ttxux +__CUDACPP_DEBUG: GridPackCmd.launch (2) refine4grid +__CUDACPP_DEBUG: GridPackCmd.refine4grid starting +Using random number seed offset = 21 +INFO: Generating 100.0 unweighted events. +__CUDACPP_DEBUG: gen_ximprove.launch starting +sum of cpu time of last step: 11 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event starting +INFO: need to improve 7 channels +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gu_ttxu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gu_ttxu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gu_ttxu/ajob1' in 3.5757 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gu_ttxu/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gu_ttxu/G1/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 0.9675s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8114s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1550s for 16384 events => throughput is 1.06E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0011s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gu_ttxu/G2/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 0.6159s + [COUNTERS] Fortran Overhead ( 0 ) : 0.4584s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1564s for 16384 events => throughput is 1.05E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0011s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gu_ttxu/G4/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 0.9866s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8304s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1551s for 16384 events => throughput is 1.06E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0011s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gu_ttxu/G5/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 0.9501s + [COUNTERS] Fortran Overhead ( 0 ) : 0.7933s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1557s for 16384 events => throughput is 1.05E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0011s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gu_ttxu/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gux_ttxux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gux_ttxux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gux_ttxux/ajob1' in 2.8133 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gux_ttxux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gux_ttxux/G1/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 0.9933s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8379s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1543s for 16384 events => throughput is 1.06E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0011s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gux_ttxux/G2/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 0.8048s + [COUNTERS] Fortran Overhead ( 0 ) : 0.6488s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1549s for 16384 events => throughput is 1.06E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0011s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gux_ttxux/G5/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 0.9721s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8154s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1556s for 16384 events => throughput is 1.05E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0011s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gux_ttxux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished in 6.3933 seconds +__CUDACPP_DEBUG: gen_ximprove.launch finished in 6.3966 seconds +__CUDACPP_DEBUG: GridPackCmd.refine4grid finished in 6.3974 seconds +__CUDACPP_DEBUG: GridPackCmd.launch (3a) combine_events +combine_events +__CUDACPP_DEBUG: GridPackCmd.launch (3b) store_events +store_events + === Results Summary for run: GridRun_21 tag: tag_1 === + + Cross-section : 0 +- 0 pb + Nb of events : 100 + +INFO: No version of lhapdf. Can not run systematics computation +decay_events -from_cards +__CUDACPP_DEBUG: GridPackCmd.launch finished in 7.1612 seconds +__CUDACPP_DEBUG: GridPackCmd.__init__ back from self.launch +DONE +write ./events.lhe.gz +END: Tue Sep 10 06:49:58 PM CEST 2024 +ELAPSED: 9 seconds diff --git a/epochX/cudacpp/tlau/fromgridpacks/gq_ttq.mad/cppsse4/output.txt b/epochX/cudacpp/tlau/fromgridpacks/gq_ttq.mad/cppsse4/output.txt new file mode 100644 index 0000000000..f00c91ca5b --- /dev/null +++ b/epochX/cudacpp/tlau/fromgridpacks/gq_ttq.mad/cppsse4/output.txt @@ -0,0 +1,111 @@ +START: Tue Sep 10 06:49:59 PM CEST 2024 +On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU (nproc=4)] [GPU: 1x Tesla V100S-PCIE-32GB]: +CUDACPP_RUNTIME_DISABLEFPE is not set +lrwxrwxrwx. 1 avalassi zg 35 Sep 10 18:49 madevent/SubProcesses/P1_gu_ttxu/madevent -> build.sse4_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 10 18:49 madevent/SubProcesses/P1_gux_ttxux/madevent -> build.sse4_m_inl0_hrd0/madevent_cpp +Now generating 100 events with random seed 21 and granularity 1 +************************************************************ +* * +* W E L C O M E to * +* M A D G R A P H 5 _ a M C @ N L O * +* M A D E V E N T * +* * +* * * * +* * * * * * +* * * * * 5 * * * * * +* * * * * * +* * * * +* * +* VERSION 3.5.3_lo_vect * +* * +* The MadGraph5_aMC@NLO Development Team - Find us at * +* https://server06.fynu.ucl.ac.be/projects/madgraph * +* * +* Type 'help' for in-line help. * +* * +************************************************************ +INFO: load configuration from /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/Cards/me5_configuration.txt +INFO: load configuration from /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/Cards/me5_configuration.txt +Using default text editor "vi". Set another one in ./input/mg5_configuration.txt +Using default eps viewer "evince". Set another one in ./input/mg5_configuration.txt +Using default web browser "firefox". Set another one in ./input/mg5_configuration.txt +WRITE GRIDCARD /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent +__CUDACPP_DEBUG: GridPackCmd.__init__ will call self.launch +__CUDACPP_DEBUG: GridPackCmd.launch starting +__CUDACPP_DEBUG: GridPackCmd.launch (1) restore_data +generate 100 events +P1_gu_ttxu +P1_gux_ttxux +__CUDACPP_DEBUG: GridPackCmd.launch (2) refine4grid +__CUDACPP_DEBUG: GridPackCmd.refine4grid starting +Using random number seed offset = 21 +INFO: Generating 100.0 unweighted events. +__CUDACPP_DEBUG: gen_ximprove.launch starting +sum of cpu time of last step: 11 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event starting +INFO: need to improve 7 channels +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gu_ttxu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gu_ttxu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gu_ttxu/ajob1' in 3.3146 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gu_ttxu/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gu_ttxu/G1/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 0.9114s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8242s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0863s for 16384 events => throughput is 1.90E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0009s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gu_ttxu/G2/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 0.5570s + [COUNTERS] Fortran Overhead ( 0 ) : 0.4658s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0903s for 16384 events => throughput is 1.81E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0009s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gu_ttxu/G4/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 0.9223s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8355s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0859s for 16384 events => throughput is 1.91E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0009s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gu_ttxu/G5/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 0.8674s + [COUNTERS] Fortran Overhead ( 0 ) : 0.7821s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0845s for 16384 events => throughput is 1.94E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0009s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gu_ttxu/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gux_ttxux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gux_ttxux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gux_ttxux/ajob1' in 2.6299 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gux_ttxux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gux_ttxux/G1/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 0.9245s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8389s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0847s for 16384 events => throughput is 1.93E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0009s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gux_ttxux/G2/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 0.7556s + [COUNTERS] Fortran Overhead ( 0 ) : 0.6633s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0915s for 16384 events => throughput is 1.79E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0008s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gux_ttxux/G5/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 0.9059s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8201s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0849s for 16384 events => throughput is 1.93E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0009s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gux_ttxux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished in 5.9486 seconds +__CUDACPP_DEBUG: gen_ximprove.launch finished in 5.9519 seconds +__CUDACPP_DEBUG: GridPackCmd.refine4grid finished in 5.9526 seconds +__CUDACPP_DEBUG: GridPackCmd.launch (3a) combine_events +combine_events +__CUDACPP_DEBUG: GridPackCmd.launch (3b) store_events +store_events + === Results Summary for run: GridRun_21 tag: tag_1 === + + Cross-section : 0 +- 0 pb + Nb of events : 100 + +INFO: No version of lhapdf. Can not run systematics computation +decay_events -from_cards +__CUDACPP_DEBUG: GridPackCmd.launch finished in 6.7104 seconds +__CUDACPP_DEBUG: GridPackCmd.__init__ back from self.launch +DONE +write ./events.lhe.gz +END: Tue Sep 10 06:50:08 PM CEST 2024 +ELAPSED: 9 seconds diff --git a/epochX/cudacpp/tlau/fromgridpacks/gq_ttq.mad/cuda/output.txt b/epochX/cudacpp/tlau/fromgridpacks/gq_ttq.mad/cuda/output.txt new file mode 100644 index 0000000000..3b5b530d88 --- /dev/null +++ b/epochX/cudacpp/tlau/fromgridpacks/gq_ttq.mad/cuda/output.txt @@ -0,0 +1,111 @@ +START: Tue Sep 10 06:49:35 PM CEST 2024 +On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU (nproc=4)] [GPU: 1x Tesla V100S-PCIE-32GB]: +CUDACPP_RUNTIME_DISABLEFPE is not set +lrwxrwxrwx. 1 avalassi zg 36 Sep 10 18:49 madevent/SubProcesses/P1_gu_ttxu/madevent -> build.cuda_m_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 10 18:49 madevent/SubProcesses/P1_gux_ttxux/madevent -> build.cuda_m_inl0_hrd0/madevent_cuda +Now generating 100 events with random seed 21 and granularity 1 +************************************************************ +* * +* W E L C O M E to * +* M A D G R A P H 5 _ a M C @ N L O * +* M A D E V E N T * +* * +* * * * +* * * * * * +* * * * * 5 * * * * * +* * * * * * +* * * * +* * +* VERSION 3.5.3_lo_vect * +* * +* The MadGraph5_aMC@NLO Development Team - Find us at * +* https://server06.fynu.ucl.ac.be/projects/madgraph * +* * +* Type 'help' for in-line help. * +* * +************************************************************ +INFO: load configuration from /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/Cards/me5_configuration.txt +INFO: load configuration from /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/Cards/me5_configuration.txt +Using default text editor "vi". Set another one in ./input/mg5_configuration.txt +Using default eps viewer "evince". Set another one in ./input/mg5_configuration.txt +Using default web browser "firefox". Set another one in ./input/mg5_configuration.txt +WRITE GRIDCARD /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent +__CUDACPP_DEBUG: GridPackCmd.__init__ will call self.launch +__CUDACPP_DEBUG: GridPackCmd.launch starting +__CUDACPP_DEBUG: GridPackCmd.launch (1) restore_data +generate 100 events +P1_gu_ttxu +P1_gux_ttxux +__CUDACPP_DEBUG: GridPackCmd.launch (2) refine4grid +__CUDACPP_DEBUG: GridPackCmd.refine4grid starting +Using random number seed offset = 21 +INFO: Generating 100.0 unweighted events. +__CUDACPP_DEBUG: gen_ximprove.launch starting +sum of cpu time of last step: 11 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event starting +INFO: need to improve 7 channels +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gu_ttxu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gu_ttxu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gu_ttxu/ajob1' in 5.6998 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gu_ttxu/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gu_ttxu/G1/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.2597s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2546s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0035s for 16384 events => throughput is 4.66E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0016s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gu_ttxu/G2/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 0.9082s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9031s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0035s for 16384 events => throughput is 4.65E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0016s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gu_ttxu/G4/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.2897s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2846s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0035s for 16384 events => throughput is 4.70E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0016s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gu_ttxu/G5/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.2350s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2298s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0036s for 16384 events => throughput is 4.59E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0016s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gu_ttxu/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gux_ttxux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gux_ttxux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gux_ttxux/ajob1' in 4.4179 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gux_ttxux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gux_ttxux/G1/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.2861s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2810s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0035s for 16384 events => throughput is 4.68E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0016s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gux_ttxux/G2/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.1069s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1018s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0035s for 16384 events => throughput is 4.66E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0016s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gux_ttxux/G5/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.2770s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2718s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0036s for 16384 events => throughput is 4.55E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0016s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gux_ttxux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished in 10.1217 seconds +__CUDACPP_DEBUG: gen_ximprove.launch finished in 10.1250 seconds +__CUDACPP_DEBUG: GridPackCmd.refine4grid finished in 10.1257 seconds +__CUDACPP_DEBUG: GridPackCmd.launch (3a) combine_events +combine_events +__CUDACPP_DEBUG: GridPackCmd.launch (3b) store_events +store_events + === Results Summary for run: GridRun_21 tag: tag_1 === + + Cross-section : 0 +- 0 pb + Nb of events : 100 + +INFO: No version of lhapdf. Can not run systematics computation +decay_events -from_cards +__CUDACPP_DEBUG: GridPackCmd.launch finished in 10.8831 seconds +__CUDACPP_DEBUG: GridPackCmd.__init__ back from self.launch +DONE +write ./events.lhe.gz +END: Tue Sep 10 06:49:48 PM CEST 2024 +ELAPSED: 13 seconds diff --git a/epochX/cudacpp/tlau/fromgridpacks/gq_ttq.mad/fortran/output.txt b/epochX/cudacpp/tlau/fromgridpacks/gq_ttq.mad/fortran/output.txt new file mode 100644 index 0000000000..872a442b5b --- /dev/null +++ b/epochX/cudacpp/tlau/fromgridpacks/gq_ttq.mad/fortran/output.txt @@ -0,0 +1,104 @@ +START: Tue Sep 10 06:49:25 PM CEST 2024 +On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU (nproc=4)] [GPU: 1x Tesla V100S-PCIE-32GB]: +CUDACPP_RUNTIME_DISABLEFPE is not set +lrwxrwxrwx. 1 avalassi zg 16 Sep 10 18:49 madevent/SubProcesses/P1_gu_ttxu/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 10 18:49 madevent/SubProcesses/P1_gux_ttxux/madevent -> madevent_fortran +Now generating 100 events with random seed 21 and granularity 1 +************************************************************ +* * +* W E L C O M E to * +* M A D G R A P H 5 _ a M C @ N L O * +* M A D E V E N T * +* * +* * * * +* * * * * * +* * * * * 5 * * * * * +* * * * * * +* * * * +* * +* VERSION 3.5.3_lo_vect * +* * +* The MadGraph5_aMC@NLO Development Team - Find us at * +* https://server06.fynu.ucl.ac.be/projects/madgraph * +* * +* Type 'help' for in-line help. * +* * +************************************************************ +INFO: load configuration from /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/Cards/me5_configuration.txt +INFO: load configuration from /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/Cards/me5_configuration.txt +Using default text editor "vi". Set another one in ./input/mg5_configuration.txt +Using default eps viewer "evince". Set another one in ./input/mg5_configuration.txt +Using default web browser "firefox". Set another one in ./input/mg5_configuration.txt +WRITE GRIDCARD /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent +__CUDACPP_DEBUG: GridPackCmd.__init__ will call self.launch +__CUDACPP_DEBUG: GridPackCmd.launch starting +__CUDACPP_DEBUG: GridPackCmd.launch (1) restore_data +generate 100 events +P1_gu_ttxu +P1_gux_ttxux +__CUDACPP_DEBUG: GridPackCmd.launch (2) refine4grid +__CUDACPP_DEBUG: GridPackCmd.refine4grid starting +Using random number seed offset = 21 +INFO: Generating 100.0 unweighted events. +__CUDACPP_DEBUG: gen_ximprove.launch starting +sum of cpu time of last step: 11 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event starting +INFO: need to improve 7 channels +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gu_ttxu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gu_ttxu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gu_ttxu/ajob1' in 3.4858 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gu_ttxu/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gu_ttxu/G1/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 0.9409s + [COUNTERS] Fortran Overhead ( 0 ) : 0.7974s + [COUNTERS] Fortran MEs ( 1 ) : 0.1435s for 16384 events => throughput is 1.14E+05 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gu_ttxu/G2/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 0.5982s + [COUNTERS] Fortran Overhead ( 0 ) : 0.4547s + [COUNTERS] Fortran MEs ( 1 ) : 0.1435s for 16384 events => throughput is 1.14E+05 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gu_ttxu/G4/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 0.9685s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8241s + [COUNTERS] Fortran MEs ( 1 ) : 0.1443s for 16384 events => throughput is 1.14E+05 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gu_ttxu/G5/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 0.9238s + [COUNTERS] Fortran Overhead ( 0 ) : 0.7792s + [COUNTERS] Fortran MEs ( 1 ) : 0.1446s for 16384 events => throughput is 1.13E+05 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gu_ttxu/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gux_ttxux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gux_ttxux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gux_ttxux/ajob1' in 2.7783 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gux_ttxux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gux_ttxux/G1/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 0.9824s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8352s + [COUNTERS] Fortran MEs ( 1 ) : 0.1471s for 16384 events => throughput is 1.11E+05 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gux_ttxux/G2/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 0.7935s + [COUNTERS] Fortran Overhead ( 0 ) : 0.6514s + [COUNTERS] Fortran MEs ( 1 ) : 0.1421s for 16384 events => throughput is 1.15E+05 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gux_ttxux/G5/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 0.9600s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8168s + [COUNTERS] Fortran MEs ( 1 ) : 0.1431s for 16384 events => throughput is 1.14E+05 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/madevent/SubProcesses/P1_gux_ttxux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished in 6.2682 seconds +__CUDACPP_DEBUG: gen_ximprove.launch finished in 6.2715 seconds +__CUDACPP_DEBUG: GridPackCmd.refine4grid finished in 6.2721 seconds +__CUDACPP_DEBUG: GridPackCmd.launch (3a) combine_events +combine_events +__CUDACPP_DEBUG: GridPackCmd.launch (3b) store_events +store_events + === Results Summary for run: GridRun_21 tag: tag_1 === + + Cross-section : 0 +- 0 pb + Nb of events : 100 + +INFO: No version of lhapdf. Can not run systematics computation +decay_events -from_cards +__CUDACPP_DEBUG: GridPackCmd.launch finished in 7.0299 seconds +__CUDACPP_DEBUG: GridPackCmd.__init__ back from self.launch +DONE +write ./events.lhe.gz +END: Tue Sep 10 06:49:34 PM CEST 2024 +ELAPSED: 9 seconds diff --git a/epochX/cudacpp/tlau/fromgridpacks/gq_ttq.mad/summary.txt b/epochX/cudacpp/tlau/fromgridpacks/gq_ttq.mad/summary.txt new file mode 100644 index 0000000000..4ee4e96148 --- /dev/null +++ b/epochX/cudacpp/tlau/fromgridpacks/gq_ttq.mad/summary.txt @@ -0,0 +1,38 @@ +gq_ttq.mad//fortran/output.txt +[GridPackCmd.launch] GRIDPCK TOTAL 6.0845 +[madevent COUNTERS] PROGRAM TOTAL 5.3412s +[madevent COUNTERS] Fortran MEs 0.8371s for 98304 events +------------------------------------------------------------------------------------------ +gq_ttq.mad//cppnone/output.txt +[GridPackCmd.launch] GRIDPCK TOTAL 6.2165 +[madevent COUNTERS] PROGRAM TOTAL 5.4627s +[madevent COUNTERS] CudaCpp MEs 0.8986s for 98304 events +------------------------------------------------------------------------------------------ +gq_ttq.mad//cppsse4/output.txt +[GridPackCmd.launch] GRIDPCK TOTAL 5.7925 +[madevent COUNTERS] PROGRAM TOTAL 5.0438s +[madevent COUNTERS] CudaCpp MEs 0.4927s for 98304 events +------------------------------------------------------------------------------------------ +gq_ttq.mad//cppavx2/output.txt +[GridPackCmd.launch] GRIDPCK TOTAL 5.6014 +[madevent COUNTERS] PROGRAM TOTAL 4.8507s +[madevent COUNTERS] CudaCpp MEs 0.2861s for 98304 events +------------------------------------------------------------------------------------------ +gq_ttq.mad//cpp512y/output.txt +[GridPackCmd.launch] GRIDPCK TOTAL 5.5618 +[madevent COUNTERS] PROGRAM TOTAL 4.8145s +[madevent COUNTERS] CudaCpp MEs 0.2568s for 98304 events +------------------------------------------------------------------------------------------ +gq_ttq.mad//cpp512z/output.txt +[GridPackCmd.launch] GRIDPCK TOTAL 5.7225 +[madevent COUNTERS] PROGRAM TOTAL 4.9635s +[madevent COUNTERS] CudaCpp MEs 0.4014s for 98304 events +------------------------------------------------------------------------------------------ +gq_ttq.mad//cuda/output.txt +[GridPackCmd.launch] GRIDPCK TOTAL 9.3365 +[madevent COUNTERS] PROGRAM TOTAL 7.1848s +[madevent COUNTERS] CudaCpp MEs 0.0066s for 98304 events +------------------------------------------------------------------------------------------ +gq_ttq.mad//hip/output.txt +File not found: SKIP backend hip +------------------------------------------------------------------------------------------ diff --git a/epochX/cudacpp/tlau/fromgridpacks/parseGridpackLogs.sh b/epochX/cudacpp/tlau/fromgridpacks/parseGridpackLogs.sh new file mode 100755 index 0000000000..2367c5dbe0 --- /dev/null +++ b/epochX/cudacpp/tlau/fromgridpacks/parseGridpackLogs.sh @@ -0,0 +1,31 @@ +#!/bin/sh + +if [ "$1" == "" ] || [ "$2" != "" ]; then + echo "Usage: $0 " + echo "Example: $0 pp_dy012j.mad" + exit 1 +fi + +procdir=$1 +if [ ! -d $procdir ]; then + echo "ERROR! Directory not found $procdir" + exit 1 +fi + +for backend in fortran cppnone cppsse4 cppavx2 cpp512y cpp512z cuda hip; do + outfile=$procdir/${backend}/output.txt + outfile=${outfile/\/\///} + if [ ! -f $outfile ]; then + echo $outfile + echo "File not found: SKIP backend ${backend}" + else + echo $outfile $(cat $outfile | awk '/events :/{printf "(#events: %d)\n", $5}') + cat $outfile | grep "__CUDACPP_DEBUG: GridPackCmd.launch finished" \ + | sed 's/__CUDACPP_DEBUG: GridPackCmd.launch finished in/[GridPackCmd.launch] OVERALL TOTAL /' + for msg in "PROGRAM TOTAL " "Fortran Overhead" "Fortran MEs " "CudaCpp MEs " "CudaCpp HEL "; do + cat $outfile | grep "\[COUNTERS\]" | grep "${msg}" | sed 's/s for.*//' | sed 's/s$//' \ + | awk -vmsg="${msg}" -vttot=0 '{jtot=$NF; ttot += jtot}; END{if ( ttot!=0 ) print "[madevent COUNTERS] ", msg, ttot}' + done + fi + echo "--------------------------------------------------------------------------------" +done diff --git a/epochX/cudacpp/tlau/fromgridpacks/pp_dy3j.mad/cpp512y/output.txt b/epochX/cudacpp/tlau/fromgridpacks/pp_dy3j.mad/cpp512y/output.txt new file mode 100644 index 0000000000..c34aa401a1 --- /dev/null +++ b/epochX/cudacpp/tlau/fromgridpacks/pp_dy3j.mad/cpp512y/output.txt @@ -0,0 +1,2336 @@ +START: Wed Sep 11 03:46:38 AM CEST 2024 +On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU (nproc=4)] [GPU: 1x Tesla V100S-PCIE-32GB]: +CUDACPP_RUNTIME_DISABLEFPE is set +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:46 madevent/SubProcesses/P0_dc_epemgdc/madevent -> build.512y_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:46 madevent/SubProcesses/P0_dc_taptamgdc/madevent -> build.512y_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:46 madevent/SubProcesses/P0_dd_epemgdd/madevent -> build.512y_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:46 madevent/SubProcesses/P0_dd_taptamgdd/madevent -> build.512y_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:46 madevent/SubProcesses/P0_ddx_epemgddx/madevent -> build.512y_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:46 madevent/SubProcesses/P0_ddx_epemggg/madevent -> build.512y_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:46 madevent/SubProcesses/P0_ddx_epemgssx/madevent -> build.512y_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:46 madevent/SubProcesses/P0_ddx_epemguux/madevent -> build.512y_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:46 madevent/SubProcesses/P0_ddx_taptamgddx/madevent -> build.512y_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:46 madevent/SubProcesses/P0_ddx_taptamggg/madevent -> build.512y_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:46 madevent/SubProcesses/P0_ddx_taptamgssx/madevent -> build.512y_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:46 madevent/SubProcesses/P0_ddx_taptamguux/madevent -> build.512y_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:46 madevent/SubProcesses/P0_ds_epemgds/madevent -> build.512y_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:46 madevent/SubProcesses/P0_ds_taptamgds/madevent -> build.512y_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:46 madevent/SubProcesses/P0_dsx_epemgdsx/madevent -> build.512y_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:46 madevent/SubProcesses/P0_dsx_taptamgdsx/madevent -> build.512y_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:46 madevent/SubProcesses/P0_dux_epemgdux/madevent -> build.512y_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:46 madevent/SubProcesses/P0_dux_taptamgdux/madevent -> build.512y_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:46 madevent/SubProcesses/P0_dxcx_epemgdxcx/madevent -> build.512y_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:46 madevent/SubProcesses/P0_dxcx_taptamgdxcx/madevent -> build.512y_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:46 madevent/SubProcesses/P0_dxdx_epemgdxdx/madevent -> build.512y_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:46 madevent/SubProcesses/P0_dxdx_taptamgdxdx/madevent -> build.512y_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:46 madevent/SubProcesses/P0_dxsx_epemgdxsx/madevent -> build.512y_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:46 madevent/SubProcesses/P0_dxsx_taptamgdxsx/madevent -> build.512y_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:46 madevent/SubProcesses/P0_gd_epemdddx/madevent -> build.512y_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:46 madevent/SubProcesses/P0_gd_epemdssx/madevent -> build.512y_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:46 madevent/SubProcesses/P0_gd_epemggd/madevent -> build.512y_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:46 madevent/SubProcesses/P0_gd_epemudux/madevent -> build.512y_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:46 madevent/SubProcesses/P0_gd_taptamdddx/madevent -> build.512y_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:46 madevent/SubProcesses/P0_gd_taptamdssx/madevent -> build.512y_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:46 madevent/SubProcesses/P0_gd_taptamggd/madevent -> build.512y_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:46 madevent/SubProcesses/P0_gd_taptamudux/madevent -> build.512y_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:46 madevent/SubProcesses/P0_gdx_epemddxdx/madevent -> build.512y_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:46 madevent/SubProcesses/P0_gdx_epemggdx/madevent -> build.512y_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:46 madevent/SubProcesses/P0_gdx_epemsdxsx/madevent -> build.512y_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:46 madevent/SubProcesses/P0_gdx_epemuuxdx/madevent -> build.512y_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:46 madevent/SubProcesses/P0_gdx_taptamddxdx/madevent -> build.512y_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:46 madevent/SubProcesses/P0_gdx_taptamggdx/madevent -> build.512y_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:46 madevent/SubProcesses/P0_gdx_taptamsdxsx/madevent -> build.512y_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:46 madevent/SubProcesses/P0_gdx_taptamuuxdx/madevent -> build.512y_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:46 madevent/SubProcesses/P0_gg_epemgddx/madevent -> build.512y_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:46 madevent/SubProcesses/P0_gg_epemguux/madevent -> build.512y_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:46 madevent/SubProcesses/P0_gg_taptamgddx/madevent -> build.512y_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:46 madevent/SubProcesses/P0_gg_taptamguux/madevent -> build.512y_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:46 madevent/SubProcesses/P0_gu_epemggu/madevent -> build.512y_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:46 madevent/SubProcesses/P0_gu_epemuccx/madevent -> build.512y_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:46 madevent/SubProcesses/P0_gu_epemuddx/madevent -> build.512y_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:46 madevent/SubProcesses/P0_gu_epemuuux/madevent -> build.512y_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:46 madevent/SubProcesses/P0_gu_taptamggu/madevent -> build.512y_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:46 madevent/SubProcesses/P0_gu_taptamuccx/madevent -> build.512y_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:46 madevent/SubProcesses/P0_gu_taptamuddx/madevent -> build.512y_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:46 madevent/SubProcesses/P0_gu_taptamuuux/madevent -> build.512y_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:46 madevent/SubProcesses/P0_gux_epemcuxcx/madevent -> build.512y_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:46 madevent/SubProcesses/P0_gux_epemduxdx/madevent -> build.512y_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:46 madevent/SubProcesses/P0_gux_epemggux/madevent -> build.512y_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:46 madevent/SubProcesses/P0_gux_epemuuxux/madevent -> build.512y_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:46 madevent/SubProcesses/P0_gux_taptamcuxcx/madevent -> build.512y_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:46 madevent/SubProcesses/P0_gux_taptamduxdx/madevent -> build.512y_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:46 madevent/SubProcesses/P0_gux_taptamggux/madevent -> build.512y_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:46 madevent/SubProcesses/P0_gux_taptamuuxux/madevent -> build.512y_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:46 madevent/SubProcesses/P0_uc_epemguc/madevent -> build.512y_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:46 madevent/SubProcesses/P0_uc_taptamguc/madevent -> build.512y_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:46 madevent/SubProcesses/P0_ucx_epemgucx/madevent -> build.512y_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:46 madevent/SubProcesses/P0_ucx_taptamgucx/madevent -> build.512y_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:46 madevent/SubProcesses/P0_ud_epemgud/madevent -> build.512y_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:46 madevent/SubProcesses/P0_ud_taptamgud/madevent -> build.512y_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:46 madevent/SubProcesses/P0_udx_epemgudx/madevent -> build.512y_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:46 madevent/SubProcesses/P0_udx_taptamgudx/madevent -> build.512y_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:46 madevent/SubProcesses/P0_uu_epemguu/madevent -> build.512y_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:46 madevent/SubProcesses/P0_uu_taptamguu/madevent -> build.512y_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:46 madevent/SubProcesses/P0_uux_epemgccx/madevent -> build.512y_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:46 madevent/SubProcesses/P0_uux_epemgddx/madevent -> build.512y_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:46 madevent/SubProcesses/P0_uux_epemggg/madevent -> build.512y_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:46 madevent/SubProcesses/P0_uux_epemguux/madevent -> build.512y_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:46 madevent/SubProcesses/P0_uux_taptamgccx/madevent -> build.512y_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:46 madevent/SubProcesses/P0_uux_taptamgddx/madevent -> build.512y_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:46 madevent/SubProcesses/P0_uux_taptamggg/madevent -> build.512y_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:46 madevent/SubProcesses/P0_uux_taptamguux/madevent -> build.512y_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:46 madevent/SubProcesses/P0_uxcx_epemguxcx/madevent -> build.512y_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:46 madevent/SubProcesses/P0_uxcx_taptamguxcx/madevent -> build.512y_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:46 madevent/SubProcesses/P0_uxdx_epemguxdx/madevent -> build.512y_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:46 madevent/SubProcesses/P0_uxdx_taptamguxdx/madevent -> build.512y_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:46 madevent/SubProcesses/P0_uxux_epemguxux/madevent -> build.512y_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:46 madevent/SubProcesses/P0_uxux_taptamguxux/madevent -> build.512y_m_inl0_hrd0/madevent_cpp +Now generating 500 events with random seed 21 and granularity 1 +************************************************************ +* * +* W E L C O M E to * +* M A D G R A P H 5 _ a M C @ N L O * +* M A D E V E N T * +* * +* * * * +* * * * * * +* * * * * 5 * * * * * +* * * * * * +* * * * +* * +* VERSION 3.5.3_lo_vect * +* * +* The MadGraph5_aMC@NLO Development Team - Find us at * +* https://server06.fynu.ucl.ac.be/projects/madgraph * +* * +* Type 'help' for in-line help. * +* * +************************************************************ +INFO: load configuration from /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/Cards/me5_configuration.txt +INFO: load configuration from /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/Cards/me5_configuration.txt +Using default text editor "vi". Set another one in ./input/mg5_configuration.txt +Using default eps viewer "evince". Set another one in ./input/mg5_configuration.txt +Using default web browser "firefox". Set another one in ./input/mg5_configuration.txt +WRITE GRIDCARD /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent +__CUDACPP_DEBUG: GridPackCmd.__init__ will call self.launch +__CUDACPP_DEBUG: GridPackCmd.launch starting +__CUDACPP_DEBUG: GridPackCmd.launch (1) restore_data +generate 500 events +P0_gg_epemguux +P0_gg_epemgddx +P0_gg_taptamguux +P0_gg_taptamgddx +P0_gu_epemggu +P0_gd_epemggd +P0_gux_epemggux +P0_gdx_epemggdx +P0_gu_taptamggu +P0_gd_taptamggd +P0_gux_taptamggux +P0_gdx_taptamggdx +P0_uux_epemggg +P0_ddx_epemggg +P0_uux_taptamggg +P0_ddx_taptamggg +P0_gu_epemuuux +P0_gd_epemdddx +P0_gux_epemuuxux +P0_gdx_epemddxdx +P0_gu_epemuddx +P0_gu_epemuccx +P0_gd_epemudux +P0_gd_epemdssx +P0_gux_epemduxdx +P0_gux_epemcuxcx +P0_gdx_epemuuxdx +P0_gdx_epemsdxsx +P0_gu_taptamuuux +P0_gd_taptamdddx +P0_gux_taptamuuxux +P0_gdx_taptamddxdx +P0_gu_taptamuddx +P0_gu_taptamuccx +P0_gd_taptamudux +P0_gd_taptamdssx +P0_gux_taptamduxdx +P0_gux_taptamcuxcx +P0_gdx_taptamuuxdx +P0_gdx_taptamsdxsx +P0_uu_epemguu +P0_uux_epemguux +P0_dd_epemgdd +P0_ddx_epemgddx +P0_uxux_epemguxux +P0_dxdx_epemgdxdx +P0_ud_epemgud +P0_uc_epemguc +P0_uux_epemgddx +P0_uux_epemgccx +P0_udx_epemgudx +P0_ucx_epemgucx +P0_dc_epemgdc +P0_ds_epemgds +P0_dux_epemgdux +P0_ddx_epemguux +P0_ddx_epemgssx +P0_dsx_epemgdsx +P0_uxdx_epemguxdx +P0_uxcx_epemguxcx +P0_dxcx_epemgdxcx +P0_dxsx_epemgdxsx +P0_uu_taptamguu +P0_uux_taptamguux +P0_dd_taptamgdd +P0_ddx_taptamgddx +P0_uxux_taptamguxux +P0_dxdx_taptamgdxdx +P0_ud_taptamgud +P0_uc_taptamguc +P0_uux_taptamgddx +P0_uux_taptamgccx +P0_udx_taptamgudx +P0_ucx_taptamgucx +P0_dc_taptamgdc +P0_ds_taptamgds +P0_dux_taptamgdux +P0_ddx_taptamguux +P0_ddx_taptamgssx +P0_dsx_taptamgdsx +P0_uxdx_taptamguxdx +P0_uxcx_taptamguxcx +P0_dxcx_taptamgdxcx +P0_dxsx_taptamgdxsx +__CUDACPP_DEBUG: GridPackCmd.launch (2) refine4grid +__CUDACPP_DEBUG: GridPackCmd.refine4grid starting +Using random number seed offset = 21 +INFO: Generating 500.0 unweighted events. +__CUDACPP_DEBUG: gen_ximprove.launch starting +sum of cpu time of last step: 5h02m29s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event starting +INFO: need to improve 362 channels +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuuux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuuux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuuux/ajob1' in 8.4696 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuuux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuuux/G12/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.6403s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1898s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4471s for 16384 events => throughput is 3.66E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuuux/G20/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.7306s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2812s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4461s for 16384 events => throughput is 3.67E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0033s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuuux/G29/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.0551s + [COUNTERS] Fortran Overhead ( 0 ) : 4.6073s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4445s for 16384 events => throughput is 3.69E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0032s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuuux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemggg/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemggg/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemggg/ajob1' in 7.9960 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemggg/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemggg/G43/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.8769s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1644s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.7087s for 16384 events => throughput is 2.31E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0037s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemggg/G10/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.7406s + [COUNTERS] Fortran Overhead ( 0 ) : 3.0914s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6458s for 16384 events => throughput is 2.54E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0035s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemggg/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.2715s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6238s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6443s for 16384 events => throughput is 2.54E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemggg/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/ajob1' in 40.4080 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G39/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.9538s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2504s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6998s for 16384 events => throughput is 2.34E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0037s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G12/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.0960s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4494s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6432s for 16384 events => throughput is 2.55E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G6/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.8148s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1700s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6415s for 16384 events => throughput is 2.55E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G13/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.1354s + [COUNTERS] Fortran Overhead ( 0 ) : 3.4915s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6405s for 16384 events => throughput is 2.56E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G11/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.9775s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3177s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6564s for 16384 events => throughput is 2.50E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.8906s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2405s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6465s for 16384 events => throughput is 2.53E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G69/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.7111s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0646s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6429s for 16384 events => throughput is 2.55E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0035s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G8/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.7247s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0791s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6421s for 16384 events => throughput is 2.55E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G7/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.6473s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0002s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6437s for 16384 events => throughput is 2.55E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G65/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.5286s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8787s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6463s for 16384 events => throughput is 2.54E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G1/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.0666s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4200s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6433s for 16384 events => throughput is 2.55E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0033s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.1161s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4662s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6464s for 16384 events => throughput is 2.53E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G82/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.3926s + [COUNTERS] Fortran Overhead ( 0 ) : 3.7372s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6518s for 16384 events => throughput is 2.51E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0035s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G80/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.4655s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8209s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6413s for 16384 events => throughput is 2.55E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0033s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G51/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.3349s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6786s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6526s for 16384 events => throughput is 2.51E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0037s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G70/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.7183s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0723s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6426s for 16384 events => throughput is 2.55E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G5/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.4913s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8454s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6425s for 16384 events => throughput is 2.55E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0033s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamggux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamggux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamggux/ajob1' in 32.6271 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamggux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamggux/G59/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.4245s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0834s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.3376s for 16384 events => throughput is 1.22E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0035s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamggux/G12/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.7647s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4739s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.2873s for 16384 events => throughput is 1.27E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0035s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamggux/G13/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.7769s + [COUNTERS] Fortran Overhead ( 0 ) : 3.4819s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.2914s for 16384 events => throughput is 1.27E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamggux/G11/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.6292s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3330s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.2928s for 16384 events => throughput is 1.27E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamggux/G9/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.4266s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1310s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.2921s for 16384 events => throughput is 1.27E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamggux/G10/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.5908s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2870s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.3005s for 16384 events => throughput is 1.26E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamggux/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.5080s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2085s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.2960s for 16384 events => throughput is 1.26E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0035s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamggux/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.8543s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5693s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.2814s for 16384 events => throughput is 1.28E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamggux/G80/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.1234s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8207s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.2989s for 16384 events => throughput is 1.26E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0038s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamggux/G70/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.3904s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1049s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.2821s for 16384 events => throughput is 1.28E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamggux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemsdxsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemsdxsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemsdxsx/ajob1' in 2.0569 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemsdxsx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemsdxsx/G8/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.0405s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8387s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1992s for 16384 events => throughput is 8.22E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0026s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemsdxsx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuccx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuccx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuccx/ajob1' in 2.1229 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuccx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuccx/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.1064s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7026s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4012s for 16384 events => throughput is 4.08E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0026s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuccx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuuux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuuux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuuux/ajob1' in 6.2992 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuuux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuuux/G81/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.4897s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5789s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.9075s for 16384 events => throughput is 1.81E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0032s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuuux/G32/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.7799s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8708s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.9058s for 16384 events => throughput is 1.81E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuuux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamggg/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamggg/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamggg/ajob1' in 5.7060 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamggg/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamggg/G56/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.7639s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4834s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.2771s for 16384 events => throughput is 1.28E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamggg/G39/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.9121s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6164s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.2922s for 16384 events => throughput is 1.27E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamggg/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamduxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamduxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamduxdx/ajob1' in 5.6771 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamduxdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamduxdx/G45/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.5235s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1209s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4000s for 16384 events => throughput is 4.10E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0026s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamduxdx/G17/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.1236s + [COUNTERS] Fortran Overhead ( 0 ) : 3.7237s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.3973s for 16384 events => throughput is 4.12E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0025s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamduxdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuccx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuccx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuccx/ajob1' in 3.6679 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuccx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuccx/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.7953s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5946s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1982s for 16384 events => throughput is 8.27E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0025s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuccx/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.8431s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6441s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1965s for 16384 events => throughput is 8.34E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0025s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuccx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuddx/ajob1' in 12.9799 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuddx/G46/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.3211s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1151s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2034s for 16384 events => throughput is 8.05E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0026s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuddx/G18/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.5796s + [COUNTERS] Fortran Overhead ( 0 ) : 3.3753s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2016s for 16384 events => throughput is 8.13E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0027s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuddx/G39/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.1767s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9765s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1977s for 16384 events => throughput is 8.29E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0025s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuddx/G45/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.2968s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0969s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1974s for 16384 events => throughput is 8.30E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0026s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuddx/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.2716s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0710s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1981s for 16384 events => throughput is 8.27E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0025s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuddx/G20/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.4541s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2545s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1969s for 16384 events => throughput is 8.32E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0027s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuddx/G47/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.4658s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2628s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2003s for 16384 events => throughput is 8.18E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0027s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuddx/G7/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.3050s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1058s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1965s for 16384 events => throughput is 8.34E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0027s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/ajob1' in 58.9715 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G14/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.5316s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8740s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6542s for 16384 events => throughput is 2.50E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0033s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G74/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.5311s + [COUNTERS] Fortran Overhead ( 0 ) : 3.8854s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6423s for 16384 events => throughput is 2.55E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G12/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.0469s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3993s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6440s for 16384 events => throughput is 2.54E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G6/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.8808s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2347s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6426s for 16384 events => throughput is 2.55E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0035s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G13/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.7696s + [COUNTERS] Fortran Overhead ( 0 ) : 4.1207s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6455s for 16384 events => throughput is 2.54E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G83/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.0777s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4132s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6610s for 16384 events => throughput is 2.48E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0035s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G33/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.0598s + [COUNTERS] Fortran Overhead ( 0 ) : 4.4115s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6450s for 16384 events => throughput is 2.54E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0033s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G11/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.8543s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2049s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6459s for 16384 events => throughput is 2.54E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0035s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G9/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.9125s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2590s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6497s for 16384 events => throughput is 2.52E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0039s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G10/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.9552s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2794s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6722s for 16384 events => throughput is 2.44E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.8572s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2116s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6423s for 16384 events => throughput is 2.55E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G69/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.6953s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0497s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6421s for 16384 events => throughput is 2.55E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0035s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G65/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.1232s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4766s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6431s for 16384 events => throughput is 2.55E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0035s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G1/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.1244s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4687s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6521s for 16384 events => throughput is 2.51E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0035s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.0614s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4177s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6402s for 16384 events => throughput is 2.56E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G72/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.9483s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3022s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6428s for 16384 events => throughput is 2.55E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G2/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.6022s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9465s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6521s for 16384 events => throughput is 2.51E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G80/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.3963s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7481s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6448s for 16384 events => throughput is 2.54E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0035s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G79/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.8530s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2098s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6397s for 16384 events => throughput is 2.56E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0035s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G70/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.7799s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1187s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6576s for 16384 events => throughput is 2.49E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0035s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G5/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.6276s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9646s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6597s for 16384 events => throughput is 2.48E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamguux/ajob1' in 1.9224 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamguux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamguux/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.9055s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4911s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4117s for 16384 events => throughput is 3.98E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0027s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamguux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/ajob1' in 59.3925 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G14/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.4769s + [COUNTERS] Fortran Overhead ( 0 ) : 3.8338s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6396s for 16384 events => throughput is 2.56E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0035s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G12/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.7853s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1384s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6435s for 16384 events => throughput is 2.55E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G6/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.6162s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9718s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6410s for 16384 events => throughput is 2.56E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0033s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G3/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.3613s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7202s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6377s for 16384 events => throughput is 2.57E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0035s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G13/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.8623s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2172s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6417s for 16384 events => throughput is 2.55E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G11/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.9293s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2873s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6384s for 16384 events => throughput is 2.57E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G9/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.9826s + [COUNTERS] Fortran Overhead ( 0 ) : 3.3002s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6789s for 16384 events => throughput is 2.41E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G10/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.4420s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7937s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6447s for 16384 events => throughput is 2.54E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0037s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G40/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.2431s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5958s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6438s for 16384 events => throughput is 2.55E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0035s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G78/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.6125s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9665s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6424s for 16384 events => throughput is 2.55E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0035s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G35/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.4280s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7847s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6400s for 16384 events => throughput is 2.56E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.9871s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3426s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6409s for 16384 events => throughput is 2.56E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G69/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.7008s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0545s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6429s for 16384 events => throughput is 2.55E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0035s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G7/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.5424s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8945s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6442s for 16384 events => throughput is 2.54E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0037s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G71/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.9296s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2857s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6405s for 16384 events => throughput is 2.56E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G1/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.1299s + [COUNTERS] Fortran Overhead ( 0 ) : 3.4844s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6420s for 16384 events => throughput is 2.55E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0035s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.8953s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2532s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6386s for 16384 events => throughput is 2.57E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G4/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.3711s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7257s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6418s for 16384 events => throughput is 2.55E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G80/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.7875s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1463s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6378s for 16384 events => throughput is 2.57E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0033s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G79/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.3352s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6900s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6416s for 16384 events => throughput is 2.55E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.2677s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6240s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6403s for 16384 events => throughput is 2.56E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0035s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G70/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.6719s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0289s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6395s for 16384 events => throughput is 2.56E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0035s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G5/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.7235s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0799s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6399s for 16384 events => throughput is 2.56E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxcx_epemguxcx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxcx_epemguxcx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxcx_epemguxcx/ajob1' in 4.9175 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxcx_epemguxcx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxcx_epemguxcx/G9/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.1289s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9285s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1979s for 16384 events => throughput is 8.28E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0025s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxcx_epemguxcx/G10/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.7587s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5606s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1955s for 16384 events => throughput is 8.38E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0026s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxcx_epemguxcx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/ajob1' in 40.1775 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/G46/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.1424s + [COUNTERS] Fortran Overhead ( 0 ) : 4.4738s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6651s for 16384 events => throughput is 2.46E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0035s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/G18/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.6337s + [COUNTERS] Fortran Overhead ( 0 ) : 3.9788s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6513s for 16384 events => throughput is 2.52E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0035s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/G39/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.4925s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8474s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6417s for 16384 events => throughput is 2.55E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.9006s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2535s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6436s for 16384 events => throughput is 2.55E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/G20/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.2106s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5632s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6438s for 16384 events => throughput is 2.55E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/G17/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.9439s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2980s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6426s for 16384 events => throughput is 2.55E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0033s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/G35/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.1347s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4849s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6464s for 16384 events => throughput is 2.53E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/G71/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.0847s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4333s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6479s for 16384 events => throughput is 2.53E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0035s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/G27/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.7393s + [COUNTERS] Fortran Overhead ( 0 ) : 4.0942s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6418s for 16384 events => throughput is 2.55E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/G92/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.0296s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3833s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6427s for 16384 events => throughput is 2.55E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/G80/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.4185s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7719s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6432s for 16384 events => throughput is 2.55E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0033s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.2110s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5660s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6415s for 16384 events => throughput is 2.55E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0035s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/G29/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.5890s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9406s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6451s for 16384 events => throughput is 2.54E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0033s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/G48/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.4562s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8121s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6404s for 16384 events => throughput is 2.56E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0037s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_epemgdux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_epemgdux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_epemgdux/ajob1' in 9.4775 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_epemgdux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_epemgdux/G46/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 0.9860s + [COUNTERS] Fortran Overhead ( 0 ) : 0.7777s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2056s for 16384 events => throughput is 7.97E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0026s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_epemgdux/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.3102s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1088s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1989s for 16384 events => throughput is 8.24E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0025s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_epemgdux/G23/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.1221s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9214s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1982s for 16384 events => throughput is 8.27E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0026s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_epemgdux/G47/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.0538s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8523s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1989s for 16384 events => throughput is 8.24E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0026s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_epemgdux/G8/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.4116s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2115s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1976s for 16384 events => throughput is 8.29E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0025s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_epemgdux/G7/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.2043s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0030s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1987s for 16384 events => throughput is 8.25E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0026s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_epemgdux/G28/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.3599s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1586s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1988s for 16384 events => throughput is 8.24E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0025s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_epemgdux/G32/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 0.9191s + [COUNTERS] Fortran Overhead ( 0 ) : 0.7181s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1985s for 16384 events => throughput is 8.25E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0025s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_epemgdux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/ajob1' in 35.3608 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/G12/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.7558s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4668s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.2854s for 16384 events => throughput is 1.27E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0037s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/G6/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.4370s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1450s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.2885s for 16384 events => throughput is 1.27E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/G11/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.5108s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2299s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.2775s for 16384 events => throughput is 1.28E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0035s +INFO: The following Floating Point Exceptions have been reported: FE_UNDERFLOW +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.4623s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1719s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.2869s for 16384 events => throughput is 1.27E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0035s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/G69/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.3500s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0601s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.2863s for 16384 events => throughput is 1.27E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/G8/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.2289s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9390s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.2864s for 16384 events => throughput is 1.27E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0035s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/G84/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.6528s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3555s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.2939s for 16384 events => throughput is 1.27E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.8477s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5579s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.2864s for 16384 events => throughput is 1.27E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0035s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/G2/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.2251s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9338s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.2877s for 16384 events => throughput is 1.27E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/G80/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.0301s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7124s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.3142s for 16384 events => throughput is 1.25E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0035s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/G51/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.0304s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7342s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.2926s for 16384 events => throughput is 1.27E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/G70/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.3732s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0823s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.2875s for 16384 events => throughput is 1.27E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/G5/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.2789s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9857s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.2898s for 16384 events => throughput is 1.27E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamguux/ajob1' in 30.5892 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamguux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamguux/G18/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.0756s + [COUNTERS] Fortran Overhead ( 0 ) : 3.7914s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.2806s for 16384 events => throughput is 1.28E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0035s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamguux/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.5157s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2354s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.2769s for 16384 events => throughput is 1.28E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamguux/G20/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.8438s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5351s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.3051s for 16384 events => throughput is 1.26E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0037s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamguux/G17/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.4791s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1922s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.2834s for 16384 events => throughput is 1.28E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0035s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamguux/G35/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.7273s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4417s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.2822s for 16384 events => throughput is 1.28E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0033s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamguux/G27/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.9133s + [COUNTERS] Fortran Overhead ( 0 ) : 3.6317s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.2781s for 16384 events => throughput is 1.28E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamguux/G34/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.8603s + [COUNTERS] Fortran Overhead ( 0 ) : 3.5679s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.2890s for 16384 events => throughput is 1.27E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamguux/G80/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.0640s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7758s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.2847s for 16384 events => throughput is 1.28E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0035s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamguux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_taptamgdsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_taptamgdsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_taptamgdsx/ajob1' in 1.5195 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_taptamgdsx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_taptamgdsx/G28/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.5027s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0971s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4030s for 16384 events => throughput is 4.07E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0026s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_taptamgdsx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_taptamggg/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_taptamggg/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_taptamggg/ajob1' in 10.7535 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_taptamggg/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_taptamggg/G12/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.9714s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6832s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.2849s for 16384 events => throughput is 1.28E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_taptamggg/G9/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.7662s + [COUNTERS] Fortran Overhead ( 0 ) : 3.4717s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.2910s for 16384 events => throughput is 1.27E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0035s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_taptamggg/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.9726s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6866s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.2825s for 16384 events => throughput is 1.28E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0035s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_taptamggg/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxdx_epemguxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxdx_epemguxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxdx_epemguxdx/ajob1' in 13.4583 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxdx_epemguxdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxdx_epemguxdx/G14/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.2751s + [COUNTERS] Fortran Overhead ( 0 ) : 3.0745s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1981s for 16384 events => throughput is 8.27E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0025s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxdx_epemguxdx/G6/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.7340s + [COUNTERS] Fortran Overhead ( 0 ) : 4.5295s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2019s for 16384 events => throughput is 8.11E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0027s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxdx_epemguxdx/G22/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.4415s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2411s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1978s for 16384 events => throughput is 8.28E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0026s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxdx_epemguxdx/G8/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.4156s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2163s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1968s for 16384 events => throughput is 8.32E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0024s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxdx_epemguxdx/G27/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.5231s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3239s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1967s for 16384 events => throughput is 8.33E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0025s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxdx_epemguxdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_epemgdsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_epemgdsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_epemgdsx/ajob1' in 11.7348 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_epemgdsx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_epemgdsx/G31/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.1075s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9065s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1984s for 16384 events => throughput is 8.26E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0026s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_epemgdsx/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.1274s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9279s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1971s for 16384 events => throughput is 8.31E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0025s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_epemgdsx/G20/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.3020s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1016s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1979s for 16384 events => throughput is 8.28E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0024s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_epemgdsx/G25/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.8729s + [COUNTERS] Fortran Overhead ( 0 ) : 3.6729s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1973s for 16384 events => throughput is 8.30E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0027s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_epemgdsx/G28/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.2203s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0179s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2000s for 16384 events => throughput is 8.19E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0025s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_epemgdsx/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.0226s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8151s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2049s for 16384 events => throughput is 8.00E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0026s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_epemgdsx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamdddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamdddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamdddx/ajob1' in 4.2100 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamdddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamdddx/G17/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.1935s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2760s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.9140s for 16384 events => throughput is 1.79E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0035s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamdddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemggg/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemggg/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemggg/ajob1' in 13.5214 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemggg/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemggg/G56/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.2660s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6190s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6433s for 16384 events => throughput is 2.55E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0037s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemggg/G12/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.4184s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7537s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6614s for 16384 events => throughput is 2.48E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0033s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemggg/G11/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.3864s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7405s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6425s for 16384 events => throughput is 2.55E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemggg/G40/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.1642s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5201s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6407s for 16384 events => throughput is 2.56E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0035s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemggg/G8/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.9719s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3255s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6428s for 16384 events => throughput is 2.55E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemggg/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.2309s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5847s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6427s for 16384 events => throughput is 2.55E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemggg/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/ajob1' in 74.9404 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G14/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.3688s + [COUNTERS] Fortran Overhead ( 0 ) : 3.0836s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.2818s for 16384 events => throughput is 1.28E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G74/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.1105s + [COUNTERS] Fortran Overhead ( 0 ) : 3.7535s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.3533s for 16384 events => throughput is 1.21E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G12/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.4737s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1838s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.2865s for 16384 events => throughput is 1.27E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G6/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.3306s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0402s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.2870s for 16384 events => throughput is 1.27E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G13/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.2857s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9998s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.2824s for 16384 events => throughput is 1.28E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0035s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G66/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.8143s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5352s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.2753s for 16384 events => throughput is 1.28E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0038s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G11/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.6598s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3713s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.2849s for 16384 events => throughput is 1.28E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G9/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.4894s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1477s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.3381s for 16384 events => throughput is 1.22E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G40/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.0703s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7581s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.3086s for 16384 events => throughput is 1.25E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0037s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G35/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.0230s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7309s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.2883s for 16384 events => throughput is 1.27E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0037s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.9540s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6402s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.3102s for 16384 events => throughput is 1.25E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G69/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.2945s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0003s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.2905s for 16384 events => throughput is 1.27E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0037s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G7/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.1459s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8617s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.2807s for 16384 events => throughput is 1.28E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G65/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.9673s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6816s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.2823s for 16384 events => throughput is 1.28E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G71/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.6221s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3236s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.2948s for 16384 events => throughput is 1.27E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0037s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G1/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.2131s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9303s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.2792s for 16384 events => throughput is 1.28E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.5431s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2478s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.2918s for 16384 events => throughput is 1.27E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0035s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G2/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.6583s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3555s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.2992s for 16384 events => throughput is 1.26E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G4/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.9159s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6179s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.2946s for 16384 events => throughput is 1.27E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G79/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.1321s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8316s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.2968s for 16384 events => throughput is 1.26E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0037s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.8814s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5815s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.2965s for 16384 events => throughput is 1.26E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0035s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G70/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.2780s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9881s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.2864s for 16384 events => throughput is 1.27E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0035s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G5/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.3993s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1160s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.2799s for 16384 events => throughput is 1.28E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0035s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/ajob1' in 52.8264 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/G46/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.7585s + [COUNTERS] Fortran Overhead ( 0 ) : 4.1134s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6417s for 16384 events => throughput is 2.55E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/G18/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.1369s + [COUNTERS] Fortran Overhead ( 0 ) : 3.4758s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6574s for 16384 events => throughput is 2.49E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0037s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.1124s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4652s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6437s for 16384 events => throughput is 2.55E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/G33/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.0922s + [COUNTERS] Fortran Overhead ( 0 ) : 3.4454s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6434s for 16384 events => throughput is 2.55E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/G20/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.2753s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6274s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6444s for 16384 events => throughput is 2.54E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/G47/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.9327s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2791s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6499s for 16384 events => throughput is 2.52E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0037s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/G17/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.8531s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2043s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6453s for 16384 events => throughput is 2.54E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0035s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/G25/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.8655s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2217s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6403s for 16384 events => throughput is 2.56E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0035s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/G35/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.9928s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3467s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6426s for 16384 events => throughput is 2.55E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0035s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/G65/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.8692s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2101s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6556s for 16384 events => throughput is 2.50E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0035s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/G27/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.4092s + [COUNTERS] Fortran Overhead ( 0 ) : 3.7657s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6400s for 16384 events => throughput is 2.56E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/G34/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.0497s + [COUNTERS] Fortran Overhead ( 0 ) : 3.4073s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6390s for 16384 events => throughput is 2.56E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/G80/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.5496s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9025s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6435s for 16384 events => throughput is 2.55E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/G67/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.3603s + [COUNTERS] Fortran Overhead ( 0 ) : 3.7154s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6414s for 16384 events => throughput is 2.55E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.1942s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5193s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6712s for 16384 events => throughput is 2.44E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0037s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/G48/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.1579s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5039s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6504s for 16384 events => throughput is 2.52E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0035s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_taptamgdux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_taptamgdux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_taptamgdux/ajob1' in 6.4950 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_taptamgdux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_taptamgdux/G14/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.8799s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4776s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.3998s for 16384 events => throughput is 4.10E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0026s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_taptamgdux/G40/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.2474s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8409s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4039s for 16384 events => throughput is 4.06E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0026s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_taptamgdux/G44/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.3238s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9215s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.3997s for 16384 events => throughput is 4.10E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0025s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_taptamgdux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemgddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemgddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemgddx/ajob1' in 7.3888 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemgddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemgddx/G56/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.5152s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0650s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4469s for 16384 events => throughput is 3.67E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0033s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemgddx/G12/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.3667s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9165s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4467s for 16384 events => throughput is 3.67E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0035s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemgddx/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.3891s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9378s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4481s for 16384 events => throughput is 3.66E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0033s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemgddx/G47/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.5705s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1194s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4480s for 16384 events => throughput is 3.66E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0032s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemgddx/G42/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.4767s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0198s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4536s for 16384 events => throughput is 3.61E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemgddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/ajob1' in 45.1244 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/G39/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.7031s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4017s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.2978s for 16384 events => throughput is 1.26E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/G12/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.5056s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2103s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.2918s for 16384 events => throughput is 1.27E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/G6/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.1832s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8932s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.2867s for 16384 events => throughput is 1.27E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/G66/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.0986s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7408s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.3541s for 16384 events => throughput is 1.21E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0037s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/G11/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.6597s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3738s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.2824s for 16384 events => throughput is 1.28E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0035s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/G9/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.8123s + [COUNTERS] Fortran Overhead ( 0 ) : 3.5324s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.2766s for 16384 events => throughput is 1.28E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.6929s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4150s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.2744s for 16384 events => throughput is 1.29E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0035s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/G69/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.3830s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0987s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.2809s for 16384 events => throughput is 1.28E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/G8/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.2452s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9626s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.2792s for 16384 events => throughput is 1.28E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/G71/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.5893s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3125s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.2733s for 16384 events => throughput is 1.29E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0035s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.4864s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2034s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.2796s for 16384 events => throughput is 1.28E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/G72/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.4722s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1867s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.2819s for 16384 events => throughput is 1.28E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0035s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/G79/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.0149s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7284s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.2830s for 16384 events => throughput is 1.28E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0035s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.2510s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8304s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.4171s for 16384 events => throughput is 1.16E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0035s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/G70/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.3114s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0150s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.2927s for 16384 events => throughput is 1.27E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0037s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/G5/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.4982s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2113s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.2834s for 16384 events => throughput is 1.28E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0035s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemdssx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemdssx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemdssx/ajob1' in 4.5355 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemdssx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemdssx/G46/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.3349s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1336s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1988s for 16384 events => throughput is 8.24E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0025s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemdssx/G17/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.1708s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9721s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1962s for 16384 events => throughput is 8.35E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0025s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemdssx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uc_taptamguc/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uc_taptamguc/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uc_taptamguc/ajob1' in 3.9618 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uc_taptamguc/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uc_taptamguc/G25/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.6803s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2774s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4003s for 16384 events => throughput is 4.09E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0026s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uc_taptamguc/G37/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.2516s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8471s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4020s for 16384 events => throughput is 4.08E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0025s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uc_taptamguc/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamgddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamgddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamgddx/ajob1' in 4.6882 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamgddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamgddx/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.5483s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6417s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.9032s for 16384 events => throughput is 1.81E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamgddx/G55/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.1100s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1935s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.9130s for 16384 events => throughput is 1.79E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamgddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemgddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemgddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemgddx/ajob1' in 1.7138 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemgddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemgddx/G8/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.6971s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4977s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1970s for 16384 events => throughput is 8.32E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0024s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemgddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamgddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamgddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamgddx/ajob1' in 27.9531 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamgddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamgddx/G31/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.4458s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1634s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.2787s for 16384 events => throughput is 1.28E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamgddx/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.5153s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2295s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.2824s for 16384 events => throughput is 1.28E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamgddx/G33/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.4387s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1548s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.2806s for 16384 events => throughput is 1.28E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamgddx/G20/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.6832s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3871s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.2923s for 16384 events => throughput is 1.27E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0039s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamgddx/G35/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.5335s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2372s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.2929s for 16384 events => throughput is 1.27E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamgddx/G71/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.3827s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0935s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.2856s for 16384 events => throughput is 1.27E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamgddx/G34/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.0769s + [COUNTERS] Fortran Overhead ( 0 ) : 3.7977s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.2759s for 16384 events => throughput is 1.28E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamgddx/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.7907s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5083s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.2791s for 16384 events => throughput is 1.28E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamgddx/G48/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.9630s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6838s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.2759s for 16384 events => throughput is 1.28E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0033s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamgddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemddxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemddxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemddxdx/ajob1' in 3.6785 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemddxdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemddxdx/G11/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.9871s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5392s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4447s for 16384 events => throughput is 3.68E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0032s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemddxdx/G71/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.6614s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2111s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4470s for 16384 events => throughput is 3.67E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0033s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemddxdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dxsx_epemgdxsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dxsx_epemgdxsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dxsx_epemgdxsx/ajob1' in 1.4549 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dxsx_epemgdxsx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dxsx_epemgdxsx/G27/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.4383s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2324s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2031s for 16384 events => throughput is 8.07E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0028s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dxsx_epemgdxsx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemduxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemduxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemduxdx/ajob1' in 4.2418 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemduxdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemduxdx/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.2687s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0636s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2026s for 16384 events => throughput is 8.09E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0025s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemduxdx/G22/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.9433s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7378s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2029s for 16384 events => throughput is 8.08E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0027s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemduxdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemguux/ajob1' in 15.1581 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemguux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemguux/G20/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.5063s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0526s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4502s for 16384 events => throughput is 3.64E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0035s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemguux/G27/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.2854s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8258s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4564s for 16384 events => throughput is 3.59E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0032s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemguux/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.8576s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4002s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4539s for 16384 events => throughput is 3.61E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0035s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemguux/G55/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.6669s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2188s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4448s for 16384 events => throughput is 3.68E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0032s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemguux/G37/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.4958s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0451s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4473s for 16384 events => throughput is 3.66E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0033s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemguux/G41/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.3462s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8940s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4489s for 16384 events => throughput is 3.65E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0033s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemguux/G57/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.1857s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7306s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4517s for 16384 events => throughput is 3.63E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemguux/G48/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.7041s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2531s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4478s for 16384 events => throughput is 3.66E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0032s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemguux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dd_taptamgdd/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dd_taptamgdd/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dd_taptamgdd/ajob1' in 4.8878 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dd_taptamgdd/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dd_taptamgdd/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.1212s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2158s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.9020s for 16384 events => throughput is 1.82E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dd_taptamgdd/G24/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.7367s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8205s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.9127s for 16384 events => throughput is 1.80E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dd_taptamgdd/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/ajob1' in 106.9295 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G14/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.1484s + [COUNTERS] Fortran Overhead ( 0 ) : 3.4527s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6920s for 16384 events => throughput is 2.37E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G38/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.3566s + [COUNTERS] Fortran Overhead ( 0 ) : 3.7001s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6531s for 16384 events => throughput is 2.51E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0035s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G39/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.0528s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4041s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6448s for 16384 events => throughput is 2.54E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0039s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G12/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.8345s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1765s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6546s for 16384 events => throughput is 2.50E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G6/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.6273s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9364s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6873s for 16384 events => throughput is 2.38E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0037s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G3/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.4274s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7695s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6542s for 16384 events => throughput is 2.50E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0037s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G13/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.9081s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2639s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6406s for 16384 events => throughput is 2.56E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G66/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.5628s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9184s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6409s for 16384 events => throughput is 2.56E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0035s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G33/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.5352s + [COUNTERS] Fortran Overhead ( 0 ) : 4.8854s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6464s for 16384 events => throughput is 2.53E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G11/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.9292s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2810s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6448s for 16384 events => throughput is 2.54E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G9/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 21.0091s + [COUNTERS] Fortran Overhead ( 0 ) : 17.7872s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.2185s for 81920 events => throughput is 2.55E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G10/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.5182s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8756s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6392s for 16384 events => throughput is 2.56E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G40/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.2377s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5932s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6409s for 16384 events => throughput is 2.56E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G35/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.2137s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5668s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6435s for 16384 events => throughput is 2.55E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.9908s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3481s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6394s for 16384 events => throughput is 2.56E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0033s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G69/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.6758s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0296s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6427s for 16384 events => throughput is 2.55E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G8/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.5488s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9069s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6384s for 16384 events => throughput is 2.57E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0035s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G7/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.6782s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0346s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6401s for 16384 events => throughput is 2.56E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0035s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G65/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.6494s + [COUNTERS] Fortran Overhead ( 0 ) : 3.0003s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6456s for 16384 events => throughput is 2.54E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0035s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G71/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.1069s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4632s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6403s for 16384 events => throughput is 2.56E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G1/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.9419s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2980s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6405s for 16384 events => throughput is 2.56E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.9013s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2556s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6423s for 16384 events => throughput is 2.55E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G2/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.7693s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1195s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6463s for 16384 events => throughput is 2.53E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0035s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G4/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.5573s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9114s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6425s for 16384 events => throughput is 2.55E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G80/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.7648s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1201s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6413s for 16384 events => throughput is 2.55E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0033s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G41/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.1899s + [COUNTERS] Fortran Overhead ( 0 ) : 3.5435s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6430s for 16384 events => throughput is 2.55E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0035s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G77/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.4652s + [COUNTERS] Fortran Overhead ( 0 ) : 3.8214s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6403s for 16384 events => throughput is 2.56E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G76/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.9464s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3008s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6422s for 16384 events => throughput is 2.55E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0035s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G79/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.4760s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8302s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6422s for 16384 events => throughput is 2.55E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0035s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.1137s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4690s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6413s for 16384 events => throughput is 2.55E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0035s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G70/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.6330s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9859s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6437s for 16384 events => throughput is 2.55E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G5/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.7300s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0842s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6423s for 16384 events => throughput is 2.55E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0035s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ds_taptamgds/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ds_taptamgds/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ds_taptamgds/ajob1' in 2.3349 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ds_taptamgds/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ds_taptamgds/G10/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.3165s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9037s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4101s for 16384 events => throughput is 3.99E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0026s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ds_taptamgds/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_epemgudx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_epemgudx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_epemgudx/ajob1' in 9.3371 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_epemgudx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_epemgudx/G12/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.0291s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8294s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1972s for 16384 events => throughput is 8.31E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0025s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_epemgudx/G11/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 0.9898s + [COUNTERS] Fortran Overhead ( 0 ) : 0.7858s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2015s for 16384 events => throughput is 8.13E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0025s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_epemgudx/G17/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.5552s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3567s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1961s for 16384 events => throughput is 8.36E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0025s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_epemgudx/G10/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.1411s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9390s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1996s for 16384 events => throughput is 8.21E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0025s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_epemgudx/G27/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.2980s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0970s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1984s for 16384 events => throughput is 8.26E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0026s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_epemgudx/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.2412s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0394s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1991s for 16384 events => throughput is 8.23E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0026s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_epemgudx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_taptamgud/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_taptamgud/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_taptamgud/ajob1' in 15.2405 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_taptamgud/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_taptamgud/G12/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.2225s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8163s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4035s for 16384 events => throughput is 4.06E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0027s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_taptamgud/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.4969s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0910s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4033s for 16384 events => throughput is 4.06E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0026s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_taptamgud/G6/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.5543s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1489s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4028s for 16384 events => throughput is 4.07E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0025s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_taptamgud/G23/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.2477s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8401s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4050s for 16384 events => throughput is 4.05E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0026s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_taptamgud/G11/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.2684s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8635s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4023s for 16384 events => throughput is 4.07E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0025s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_taptamgud/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.6014s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1936s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4051s for 16384 events => throughput is 4.04E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0026s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_taptamgud/G22/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.7774s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3697s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4050s for 16384 events => throughput is 4.05E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0028s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_taptamgud/G7/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.6457s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2383s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4047s for 16384 events => throughput is 4.05E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0026s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_taptamgud/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.3029s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8970s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4033s for 16384 events => throughput is 4.06E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0026s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_taptamgud/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemdddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemdddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemdddx/ajob1' in 3.9158 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemdddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemdddx/G31/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.0977s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6430s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4515s for 16384 events => throughput is 3.63E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0032s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemdddx/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.7881s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3316s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4532s for 16384 events => throughput is 3.62E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0033s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemdddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamdssx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamdssx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamdssx/ajob1' in 3.1621 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamdssx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamdssx/G45/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.4795s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0774s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.3993s for 16384 events => throughput is 4.10E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0027s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamdssx/G20/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.6527s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2489s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4011s for 16384 events => throughput is 4.08E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0027s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamdssx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuddx/ajob1' in 4.5146 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuddx/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.4452s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0414s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4012s for 16384 events => throughput is 4.08E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0026s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuddx/G20/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.5158s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1165s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.3968s for 16384 events => throughput is 4.13E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0025s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuddx/G47/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.5104s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1108s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.3970s for 16384 events => throughput is 4.13E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0026s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemuuxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemuuxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemuuxdx/ajob1' in 4.8335 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemuuxdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemuuxdx/G7/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.5077s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3088s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1964s for 16384 events => throughput is 8.34E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0026s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemuuxdx/G41/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.2955s + [COUNTERS] Fortran Overhead ( 0 ) : 3.0955s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1975s for 16384 events => throughput is 8.29E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0025s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemuuxdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/ajob1' in 32.1601 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G14/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.2454s + [COUNTERS] Fortran Overhead ( 0 ) : 3.0440s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1988s for 16384 events => throughput is 8.24E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0026s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G18/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.7236s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5226s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1984s for 16384 events => throughput is 8.26E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0025s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G39/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.0124s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8126s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1973s for 16384 events => throughput is 8.30E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0025s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G30/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.3188s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1206s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1956s for 16384 events => throughput is 8.38E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0026s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G12/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.0099s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8084s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1990s for 16384 events => throughput is 8.23E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0025s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.3812s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1774s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2013s for 16384 events => throughput is 8.14E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0025s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G6/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.6025s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4011s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1989s for 16384 events => throughput is 8.24E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0025s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G3/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.0488s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8467s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1995s for 16384 events => throughput is 8.21E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0026s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G23/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.0609s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8595s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1990s for 16384 events => throughput is 8.23E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0025s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G24/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.0461s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8452s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1984s for 16384 events => throughput is 8.26E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0025s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G13/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.5298s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3127s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2143s for 16384 events => throughput is 7.65E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0028s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G20/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.0857s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8836s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1995s for 16384 events => throughput is 8.21E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0025s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G25/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.6313s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4299s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1988s for 16384 events => throughput is 8.24E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0026s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.3259s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1250s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1984s for 16384 events => throughput is 8.26E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0025s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G8/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.1275s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9266s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1983s for 16384 events => throughput is 8.26E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0026s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G7/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.2676s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0521s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2129s for 16384 events => throughput is 7.70E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0027s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G4/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 0.8724s + [COUNTERS] Fortran Overhead ( 0 ) : 0.6713s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1986s for 16384 events => throughput is 8.25E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0026s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G41/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.5235s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3213s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1994s for 16384 events => throughput is 8.21E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0027s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G32/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.0902s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8863s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2014s for 16384 events => throughput is 8.14E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0025s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ucx_epemgucx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ucx_epemgucx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ucx_epemgucx/ajob1' in 3.6515 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ucx_epemgucx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ucx_epemgucx/G3/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 0.8532s + [COUNTERS] Fortran Overhead ( 0 ) : 0.6526s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1980s for 16384 events => throughput is 8.27E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0025s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ucx_epemgucx/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.4640s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2643s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1973s for 16384 events => throughput is 8.31E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0025s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ucx_epemgucx/G7/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.2909s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0890s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1995s for 16384 events => throughput is 8.21E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0024s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ucx_epemgucx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_taptamgudx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_taptamgudx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_taptamgudx/ajob1' in 8.6842 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_taptamgudx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_taptamgudx/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.3598s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9494s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4078s for 16384 events => throughput is 4.02E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0026s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_taptamgudx/G11/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.2838s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8738s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4074s for 16384 events => throughput is 4.02E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0026s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_taptamgudx/G17/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.8028s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3970s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4033s for 16384 events => throughput is 4.06E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0025s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_taptamgudx/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.6527s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2367s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4133s for 16384 events => throughput is 3.96E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0027s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_taptamgudx/G27/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.5153s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1032s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4095s for 16384 events => throughput is 4.00E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0026s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_taptamgudx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamuuxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamuuxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamuuxdx/ajob1' in 3.1848 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamuuxdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamuuxdx/G6/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.1682s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7614s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4042s for 16384 events => throughput is 4.05E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0027s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamuuxdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_taptamguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_taptamguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_taptamguux/ajob1' in 2.1641 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_taptamguux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_taptamguux/G48/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.1474s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2175s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.9267s for 16384 events => throughput is 1.77E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0032s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_taptamguux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemudux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemudux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemudux/ajob1' in 8.5849 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemudux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemudux/G43/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.6939s + [COUNTERS] Fortran Overhead ( 0 ) : 3.4943s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1971s for 16384 events => throughput is 8.31E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0025s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemudux/G8/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.3490s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1468s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1996s for 16384 events => throughput is 8.21E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0027s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemudux/G5/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.4990s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2989s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1976s for 16384 events => throughput is 8.29E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0025s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemudux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamsdxsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamsdxsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamsdxsx/ajob1' in 1.6878 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamsdxsx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamsdxsx/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.6712s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2698s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.3989s for 16384 events => throughput is 4.11E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0025s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamsdxsx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_epemguu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_epemguu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_epemguu/ajob1' in 19.2323 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_epemguu/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_epemguu/G30/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.4378s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9856s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4490s for 16384 events => throughput is 3.65E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0032s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_epemguu/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.4760s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9945s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4779s for 16384 events => throughput is 3.43E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_epemguu/G9/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.7429s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2901s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4496s for 16384 events => throughput is 3.64E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0032s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_epemguu/G49/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.4319s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9396s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4888s for 16384 events => throughput is 3.35E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0035s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_epemguu/G10/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.6911s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2362s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4517s for 16384 events => throughput is 3.63E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0032s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_epemguu/G81/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.4849s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0317s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4499s for 16384 events => throughput is 3.64E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0033s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_epemguu/G62/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.2616s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8037s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4547s for 16384 events => throughput is 3.60E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0032s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_epemguu/G51/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.3132s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8606s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4494s for 16384 events => throughput is 3.65E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0033s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_epemguu/G28/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.2694s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8186s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4473s for 16384 events => throughput is 3.66E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0035s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_epemguu/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uc_epemguc/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uc_epemguc/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uc_epemguc/ajob1' in 2.2564 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uc_epemguc/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uc_epemguc/G12/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 0.9710s + [COUNTERS] Fortran Overhead ( 0 ) : 0.7704s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1982s for 16384 events => throughput is 8.27E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0024s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uc_epemguc/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.2555s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0544s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1987s for 16384 events => throughput is 8.25E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0025s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uc_epemguc/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ds_epemgds/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ds_epemgds/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ds_epemgds/ajob1' in 1.2850 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ds_epemgds/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ds_epemgds/G27/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.2683s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0697s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1960s for 16384 events => throughput is 8.36E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0026s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ds_epemgds/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_taptamguu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_taptamguu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_taptamguu/ajob1' in 6.3125 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_taptamguu/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_taptamguu/G52/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.8610s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9421s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.9155s for 16384 events => throughput is 1.79E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_taptamguu/G49/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.5393s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6252s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.9108s for 16384 events => throughput is 1.80E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0033s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_taptamguu/G32/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.8696s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9531s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.9133s for 16384 events => throughput is 1.79E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0032s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_taptamguu/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished in 922.7042 seconds +__CUDACPP_DEBUG: gen_ximprove.launch finished in 923.1233 seconds +__CUDACPP_DEBUG: GridPackCmd.refine4grid finished in 923.1240 seconds +__CUDACPP_DEBUG: GridPackCmd.launch (3a) combine_events +combine_events +__CUDACPP_DEBUG: GridPackCmd.launch (3b) store_events +store_events + === Results Summary for run: GridRun_21 tag: tag_1 === + + Cross-section : 0 +- 0 pb + Nb of events : 500 + +INFO: No version of lhapdf. Can not run systematics computation +decay_events -from_cards +__CUDACPP_DEBUG: GridPackCmd.launch finished in 940.0347 seconds +__CUDACPP_DEBUG: GridPackCmd.__init__ back from self.launch +DONE +write ./events.lhe.gz +END: Wed Sep 11 04:02:20 AM CEST 2024 +ELAPSED: 942 seconds diff --git a/epochX/cudacpp/tlau/fromgridpacks/pp_dy3j.mad/cpp512z/output.txt b/epochX/cudacpp/tlau/fromgridpacks/pp_dy3j.mad/cpp512z/output.txt new file mode 100644 index 0000000000..3f66d45300 --- /dev/null +++ b/epochX/cudacpp/tlau/fromgridpacks/pp_dy3j.mad/cpp512z/output.txt @@ -0,0 +1,2336 @@ +START: Wed Sep 11 04:02:59 AM CEST 2024 +On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU (nproc=4)] [GPU: 1x Tesla V100S-PCIE-32GB]: +CUDACPP_RUNTIME_DISABLEFPE is set +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 04:02 madevent/SubProcesses/P0_dc_epemgdc/madevent -> build.512z_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 04:02 madevent/SubProcesses/P0_dc_taptamgdc/madevent -> build.512z_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 04:02 madevent/SubProcesses/P0_dd_epemgdd/madevent -> build.512z_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 04:02 madevent/SubProcesses/P0_dd_taptamgdd/madevent -> build.512z_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 04:02 madevent/SubProcesses/P0_ddx_epemgddx/madevent -> build.512z_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 04:02 madevent/SubProcesses/P0_ddx_epemggg/madevent -> build.512z_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 04:02 madevent/SubProcesses/P0_ddx_epemgssx/madevent -> build.512z_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 04:02 madevent/SubProcesses/P0_ddx_epemguux/madevent -> build.512z_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 04:02 madevent/SubProcesses/P0_ddx_taptamgddx/madevent -> build.512z_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 04:02 madevent/SubProcesses/P0_ddx_taptamggg/madevent -> build.512z_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 04:02 madevent/SubProcesses/P0_ddx_taptamgssx/madevent -> build.512z_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 04:02 madevent/SubProcesses/P0_ddx_taptamguux/madevent -> build.512z_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 04:02 madevent/SubProcesses/P0_ds_epemgds/madevent -> build.512z_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 04:02 madevent/SubProcesses/P0_ds_taptamgds/madevent -> build.512z_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 04:02 madevent/SubProcesses/P0_dsx_epemgdsx/madevent -> build.512z_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 04:02 madevent/SubProcesses/P0_dsx_taptamgdsx/madevent -> build.512z_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 04:02 madevent/SubProcesses/P0_dux_epemgdux/madevent -> build.512z_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 04:02 madevent/SubProcesses/P0_dux_taptamgdux/madevent -> build.512z_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 04:02 madevent/SubProcesses/P0_dxcx_epemgdxcx/madevent -> build.512z_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 04:02 madevent/SubProcesses/P0_dxcx_taptamgdxcx/madevent -> build.512z_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 04:02 madevent/SubProcesses/P0_dxdx_epemgdxdx/madevent -> build.512z_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 04:02 madevent/SubProcesses/P0_dxdx_taptamgdxdx/madevent -> build.512z_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 04:02 madevent/SubProcesses/P0_dxsx_epemgdxsx/madevent -> build.512z_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 04:02 madevent/SubProcesses/P0_dxsx_taptamgdxsx/madevent -> build.512z_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 04:02 madevent/SubProcesses/P0_gd_epemdddx/madevent -> build.512z_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 04:02 madevent/SubProcesses/P0_gd_epemdssx/madevent -> build.512z_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 04:02 madevent/SubProcesses/P0_gd_epemggd/madevent -> build.512z_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 04:02 madevent/SubProcesses/P0_gd_epemudux/madevent -> build.512z_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 04:02 madevent/SubProcesses/P0_gd_taptamdddx/madevent -> build.512z_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 04:02 madevent/SubProcesses/P0_gd_taptamdssx/madevent -> build.512z_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 04:02 madevent/SubProcesses/P0_gd_taptamggd/madevent -> build.512z_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 04:02 madevent/SubProcesses/P0_gd_taptamudux/madevent -> build.512z_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 04:02 madevent/SubProcesses/P0_gdx_epemddxdx/madevent -> build.512z_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 04:02 madevent/SubProcesses/P0_gdx_epemggdx/madevent -> build.512z_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 04:02 madevent/SubProcesses/P0_gdx_epemsdxsx/madevent -> build.512z_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 04:02 madevent/SubProcesses/P0_gdx_epemuuxdx/madevent -> build.512z_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 04:02 madevent/SubProcesses/P0_gdx_taptamddxdx/madevent -> build.512z_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 04:02 madevent/SubProcesses/P0_gdx_taptamggdx/madevent -> build.512z_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 04:02 madevent/SubProcesses/P0_gdx_taptamsdxsx/madevent -> build.512z_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 04:02 madevent/SubProcesses/P0_gdx_taptamuuxdx/madevent -> build.512z_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 04:02 madevent/SubProcesses/P0_gg_epemgddx/madevent -> build.512z_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 04:02 madevent/SubProcesses/P0_gg_epemguux/madevent -> build.512z_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 04:02 madevent/SubProcesses/P0_gg_taptamgddx/madevent -> build.512z_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 04:02 madevent/SubProcesses/P0_gg_taptamguux/madevent -> build.512z_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 04:02 madevent/SubProcesses/P0_gu_epemggu/madevent -> build.512z_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 04:02 madevent/SubProcesses/P0_gu_epemuccx/madevent -> build.512z_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 04:02 madevent/SubProcesses/P0_gu_epemuddx/madevent -> build.512z_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 04:02 madevent/SubProcesses/P0_gu_epemuuux/madevent -> build.512z_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 04:02 madevent/SubProcesses/P0_gu_taptamggu/madevent -> build.512z_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 04:02 madevent/SubProcesses/P0_gu_taptamuccx/madevent -> build.512z_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 04:02 madevent/SubProcesses/P0_gu_taptamuddx/madevent -> build.512z_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 04:02 madevent/SubProcesses/P0_gu_taptamuuux/madevent -> build.512z_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 04:02 madevent/SubProcesses/P0_gux_epemcuxcx/madevent -> build.512z_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 04:02 madevent/SubProcesses/P0_gux_epemduxdx/madevent -> build.512z_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 04:02 madevent/SubProcesses/P0_gux_epemggux/madevent -> build.512z_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 04:02 madevent/SubProcesses/P0_gux_epemuuxux/madevent -> build.512z_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 04:02 madevent/SubProcesses/P0_gux_taptamcuxcx/madevent -> build.512z_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 04:02 madevent/SubProcesses/P0_gux_taptamduxdx/madevent -> build.512z_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 04:02 madevent/SubProcesses/P0_gux_taptamggux/madevent -> build.512z_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 04:02 madevent/SubProcesses/P0_gux_taptamuuxux/madevent -> build.512z_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 04:02 madevent/SubProcesses/P0_uc_epemguc/madevent -> build.512z_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 04:02 madevent/SubProcesses/P0_uc_taptamguc/madevent -> build.512z_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 04:02 madevent/SubProcesses/P0_ucx_epemgucx/madevent -> build.512z_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 04:02 madevent/SubProcesses/P0_ucx_taptamgucx/madevent -> build.512z_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 04:02 madevent/SubProcesses/P0_ud_epemgud/madevent -> build.512z_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 04:02 madevent/SubProcesses/P0_ud_taptamgud/madevent -> build.512z_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 04:02 madevent/SubProcesses/P0_udx_epemgudx/madevent -> build.512z_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 04:02 madevent/SubProcesses/P0_udx_taptamgudx/madevent -> build.512z_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 04:02 madevent/SubProcesses/P0_uu_epemguu/madevent -> build.512z_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 04:02 madevent/SubProcesses/P0_uu_taptamguu/madevent -> build.512z_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 04:02 madevent/SubProcesses/P0_uux_epemgccx/madevent -> build.512z_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 04:02 madevent/SubProcesses/P0_uux_epemgddx/madevent -> build.512z_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 04:02 madevent/SubProcesses/P0_uux_epemggg/madevent -> build.512z_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 04:02 madevent/SubProcesses/P0_uux_epemguux/madevent -> build.512z_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 04:02 madevent/SubProcesses/P0_uux_taptamgccx/madevent -> build.512z_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 04:02 madevent/SubProcesses/P0_uux_taptamgddx/madevent -> build.512z_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 04:02 madevent/SubProcesses/P0_uux_taptamggg/madevent -> build.512z_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 04:02 madevent/SubProcesses/P0_uux_taptamguux/madevent -> build.512z_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 04:02 madevent/SubProcesses/P0_uxcx_epemguxcx/madevent -> build.512z_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 04:02 madevent/SubProcesses/P0_uxcx_taptamguxcx/madevent -> build.512z_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 04:02 madevent/SubProcesses/P0_uxdx_epemguxdx/madevent -> build.512z_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 04:02 madevent/SubProcesses/P0_uxdx_taptamguxdx/madevent -> build.512z_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 04:02 madevent/SubProcesses/P0_uxux_epemguxux/madevent -> build.512z_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 04:02 madevent/SubProcesses/P0_uxux_taptamguxux/madevent -> build.512z_m_inl0_hrd0/madevent_cpp +Now generating 500 events with random seed 21 and granularity 1 +************************************************************ +* * +* W E L C O M E to * +* M A D G R A P H 5 _ a M C @ N L O * +* M A D E V E N T * +* * +* * * * +* * * * * * +* * * * * 5 * * * * * +* * * * * * +* * * * +* * +* VERSION 3.5.3_lo_vect * +* * +* The MadGraph5_aMC@NLO Development Team - Find us at * +* https://server06.fynu.ucl.ac.be/projects/madgraph * +* * +* Type 'help' for in-line help. * +* * +************************************************************ +INFO: load configuration from /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/Cards/me5_configuration.txt +INFO: load configuration from /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/Cards/me5_configuration.txt +Using default text editor "vi". Set another one in ./input/mg5_configuration.txt +Using default eps viewer "evince". Set another one in ./input/mg5_configuration.txt +Using default web browser "firefox". Set another one in ./input/mg5_configuration.txt +WRITE GRIDCARD /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent +__CUDACPP_DEBUG: GridPackCmd.__init__ will call self.launch +__CUDACPP_DEBUG: GridPackCmd.launch starting +__CUDACPP_DEBUG: GridPackCmd.launch (1) restore_data +generate 500 events +P0_gg_epemguux +P0_gg_epemgddx +P0_gg_taptamguux +P0_gg_taptamgddx +P0_gu_epemggu +P0_gd_epemggd +P0_gux_epemggux +P0_gdx_epemggdx +P0_gu_taptamggu +P0_gd_taptamggd +P0_gux_taptamggux +P0_gdx_taptamggdx +P0_uux_epemggg +P0_ddx_epemggg +P0_uux_taptamggg +P0_ddx_taptamggg +P0_gu_epemuuux +P0_gd_epemdddx +P0_gux_epemuuxux +P0_gdx_epemddxdx +P0_gu_epemuddx +P0_gu_epemuccx +P0_gd_epemudux +P0_gd_epemdssx +P0_gux_epemduxdx +P0_gux_epemcuxcx +P0_gdx_epemuuxdx +P0_gdx_epemsdxsx +P0_gu_taptamuuux +P0_gd_taptamdddx +P0_gux_taptamuuxux +P0_gdx_taptamddxdx +P0_gu_taptamuddx +P0_gu_taptamuccx +P0_gd_taptamudux +P0_gd_taptamdssx +P0_gux_taptamduxdx +P0_gux_taptamcuxcx +P0_gdx_taptamuuxdx +P0_gdx_taptamsdxsx +P0_uu_epemguu +P0_uux_epemguux +P0_dd_epemgdd +P0_ddx_epemgddx +P0_uxux_epemguxux +P0_dxdx_epemgdxdx +P0_ud_epemgud +P0_uc_epemguc +P0_uux_epemgddx +P0_uux_epemgccx +P0_udx_epemgudx +P0_ucx_epemgucx +P0_dc_epemgdc +P0_ds_epemgds +P0_dux_epemgdux +P0_ddx_epemguux +P0_ddx_epemgssx +P0_dsx_epemgdsx +P0_uxdx_epemguxdx +P0_uxcx_epemguxcx +P0_dxcx_epemgdxcx +P0_dxsx_epemgdxsx +P0_uu_taptamguu +P0_uux_taptamguux +P0_dd_taptamgdd +P0_ddx_taptamgddx +P0_uxux_taptamguxux +P0_dxdx_taptamgdxdx +P0_ud_taptamgud +P0_uc_taptamguc +P0_uux_taptamgddx +P0_uux_taptamgccx +P0_udx_taptamgudx +P0_ucx_taptamgucx +P0_dc_taptamgdc +P0_ds_taptamgds +P0_dux_taptamgdux +P0_ddx_taptamguux +P0_ddx_taptamgssx +P0_dsx_taptamgdsx +P0_uxdx_taptamguxdx +P0_uxcx_taptamguxcx +P0_dxcx_taptamgdxcx +P0_dxsx_taptamgdxsx +__CUDACPP_DEBUG: GridPackCmd.launch (2) refine4grid +__CUDACPP_DEBUG: GridPackCmd.refine4grid starting +Using random number seed offset = 21 +INFO: Generating 500.0 unweighted events. +__CUDACPP_DEBUG: gen_ximprove.launch starting +sum of cpu time of last step: 5h02m29s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event starting +INFO: need to improve 362 channels +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuuux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuuux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuuux/ajob1' in 8.8801 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuuux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuuux/G12/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.7616s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1721s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5852s for 16384 events => throughput is 2.80E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0043s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuuux/G20/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.8631s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2760s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5830s for 16384 events => throughput is 2.81E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0041s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuuux/G29/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.2120s + [COUNTERS] Fortran Overhead ( 0 ) : 4.6246s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5830s for 16384 events => throughput is 2.81E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0044s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuuux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemggg/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemggg/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemggg/ajob1' in 8.6738 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemggg/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemggg/G43/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.0389s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1001s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.9336s for 16384 events => throughput is 1.75E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0051s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemggg/G10/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.9267s + [COUNTERS] Fortran Overhead ( 0 ) : 3.0763s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8454s for 16384 events => throughput is 1.94E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0050s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemggg/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.4664s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6154s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8467s for 16384 events => throughput is 1.94E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0043s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemggg/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/ajob1' in 43.7642 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G39/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.0675s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2335s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8296s for 16384 events => throughput is 1.98E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0045s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G12/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.3136s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4754s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8335s for 16384 events => throughput is 1.97E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0047s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G6/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.0251s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1854s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8353s for 16384 events => throughput is 1.96E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0044s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G13/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.4177s + [COUNTERS] Fortran Overhead ( 0 ) : 3.5271s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8858s for 16384 events => throughput is 1.85E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0048s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G11/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.1647s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3175s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8428s for 16384 events => throughput is 1.94E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0044s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.0768s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2365s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8358s for 16384 events => throughput is 1.96E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0044s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G69/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.9063s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0605s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8415s for 16384 events => throughput is 1.95E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0043s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G8/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.9099s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0712s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8341s for 16384 events => throughput is 1.96E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0045s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G7/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.8354s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9965s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8345s for 16384 events => throughput is 1.96E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0044s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G65/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.7075s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8653s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8378s for 16384 events => throughput is 1.96E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0044s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G1/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.2648s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4193s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8408s for 16384 events => throughput is 1.95E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0048s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.2816s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4458s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8311s for 16384 events => throughput is 1.97E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0047s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G82/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.7566s + [COUNTERS] Fortran Overhead ( 0 ) : 3.9189s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8331s for 16384 events => throughput is 1.97E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0046s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G80/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.6627s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8269s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8316s for 16384 events => throughput is 1.97E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0042s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G51/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.4394s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5988s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8361s for 16384 events => throughput is 1.96E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0044s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G70/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.9392s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0731s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8617s for 16384 events => throughput is 1.90E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0044s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G5/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.6942s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8501s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8395s for 16384 events => throughput is 1.95E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0046s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamggux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamggux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamggux/ajob1' in 36.7815 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamggux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamggux/G59/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.7812s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0679s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7088s for 16384 events => throughput is 9.59E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0045s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamggux/G12/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.1873s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4794s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7035s for 16384 events => throughput is 9.62E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0043s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamggux/G13/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.2049s + [COUNTERS] Fortran Overhead ( 0 ) : 3.4774s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7226s for 16384 events => throughput is 9.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0049s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamggux/G11/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.0483s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3279s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7160s for 16384 events => throughput is 9.55E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0044s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamggux/G9/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.8488s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1416s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7025s for 16384 events => throughput is 9.62E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0047s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamggux/G10/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.9912s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2636s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7232s for 16384 events => throughput is 9.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0043s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamggux/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.9214s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2017s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7153s for 16384 events => throughput is 9.55E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0044s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamggux/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.2889s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5698s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7142s for 16384 events => throughput is 9.56E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0050s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamggux/G80/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.5358s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8136s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7177s for 16384 events => throughput is 9.54E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0044s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamggux/G70/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.8365s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1196s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7123s for 16384 events => throughput is 9.57E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0046s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamggux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemsdxsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemsdxsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemsdxsx/ajob1' in 2.1029 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemsdxsx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemsdxsx/G8/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.0861s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8194s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2633s for 16384 events => throughput is 6.22E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemsdxsx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuccx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuccx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuccx/ajob1' in 2.2581 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuccx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuccx/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.2414s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6872s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5509s for 16384 events => throughput is 2.97E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuccx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuuux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuuux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuuux/ajob1' in 6.9336 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuuux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuuux/G81/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.8376s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6380s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.1954s for 16384 events => throughput is 1.37E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0041s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuuux/G32/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.0661s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8709s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.1909s for 16384 events => throughput is 1.38E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0043s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuuux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamggg/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamggg/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamggg/ajob1' in 6.6689 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamggg/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamggg/G56/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.3112s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4983s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.8084s for 16384 events => throughput is 9.06E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0045s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamggg/G39/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.3270s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6104s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7119s for 16384 events => throughput is 9.57E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0046s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamggg/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamduxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamduxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamduxdx/ajob1' in 6.0001 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamduxdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamduxdx/G45/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.6890s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1224s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5634s for 16384 events => throughput is 2.91E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0032s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamduxdx/G17/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.2813s + [COUNTERS] Fortran Overhead ( 0 ) : 3.7306s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5474s for 16384 events => throughput is 2.99E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamduxdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuccx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuccx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuccx/ajob1' in 3.7936 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuccx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuccx/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.8570s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5921s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2617s for 16384 events => throughput is 6.26E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0032s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuccx/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.9067s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6398s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2635s for 16384 events => throughput is 6.22E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuccx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuddx/ajob1' in 13.4986 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuddx/G46/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.3869s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1153s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2683s for 16384 events => throughput is 6.11E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0033s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuddx/G18/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.6525s + [COUNTERS] Fortran Overhead ( 0 ) : 3.3858s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2635s for 16384 events => throughput is 6.22E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0032s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuddx/G39/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.2530s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9811s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2687s for 16384 events => throughput is 6.10E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0032s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuddx/G45/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.3660s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0974s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2655s for 16384 events => throughput is 6.17E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0031s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuddx/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.3440s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0755s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2650s for 16384 events => throughput is 6.18E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuddx/G20/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.5038s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2349s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2657s for 16384 events => throughput is 6.17E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0032s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuddx/G47/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.5238s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2542s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2663s for 16384 events => throughput is 6.15E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0033s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuddx/G7/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.3587s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0913s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2640s for 16384 events => throughput is 6.21E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/ajob1' in 62.8960 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G14/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.7164s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8775s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8345s for 16384 events => throughput is 1.96E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0044s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G74/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.7381s + [COUNTERS] Fortran Overhead ( 0 ) : 3.8929s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8409s for 16384 events => throughput is 1.95E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0043s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G12/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.2482s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4063s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8375s for 16384 events => throughput is 1.96E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0044s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G6/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.1171s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2611s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8518s for 16384 events => throughput is 1.92E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0043s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G13/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.9172s + [COUNTERS] Fortran Overhead ( 0 ) : 4.0805s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8322s for 16384 events => throughput is 1.97E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0045s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G83/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.2408s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4011s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8351s for 16384 events => throughput is 1.96E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0046s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G33/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.1758s + [COUNTERS] Fortran Overhead ( 0 ) : 4.3357s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8357s for 16384 events => throughput is 1.96E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0043s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G11/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.0531s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2107s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8378s for 16384 events => throughput is 1.96E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0046s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G9/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.0676s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2268s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8359s for 16384 events => throughput is 1.96E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0049s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G10/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.1035s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2645s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8344s for 16384 events => throughput is 1.96E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0047s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.0565s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2231s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8291s for 16384 events => throughput is 1.98E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0043s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G69/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.8819s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0442s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8334s for 16384 events => throughput is 1.97E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0043s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G65/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.3143s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4768s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8332s for 16384 events => throughput is 1.97E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0043s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G1/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.3183s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4819s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8320s for 16384 events => throughput is 1.97E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0044s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.2547s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4180s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8321s for 16384 events => throughput is 1.97E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0046s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G72/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.1293s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2924s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8326s for 16384 events => throughput is 1.97E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0043s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G2/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.8559s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9937s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8574s for 16384 events => throughput is 1.91E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0048s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G80/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.5810s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7385s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8376s for 16384 events => throughput is 1.96E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0049s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G79/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.0596s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2143s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8409s for 16384 events => throughput is 1.95E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0044s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G70/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.9844s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1250s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8546s for 16384 events => throughput is 1.92E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0047s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G5/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.7979s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9567s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8367s for 16384 events => throughput is 1.96E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0045s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamguux/ajob1' in 2.0285 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamguux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamguux/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.0118s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4634s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5451s for 16384 events => throughput is 3.01E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamguux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/ajob1' in 64.0201 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G14/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.6642s + [COUNTERS] Fortran Overhead ( 0 ) : 3.8243s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8351s for 16384 events => throughput is 1.96E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0048s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G12/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.9721s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1363s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8313s for 16384 events => throughput is 1.97E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0045s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G6/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.8056s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9732s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8279s for 16384 events => throughput is 1.98E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0045s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G3/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.6066s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7497s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8523s for 16384 events => throughput is 1.92E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0046s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G13/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.1419s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2940s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8436s for 16384 events => throughput is 1.94E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0043s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G11/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.1312s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2923s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8345s for 16384 events => throughput is 1.96E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0044s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G9/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.1338s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2935s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8360s for 16384 events => throughput is 1.96E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0043s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G10/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.6408s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8009s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8355s for 16384 events => throughput is 1.96E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0045s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G40/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.4279s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5941s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8295s for 16384 events => throughput is 1.98E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0043s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G78/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.7569s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9157s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8366s for 16384 events => throughput is 1.96E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0045s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G35/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.6309s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7885s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8381s for 16384 events => throughput is 1.95E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0044s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.1923s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3475s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8405s for 16384 events => throughput is 1.95E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0042s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G69/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.9066s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0554s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8469s for 16384 events => throughput is 1.93E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0042s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G7/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.7492s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9064s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8380s for 16384 events => throughput is 1.96E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0047s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G71/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.1247s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2895s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8308s for 16384 events => throughput is 1.97E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0044s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G1/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.2972s + [COUNTERS] Fortran Overhead ( 0 ) : 3.4654s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8274s for 16384 events => throughput is 1.98E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0044s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.1267s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2677s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8542s for 16384 events => throughput is 1.92E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0049s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G4/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.5747s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7226s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8478s for 16384 events => throughput is 1.93E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0043s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G80/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.9879s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1524s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8310s for 16384 events => throughput is 1.97E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0046s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G79/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.5305s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6884s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8378s for 16384 events => throughput is 1.96E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0042s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.4751s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6379s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8326s for 16384 events => throughput is 1.97E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0046s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G70/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.8842s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0370s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8428s for 16384 events => throughput is 1.94E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0044s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G5/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.9487s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0931s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8512s for 16384 events => throughput is 1.92E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0043s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxcx_epemguxcx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxcx_epemguxcx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxcx_epemguxcx/ajob1' in 5.0368 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxcx_epemguxcx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxcx_epemguxcx/G9/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.1992s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9294s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2665s for 16384 events => throughput is 6.15E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0033s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxcx_epemguxcx/G10/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.8078s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5363s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2683s for 16384 events => throughput is 6.11E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0032s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxcx_epemguxcx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/ajob1' in 43.0360 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/G46/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.3457s + [COUNTERS] Fortran Overhead ( 0 ) : 4.5038s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8375s for 16384 events => throughput is 1.96E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0045s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/G18/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.8301s + [COUNTERS] Fortran Overhead ( 0 ) : 3.9966s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8292s for 16384 events => throughput is 1.98E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0042s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/G39/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.7019s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8622s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8355s for 16384 events => throughput is 1.96E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0043s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.0952s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2528s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8379s for 16384 events => throughput is 1.96E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0045s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/G20/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.4048s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5732s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8270s for 16384 events => throughput is 1.98E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0046s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/G17/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.1639s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2891s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8702s for 16384 events => throughput is 1.88E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0046s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/G35/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.3323s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4948s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8333s for 16384 events => throughput is 1.97E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0042s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/G71/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.2660s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4222s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8393s for 16384 events => throughput is 1.95E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0045s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/G27/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.9056s + [COUNTERS] Fortran Overhead ( 0 ) : 4.0732s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8282s for 16384 events => throughput is 1.98E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0042s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/G92/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.2285s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3869s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8372s for 16384 events => throughput is 1.96E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0044s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/G80/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.6528s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8018s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8461s for 16384 events => throughput is 1.94E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0049s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.4252s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5852s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8352s for 16384 events => throughput is 1.96E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0048s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/G29/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.8335s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9860s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8428s for 16384 events => throughput is 1.94E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0046s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/G48/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.6588s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8164s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8378s for 16384 events => throughput is 1.96E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0046s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_epemgdux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_epemgdux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_epemgdux/ajob1' in 10.0186 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_epemgdux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_epemgdux/G46/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.0456s + [COUNTERS] Fortran Overhead ( 0 ) : 0.7802s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2621s for 16384 events => throughput is 6.25E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0032s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_epemgdux/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.3882s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1102s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2747s for 16384 events => throughput is 5.96E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0033s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_epemgdux/G23/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.1929s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9245s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2652s for 16384 events => throughput is 6.18E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0032s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_epemgdux/G47/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.1286s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8555s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2693s for 16384 events => throughput is 6.08E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0038s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_epemgdux/G8/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.4882s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2192s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2657s for 16384 events => throughput is 6.17E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0033s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_epemgdux/G7/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.2684s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9966s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2685s for 16384 events => throughput is 6.10E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0033s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_epemgdux/G28/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.4142s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1457s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2653s for 16384 events => throughput is 6.18E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0031s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_epemgdux/G32/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 0.9826s + [COUNTERS] Fortran Overhead ( 0 ) : 0.7149s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2645s for 16384 events => throughput is 6.20E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0033s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_epemgdux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/ajob1' in 41.0568 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/G12/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.1777s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4617s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7115s for 16384 events => throughput is 9.57E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0045s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/G6/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.8735s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1476s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7212s for 16384 events => throughput is 9.52E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0046s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/G11/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.9626s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2433s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7146s for 16384 events => throughput is 9.56E+03 events/s +INFO: The following Floating Point Exceptions have been reported: FE_UNDERFLOW + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0048s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.8868s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1655s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7165s for 16384 events => throughput is 9.54E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0048s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/G69/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.8086s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0711s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7328s for 16384 events => throughput is 9.46E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0048s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/G8/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.6600s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9410s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7144s for 16384 events => throughput is 9.56E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0046s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/G84/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.0453s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3223s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7184s for 16384 events => throughput is 9.53E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0046s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.2656s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5501s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7110s for 16384 events => throughput is 9.58E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0045s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/G2/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.7528s + [COUNTERS] Fortran Overhead ( 0 ) : 3.0214s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7267s for 16384 events => throughput is 9.49E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0047s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/G80/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.4262s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6898s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7318s for 16384 events => throughput is 9.46E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0046s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/G51/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.4412s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7323s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7042s for 16384 events => throughput is 9.61E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0047s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/G70/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.8190s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0700s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7444s for 16384 events => throughput is 9.39E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0046s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/G5/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.7609s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0006s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7557s for 16384 events => throughput is 9.33E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0047s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamguux/ajob1' in 34.0469 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamguux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamguux/G18/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.5221s + [COUNTERS] Fortran Overhead ( 0 ) : 3.8093s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7081s for 16384 events => throughput is 9.59E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0047s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamguux/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.9518s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2285s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7184s for 16384 events => throughput is 9.53E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0049s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamguux/G20/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.2231s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5012s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7174s for 16384 events => throughput is 9.54E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0045s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamguux/G17/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.9084s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1912s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7128s for 16384 events => throughput is 9.57E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0044s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamguux/G35/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.1594s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4398s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7151s for 16384 events => throughput is 9.55E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0045s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamguux/G27/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.3871s + [COUNTERS] Fortran Overhead ( 0 ) : 3.6577s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7247s for 16384 events => throughput is 9.50E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0047s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamguux/G34/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.2829s + [COUNTERS] Fortran Overhead ( 0 ) : 3.5612s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7173s for 16384 events => throughput is 9.54E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0044s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamguux/G80/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.5019s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7769s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7204s for 16384 events => throughput is 9.52E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0046s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamguux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_taptamgdsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_taptamgdsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_taptamgdsx/ajob1' in 1.6420 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_taptamgdsx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_taptamgdsx/G28/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.6254s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0798s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5424s for 16384 events => throughput is 3.02E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0033s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_taptamgdsx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_taptamggg/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_taptamggg/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_taptamggg/ajob1' in 12.1134 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_taptamggg/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_taptamggg/G12/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.4083s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6772s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7263s for 16384 events => throughput is 9.49E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0048s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_taptamggg/G9/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.2125s + [COUNTERS] Fortran Overhead ( 0 ) : 3.4972s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7109s for 16384 events => throughput is 9.58E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0045s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_taptamggg/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.4487s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7238s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7203s for 16384 events => throughput is 9.52E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0046s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_taptamggg/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxdx_epemguxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxdx_epemguxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxdx_epemguxdx/ajob1' in 13.9029 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxdx_epemguxdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxdx_epemguxdx/G14/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.3756s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1079s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2645s for 16384 events => throughput is 6.20E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0032s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxdx_epemguxdx/G6/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.8103s + [COUNTERS] Fortran Overhead ( 0 ) : 4.5406s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2662s for 16384 events => throughput is 6.15E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0035s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxdx_epemguxdx/G22/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.5025s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2290s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2703s for 16384 events => throughput is 6.06E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0033s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxdx_epemguxdx/G8/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.4908s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2187s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2689s for 16384 events => throughput is 6.09E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0032s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxdx_epemguxdx/G27/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.6534s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3587s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2914s for 16384 events => throughput is 5.62E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0032s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxdx_epemguxdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_epemgdsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_epemgdsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_epemgdsx/ajob1' in 12.2227 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_epemgdsx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_epemgdsx/G31/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.1795s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9083s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2680s for 16384 events => throughput is 6.11E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0033s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_epemgdsx/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.2108s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9405s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2668s for 16384 events => throughput is 6.14E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0035s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_epemgdsx/G20/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.3944s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1227s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2684s for 16384 events => throughput is 6.10E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0032s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_epemgdsx/G25/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.9758s + [COUNTERS] Fortran Overhead ( 0 ) : 3.6994s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2730s for 16384 events => throughput is 6.00E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_epemgdsx/G28/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.2819s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0151s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2635s for 16384 events => throughput is 6.22E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0033s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_epemgdsx/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.0978s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8284s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2661s for 16384 events => throughput is 6.16E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0033s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_epemgdsx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamdddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamdddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamdddx/ajob1' in 4.5139 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamdddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamdddx/G17/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.4973s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2920s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.2009s for 16384 events => throughput is 1.36E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0045s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamdddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemggg/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemggg/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemggg/ajob1' in 14.6688 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemggg/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemggg/G56/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.4523s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6139s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8342s for 16384 events => throughput is 1.96E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0043s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemggg/G12/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.6020s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7584s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8390s for 16384 events => throughput is 1.95E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0046s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemggg/G11/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.5884s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7450s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8390s for 16384 events => throughput is 1.95E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0044s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemggg/G40/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.3574s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5213s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8314s for 16384 events => throughput is 1.97E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0047s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemggg/G8/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.1663s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3259s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8362s for 16384 events => throughput is 1.96E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0043s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemggg/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.4190s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5828s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8317s for 16384 events => throughput is 1.97E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0045s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemggg/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/ajob1' in 84.9683 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G14/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.8107s + [COUNTERS] Fortran Overhead ( 0 ) : 3.0835s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7223s for 16384 events => throughput is 9.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0049s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G74/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.4598s + [COUNTERS] Fortran Overhead ( 0 ) : 3.7339s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7215s for 16384 events => throughput is 9.52E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0045s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G12/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.9075s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1745s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7286s for 16384 events => throughput is 9.48E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0045s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G6/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.7853s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0362s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7445s for 16384 events => throughput is 9.39E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0046s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G13/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.7139s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9870s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7224s for 16384 events => throughput is 9.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0045s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G66/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.2767s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5398s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7322s for 16384 events => throughput is 9.46E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0047s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G11/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.0998s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3758s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7194s for 16384 events => throughput is 9.53E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0046s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G9/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.8588s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1446s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7096s for 16384 events => throughput is 9.58E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0046s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G40/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.4876s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7544s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7287s for 16384 events => throughput is 9.48E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0045s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G35/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.4948s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7347s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7556s for 16384 events => throughput is 9.33E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0044s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.3400s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6134s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7221s for 16384 events => throughput is 9.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0045s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G69/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.7068s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0009s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7016s for 16384 events => throughput is 9.63E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0043s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G7/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.6258s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8615s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7600s for 16384 events => throughput is 9.31E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0043s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G65/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.3658s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6500s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7110s for 16384 events => throughput is 9.58E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0048s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G71/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.0266s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3127s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7095s for 16384 events => throughput is 9.58E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0044s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G1/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.6578s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9308s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7223s for 16384 events => throughput is 9.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0047s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.9528s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2345s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7137s for 16384 events => throughput is 9.56E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0047s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G2/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.0710s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3477s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7187s for 16384 events => throughput is 9.53E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0046s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G4/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.3374s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6182s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7143s for 16384 events => throughput is 9.56E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0048s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G79/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.5334s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8207s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7084s for 16384 events => throughput is 9.59E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0044s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.4123s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5850s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.8223s for 16384 events => throughput is 8.99E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0050s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G70/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.7167s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9904s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7217s for 16384 events => throughput is 9.52E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0046s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G5/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.0177s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1728s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.8401s for 16384 events => throughput is 8.90E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0047s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/ajob1' in 55.6980 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/G46/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.9805s + [COUNTERS] Fortran Overhead ( 0 ) : 4.1419s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8337s for 16384 events => throughput is 1.97E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0049s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/G18/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.2767s + [COUNTERS] Fortran Overhead ( 0 ) : 3.4373s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8347s for 16384 events => throughput is 1.96E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0048s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.3046s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4649s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8351s for 16384 events => throughput is 1.96E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0046s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/G33/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.2890s + [COUNTERS] Fortran Overhead ( 0 ) : 3.4498s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8346s for 16384 events => throughput is 1.96E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0046s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/G20/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.4769s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6369s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8355s for 16384 events => throughput is 1.96E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0045s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/G47/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.1158s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2704s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8408s for 16384 events => throughput is 1.95E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0047s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/G17/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.0290s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1902s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8343s for 16384 events => throughput is 1.96E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0045s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/G25/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.0695s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2356s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8295s for 16384 events => throughput is 1.98E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0044s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/G35/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.1806s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3442s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8319s for 16384 events => throughput is 1.97E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0045s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/G65/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.0274s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1939s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8289s for 16384 events => throughput is 1.98E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0046s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/G27/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.5342s + [COUNTERS] Fortran Overhead ( 0 ) : 3.6953s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8345s for 16384 events => throughput is 1.96E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0044s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/G34/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.2320s + [COUNTERS] Fortran Overhead ( 0 ) : 3.4014s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8263s for 16384 events => throughput is 1.98E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0042s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/G80/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.7412s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9050s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8317s for 16384 events => throughput is 1.97E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0045s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/G67/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.5535s + [COUNTERS] Fortran Overhead ( 0 ) : 3.7279s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8213s for 16384 events => throughput is 1.99E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0043s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.3192s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4918s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8232s for 16384 events => throughput is 1.99E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0042s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/G48/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.3491s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4973s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8475s for 16384 events => throughput is 1.93E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0043s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_taptamgdux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_taptamgdux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_taptamgdux/ajob1' in 7.0772 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_taptamgdux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_taptamgdux/G14/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.0833s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5232s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5567s for 16384 events => throughput is 2.94E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0033s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_taptamgdux/G40/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.4714s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8731s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5947s for 16384 events => throughput is 2.75E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_taptamgdux/G44/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.4790s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9331s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5426s for 16384 events => throughput is 3.02E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_taptamgdux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemgddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemgddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemgddx/ajob1' in 8.1214 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemgddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemgddx/G56/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.6800s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0627s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6126s for 16384 events => throughput is 2.67E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0047s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemgddx/G12/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.5144s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9181s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5920s for 16384 events => throughput is 2.77E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0043s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemgddx/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.5339s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9436s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5861s for 16384 events => throughput is 2.80E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0042s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemgddx/G47/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.7101s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1178s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5879s for 16384 events => throughput is 2.79E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0044s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemgddx/G42/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.6132s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0213s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5877s for 16384 events => throughput is 2.79E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0042s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemgddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/ajob1' in 52.0835 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/G39/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.1259s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4043s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7171s for 16384 events => throughput is 9.54E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0045s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/G12/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.9598s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2182s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7367s for 16384 events => throughput is 9.43E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0048s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/G6/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.6098s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8912s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7139s for 16384 events => throughput is 9.56E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0048s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/G66/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.5612s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7171s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.8393s for 16384 events => throughput is 8.91E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0048s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/G11/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.0970s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3719s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7208s for 16384 events => throughput is 9.52E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0043s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/G9/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.2778s + [COUNTERS] Fortran Overhead ( 0 ) : 3.5376s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7358s for 16384 events => throughput is 9.44E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0044s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.1381s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4216s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7121s for 16384 events => throughput is 9.57E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0043s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/G69/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.8110s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0994s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7072s for 16384 events => throughput is 9.60E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0044s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/G8/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.6743s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9560s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7138s for 16384 events => throughput is 9.56E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0045s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/G71/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.0371s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3222s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7102s for 16384 events => throughput is 9.58E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0048s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.9209s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2080s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7085s for 16384 events => throughput is 9.59E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0044s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/G72/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.9291s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1800s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7441s for 16384 events => throughput is 9.39E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0050s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/G79/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.4617s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7327s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7243s for 16384 events => throughput is 9.50E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0047s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.5637s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8387s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7205s for 16384 events => throughput is 9.52E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0046s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/G70/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.7358s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0075s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7237s for 16384 events => throughput is 9.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0046s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/G5/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.9635s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2201s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7383s for 16384 events => throughput is 9.43E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0050s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemdssx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemdssx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemdssx/ajob1' in 4.6708 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemdssx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemdssx/G46/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.3876s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1207s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2636s for 16384 events => throughput is 6.22E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0033s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemdssx/G17/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.2532s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9842s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2657s for 16384 events => throughput is 6.17E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0033s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemdssx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uc_taptamguc/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uc_taptamguc/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uc_taptamguc/ajob1' in 4.2583 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uc_taptamguc/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uc_taptamguc/G25/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.8202s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2660s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5505s for 16384 events => throughput is 2.98E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0037s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uc_taptamguc/G37/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.4085s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8466s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5584s for 16384 events => throughput is 2.93E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0035s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uc_taptamguc/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamgddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamgddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamgddx/ajob1' in 5.2722 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamgddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamgddx/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.8502s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6517s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.1942s for 16384 events => throughput is 1.37E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0043s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamgddx/G55/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.3920s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1990s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.1889s for 16384 events => throughput is 1.38E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0041s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamgddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemgddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemgddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemgddx/ajob1' in 1.7767 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemgddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemgddx/G8/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.7599s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4942s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2625s for 16384 events => throughput is 6.24E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0032s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemgddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamgddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamgddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamgddx/ajob1' in 31.9174 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamgddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamgddx/G31/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.8800s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1639s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7114s for 16384 events => throughput is 9.57E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0047s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamgddx/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.9543s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2239s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7258s for 16384 events => throughput is 9.49E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0046s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamgddx/G33/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.8514s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1328s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7142s for 16384 events => throughput is 9.56E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0044s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamgddx/G20/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.1142s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3837s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7261s for 16384 events => throughput is 9.49E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0044s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamgddx/G35/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.9646s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2337s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7264s for 16384 events => throughput is 9.49E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0045s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamgddx/G71/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.8036s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0946s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7044s for 16384 events => throughput is 9.61E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0046s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamgddx/G34/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.5664s + [COUNTERS] Fortran Overhead ( 0 ) : 3.7913s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7708s for 16384 events => throughput is 9.25E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0044s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamgddx/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.2510s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5294s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7172s for 16384 events => throughput is 9.54E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0043s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamgddx/G48/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.4070s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6788s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7236s for 16384 events => throughput is 9.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0046s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamgddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemddxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemddxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemddxdx/ajob1' in 3.9468 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemddxdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemddxdx/G11/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.1289s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5413s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5835s for 16384 events => throughput is 2.81E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0042s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemddxdx/G71/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.7875s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2006s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5824s for 16384 events => throughput is 2.81E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0046s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemddxdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dxsx_epemgdxsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dxsx_epemgdxsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dxsx_epemgdxsx/ajob1' in 1.5159 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dxsx_epemgdxsx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dxsx_epemgdxsx/G27/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.4994s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2290s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2673s for 16384 events => throughput is 6.13E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0031s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dxsx_epemgdxsx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemduxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemduxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemduxdx/ajob1' in 4.3791 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemduxdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemduxdx/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.3742s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0827s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2879s for 16384 events => throughput is 5.69E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0035s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemduxdx/G22/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.9751s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7065s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2654s for 16384 events => throughput is 6.17E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0031s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemduxdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemguux/ajob1' in 16.2056 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemguux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemguux/G20/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.6411s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0396s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5973s for 16384 events => throughput is 2.74E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0042s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemguux/G27/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.4132s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8245s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5844s for 16384 events => throughput is 2.80E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0042s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemguux/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.0368s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4499s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5823s for 16384 events => throughput is 2.81E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0046s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemguux/G55/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.7911s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2018s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5850s for 16384 events => throughput is 2.80E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0043s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemguux/G37/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.6487s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0557s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5889s for 16384 events => throughput is 2.78E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0042s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemguux/G41/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.4985s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9018s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5924s for 16384 events => throughput is 2.77E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0043s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemguux/G57/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.2218s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6343s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5834s for 16384 events => throughput is 2.81E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0040s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemguux/G48/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.8446s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2549s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5856s for 16384 events => throughput is 2.80E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0040s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemguux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dd_taptamgdd/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dd_taptamgdd/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dd_taptamgdd/ajob1' in 5.4480 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dd_taptamgdd/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dd_taptamgdd/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.4106s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2155s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.1908s for 16384 events => throughput is 1.38E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0043s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dd_taptamgdd/G24/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.0072s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8088s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.1944s for 16384 events => throughput is 1.37E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0041s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dd_taptamgdd/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/ajob1' in 113.8726 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G14/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.2386s + [COUNTERS] Fortran Overhead ( 0 ) : 3.4015s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8328s for 16384 events => throughput is 1.97E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0044s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G38/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.5488s + [COUNTERS] Fortran Overhead ( 0 ) : 3.7145s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8300s for 16384 events => throughput is 1.97E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0043s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G39/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.2309s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3980s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8286s for 16384 events => throughput is 1.98E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0043s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G12/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.0103s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1721s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8339s for 16384 events => throughput is 1.96E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0043s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G6/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.7382s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8911s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8426s for 16384 events => throughput is 1.94E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0045s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G3/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.6063s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7626s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8393s for 16384 events => throughput is 1.95E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0044s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G13/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.0818s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2480s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8294s for 16384 events => throughput is 1.98E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0044s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G66/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.7819s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9462s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8312s for 16384 events => throughput is 1.97E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0045s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G33/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.8273s + [COUNTERS] Fortran Overhead ( 0 ) : 4.9882s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8348s for 16384 events => throughput is 1.96E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0043s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G11/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.1212s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2822s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8344s for 16384 events => throughput is 1.96E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0046s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G9/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 22.0778s + [COUNTERS] Fortran Overhead ( 0 ) : 17.8729s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.2007s for 81920 events => throughput is 1.95E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0043s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G10/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.7098s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8731s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8323s for 16384 events => throughput is 1.97E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0044s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G40/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.4165s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5853s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8265s for 16384 events => throughput is 1.98E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0047s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G35/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.4486s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5840s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8602s for 16384 events => throughput is 1.90E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0044s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.1662s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3314s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8297s for 16384 events => throughput is 1.97E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0051s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G69/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.8814s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0437s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8334s for 16384 events => throughput is 1.97E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0043s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G8/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.7443s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9046s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8353s for 16384 events => throughput is 1.96E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0044s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G7/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.8745s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0373s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8324s for 16384 events => throughput is 1.97E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0048s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G65/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.8350s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9943s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8362s for 16384 events => throughput is 1.96E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0044s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G71/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.2908s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4589s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8274s for 16384 events => throughput is 1.98E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0045s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G1/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.1303s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2975s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8285s for 16384 events => throughput is 1.98E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0043s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.0765s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2439s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8282s for 16384 events => throughput is 1.98E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0045s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G2/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.9455s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1105s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8307s for 16384 events => throughput is 1.97E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0043s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G4/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.7219s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8886s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8288s for 16384 events => throughput is 1.98E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0045s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G80/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.0294s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1823s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8422s for 16384 events => throughput is 1.95E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0050s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G41/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.3883s + [COUNTERS] Fortran Overhead ( 0 ) : 3.5586s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8253s for 16384 events => throughput is 1.99E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0043s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G77/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.6587s + [COUNTERS] Fortran Overhead ( 0 ) : 3.8241s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8302s for 16384 events => throughput is 1.97E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0044s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G76/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.1448s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3044s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8359s for 16384 events => throughput is 1.96E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0045s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G79/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.6760s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8320s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8395s for 16384 events => throughput is 1.95E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0045s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.3048s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4710s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8291s for 16384 events => throughput is 1.98E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0047s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G70/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.8393s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9867s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8482s for 16384 events => throughput is 1.93E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0044s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G5/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.8962s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0619s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8298s for 16384 events => throughput is 1.97E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0045s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ds_taptamgds/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ds_taptamgds/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ds_taptamgds/ajob1' in 2.3598 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ds_taptamgds/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ds_taptamgds/G10/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.3431s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7952s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5445s for 16384 events => throughput is 3.01E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ds_taptamgds/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_epemgudx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_epemgudx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_epemgudx/ajob1' in 9.7325 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_epemgudx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_epemgudx/G12/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.1014s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8310s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2673s for 16384 events => throughput is 6.13E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0031s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_epemgudx/G11/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.0562s + [COUNTERS] Fortran Overhead ( 0 ) : 0.7868s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2660s for 16384 events => throughput is 6.16E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_epemgudx/G17/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.6270s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3565s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2674s for 16384 events => throughput is 6.13E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0032s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_epemgudx/G10/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.2009s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9329s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2648s for 16384 events => throughput is 6.19E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0032s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_epemgudx/G27/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.3456s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0755s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2670s for 16384 events => throughput is 6.14E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0031s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_epemgudx/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.3185s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0482s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2670s for 16384 events => throughput is 6.14E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0033s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_epemgudx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_taptamgud/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_taptamgud/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_taptamgud/ajob1' in 16.6024 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_taptamgud/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_taptamgud/G12/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.3696s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8183s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5480s for 16384 events => throughput is 2.99E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0033s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_taptamgud/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.6664s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1118s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5507s for 16384 events => throughput is 2.97E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0039s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_taptamgud/G6/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.7145s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1601s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5511s for 16384 events => throughput is 2.97E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0032s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_taptamgud/G23/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.3834s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8357s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5442s for 16384 events => throughput is 3.01E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0035s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_taptamgud/G11/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.4260s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8649s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5576s for 16384 events => throughput is 2.94E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0035s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_taptamgud/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.7368s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1899s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5435s for 16384 events => throughput is 3.01E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0033s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_taptamgud/G22/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.9101s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3659s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5408s for 16384 events => throughput is 3.03E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_taptamgud/G7/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.8214s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2636s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5544s for 16384 events => throughput is 2.96E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_taptamgud/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.4492s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8990s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5468s for 16384 events => throughput is 3.00E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_taptamgud/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemdddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemdddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemdddx/ajob1' in 4.1870 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemdddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemdddx/G31/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.2326s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6440s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5844s for 16384 events => throughput is 2.80E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0042s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemdddx/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.9244s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3308s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5893s for 16384 events => throughput is 2.78E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0043s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemdddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamdssx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamdssx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamdssx/ajob1' in 3.4543 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamdssx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamdssx/G45/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.6323s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0862s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5426s for 16384 events => throughput is 3.02E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamdssx/G20/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.7920s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2479s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5406s for 16384 events => throughput is 3.03E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0035s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamdssx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuddx/ajob1' in 4.9663 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuddx/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.5925s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0433s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5458s for 16384 events => throughput is 3.00E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuddx/G20/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.6660s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1192s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5433s for 16384 events => throughput is 3.02E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuddx/G47/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.6646s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1139s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5470s for 16384 events => throughput is 3.00E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0037s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemuuxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemuuxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemuuxdx/ajob1' in 4.9801 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemuuxdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemuuxdx/G7/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.5765s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3080s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2651s for 16384 events => throughput is 6.18E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0033s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemuuxdx/G41/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.3740s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1052s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2655s for 16384 events => throughput is 6.17E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0032s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemuuxdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/ajob1' in 33.4641 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G14/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.3094s + [COUNTERS] Fortran Overhead ( 0 ) : 3.0379s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2684s for 16384 events => throughput is 6.10E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0031s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G18/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.8468s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5594s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2840s for 16384 events => throughput is 5.77E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G39/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.1035s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8323s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2681s for 16384 events => throughput is 6.11E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0032s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G30/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.4019s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1245s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2739s for 16384 events => throughput is 5.98E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0035s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G12/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.0828s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8130s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2663s for 16384 events => throughput is 6.15E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0035s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.4526s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1829s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2666s for 16384 events => throughput is 6.15E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0031s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G6/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.6551s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3825s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2690s for 16384 events => throughput is 6.09E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G3/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.1211s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8472s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2706s for 16384 events => throughput is 6.06E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G23/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.1320s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8613s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2675s for 16384 events => throughput is 6.13E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0032s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G24/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.1164s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8474s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2658s for 16384 events => throughput is 6.16E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0032s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G13/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.5492s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2791s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2670s for 16384 events => throughput is 6.14E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0032s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G20/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.1585s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8856s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2694s for 16384 events => throughput is 6.08E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G25/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.6966s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4254s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2679s for 16384 events => throughput is 6.12E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0033s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.3935s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1216s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2685s for 16384 events => throughput is 6.10E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0033s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G8/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.2084s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9341s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2711s for 16384 events => throughput is 6.04E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0032s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G7/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.2837s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0159s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2644s for 16384 events => throughput is 6.20E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G4/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 0.9433s + [COUNTERS] Fortran Overhead ( 0 ) : 0.6762s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2639s for 16384 events => throughput is 6.21E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0031s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G41/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.5902s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3228s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2642s for 16384 events => throughput is 6.20E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0032s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G32/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.1620s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8887s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2699s for 16384 events => throughput is 6.07E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ucx_epemgucx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ucx_epemgucx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ucx_epemgucx/ajob1' in 3.8717 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ucx_epemgucx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ucx_epemgucx/G3/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 0.9231s + [COUNTERS] Fortran Overhead ( 0 ) : 0.6489s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2709s for 16384 events => throughput is 6.05E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0033s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ucx_epemgucx/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.5433s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2637s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2760s for 16384 events => throughput is 5.94E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ucx_epemgucx/G7/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.3623s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0935s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2656s for 16384 events => throughput is 6.17E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0032s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ucx_epemgucx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_taptamgudx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_taptamgudx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_taptamgudx/ajob1' in 9.3887 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_taptamgudx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_taptamgudx/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.4872s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9463s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5378s for 16384 events => throughput is 3.05E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0032s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_taptamgudx/G11/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.4161s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8625s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5500s for 16384 events => throughput is 2.98E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_taptamgudx/G17/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.9856s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4207s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5615s for 16384 events => throughput is 2.92E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0035s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_taptamgudx/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.7857s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2307s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5516s for 16384 events => throughput is 2.97E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_taptamgudx/G27/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.6443s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0933s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5476s for 16384 events => throughput is 2.99E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_taptamgudx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamuuxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamuuxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamuuxdx/ajob1' in 3.3204 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamuuxdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamuuxdx/G6/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.3038s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7548s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5458s for 16384 events => throughput is 3.00E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0032s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamuuxdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_taptamguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_taptamguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_taptamguux/ajob1' in 2.4217 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_taptamguux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_taptamguux/G48/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.4054s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2132s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.1877s for 16384 events => throughput is 1.38E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0046s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_taptamguux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemudux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemudux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemudux/ajob1' in 8.7777 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemudux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemudux/G43/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.7719s + [COUNTERS] Fortran Overhead ( 0 ) : 3.5015s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2670s for 16384 events => throughput is 6.14E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemudux/G8/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.4123s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1417s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2672s for 16384 events => throughput is 6.13E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemudux/G5/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.5508s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2791s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2685s for 16384 events => throughput is 6.10E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0032s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemudux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamsdxsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamsdxsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamsdxsx/ajob1' in 1.8744 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamsdxsx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamsdxsx/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.8569s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2802s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5731s for 16384 events => throughput is 2.86E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamsdxsx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_epemguu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_epemguu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_epemguu/ajob1' in 20.3744 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_epemguu/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_epemguu/G30/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.5763s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9935s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5788s for 16384 events => throughput is 2.83E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0040s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_epemguu/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.5486s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9580s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5864s for 16384 events => throughput is 2.79E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0042s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_epemguu/G9/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.8904s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3001s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5863s for 16384 events => throughput is 2.79E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0040s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_epemguu/G49/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.4988s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9115s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5832s for 16384 events => throughput is 2.81E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0040s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_epemguu/G10/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.8360s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2400s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5918s for 16384 events => throughput is 2.77E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0043s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_epemguu/G81/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.6061s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0196s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5823s for 16384 events => throughput is 2.81E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0042s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_epemguu/G62/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.3802s + [COUNTERS] Fortran Overhead ( 0 ) : 0.7971s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5789s for 16384 events => throughput is 2.83E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0042s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_epemguu/G51/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.4505s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8632s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5830s for 16384 events => throughput is 2.81E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0043s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_epemguu/G28/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.4643s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8462s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6135s for 16384 events => throughput is 2.67E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0046s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_epemguu/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uc_epemguc/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uc_epemguc/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uc_epemguc/ajob1' in 2.3881 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uc_epemguc/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uc_epemguc/G12/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.0406s + [COUNTERS] Fortran Overhead ( 0 ) : 0.7706s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2667s for 16384 events => throughput is 6.14E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0033s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uc_epemguc/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.3175s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0497s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2643s for 16384 events => throughput is 6.20E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0035s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uc_epemguc/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ds_epemgds/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ds_epemgds/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ds_epemgds/ajob1' in 1.3593 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ds_epemgds/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ds_epemgds/G27/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.3427s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0684s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2707s for 16384 events => throughput is 6.05E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ds_epemgds/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_taptamguu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_taptamguu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_taptamguu/ajob1' in 7.1320 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_taptamguu/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_taptamguu/G52/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.1308s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9421s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.1844s for 16384 events => throughput is 1.38E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0043s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_taptamguu/G49/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.8173s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6242s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.1887s for 16384 events => throughput is 1.38E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0043s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_taptamguu/G32/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.1409s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9518s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.1849s for 16384 events => throughput is 1.38E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0043s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_taptamguu/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished in 1002.9712 seconds +__CUDACPP_DEBUG: gen_ximprove.launch finished in 1003.3960 seconds +__CUDACPP_DEBUG: GridPackCmd.refine4grid finished in 1003.3967 seconds +__CUDACPP_DEBUG: GridPackCmd.launch (3a) combine_events +combine_events +__CUDACPP_DEBUG: GridPackCmd.launch (3b) store_events +store_events + === Results Summary for run: GridRun_21 tag: tag_1 === + + Cross-section : 0 +- 0 pb + Nb of events : 500 + +INFO: No version of lhapdf. Can not run systematics computation +decay_events -from_cards +__CUDACPP_DEBUG: GridPackCmd.launch finished in 1022.0703 seconds +__CUDACPP_DEBUG: GridPackCmd.__init__ back from self.launch +DONE +write ./events.lhe.gz +END: Wed Sep 11 04:20:03 AM CEST 2024 +ELAPSED: 1024 seconds diff --git a/epochX/cudacpp/tlau/fromgridpacks/pp_dy3j.mad/cppavx2/output.txt b/epochX/cudacpp/tlau/fromgridpacks/pp_dy3j.mad/cppavx2/output.txt new file mode 100644 index 0000000000..78f56dd239 --- /dev/null +++ b/epochX/cudacpp/tlau/fromgridpacks/pp_dy3j.mad/cppavx2/output.txt @@ -0,0 +1,2336 @@ +START: Wed Sep 11 03:29:56 AM CEST 2024 +On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU (nproc=4)] [GPU: 1x Tesla V100S-PCIE-32GB]: +CUDACPP_RUNTIME_DISABLEFPE is set +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:29 madevent/SubProcesses/P0_dc_epemgdc/madevent -> build.avx2_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:29 madevent/SubProcesses/P0_dc_taptamgdc/madevent -> build.avx2_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:29 madevent/SubProcesses/P0_dd_epemgdd/madevent -> build.avx2_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:29 madevent/SubProcesses/P0_dd_taptamgdd/madevent -> build.avx2_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:29 madevent/SubProcesses/P0_ddx_epemgddx/madevent -> build.avx2_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:29 madevent/SubProcesses/P0_ddx_epemggg/madevent -> build.avx2_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:29 madevent/SubProcesses/P0_ddx_epemgssx/madevent -> build.avx2_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:29 madevent/SubProcesses/P0_ddx_epemguux/madevent -> build.avx2_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:29 madevent/SubProcesses/P0_ddx_taptamgddx/madevent -> build.avx2_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:29 madevent/SubProcesses/P0_ddx_taptamggg/madevent -> build.avx2_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:29 madevent/SubProcesses/P0_ddx_taptamgssx/madevent -> build.avx2_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:29 madevent/SubProcesses/P0_ddx_taptamguux/madevent -> build.avx2_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:29 madevent/SubProcesses/P0_ds_epemgds/madevent -> build.avx2_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:29 madevent/SubProcesses/P0_ds_taptamgds/madevent -> build.avx2_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:29 madevent/SubProcesses/P0_dsx_epemgdsx/madevent -> build.avx2_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:29 madevent/SubProcesses/P0_dsx_taptamgdsx/madevent -> build.avx2_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:29 madevent/SubProcesses/P0_dux_epemgdux/madevent -> build.avx2_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:29 madevent/SubProcesses/P0_dux_taptamgdux/madevent -> build.avx2_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:29 madevent/SubProcesses/P0_dxcx_epemgdxcx/madevent -> build.avx2_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:29 madevent/SubProcesses/P0_dxcx_taptamgdxcx/madevent -> build.avx2_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:29 madevent/SubProcesses/P0_dxdx_epemgdxdx/madevent -> build.avx2_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:29 madevent/SubProcesses/P0_dxdx_taptamgdxdx/madevent -> build.avx2_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:29 madevent/SubProcesses/P0_dxsx_epemgdxsx/madevent -> build.avx2_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:29 madevent/SubProcesses/P0_dxsx_taptamgdxsx/madevent -> build.avx2_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:29 madevent/SubProcesses/P0_gd_epemdddx/madevent -> build.avx2_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:29 madevent/SubProcesses/P0_gd_epemdssx/madevent -> build.avx2_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:29 madevent/SubProcesses/P0_gd_epemggd/madevent -> build.avx2_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:29 madevent/SubProcesses/P0_gd_epemudux/madevent -> build.avx2_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:29 madevent/SubProcesses/P0_gd_taptamdddx/madevent -> build.avx2_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:29 madevent/SubProcesses/P0_gd_taptamdssx/madevent -> build.avx2_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:29 madevent/SubProcesses/P0_gd_taptamggd/madevent -> build.avx2_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:29 madevent/SubProcesses/P0_gd_taptamudux/madevent -> build.avx2_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:29 madevent/SubProcesses/P0_gdx_epemddxdx/madevent -> build.avx2_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:29 madevent/SubProcesses/P0_gdx_epemggdx/madevent -> build.avx2_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:29 madevent/SubProcesses/P0_gdx_epemsdxsx/madevent -> build.avx2_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:29 madevent/SubProcesses/P0_gdx_epemuuxdx/madevent -> build.avx2_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:29 madevent/SubProcesses/P0_gdx_taptamddxdx/madevent -> build.avx2_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:29 madevent/SubProcesses/P0_gdx_taptamggdx/madevent -> build.avx2_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:29 madevent/SubProcesses/P0_gdx_taptamsdxsx/madevent -> build.avx2_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:29 madevent/SubProcesses/P0_gdx_taptamuuxdx/madevent -> build.avx2_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:29 madevent/SubProcesses/P0_gg_epemgddx/madevent -> build.avx2_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:29 madevent/SubProcesses/P0_gg_epemguux/madevent -> build.avx2_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:29 madevent/SubProcesses/P0_gg_taptamgddx/madevent -> build.avx2_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:29 madevent/SubProcesses/P0_gg_taptamguux/madevent -> build.avx2_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:29 madevent/SubProcesses/P0_gu_epemggu/madevent -> build.avx2_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:29 madevent/SubProcesses/P0_gu_epemuccx/madevent -> build.avx2_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:29 madevent/SubProcesses/P0_gu_epemuddx/madevent -> build.avx2_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:29 madevent/SubProcesses/P0_gu_epemuuux/madevent -> build.avx2_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:29 madevent/SubProcesses/P0_gu_taptamggu/madevent -> build.avx2_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:29 madevent/SubProcesses/P0_gu_taptamuccx/madevent -> build.avx2_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:29 madevent/SubProcesses/P0_gu_taptamuddx/madevent -> build.avx2_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:29 madevent/SubProcesses/P0_gu_taptamuuux/madevent -> build.avx2_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:29 madevent/SubProcesses/P0_gux_epemcuxcx/madevent -> build.avx2_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:29 madevent/SubProcesses/P0_gux_epemduxdx/madevent -> build.avx2_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:29 madevent/SubProcesses/P0_gux_epemggux/madevent -> build.avx2_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:29 madevent/SubProcesses/P0_gux_epemuuxux/madevent -> build.avx2_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:29 madevent/SubProcesses/P0_gux_taptamcuxcx/madevent -> build.avx2_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:29 madevent/SubProcesses/P0_gux_taptamduxdx/madevent -> build.avx2_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:29 madevent/SubProcesses/P0_gux_taptamggux/madevent -> build.avx2_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:29 madevent/SubProcesses/P0_gux_taptamuuxux/madevent -> build.avx2_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:29 madevent/SubProcesses/P0_uc_epemguc/madevent -> build.avx2_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:29 madevent/SubProcesses/P0_uc_taptamguc/madevent -> build.avx2_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:29 madevent/SubProcesses/P0_ucx_epemgucx/madevent -> build.avx2_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:29 madevent/SubProcesses/P0_ucx_taptamgucx/madevent -> build.avx2_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:29 madevent/SubProcesses/P0_ud_epemgud/madevent -> build.avx2_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:29 madevent/SubProcesses/P0_ud_taptamgud/madevent -> build.avx2_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:29 madevent/SubProcesses/P0_udx_epemgudx/madevent -> build.avx2_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:29 madevent/SubProcesses/P0_udx_taptamgudx/madevent -> build.avx2_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:29 madevent/SubProcesses/P0_uu_epemguu/madevent -> build.avx2_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:29 madevent/SubProcesses/P0_uu_taptamguu/madevent -> build.avx2_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:29 madevent/SubProcesses/P0_uux_epemgccx/madevent -> build.avx2_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:29 madevent/SubProcesses/P0_uux_epemgddx/madevent -> build.avx2_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:29 madevent/SubProcesses/P0_uux_epemggg/madevent -> build.avx2_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:29 madevent/SubProcesses/P0_uux_epemguux/madevent -> build.avx2_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:29 madevent/SubProcesses/P0_uux_taptamgccx/madevent -> build.avx2_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:29 madevent/SubProcesses/P0_uux_taptamgddx/madevent -> build.avx2_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:29 madevent/SubProcesses/P0_uux_taptamggg/madevent -> build.avx2_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:29 madevent/SubProcesses/P0_uux_taptamguux/madevent -> build.avx2_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:29 madevent/SubProcesses/P0_uxcx_epemguxcx/madevent -> build.avx2_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:29 madevent/SubProcesses/P0_uxcx_taptamguxcx/madevent -> build.avx2_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:29 madevent/SubProcesses/P0_uxdx_epemguxdx/madevent -> build.avx2_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:29 madevent/SubProcesses/P0_uxdx_taptamguxdx/madevent -> build.avx2_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:29 madevent/SubProcesses/P0_uxux_epemguxux/madevent -> build.avx2_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:29 madevent/SubProcesses/P0_uxux_taptamguxux/madevent -> build.avx2_m_inl0_hrd0/madevent_cpp +Now generating 500 events with random seed 21 and granularity 1 +************************************************************ +* * +* W E L C O M E to * +* M A D G R A P H 5 _ a M C @ N L O * +* M A D E V E N T * +* * +* * * * +* * * * * * +* * * * * 5 * * * * * +* * * * * * +* * * * +* * +* VERSION 3.5.3_lo_vect * +* * +* The MadGraph5_aMC@NLO Development Team - Find us at * +* https://server06.fynu.ucl.ac.be/projects/madgraph * +* * +* Type 'help' for in-line help. * +* * +************************************************************ +INFO: load configuration from /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/Cards/me5_configuration.txt +INFO: load configuration from /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/Cards/me5_configuration.txt +Using default text editor "vi". Set another one in ./input/mg5_configuration.txt +Using default eps viewer "evince". Set another one in ./input/mg5_configuration.txt +Using default web browser "firefox". Set another one in ./input/mg5_configuration.txt +WRITE GRIDCARD /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent +__CUDACPP_DEBUG: GridPackCmd.__init__ will call self.launch +__CUDACPP_DEBUG: GridPackCmd.launch starting +__CUDACPP_DEBUG: GridPackCmd.launch (1) restore_data +generate 500 events +P0_gg_epemguux +P0_gg_epemgddx +P0_gg_taptamguux +P0_gg_taptamgddx +P0_gu_epemggu +P0_gd_epemggd +P0_gux_epemggux +P0_gdx_epemggdx +P0_gu_taptamggu +P0_gd_taptamggd +P0_gux_taptamggux +P0_gdx_taptamggdx +P0_uux_epemggg +P0_ddx_epemggg +P0_uux_taptamggg +P0_ddx_taptamggg +P0_gu_epemuuux +P0_gd_epemdddx +P0_gux_epemuuxux +P0_gdx_epemddxdx +P0_gu_epemuddx +P0_gu_epemuccx +P0_gd_epemudux +P0_gd_epemdssx +P0_gux_epemduxdx +P0_gux_epemcuxcx +P0_gdx_epemuuxdx +P0_gdx_epemsdxsx +P0_gu_taptamuuux +P0_gd_taptamdddx +P0_gux_taptamuuxux +P0_gdx_taptamddxdx +P0_gu_taptamuddx +P0_gu_taptamuccx +P0_gd_taptamudux +P0_gd_taptamdssx +P0_gux_taptamduxdx +P0_gux_taptamcuxcx +P0_gdx_taptamuuxdx +P0_gdx_taptamsdxsx +P0_uu_epemguu +P0_uux_epemguux +P0_dd_epemgdd +P0_ddx_epemgddx +P0_uxux_epemguxux +P0_dxdx_epemgdxdx +P0_ud_epemgud +P0_uc_epemguc +P0_uux_epemgddx +P0_uux_epemgccx +P0_udx_epemgudx +P0_ucx_epemgucx +P0_dc_epemgdc +P0_ds_epemgds +P0_dux_epemgdux +P0_ddx_epemguux +P0_ddx_epemgssx +P0_dsx_epemgdsx +P0_uxdx_epemguxdx +P0_uxcx_epemguxcx +P0_dxcx_epemgdxcx +P0_dxsx_epemgdxsx +P0_uu_taptamguu +P0_uux_taptamguux +P0_dd_taptamgdd +P0_ddx_taptamgddx +P0_uxux_taptamguxux +P0_dxdx_taptamgdxdx +P0_ud_taptamgud +P0_uc_taptamguc +P0_uux_taptamgddx +P0_uux_taptamgccx +P0_udx_taptamgudx +P0_ucx_taptamgucx +P0_dc_taptamgdc +P0_ds_taptamgds +P0_dux_taptamgdux +P0_ddx_taptamguux +P0_ddx_taptamgssx +P0_dsx_taptamgdsx +P0_uxdx_taptamguxdx +P0_uxcx_taptamguxcx +P0_dxcx_taptamgdxcx +P0_dxsx_taptamgdxsx +__CUDACPP_DEBUG: GridPackCmd.launch (2) refine4grid +__CUDACPP_DEBUG: GridPackCmd.refine4grid starting +Using random number seed offset = 21 +INFO: Generating 500.0 unweighted events. +__CUDACPP_DEBUG: gen_ximprove.launch starting +sum of cpu time of last step: 5h02m29s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event starting +INFO: need to improve 362 channels +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuuux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuuux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuuux/ajob1' in 8.5974 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuuux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuuux/G12/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.6689s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1760s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4894s for 16384 events => throughput is 3.35E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0035s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuuux/G20/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.7664s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2755s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4872s for 16384 events => throughput is 3.36E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuuux/G29/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.1188s + [COUNTERS] Fortran Overhead ( 0 ) : 4.6299s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4854s for 16384 events => throughput is 3.38E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0035s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuuux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemggg/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemggg/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemggg/ajob1' in 8.0752 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemggg/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemggg/G43/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.9078s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1560s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.7479s for 16384 events => throughput is 2.19E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0039s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemggg/G10/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.7705s + [COUNTERS] Fortran Overhead ( 0 ) : 3.0757s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6912s for 16384 events => throughput is 2.37E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0037s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemggg/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.3534s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6395s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.7103s for 16384 events => throughput is 2.31E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemggg/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/ajob1' in 41.3121 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G39/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.0539s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2804s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.7696s for 16384 events => throughput is 2.13E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0039s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G12/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.1403s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4455s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6912s for 16384 events => throughput is 2.37E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0035s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G6/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.8767s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1795s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6934s for 16384 events => throughput is 2.36E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0038s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G13/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.1863s + [COUNTERS] Fortran Overhead ( 0 ) : 3.4890s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6938s for 16384 events => throughput is 2.36E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0035s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G11/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.0178s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3196s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6946s for 16384 events => throughput is 2.36E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.9405s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2327s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.7041s for 16384 events => throughput is 2.33E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0037s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G69/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.7781s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0824s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6921s for 16384 events => throughput is 2.37E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G8/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.7801s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0767s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6998s for 16384 events => throughput is 2.34E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0037s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G7/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.6969s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0018s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6915s for 16384 events => throughput is 2.37E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G65/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.5638s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8698s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6901s for 16384 events => throughput is 2.37E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0038s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G1/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.1088s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4071s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6981s for 16384 events => throughput is 2.35E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0035s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.1423s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4453s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6932s for 16384 events => throughput is 2.36E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0039s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G82/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.4225s + [COUNTERS] Fortran Overhead ( 0 ) : 3.7253s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6936s for 16384 events => throughput is 2.36E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G80/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.5106s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8164s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6907s for 16384 events => throughput is 2.37E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0035s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G51/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.3474s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6543s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6894s for 16384 events => throughput is 2.38E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0038s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G70/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.7657s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0719s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6903s for 16384 events => throughput is 2.37E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0035s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G5/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.5485s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8520s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6930s for 16384 events => throughput is 2.36E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0035s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamggux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamggux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamggux/ajob1' in 33.8054 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamggux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamggux/G59/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.4818s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0790s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.3991s for 16384 events => throughput is 1.17E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0037s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamggux/G12/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.8854s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4838s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.3976s for 16384 events => throughput is 1.17E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0039s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamggux/G13/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.8836s + [COUNTERS] Fortran Overhead ( 0 ) : 3.4719s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.4080s for 16384 events => throughput is 1.16E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0037s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamggux/G11/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.7341s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3267s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.4038s for 16384 events => throughput is 1.17E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamggux/G9/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.6800s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1784s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.4977s for 16384 events => throughput is 1.09E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0038s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamggux/G10/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.6794s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2747s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.4008s for 16384 events => throughput is 1.17E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0039s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamggux/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.6043s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1987s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.4020s for 16384 events => throughput is 1.17E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0037s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamggux/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.9688s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5664s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.3988s for 16384 events => throughput is 1.17E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamggux/G80/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.2323s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8259s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.4028s for 16384 events => throughput is 1.17E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamggux/G70/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.5192s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1066s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.4090s for 16384 events => throughput is 1.16E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0037s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamggux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemsdxsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemsdxsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemsdxsx/ajob1' in 2.1071 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemsdxsx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemsdxsx/G8/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.0903s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8683s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2192s for 16384 events => throughput is 7.47E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0028s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemsdxsx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuccx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuccx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuccx/ajob1' in 2.1803 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuccx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuccx/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.1637s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7199s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4410s for 16384 events => throughput is 3.72E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0028s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuccx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuuux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuuux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuuux/ajob1' in 6.4746 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuuux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuuux/G81/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.5910s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5918s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.9956s for 16384 events => throughput is 1.65E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0037s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuuux/G32/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.8537s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8571s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.9931s for 16384 events => throughput is 1.65E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0035s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuuux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamggg/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamggg/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamggg/ajob1' in 5.9295 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamggg/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamggg/G56/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.8858s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4892s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.3928s for 16384 events => throughput is 1.18E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0039s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamggg/G39/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.0135s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6107s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.3993s for 16384 events => throughput is 1.17E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamggg/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamduxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamduxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamduxdx/ajob1' in 5.8120 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamduxdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamduxdx/G45/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.5712s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1334s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4350s for 16384 events => throughput is 3.77E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0027s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamduxdx/G17/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.2107s + [COUNTERS] Fortran Overhead ( 0 ) : 3.7728s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4353s for 16384 events => throughput is 3.76E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0026s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamduxdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuccx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuccx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuccx/ajob1' in 3.6950 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuccx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuccx/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.8137s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5997s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2112s for 16384 events => throughput is 7.76E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0027s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuccx/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.8513s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6354s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2133s for 16384 events => throughput is 7.68E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0027s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuccx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuddx/ajob1' in 13.0112 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuddx/G46/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.3225s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1069s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2128s for 16384 events => throughput is 7.70E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0028s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuddx/G18/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.5502s + [COUNTERS] Fortran Overhead ( 0 ) : 3.3319s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2156s for 16384 events => throughput is 7.60E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0027s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuddx/G39/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.1886s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9736s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2124s for 16384 events => throughput is 7.71E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0026s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuddx/G45/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.3132s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0966s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2139s for 16384 events => throughput is 7.66E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0026s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuddx/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.2805s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0652s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2127s for 16384 events => throughput is 7.70E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0026s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuddx/G20/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.4614s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2463s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2124s for 16384 events => throughput is 7.72E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0027s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuddx/G47/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.4781s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2589s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2167s for 16384 events => throughput is 7.56E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0026s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuddx/G7/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.3069s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0898s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2144s for 16384 events => throughput is 7.64E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0027s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/ajob1' in 59.9570 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G14/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.5851s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8839s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6976s for 16384 events => throughput is 2.35E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G74/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.6074s + [COUNTERS] Fortran Overhead ( 0 ) : 3.9082s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6955s for 16384 events => throughput is 2.36E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0037s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G12/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.0998s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3998s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6962s for 16384 events => throughput is 2.35E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0038s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G6/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.9482s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2463s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6983s for 16384 events => throughput is 2.35E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G13/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.7562s + [COUNTERS] Fortran Overhead ( 0 ) : 4.0532s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6994s for 16384 events => throughput is 2.34E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G83/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.0985s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3951s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6998s for 16384 events => throughput is 2.34E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G33/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.0373s + [COUNTERS] Fortran Overhead ( 0 ) : 4.3384s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6953s for 16384 events => throughput is 2.36E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0037s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G11/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.9004s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2025s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6940s for 16384 events => throughput is 2.36E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0038s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G9/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.9447s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2336s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.7074s for 16384 events => throughput is 2.32E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0037s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G10/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.9721s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2734s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6948s for 16384 events => throughput is 2.36E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0039s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.9224s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2247s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6938s for 16384 events => throughput is 2.36E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0038s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G69/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.7671s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0617s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.7015s for 16384 events => throughput is 2.34E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0039s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G65/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.1776s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4741s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6999s for 16384 events => throughput is 2.34E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G1/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.1758s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4729s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6994s for 16384 events => throughput is 2.34E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0035s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.1169s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4165s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6966s for 16384 events => throughput is 2.35E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0038s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G72/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.0904s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3429s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.7434s for 16384 events => throughput is 2.20E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0041s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G2/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.6424s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9418s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6969s for 16384 events => throughput is 2.35E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G80/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.4515s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7532s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6947s for 16384 events => throughput is 2.36E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G79/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.9085s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2098s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6951s for 16384 events => throughput is 2.36E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0035s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G70/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.8167s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1151s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6979s for 16384 events => throughput is 2.35E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0038s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G5/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.6555s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9523s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6993s for 16384 events => throughput is 2.34E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0039s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamguux/ajob1' in 1.9204 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamguux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamguux/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.9037s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4602s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4408s for 16384 events => throughput is 3.72E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0027s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamguux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/ajob1' in 60.6462 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G14/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.5509s + [COUNTERS] Fortran Overhead ( 0 ) : 3.8549s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6924s for 16384 events => throughput is 2.37E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G12/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.8336s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1376s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6924s for 16384 events => throughput is 2.37E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G6/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.6681s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9743s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6902s for 16384 events => throughput is 2.37E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G3/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.4343s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7399s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6907s for 16384 events => throughput is 2.37E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0037s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G13/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.9106s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2178s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6889s for 16384 events => throughput is 2.38E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0038s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G11/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.9745s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2825s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6884s for 16384 events => throughput is 2.38E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G9/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.9653s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2746s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6870s for 16384 events => throughput is 2.38E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0037s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G10/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.4969s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8011s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6922s for 16384 events => throughput is 2.37E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G40/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.2808s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5889s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6883s for 16384 events => throughput is 2.38E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G78/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.6258s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9302s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6917s for 16384 events => throughput is 2.37E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0038s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G35/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.5722s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8268s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.7416s for 16384 events => throughput is 2.21E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0038s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.0371s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3448s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6887s for 16384 events => throughput is 2.38E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G69/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.7605s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0539s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.7030s for 16384 events => throughput is 2.33E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G7/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.5915s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8952s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6928s for 16384 events => throughput is 2.36E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0035s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G71/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.9910s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2911s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6961s for 16384 events => throughput is 2.35E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0037s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G1/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.2101s + [COUNTERS] Fortran Overhead ( 0 ) : 3.5097s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6963s for 16384 events => throughput is 2.35E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0040s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.9569s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2557s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6974s for 16384 events => throughput is 2.35E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0038s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G4/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.4150s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7161s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6953s for 16384 events => throughput is 2.36E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0035s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G80/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.8381s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1397s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6948s for 16384 events => throughput is 2.36E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G79/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.3954s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6868s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.7050s for 16384 events => throughput is 2.32E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.3212s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6267s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6910s for 16384 events => throughput is 2.37E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0035s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G70/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.7275s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0288s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6950s for 16384 events => throughput is 2.36E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G5/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.7807s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0882s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6887s for 16384 events => throughput is 2.38E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0037s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxcx_epemguxcx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxcx_epemguxcx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxcx_epemguxcx/ajob1' in 4.9237 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxcx_epemguxcx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxcx_epemguxcx/G9/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.1482s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9321s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2135s for 16384 events => throughput is 7.67E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0026s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxcx_epemguxcx/G10/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.7457s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5274s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2156s for 16384 events => throughput is 7.60E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0027s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxcx_epemguxcx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/ajob1' in 40.9457 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/G46/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.1694s + [COUNTERS] Fortran Overhead ( 0 ) : 4.4756s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6903s for 16384 events => throughput is 2.37E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0035s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/G18/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.6860s + [COUNTERS] Fortran Overhead ( 0 ) : 3.9910s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6914s for 16384 events => throughput is 2.37E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/G39/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.5324s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8386s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6901s for 16384 events => throughput is 2.37E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0037s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.9488s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2525s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6929s for 16384 events => throughput is 2.36E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0035s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/G20/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.3654s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5898s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.7720s for 16384 events => throughput is 2.12E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/G17/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.9654s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2700s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6918s for 16384 events => throughput is 2.37E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/G35/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.1769s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4744s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6987s for 16384 events => throughput is 2.34E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0037s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/G71/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.1228s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4295s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6895s for 16384 events => throughput is 2.38E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0038s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/G27/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.7560s + [COUNTERS] Fortran Overhead ( 0 ) : 4.0624s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6899s for 16384 events => throughput is 2.37E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/G92/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.1636s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4262s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.7335s for 16384 events => throughput is 2.23E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0040s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/G80/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.4774s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7862s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6877s for 16384 events => throughput is 2.38E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0035s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.2564s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5638s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6890s for 16384 events => throughput is 2.38E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/G29/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.6373s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9441s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6896s for 16384 events => throughput is 2.38E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/G48/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.4967s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7989s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6942s for 16384 events => throughput is 2.36E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_epemgdux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_epemgdux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_epemgdux/ajob1' in 9.5994 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_epemgdux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_epemgdux/G46/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 0.9998s + [COUNTERS] Fortran Overhead ( 0 ) : 0.7845s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2127s for 16384 events => throughput is 7.70E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0026s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_epemgdux/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.3224s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1034s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2163s for 16384 events => throughput is 7.58E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0027s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_epemgdux/G23/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.1387s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9219s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2141s for 16384 events => throughput is 7.65E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0027s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_epemgdux/G47/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.0655s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8499s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2127s for 16384 events => throughput is 7.70E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0029s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_epemgdux/G8/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.4593s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2429s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2138s for 16384 events => throughput is 7.66E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0026s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_epemgdux/G7/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.2098s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9939s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2132s for 16384 events => throughput is 7.68E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0028s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_epemgdux/G28/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.3651s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1491s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2133s for 16384 events => throughput is 7.68E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0027s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_epemgdux/G32/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 0.9295s + [COUNTERS] Fortran Overhead ( 0 ) : 0.7143s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2126s for 16384 events => throughput is 7.71E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0026s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_epemgdux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/ajob1' in 36.8332 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/G12/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.8699s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4586s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.4077s for 16384 events => throughput is 1.16E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/G6/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.6337s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1651s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.4648s for 16384 events => throughput is 1.12E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0038s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/G11/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.6198s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2168s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.3993s for 16384 events => throughput is 1.17E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +INFO: The following Floating Point Exceptions have been reported: FE_UNDERFLOW +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.5735s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1675s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.4024s for 16384 events => throughput is 1.17E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/G69/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.4713s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0628s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.4047s for 16384 events => throughput is 1.17E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0038s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/G8/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.3531s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9376s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.4120s for 16384 events => throughput is 1.16E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/G84/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.7227s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3223s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.3968s for 16384 events => throughput is 1.17E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.9546s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5484s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.4026s for 16384 events => throughput is 1.17E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/G2/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.3384s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9312s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.4036s for 16384 events => throughput is 1.17E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/G80/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.1157s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7050s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.4067s for 16384 events => throughput is 1.16E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0040s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/G51/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.1410s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7319s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.4051s for 16384 events => throughput is 1.17E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0040s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/G70/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.4700s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0662s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.3999s for 16384 events => throughput is 1.17E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0038s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/G5/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.3933s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9831s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.4064s for 16384 events => throughput is 1.16E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0038s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamguux/ajob1' in 31.4805 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamguux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamguux/G18/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.1983s + [COUNTERS] Fortran Overhead ( 0 ) : 3.8041s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.3903s for 16384 events => throughput is 1.18E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0038s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamguux/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.6301s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2317s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.3946s for 16384 events => throughput is 1.17E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0038s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamguux/G20/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.9195s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5041s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.4118s for 16384 events => throughput is 1.16E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamguux/G17/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.5828s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1811s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.3981s for 16384 events => throughput is 1.17E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamguux/G35/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.8229s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4203s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.3988s for 16384 events => throughput is 1.17E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0038s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamguux/G27/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.0669s + [COUNTERS] Fortran Overhead ( 0 ) : 3.6438s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.4195s for 16384 events => throughput is 1.15E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamguux/G34/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.9589s + [COUNTERS] Fortran Overhead ( 0 ) : 3.5642s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.3908s for 16384 events => throughput is 1.18E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0038s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamguux/G80/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.1908s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7828s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.4043s for 16384 events => throughput is 1.17E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0037s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamguux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_taptamgdsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_taptamgdsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_taptamgdsx/ajob1' in 1.5350 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_taptamgdsx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_taptamgdsx/G28/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.5182s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0812s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4343s for 16384 events => throughput is 3.77E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0026s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_taptamgdsx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_taptamggg/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_taptamggg/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_taptamggg/ajob1' in 11.1739 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_taptamggg/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_taptamggg/G12/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.0815s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6686s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.4093s for 16384 events => throughput is 1.16E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_taptamggg/G9/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.9288s + [COUNTERS] Fortran Overhead ( 0 ) : 3.5272s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.3978s for 16384 events => throughput is 1.17E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0038s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_taptamggg/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.1197s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6947s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.4213s for 16384 events => throughput is 1.15E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0037s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_taptamggg/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxdx_epemguxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxdx_epemguxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxdx_epemguxdx/ajob1' in 13.5109 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxdx_epemguxdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxdx_epemguxdx/G14/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.3090s + [COUNTERS] Fortran Overhead ( 0 ) : 3.0951s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2110s for 16384 events => throughput is 7.76E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0028s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxdx_epemguxdx/G6/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.7299s + [COUNTERS] Fortran Overhead ( 0 ) : 4.5141s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2131s for 16384 events => throughput is 7.69E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0026s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxdx_epemguxdx/G22/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.4459s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2230s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2203s for 16384 events => throughput is 7.44E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0026s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxdx_epemguxdx/G8/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.4275s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2124s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2124s for 16384 events => throughput is 7.71E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0027s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxdx_epemguxdx/G27/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.5290s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3077s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2186s for 16384 events => throughput is 7.49E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0026s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxdx_epemguxdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_epemgdsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_epemgdsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_epemgdsx/ajob1' in 11.8803 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_epemgdsx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_epemgdsx/G31/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.1223s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9066s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2130s for 16384 events => throughput is 7.69E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0027s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_epemgdsx/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.1579s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9383s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2170s for 16384 events => throughput is 7.55E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0026s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_epemgdsx/G20/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.3224s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1069s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2129s for 16384 events => throughput is 7.69E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0026s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_epemgdsx/G25/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.9403s + [COUNTERS] Fortran Overhead ( 0 ) : 3.7216s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2161s for 16384 events => throughput is 7.58E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0027s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_epemgdsx/G28/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.2335s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0181s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2128s for 16384 events => throughput is 7.70E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0026s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_epemgdsx/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.0218s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8066s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2126s for 16384 events => throughput is 7.71E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0026s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_epemgdsx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamdddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamdddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamdddx/ajob1' in 4.3168 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamdddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamdddx/G17/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.3003s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2904s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.0060s for 16384 events => throughput is 1.63E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0039s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamdddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemggg/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemggg/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemggg/ajob1' in 13.8289 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemggg/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemggg/G56/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.3268s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6232s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6999s for 16384 events => throughput is 2.34E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0037s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemggg/G12/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.4456s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7477s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6941s for 16384 events => throughput is 2.36E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0038s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemggg/G11/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.4365s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7331s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6997s for 16384 events => throughput is 2.34E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0038s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemggg/G40/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.2196s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5243s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6917s for 16384 events => throughput is 2.37E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemggg/G8/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.0394s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3359s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6999s for 16384 events => throughput is 2.34E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0037s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemggg/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.2775s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5875s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6861s for 16384 events => throughput is 2.39E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0040s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemggg/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/ajob1' in 77.4836 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G14/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.4968s + [COUNTERS] Fortran Overhead ( 0 ) : 3.0846s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.4082s for 16384 events => throughput is 1.16E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0040s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G74/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.1735s + [COUNTERS] Fortran Overhead ( 0 ) : 3.7703s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.3995s for 16384 events => throughput is 1.17E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0037s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G12/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.5778s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1711s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.4030s for 16384 events => throughput is 1.17E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0037s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G6/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.4414s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0290s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.4088s for 16384 events => throughput is 1.16E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G13/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.3796s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9734s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.4022s for 16384 events => throughput is 1.17E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0040s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G66/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.9361s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5328s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.3996s for 16384 events => throughput is 1.17E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0037s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G11/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.7848s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3803s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.4010s for 16384 events => throughput is 1.17E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G9/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.5666s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1523s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.4106s for 16384 events => throughput is 1.16E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0037s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G40/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.1425s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7407s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.3982s for 16384 events => throughput is 1.17E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G35/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.2590s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7626s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.4924s for 16384 events => throughput is 1.10E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0040s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.0038s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5990s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.4011s for 16384 events => throughput is 1.17E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0037s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G69/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.4189s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9941s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.4212s for 16384 events => throughput is 1.15E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G7/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.2629s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8499s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.4093s for 16384 events => throughput is 1.16E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0037s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G65/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.0565s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6521s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.4008s for 16384 events => throughput is 1.17E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G71/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.7183s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3100s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.4046s for 16384 events => throughput is 1.17E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0037s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G1/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.3396s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9298s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.4061s for 16384 events => throughput is 1.17E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0037s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.6471s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2354s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.4079s for 16384 events => throughput is 1.16E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0038s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G2/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.7506s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3446s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.4022s for 16384 events => throughput is 1.17E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0038s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G4/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.0078s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6030s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.4010s for 16384 events => throughput is 1.17E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0037s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G79/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.2391s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8203s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.4149s for 16384 events => throughput is 1.16E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0038s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.9678s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5660s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.3981s for 16384 events => throughput is 1.17E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0037s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G70/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.3899s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9809s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.4050s for 16384 events => throughput is 1.17E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0039s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G5/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.6162s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1325s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.4798s for 16384 events => throughput is 1.11E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0039s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/ajob1' in 53.3761 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/G46/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.8118s + [COUNTERS] Fortran Overhead ( 0 ) : 4.1181s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6901s for 16384 events => throughput is 2.37E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0037s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/G18/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.0983s + [COUNTERS] Fortran Overhead ( 0 ) : 3.3930s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.7013s for 16384 events => throughput is 2.34E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0041s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.1595s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4653s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6906s for 16384 events => throughput is 2.37E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/G33/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.1545s + [COUNTERS] Fortran Overhead ( 0 ) : 3.4616s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6893s for 16384 events => throughput is 2.38E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/G20/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.3275s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6338s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6901s for 16384 events => throughput is 2.37E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/G47/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.9709s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2678s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6993s for 16384 events => throughput is 2.34E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0038s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/G17/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.8828s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1890s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6903s for 16384 events => throughput is 2.37E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/G25/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.9148s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2220s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6889s for 16384 events => throughput is 2.38E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0039s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/G35/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.0377s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3422s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6919s for 16384 events => throughput is 2.37E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/G65/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.9163s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2091s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.7030s for 16384 events => throughput is 2.33E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0042s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/G27/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.3894s + [COUNTERS] Fortran Overhead ( 0 ) : 3.6961s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6896s for 16384 events => throughput is 2.38E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0037s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/G34/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.1122s + [COUNTERS] Fortran Overhead ( 0 ) : 3.4055s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.7031s for 16384 events => throughput is 2.33E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/G80/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.5946s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9014s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6895s for 16384 events => throughput is 2.38E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/G67/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.3861s + [COUNTERS] Fortran Overhead ( 0 ) : 3.6918s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6906s for 16384 events => throughput is 2.37E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0037s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.1979s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5049s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6894s for 16384 events => throughput is 2.38E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/G48/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.2063s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5050s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6974s for 16384 events => throughput is 2.35E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0039s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_taptamgdux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_taptamgdux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_taptamgdux/ajob1' in 6.6903 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_taptamgdux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_taptamgdux/G14/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.9251s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4843s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4382s for 16384 events => throughput is 3.74E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0026s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_taptamgdux/G40/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.2784s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8383s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4374s for 16384 events => throughput is 3.75E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0026s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_taptamgdux/G44/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.4440s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0056s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4358s for 16384 events => throughput is 3.76E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0026s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_taptamgdux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemgddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemgddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemgddx/ajob1' in 7.5959 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemgddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemgddx/G56/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.5468s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0562s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4873s for 16384 events => throughput is 3.36E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemgddx/G12/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.4074s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9143s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4896s for 16384 events => throughput is 3.35E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0035s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemgddx/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.4282s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9372s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4875s for 16384 events => throughput is 3.36E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0035s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemgddx/G47/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.6160s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1148s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4978s for 16384 events => throughput is 3.29E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemgddx/G42/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.5279s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0296s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4946s for 16384 events => throughput is 3.31E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0037s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemgddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/ajob1' in 46.7998 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/G39/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.8081s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4093s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.3952s for 16384 events => throughput is 1.17E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/G12/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.6520s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2087s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.4398s for 16384 events => throughput is 1.14E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/G6/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.2996s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8948s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.4013s for 16384 events => throughput is 1.17E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/G66/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.1145s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7115s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.3992s for 16384 events => throughput is 1.17E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0038s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/G11/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.7718s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3757s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.3925s for 16384 events => throughput is 1.18E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/G9/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.9450s + [COUNTERS] Fortran Overhead ( 0 ) : 3.5414s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.3999s for 16384 events => throughput is 1.17E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0037s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.8145s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4135s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.3972s for 16384 events => throughput is 1.17E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0037s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/G69/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.5092s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1057s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.3998s for 16384 events => throughput is 1.17E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0037s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/G8/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.3573s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9583s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.3954s for 16384 events => throughput is 1.17E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/G71/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.7118s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3093s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.3989s for 16384 events => throughput is 1.17E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.6100s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2030s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.4031s for 16384 events => throughput is 1.17E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0038s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/G72/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.5791s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1732s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.4023s for 16384 events => throughput is 1.17E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/G79/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.1345s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7368s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.3938s for 16384 events => throughput is 1.18E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0040s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.2465s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8397s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.4029s for 16384 events => throughput is 1.17E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0039s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/G70/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.4211s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0124s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.4049s for 16384 events => throughput is 1.17E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0038s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/G5/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.6091s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2088s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.3968s for 16384 events => throughput is 1.17E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemdssx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemdssx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemdssx/ajob1' in 4.5487 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemdssx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemdssx/G46/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.3304s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1155s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2123s for 16384 events => throughput is 7.72E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0026s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemdssx/G17/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.1886s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9620s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2239s for 16384 events => throughput is 7.32E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0027s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemdssx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uc_taptamguc/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uc_taptamguc/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uc_taptamguc/ajob1' in 4.0682 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uc_taptamguc/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uc_taptamguc/G25/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.6978s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2610s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4341s for 16384 events => throughput is 3.77E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0027s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uc_taptamguc/G37/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.3406s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8684s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4693s for 16384 events => throughput is 3.49E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0029s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uc_taptamguc/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamgddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamgddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamgddx/ajob1' in 4.8400 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamgddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamgddx/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.6253s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6285s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.9930s for 16384 events => throughput is 1.65E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0037s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamgddx/G55/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.1846s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1897s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.9911s for 16384 events => throughput is 1.65E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0037s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamgddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemgddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemgddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemgddx/ajob1' in 1.7257 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemgddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemgddx/G8/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.7092s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4884s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2182s for 16384 events => throughput is 7.51E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0027s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemgddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamgddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamgddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamgddx/ajob1' in 28.9919 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamgddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamgddx/G31/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.5586s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1630s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.3920s for 16384 events => throughput is 1.18E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamgddx/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.6259s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2239s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.3970s for 16384 events => throughput is 1.17E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0050s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamgddx/G33/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.5573s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1625s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.3912s for 16384 events => throughput is 1.18E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamgddx/G20/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.7893s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3811s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.4046s for 16384 events => throughput is 1.17E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamgddx/G35/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.6638s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2457s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.4144s for 16384 events => throughput is 1.16E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0037s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamgddx/G71/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.4889s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0955s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.3898s for 16384 events => throughput is 1.18E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamgddx/G34/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.1963s + [COUNTERS] Fortran Overhead ( 0 ) : 3.7991s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.3937s for 16384 events => throughput is 1.18E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamgddx/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.9059s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5072s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.3950s for 16384 events => throughput is 1.17E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0037s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamgddx/G48/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.0828s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6778s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.4013s for 16384 events => throughput is 1.17E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamgddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemddxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemddxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemddxdx/ajob1' in 3.7504 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemddxdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemddxdx/G11/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.0293s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5385s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4871s for 16384 events => throughput is 3.36E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0037s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemddxdx/G71/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.6913s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2046s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4833s for 16384 events => throughput is 3.39E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemddxdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dxsx_epemgdxsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dxsx_epemgdxsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dxsx_epemgdxsx/ajob1' in 1.4580 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dxsx_epemgdxsx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dxsx_epemgdxsx/G27/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.4413s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2264s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2122s for 16384 events => throughput is 7.72E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0027s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dxsx_epemgdxsx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemduxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemduxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemduxdx/ajob1' in 4.2303 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemduxdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemduxdx/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.2736s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0568s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2141s for 16384 events => throughput is 7.65E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0027s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemduxdx/G22/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.9266s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7124s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2116s for 16384 events => throughput is 7.74E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0026s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemduxdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemguux/ajob1' in 15.4043 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemguux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemguux/G20/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.5335s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0427s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4873s for 16384 events => throughput is 3.36E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0035s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemguux/G27/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.3147s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8187s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4924s for 16384 events => throughput is 3.33E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemguux/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.8932s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4001s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4894s for 16384 events => throughput is 3.35E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0037s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemguux/G55/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.6883s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1961s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4885s for 16384 events => throughput is 3.35E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0037s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemguux/G37/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.5793s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0729s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5026s for 16384 events => throughput is 3.26E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0037s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemguux/G41/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.3919s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8982s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4903s for 16384 events => throughput is 3.34E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemguux/G57/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.1136s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6209s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4890s for 16384 events => throughput is 3.35E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0038s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemguux/G48/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.7801s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2867s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4900s for 16384 events => throughput is 3.34E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemguux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dd_taptamgdd/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dd_taptamgdd/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dd_taptamgdd/ajob1' in 5.0505 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dd_taptamgdd/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dd_taptamgdd/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.2156s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2161s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.9959s for 16384 events => throughput is 1.65E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0037s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dd_taptamgdd/G24/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.8047s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8121s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.9889s for 16384 events => throughput is 1.66E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0037s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dd_taptamgdd/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/ajob1' in 108.4156 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G14/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.0796s + [COUNTERS] Fortran Overhead ( 0 ) : 3.3848s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6911s for 16384 events => throughput is 2.37E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0037s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G38/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.4005s + [COUNTERS] Fortran Overhead ( 0 ) : 3.7065s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6905s for 16384 events => throughput is 2.37E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0035s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G39/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.0998s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4047s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6911s for 16384 events => throughput is 2.37E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0040s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G12/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.8692s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1741s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6913s for 16384 events => throughput is 2.37E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0038s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G6/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.5853s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8869s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6948s for 16384 events => throughput is 2.36E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0037s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G3/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.4789s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7749s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.7005s for 16384 events => throughput is 2.34E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0035s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G13/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.9412s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2408s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6968s for 16384 events => throughput is 2.35E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0035s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G66/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.6022s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9099s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6886s for 16384 events => throughput is 2.38E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0037s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G33/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.5837s + [COUNTERS] Fortran Overhead ( 0 ) : 4.8884s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6915s for 16384 events => throughput is 2.37E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0038s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G11/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.9821s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2838s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6947s for 16384 events => throughput is 2.36E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0035s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G9/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 21.2165s + [COUNTERS] Fortran Overhead ( 0 ) : 17.7585s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.4544s for 81920 events => throughput is 2.37E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G10/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.5714s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8729s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6949s for 16384 events => throughput is 2.36E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G40/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.2808s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5904s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6869s for 16384 events => throughput is 2.39E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G35/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.2599s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5663s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6898s for 16384 events => throughput is 2.38E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0038s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.0145s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3226s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6884s for 16384 events => throughput is 2.38E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0035s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G69/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.7278s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0301s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6938s for 16384 events => throughput is 2.36E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0039s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G8/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.5984s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9032s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6916s for 16384 events => throughput is 2.37E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G7/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.7304s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0362s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6905s for 16384 events => throughput is 2.37E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0037s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G65/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.6811s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9880s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6893s for 16384 events => throughput is 2.38E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0038s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G71/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.1373s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4436s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6900s for 16384 events => throughput is 2.37E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0037s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G1/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.9747s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2839s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6872s for 16384 events => throughput is 2.38E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.0517s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2948s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.7530s for 16384 events => throughput is 2.18E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0039s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G2/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.7959s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0990s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6931s for 16384 events => throughput is 2.36E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0038s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G4/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.5956s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9043s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6878s for 16384 events => throughput is 2.38E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0035s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G80/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.8188s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1232s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6919s for 16384 events => throughput is 2.37E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0037s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G41/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.2548s + [COUNTERS] Fortran Overhead ( 0 ) : 3.5528s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6983s for 16384 events => throughput is 2.35E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G77/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.5059s + [COUNTERS] Fortran Overhead ( 0 ) : 3.8127s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6895s for 16384 events => throughput is 2.38E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0037s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G76/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.9986s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3057s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6893s for 16384 events => throughput is 2.38E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G79/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.5305s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8410s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6860s for 16384 events => throughput is 2.39E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.1680s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4770s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6873s for 16384 events => throughput is 2.38E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0037s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G70/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.7008s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0048s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6925s for 16384 events => throughput is 2.37E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G5/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.7498s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0572s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6890s for 16384 events => throughput is 2.38E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ds_taptamgds/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ds_taptamgds/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ds_taptamgds/ajob1' in 2.2462 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ds_taptamgds/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ds_taptamgds/G10/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.2295s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7941s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4328s for 16384 events => throughput is 3.79E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0027s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ds_taptamgds/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_epemgudx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_epemgudx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_epemgudx/ajob1' in 9.4611 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_epemgudx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_epemgudx/G12/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.0494s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8235s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2231s for 16384 events => throughput is 7.34E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0028s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_epemgudx/G11/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.0027s + [COUNTERS] Fortran Overhead ( 0 ) : 0.7855s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2143s for 16384 events => throughput is 7.65E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0029s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_epemgudx/G17/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.6391s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4216s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2150s for 16384 events => throughput is 7.62E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0026s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_epemgudx/G10/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.1475s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9234s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2214s for 16384 events => throughput is 7.40E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0026s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_epemgudx/G27/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.2863s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0712s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2124s for 16384 events => throughput is 7.71E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0028s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_epemgudx/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.2537s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0368s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2141s for 16384 events => throughput is 7.65E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0027s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_epemgudx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_taptamgud/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_taptamgud/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_taptamgud/ajob1' in 15.5487 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_taptamgud/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_taptamgud/G12/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.2494s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8115s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4353s for 16384 events => throughput is 3.76E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0027s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_taptamgud/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.5244s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0912s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4304s for 16384 events => throughput is 3.81E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0028s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_taptamgud/G6/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.5845s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1446s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4372s for 16384 events => throughput is 3.75E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0027s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_taptamgud/G23/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.2841s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8357s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4457s for 16384 events => throughput is 3.68E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0027s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_taptamgud/G11/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.3037s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8615s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4396s for 16384 events => throughput is 3.73E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0026s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_taptamgud/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.6266s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1887s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4352s for 16384 events => throughput is 3.77E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0028s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_taptamgud/G22/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.8247s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3803s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4415s for 16384 events => throughput is 3.71E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0029s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_taptamgud/G7/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.6944s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2512s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4404s for 16384 events => throughput is 3.72E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0027s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_taptamgud/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.3343s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8961s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4355s for 16384 events => throughput is 3.76E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0028s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_taptamgud/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemdddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemdddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemdddx/ajob1' in 3.9876 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemdddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemdddx/G31/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.1398s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6469s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4895s for 16384 events => throughput is 3.35E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemdddx/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.8179s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3277s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4866s for 16384 events => throughput is 3.37E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemdddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamdssx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamdssx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamdssx/ajob1' in 3.2436 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamdssx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamdssx/G45/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.5203s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0767s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4409s for 16384 events => throughput is 3.72E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0026s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamdssx/G20/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.6932s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2567s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4340s for 16384 events => throughput is 3.78E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0026s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamdssx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuddx/ajob1' in 4.6254 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuddx/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.4851s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0448s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4376s for 16384 events => throughput is 3.74E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0026s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuddx/G20/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.5538s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1146s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4365s for 16384 events => throughput is 3.75E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0027s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuddx/G47/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.5433s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1072s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4335s for 16384 events => throughput is 3.78E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0026s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemuuxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemuuxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemuuxdx/ajob1' in 4.8494 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemuuxdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemuuxdx/G7/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.5199s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3018s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2155s for 16384 events => throughput is 7.60E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0026s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemuuxdx/G41/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.2999s + [COUNTERS] Fortran Overhead ( 0 ) : 3.0843s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2129s for 16384 events => throughput is 7.69E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0026s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemuuxdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/ajob1' in 32.4116 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G14/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.2869s + [COUNTERS] Fortran Overhead ( 0 ) : 3.0696s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2147s for 16384 events => throughput is 7.63E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0026s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G18/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.7300s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5133s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2140s for 16384 events => throughput is 7.66E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0027s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G39/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.0273s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8094s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2152s for 16384 events => throughput is 7.61E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0027s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G30/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.3369s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1217s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2125s for 16384 events => throughput is 7.71E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0027s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G12/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.0292s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8101s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2165s for 16384 events => throughput is 7.57E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0026s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.3991s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1850s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2114s for 16384 events => throughput is 7.75E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0027s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G6/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.6543s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4382s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2135s for 16384 events => throughput is 7.67E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0026s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G3/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.0562s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8397s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2137s for 16384 events => throughput is 7.67E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0028s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G23/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.0798s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8645s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2125s for 16384 events => throughput is 7.71E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0028s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G24/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.0614s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8457s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2131s for 16384 events => throughput is 7.69E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0026s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G13/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.4807s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2600s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2181s for 16384 events => throughput is 7.51E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0027s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G20/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.1039s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8889s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2123s for 16384 events => throughput is 7.72E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0027s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G25/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.6450s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4278s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2145s for 16384 events => throughput is 7.64E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0027s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.3388s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1238s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2124s for 16384 events => throughput is 7.71E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0026s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G8/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.1591s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9338s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2225s for 16384 events => throughput is 7.36E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0028s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G7/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.2342s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0186s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2130s for 16384 events => throughput is 7.69E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0026s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G4/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 0.8886s + [COUNTERS] Fortran Overhead ( 0 ) : 0.6729s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2132s for 16384 events => throughput is 7.68E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0026s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G41/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.5413s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3240s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2146s for 16384 events => throughput is 7.63E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0027s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G32/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.1047s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8905s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2115s for 16384 events => throughput is 7.75E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0027s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ucx_epemgucx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ucx_epemgucx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ucx_epemgucx/ajob1' in 3.6868 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ucx_epemgucx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ucx_epemgucx/G3/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 0.8567s + [COUNTERS] Fortran Overhead ( 0 ) : 0.6421s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2119s for 16384 events => throughput is 7.73E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0027s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ucx_epemgucx/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.4783s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2635s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2121s for 16384 events => throughput is 7.73E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0027s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ucx_epemgucx/G7/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.3083s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0920s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2136s for 16384 events => throughput is 7.67E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0027s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ucx_epemgucx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_taptamgudx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_taptamgudx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_taptamgudx/ajob1' in 8.8469 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_taptamgudx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_taptamgudx/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.3817s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9448s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4342s for 16384 events => throughput is 3.77E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0027s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_taptamgudx/G11/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.3407s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8932s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4447s for 16384 events => throughput is 3.68E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0028s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_taptamgudx/G17/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.8369s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3988s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4352s for 16384 events => throughput is 3.76E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0028s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_taptamgudx/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.6775s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2381s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4366s for 16384 events => throughput is 3.75E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0028s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_taptamgudx/G27/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.5396s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0998s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4371s for 16384 events => throughput is 3.75E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0027s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_taptamgudx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamuuxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamuuxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamuuxdx/ajob1' in 3.1910 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamuuxdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamuuxdx/G6/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.1742s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7410s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4305s for 16384 events => throughput is 3.81E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0027s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamuuxdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_taptamguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_taptamguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_taptamguux/ajob1' in 2.2382 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_taptamguux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_taptamguux/G48/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.2219s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2204s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.9978s for 16384 events => throughput is 1.64E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_taptamguux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemudux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemudux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemudux/ajob1' in 8.6467 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemudux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemudux/G43/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.7273s + [COUNTERS] Fortran Overhead ( 0 ) : 3.5100s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2147s for 16384 events => throughput is 7.63E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0026s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemudux/G8/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.3741s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1547s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2165s for 16384 events => throughput is 7.57E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0029s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemudux/G5/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.5020s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2868s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2126s for 16384 events => throughput is 7.71E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0027s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemudux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamsdxsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamsdxsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamsdxsx/ajob1' in 1.7322 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamsdxsx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamsdxsx/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.7157s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2670s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4459s for 16384 events => throughput is 3.67E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0027s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamsdxsx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_epemguu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_epemguu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_epemguu/ajob1' in 19.4291 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_epemguu/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_epemguu/G30/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.4800s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9885s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4880s for 16384 events => throughput is 3.36E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0035s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_epemguu/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.4394s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9486s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4874s for 16384 events => throughput is 3.36E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_epemguu/G9/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.8002s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3056s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4908s for 16384 events => throughput is 3.34E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0037s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_epemguu/G49/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.3988s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9013s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4938s for 16384 events => throughput is 3.32E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0037s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_epemguu/G10/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.7212s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2283s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4895s for 16384 events => throughput is 3.35E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_epemguu/G81/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.5186s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0237s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4916s for 16384 events => throughput is 3.33E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_epemguu/G62/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.2906s + [COUNTERS] Fortran Overhead ( 0 ) : 0.7984s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4887s for 16384 events => throughput is 3.35E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0034s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_epemguu/G51/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.3573s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8568s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4968s for 16384 events => throughput is 3.30E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_epemguu/G28/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.3000s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8070s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4894s for 16384 events => throughput is 3.35E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_epemguu/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uc_epemguc/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uc_epemguc/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uc_epemguc/ajob1' in 2.2782 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uc_epemguc/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uc_epemguc/G12/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 0.9842s + [COUNTERS] Fortran Overhead ( 0 ) : 0.7679s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2137s for 16384 events => throughput is 7.67E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0026s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uc_epemguc/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.2641s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0457s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2157s for 16384 events => throughput is 7.60E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0026s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uc_epemguc/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ds_epemgds/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ds_epemgds/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ds_epemgds/ajob1' in 1.3083 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ds_epemgds/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ds_epemgds/G27/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.2915s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0678s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2209s for 16384 events => throughput is 7.42E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0027s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ds_epemgds/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_taptamguu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_taptamguu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_taptamguu/ajob1' in 6.6248 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_taptamguu/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_taptamguu/G52/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.9394s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9425s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.9934s for 16384 events => throughput is 1.65E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0035s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_taptamguu/G49/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.6942s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6965s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.9941s for 16384 events => throughput is 1.65E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0036s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_taptamguu/G32/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.9482s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9491s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.9956s for 16384 events => throughput is 1.65E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0035s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_taptamguu/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished in 942.5023 seconds +__CUDACPP_DEBUG: gen_ximprove.launch finished in 942.9244 seconds +__CUDACPP_DEBUG: GridPackCmd.refine4grid finished in 942.9252 seconds +__CUDACPP_DEBUG: GridPackCmd.launch (3a) combine_events +combine_events +__CUDACPP_DEBUG: GridPackCmd.launch (3b) store_events +store_events + === Results Summary for run: GridRun_21 tag: tag_1 === + + Cross-section : 0 +- 0 pb + Nb of events : 500 + +INFO: No version of lhapdf. Can not run systematics computation +decay_events -from_cards +__CUDACPP_DEBUG: GridPackCmd.launch finished in 960.2111 seconds +__CUDACPP_DEBUG: GridPackCmd.__init__ back from self.launch +DONE +write ./events.lhe.gz +END: Wed Sep 11 03:45:58 AM CEST 2024 +ELAPSED: 962 seconds diff --git a/epochX/cudacpp/tlau/fromgridpacks/pp_dy3j.mad/cppnone/output.txt b/epochX/cudacpp/tlau/fromgridpacks/pp_dy3j.mad/cppnone/output.txt new file mode 100644 index 0000000000..4c6918caa0 --- /dev/null +++ b/epochX/cudacpp/tlau/fromgridpacks/pp_dy3j.mad/cppnone/output.txt @@ -0,0 +1,2336 @@ +START: Wed Sep 11 02:34:16 AM CEST 2024 +On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU (nproc=4)] [GPU: 1x Tesla V100S-PCIE-32GB]: +CUDACPP_RUNTIME_DISABLEFPE is set +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 02:34 madevent/SubProcesses/P0_dc_epemgdc/madevent -> build.none_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 02:34 madevent/SubProcesses/P0_dc_taptamgdc/madevent -> build.none_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 02:34 madevent/SubProcesses/P0_dd_epemgdd/madevent -> build.none_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 02:34 madevent/SubProcesses/P0_dd_taptamgdd/madevent -> build.none_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 02:34 madevent/SubProcesses/P0_ddx_epemgddx/madevent -> build.none_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 02:34 madevent/SubProcesses/P0_ddx_epemggg/madevent -> build.none_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 02:34 madevent/SubProcesses/P0_ddx_epemgssx/madevent -> build.none_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 02:34 madevent/SubProcesses/P0_ddx_epemguux/madevent -> build.none_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 02:34 madevent/SubProcesses/P0_ddx_taptamgddx/madevent -> build.none_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 02:34 madevent/SubProcesses/P0_ddx_taptamggg/madevent -> build.none_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 02:34 madevent/SubProcesses/P0_ddx_taptamgssx/madevent -> build.none_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 02:34 madevent/SubProcesses/P0_ddx_taptamguux/madevent -> build.none_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 02:34 madevent/SubProcesses/P0_ds_epemgds/madevent -> build.none_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 02:34 madevent/SubProcesses/P0_ds_taptamgds/madevent -> build.none_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 02:34 madevent/SubProcesses/P0_dsx_epemgdsx/madevent -> build.none_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 02:34 madevent/SubProcesses/P0_dsx_taptamgdsx/madevent -> build.none_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 02:34 madevent/SubProcesses/P0_dux_epemgdux/madevent -> build.none_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 02:34 madevent/SubProcesses/P0_dux_taptamgdux/madevent -> build.none_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 02:34 madevent/SubProcesses/P0_dxcx_epemgdxcx/madevent -> build.none_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 02:34 madevent/SubProcesses/P0_dxcx_taptamgdxcx/madevent -> build.none_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 02:34 madevent/SubProcesses/P0_dxdx_epemgdxdx/madevent -> build.none_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 02:34 madevent/SubProcesses/P0_dxdx_taptamgdxdx/madevent -> build.none_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 02:34 madevent/SubProcesses/P0_dxsx_epemgdxsx/madevent -> build.none_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 02:34 madevent/SubProcesses/P0_dxsx_taptamgdxsx/madevent -> build.none_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 02:34 madevent/SubProcesses/P0_gd_epemdddx/madevent -> build.none_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 02:34 madevent/SubProcesses/P0_gd_epemdssx/madevent -> build.none_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 02:34 madevent/SubProcesses/P0_gd_epemggd/madevent -> build.none_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 02:34 madevent/SubProcesses/P0_gd_epemudux/madevent -> build.none_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 02:34 madevent/SubProcesses/P0_gd_taptamdddx/madevent -> build.none_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 02:34 madevent/SubProcesses/P0_gd_taptamdssx/madevent -> build.none_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 02:34 madevent/SubProcesses/P0_gd_taptamggd/madevent -> build.none_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 02:34 madevent/SubProcesses/P0_gd_taptamudux/madevent -> build.none_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 02:34 madevent/SubProcesses/P0_gdx_epemddxdx/madevent -> build.none_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 02:34 madevent/SubProcesses/P0_gdx_epemggdx/madevent -> build.none_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 02:34 madevent/SubProcesses/P0_gdx_epemsdxsx/madevent -> build.none_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 02:34 madevent/SubProcesses/P0_gdx_epemuuxdx/madevent -> build.none_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 02:34 madevent/SubProcesses/P0_gdx_taptamddxdx/madevent -> build.none_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 02:34 madevent/SubProcesses/P0_gdx_taptamggdx/madevent -> build.none_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 02:34 madevent/SubProcesses/P0_gdx_taptamsdxsx/madevent -> build.none_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 02:34 madevent/SubProcesses/P0_gdx_taptamuuxdx/madevent -> build.none_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 02:34 madevent/SubProcesses/P0_gg_epemgddx/madevent -> build.none_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 02:34 madevent/SubProcesses/P0_gg_epemguux/madevent -> build.none_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 02:34 madevent/SubProcesses/P0_gg_taptamgddx/madevent -> build.none_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 02:34 madevent/SubProcesses/P0_gg_taptamguux/madevent -> build.none_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 02:34 madevent/SubProcesses/P0_gu_epemggu/madevent -> build.none_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 02:34 madevent/SubProcesses/P0_gu_epemuccx/madevent -> build.none_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 02:34 madevent/SubProcesses/P0_gu_epemuddx/madevent -> build.none_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 02:34 madevent/SubProcesses/P0_gu_epemuuux/madevent -> build.none_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 02:34 madevent/SubProcesses/P0_gu_taptamggu/madevent -> build.none_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 02:34 madevent/SubProcesses/P0_gu_taptamuccx/madevent -> build.none_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 02:34 madevent/SubProcesses/P0_gu_taptamuddx/madevent -> build.none_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 02:34 madevent/SubProcesses/P0_gu_taptamuuux/madevent -> build.none_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 02:34 madevent/SubProcesses/P0_gux_epemcuxcx/madevent -> build.none_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 02:34 madevent/SubProcesses/P0_gux_epemduxdx/madevent -> build.none_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 02:34 madevent/SubProcesses/P0_gux_epemggux/madevent -> build.none_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 02:34 madevent/SubProcesses/P0_gux_epemuuxux/madevent -> build.none_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 02:34 madevent/SubProcesses/P0_gux_taptamcuxcx/madevent -> build.none_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 02:34 madevent/SubProcesses/P0_gux_taptamduxdx/madevent -> build.none_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 02:34 madevent/SubProcesses/P0_gux_taptamggux/madevent -> build.none_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 02:34 madevent/SubProcesses/P0_gux_taptamuuxux/madevent -> build.none_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 02:34 madevent/SubProcesses/P0_uc_epemguc/madevent -> build.none_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 02:34 madevent/SubProcesses/P0_uc_taptamguc/madevent -> build.none_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 02:34 madevent/SubProcesses/P0_ucx_epemgucx/madevent -> build.none_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 02:34 madevent/SubProcesses/P0_ucx_taptamgucx/madevent -> build.none_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 02:34 madevent/SubProcesses/P0_ud_epemgud/madevent -> build.none_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 02:34 madevent/SubProcesses/P0_ud_taptamgud/madevent -> build.none_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 02:34 madevent/SubProcesses/P0_udx_epemgudx/madevent -> build.none_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 02:34 madevent/SubProcesses/P0_udx_taptamgudx/madevent -> build.none_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 02:34 madevent/SubProcesses/P0_uu_epemguu/madevent -> build.none_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 02:34 madevent/SubProcesses/P0_uu_taptamguu/madevent -> build.none_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 02:34 madevent/SubProcesses/P0_uux_epemgccx/madevent -> build.none_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 02:34 madevent/SubProcesses/P0_uux_epemgddx/madevent -> build.none_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 02:34 madevent/SubProcesses/P0_uux_epemggg/madevent -> build.none_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 02:34 madevent/SubProcesses/P0_uux_epemguux/madevent -> build.none_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 02:34 madevent/SubProcesses/P0_uux_taptamgccx/madevent -> build.none_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 02:34 madevent/SubProcesses/P0_uux_taptamgddx/madevent -> build.none_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 02:34 madevent/SubProcesses/P0_uux_taptamggg/madevent -> build.none_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 02:34 madevent/SubProcesses/P0_uux_taptamguux/madevent -> build.none_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 02:34 madevent/SubProcesses/P0_uxcx_epemguxcx/madevent -> build.none_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 02:34 madevent/SubProcesses/P0_uxcx_taptamguxcx/madevent -> build.none_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 02:34 madevent/SubProcesses/P0_uxdx_epemguxdx/madevent -> build.none_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 02:34 madevent/SubProcesses/P0_uxdx_taptamguxdx/madevent -> build.none_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 02:34 madevent/SubProcesses/P0_uxux_epemguxux/madevent -> build.none_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 02:34 madevent/SubProcesses/P0_uxux_taptamguxux/madevent -> build.none_m_inl0_hrd0/madevent_cpp +Now generating 500 events with random seed 21 and granularity 1 +************************************************************ +* * +* W E L C O M E to * +* M A D G R A P H 5 _ a M C @ N L O * +* M A D E V E N T * +* * +* * * * +* * * * * * +* * * * * 5 * * * * * +* * * * * * +* * * * +* * +* VERSION 3.5.3_lo_vect * +* * +* The MadGraph5_aMC@NLO Development Team - Find us at * +* https://server06.fynu.ucl.ac.be/projects/madgraph * +* * +* Type 'help' for in-line help. * +* * +************************************************************ +INFO: load configuration from /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/Cards/me5_configuration.txt +INFO: load configuration from /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/Cards/me5_configuration.txt +Using default text editor "vi". Set another one in ./input/mg5_configuration.txt +Using default eps viewer "evince". Set another one in ./input/mg5_configuration.txt +Using default web browser "firefox". Set another one in ./input/mg5_configuration.txt +WRITE GRIDCARD /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent +__CUDACPP_DEBUG: GridPackCmd.__init__ will call self.launch +__CUDACPP_DEBUG: GridPackCmd.launch starting +__CUDACPP_DEBUG: GridPackCmd.launch (1) restore_data +generate 500 events +P0_gg_epemguux +P0_gg_epemgddx +P0_gg_taptamguux +P0_gg_taptamgddx +P0_gu_epemggu +P0_gd_epemggd +P0_gux_epemggux +P0_gdx_epemggdx +P0_gu_taptamggu +P0_gd_taptamggd +P0_gux_taptamggux +P0_gdx_taptamggdx +P0_uux_epemggg +P0_ddx_epemggg +P0_uux_taptamggg +P0_ddx_taptamggg +P0_gu_epemuuux +P0_gd_epemdddx +P0_gux_epemuuxux +P0_gdx_epemddxdx +P0_gu_epemuddx +P0_gu_epemuccx +P0_gd_epemudux +P0_gd_epemdssx +P0_gux_epemduxdx +P0_gux_epemcuxcx +P0_gdx_epemuuxdx +P0_gdx_epemsdxsx +P0_gu_taptamuuux +P0_gd_taptamdddx +P0_gux_taptamuuxux +P0_gdx_taptamddxdx +P0_gu_taptamuddx +P0_gu_taptamuccx +P0_gd_taptamudux +P0_gd_taptamdssx +P0_gux_taptamduxdx +P0_gux_taptamcuxcx +P0_gdx_taptamuuxdx +P0_gdx_taptamsdxsx +P0_uu_epemguu +P0_uux_epemguux +P0_dd_epemgdd +P0_ddx_epemgddx +P0_uxux_epemguxux +P0_dxdx_epemgdxdx +P0_ud_epemgud +P0_uc_epemguc +P0_uux_epemgddx +P0_uux_epemgccx +P0_udx_epemgudx +P0_ucx_epemgucx +P0_dc_epemgdc +P0_ds_epemgds +P0_dux_epemgdux +P0_ddx_epemguux +P0_ddx_epemgssx +P0_dsx_epemgdsx +P0_uxdx_epemguxdx +P0_uxcx_epemguxcx +P0_dxcx_epemgdxcx +P0_dxsx_epemgdxsx +P0_uu_taptamguu +P0_uux_taptamguux +P0_dd_taptamgdd +P0_ddx_taptamgddx +P0_uxux_taptamguxux +P0_dxdx_taptamgdxdx +P0_ud_taptamgud +P0_uc_taptamguc +P0_uux_taptamgddx +P0_uux_taptamgccx +P0_udx_taptamgudx +P0_ucx_taptamgucx +P0_dc_taptamgdc +P0_ds_taptamgds +P0_dux_taptamgdux +P0_ddx_taptamguux +P0_ddx_taptamgssx +P0_dsx_taptamgdsx +P0_uxdx_taptamguxdx +P0_uxcx_taptamguxcx +P0_dxcx_taptamgdxcx +P0_dxsx_taptamgdxsx +__CUDACPP_DEBUG: GridPackCmd.launch (2) refine4grid +__CUDACPP_DEBUG: GridPackCmd.refine4grid starting +Using random number seed offset = 21 +INFO: Generating 500.0 unweighted events. +__CUDACPP_DEBUG: gen_ximprove.launch starting +sum of cpu time of last step: 5h02m29s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event starting +INFO: need to improve 362 channels +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuuux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuuux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuuux/ajob1' in 14.7776 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuuux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuuux/G12/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.6523s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2318s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.4073s for 16384 events => throughput is 6.81E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0132s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuuux/G20/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.7888s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3536s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.4223s for 16384 events => throughput is 6.76E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0130s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuuux/G29/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.2637s + [COUNTERS] Fortran Overhead ( 0 ) : 4.8297s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.4206s for 16384 events => throughput is 6.77E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0133s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuuux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemggg/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemggg/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemggg/ajob1' in 15.7928 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemggg/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemggg/G43/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.3362s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0846s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.2385s for 16384 events => throughput is 5.06E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0131s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemggg/G10/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.3969s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1481s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.2353s for 16384 events => throughput is 5.06E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0134s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemggg/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.0156s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7202s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.2815s for 16384 events => throughput is 4.99E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0140s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemggg/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/ajob1' in 83.5353 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G39/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.4345s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2297s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.1921s for 16384 events => throughput is 5.13E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0127s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G12/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.6749s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4483s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.2133s for 16384 events => throughput is 5.10E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0132s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G6/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.3722s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1676s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.1919s for 16384 events => throughput is 5.13E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0128s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G13/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.6960s + [COUNTERS] Fortran Overhead ( 0 ) : 3.4964s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.1871s for 16384 events => throughput is 5.14E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0126s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G11/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.5152s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3190s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.1830s for 16384 events => throughput is 5.15E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0132s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.4373s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2297s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.1940s for 16384 events => throughput is 5.13E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0135s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G69/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.2640s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0667s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.1837s for 16384 events => throughput is 5.15E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0137s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G8/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.2609s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0633s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.1838s for 16384 events => throughput is 5.15E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0139s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G7/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.2035s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0003s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.1899s for 16384 events => throughput is 5.14E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0133s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G65/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.0755s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8819s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.1806s for 16384 events => throughput is 5.15E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0129s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G1/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.6528s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4434s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.1964s for 16384 events => throughput is 5.13E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0131s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.6436s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4523s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.1786s for 16384 events => throughput is 5.15E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0127s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G82/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.9351s + [COUNTERS] Fortran Overhead ( 0 ) : 3.7432s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.1791s for 16384 events => throughput is 5.15E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0129s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G80/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.0169s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8187s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.1849s for 16384 events => throughput is 5.14E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0133s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G51/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.7769s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5799s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.1841s for 16384 events => throughput is 5.15E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0129s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G70/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.2836s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0776s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.1925s for 16384 events => throughput is 5.13E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0135s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G5/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.0604s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8531s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.1942s for 16384 events => throughput is 5.13E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0131s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamggux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamggux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamggux/ajob1' in 83.6422 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamggux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamggux/G59/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.4833s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0817s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.3879s for 16384 events => throughput is 2.56E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0138s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamggux/G12/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.8640s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4797s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.3713s for 16384 events => throughput is 2.57E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0129s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamggux/G13/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 9.8729s + [COUNTERS] Fortran Overhead ( 0 ) : 3.4776s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.3817s for 16384 events => throughput is 2.57E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0135s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamggux/G11/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.6977s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3201s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.3647s for 16384 events => throughput is 2.57E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0129s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamggux/G9/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 9.5408s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1349s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.3925s for 16384 events => throughput is 2.56E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0133s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamggux/G10/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 9.6734s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2548s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.4055s for 16384 events => throughput is 2.56E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0131s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamggux/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.6097s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2199s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.3761s for 16384 events => throughput is 2.57E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0137s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamggux/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.9643s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5800s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.3713s for 16384 events => throughput is 2.57E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0129s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamggux/G80/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.2593s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8282s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.4178s for 16384 events => throughput is 2.55E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0133s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamggux/G70/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.5391s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1033s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.4228s for 16384 events => throughput is 2.55E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0131s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamggux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemsdxsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemsdxsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemsdxsx/ajob1' in 2.8203 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemsdxsx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemsdxsx/G8/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.8035s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8305s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.9646s for 16384 events => throughput is 1.70E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0084s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemsdxsx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuccx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuccx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuccx/ajob1' in 3.7149 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuccx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuccx/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.6983s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7060s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.9836s for 16384 events => throughput is 8.26E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0087s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuccx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuuux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuuux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuuux/ajob1' in 13.8665 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuuux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuuux/G81/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.2902s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5960s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.6814s for 16384 events => throughput is 3.50E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0128s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuuux/G32/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.5462s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8659s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.6676s for 16384 events => throughput is 3.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0127s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuuux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamggg/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamggg/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamggg/ajob1' in 15.9472 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamggg/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamggg/G56/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.8846s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4790s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.3920s for 16384 events => throughput is 2.56E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0136s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamggg/G39/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.0325s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6224s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.3970s for 16384 events => throughput is 2.56E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0131s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamggg/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamduxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamduxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamduxdx/ajob1' in 8.8300 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamduxdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamduxdx/G45/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.0930s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1309s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.9537s for 16384 events => throughput is 8.39E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0083s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamduxdx/G17/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.7070s + [COUNTERS] Fortran Overhead ( 0 ) : 3.7449s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.9537s for 16384 events => throughput is 8.39E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0083s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamduxdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuccx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuccx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuccx/ajob1' in 5.2142 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuccx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuccx/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.5726s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5978s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.9661s for 16384 events => throughput is 1.70E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0087s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuccx/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.6115s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6395s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.9635s for 16384 events => throughput is 1.70E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0085s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuccx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuddx/ajob1' in 19.1646 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuddx/G46/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.0843s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1053s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.9708s for 16384 events => throughput is 1.69E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0082s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuddx/G18/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.3066s + [COUNTERS] Fortran Overhead ( 0 ) : 3.3296s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.9685s for 16384 events => throughput is 1.69E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0085s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuddx/G39/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.9542s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9743s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.9715s for 16384 events => throughput is 1.69E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0083s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuddx/G45/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.1114s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1307s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.9721s for 16384 events => throughput is 1.69E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0086s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuddx/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.0858s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0965s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.9811s for 16384 events => throughput is 1.67E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0082s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuddx/G20/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.2156s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2379s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.9690s for 16384 events => throughput is 1.69E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0087s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuddx/G47/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.2364s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2626s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.9656s for 16384 events => throughput is 1.70E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0082s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuddx/G7/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.0605s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0852s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.9667s for 16384 events => throughput is 1.69E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0086s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/ajob1' in 112.3617 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G14/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.0668s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8809s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.1728s for 16384 events => throughput is 5.16E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0131s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G74/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.1228s + [COUNTERS] Fortran Overhead ( 0 ) : 3.9282s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.1815s for 16384 events => throughput is 5.15E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0131s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G12/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.6013s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4000s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.1882s for 16384 events => throughput is 5.14E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0131s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G6/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.4349s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2451s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.1769s for 16384 events => throughput is 5.16E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0129s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G13/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.2471s + [COUNTERS] Fortran Overhead ( 0 ) : 4.0572s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.1769s for 16384 events => throughput is 5.16E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0130s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G83/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.6403s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4125s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.2150s for 16384 events => throughput is 5.10E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0128s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G33/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.5082s + [COUNTERS] Fortran Overhead ( 0 ) : 4.3150s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.1806s for 16384 events => throughput is 5.15E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0127s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G11/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.4066s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2047s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.1887s for 16384 events => throughput is 5.14E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0131s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G9/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.4258s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2236s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.1888s for 16384 events => throughput is 5.14E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0133s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G10/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.4781s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2809s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.1834s for 16384 events => throughput is 5.15E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0139s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.4104s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2204s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.1768s for 16384 events => throughput is 5.16E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0133s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G69/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.2434s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0520s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.1782s for 16384 events => throughput is 5.16E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0132s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G65/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.7003s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5056s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.1813s for 16384 events => throughput is 5.15E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0134s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G1/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.6677s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4785s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.1759s for 16384 events => throughput is 5.16E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0133s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.6109s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4109s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.1871s for 16384 events => throughput is 5.14E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0130s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G72/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.4889s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2957s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.1805s for 16384 events => throughput is 5.15E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0127s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G2/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.1625s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9689s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.1809s for 16384 events => throughput is 5.15E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0127s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G80/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.9576s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7546s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.1894s for 16384 events => throughput is 5.14E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0136s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G79/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.4069s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2085s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.1857s for 16384 events => throughput is 5.14E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0126s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G70/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.3077s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1097s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.1850s for 16384 events => throughput is 5.14E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0130s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G5/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.1900s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9646s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.2124s for 16384 events => throughput is 5.10E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0130s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamguux/ajob1' in 3.4503 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamguux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamguux/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.4336s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4716s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.9530s for 16384 events => throughput is 8.39E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0090s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamguux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/ajob1' in 118.2230 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G14/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.0978s + [COUNTERS] Fortran Overhead ( 0 ) : 3.8385s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.2465s for 16384 events => throughput is 5.05E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0128s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G12/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.3492s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1525s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.1838s for 16384 events => throughput is 5.15E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0129s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G6/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.1777s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9843s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.1805s for 16384 events => throughput is 5.15E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0129s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G3/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.9246s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7238s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.1878s for 16384 events => throughput is 5.14E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0130s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G13/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.4704s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2474s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.2102s for 16384 events => throughput is 5.10E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0128s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G11/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.4861s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2862s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.1868s for 16384 events => throughput is 5.14E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0131s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G9/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.4670s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2745s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.1797s for 16384 events => throughput is 5.15E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0128s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G10/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.0014s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8134s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.1747s for 16384 events => throughput is 5.16E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0133s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G40/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.8156s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6110s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.1907s for 16384 events => throughput is 5.13E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0140s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G78/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.1635s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9622s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.1881s for 16384 events => throughput is 5.14E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0132s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G35/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.9717s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7795s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.1789s for 16384 events => throughput is 5.15E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0133s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.5270s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3405s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.1729s for 16384 events => throughput is 5.16E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0135s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G69/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.2597s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0567s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.1897s for 16384 events => throughput is 5.14E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0133s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G7/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.0869s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8936s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.1803s for 16384 events => throughput is 5.15E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0130s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G71/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.4843s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2907s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.1805s for 16384 events => throughput is 5.15E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0131s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G1/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.6566s + [COUNTERS] Fortran Overhead ( 0 ) : 3.4709s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.1725s for 16384 events => throughput is 5.16E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0131s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.4487s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2502s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.1852s for 16384 events => throughput is 5.14E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0133s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G4/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.9139s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7137s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.1873s for 16384 events => throughput is 5.14E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0129s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G80/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.3469s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1464s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.1872s for 16384 events => throughput is 5.14E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0133s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G79/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.9476s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7162s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.2181s for 16384 events => throughput is 5.09E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0133s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.8147s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6182s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.1833s for 16384 events => throughput is 5.15E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0133s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G70/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.2192s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0272s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.1793s for 16384 events => throughput is 5.15E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0127s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G5/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.2814s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0869s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.1816s for 16384 events => throughput is 5.15E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0128s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxcx_epemguxcx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxcx_epemguxcx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxcx_epemguxcx/ajob1' in 6.4908 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxcx_epemguxcx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxcx_epemguxcx/G9/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.9517s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9727s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.9704s for 16384 events => throughput is 1.69E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0086s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxcx_epemguxcx/G10/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.5091s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5329s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.9677s for 16384 events => throughput is 1.69E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0085s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxcx_epemguxcx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/ajob1' in 75.7638 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/G46/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.6801s + [COUNTERS] Fortran Overhead ( 0 ) : 4.4959s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.1712s for 16384 events => throughput is 5.17E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0130s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/G18/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.2238s + [COUNTERS] Fortran Overhead ( 0 ) : 4.0415s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.1693s for 16384 events => throughput is 5.17E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0130s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/G39/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.0272s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8450s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.1689s for 16384 events => throughput is 5.17E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0134s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.4387s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2462s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.1797s for 16384 events => throughput is 5.15E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0128s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/G20/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.7811s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5817s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.1861s for 16384 events => throughput is 5.14E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0133s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/G17/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.4472s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2684s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.1658s for 16384 events => throughput is 5.18E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0130s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/G35/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.7107s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5017s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.1955s for 16384 events => throughput is 5.13E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0134s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/G71/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.5941s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4146s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.1669s for 16384 events => throughput is 5.17E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0127s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/G27/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.2575s + [COUNTERS] Fortran Overhead ( 0 ) : 4.0737s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.1707s for 16384 events => throughput is 5.17E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0132s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/G92/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.5742s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3825s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.1788s for 16384 events => throughput is 5.15E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0128s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/G80/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.9793s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7874s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.1791s for 16384 events => throughput is 5.15E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0128s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.7564s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5774s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.1661s for 16384 events => throughput is 5.17E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0128s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/G29/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.1201s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9439s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.1631s for 16384 events => throughput is 5.18E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0131s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/G48/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.9831s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7997s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.1704s for 16384 events => throughput is 5.17E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0130s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_epemgdux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_epemgdux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_epemgdux/ajob1' in 15.7335 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_epemgdux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_epemgdux/G46/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.7738s + [COUNTERS] Fortran Overhead ( 0 ) : 0.7828s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.9826s for 16384 events => throughput is 1.67E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0085s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_epemgdux/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.0829s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1031s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.9712s for 16384 events => throughput is 1.69E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0086s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_epemgdux/G23/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.9116s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9216s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.9814s for 16384 events => throughput is 1.67E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0086s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_epemgdux/G47/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.8337s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8556s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.9698s for 16384 events => throughput is 1.69E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0083s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_epemgdux/G8/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.2035s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2190s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.9756s for 16384 events => throughput is 1.68E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0088s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_epemgdux/G7/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.9772s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0005s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.9683s for 16384 events => throughput is 1.69E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0085s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_epemgdux/G28/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.1336s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1500s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.9752s for 16384 events => throughput is 1.68E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0084s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_epemgdux/G32/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.7082s + [COUNTERS] Fortran Overhead ( 0 ) : 0.7247s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.9751s for 16384 events => throughput is 1.68E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0084s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_epemgdux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/ajob1' in 101.7502 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/G12/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.8302s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4588s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.3582s for 16384 events => throughput is 2.58E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0133s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/G6/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.5496s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1497s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.3870s for 16384 events => throughput is 2.57E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0130s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/G11/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.6273s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2302s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.3836s for 16384 events => throughput is 2.57E+03 events/s +INFO: The following Floating Point Exceptions have been reported: FE_UNDERFLOW + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0135s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.5635s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1929s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.3575s for 16384 events => throughput is 2.58E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0130s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/G69/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.4534s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0629s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.3773s for 16384 events => throughput is 2.57E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0132s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/G8/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.4305s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9440s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.4730s for 16384 events => throughput is 2.53E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0135s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/G84/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.7068s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3260s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.3675s for 16384 events => throughput is 2.57E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0133s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.9421s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5573s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.3718s for 16384 events => throughput is 2.57E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0130s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/G2/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 9.4273s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9709s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.4433s for 16384 events => throughput is 2.54E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0130s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/G80/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.0876s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6992s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.3754s for 16384 events => throughput is 2.57E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0130s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/G51/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.1066s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7319s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.3611s for 16384 events => throughput is 2.58E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0135s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/G70/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.4645s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0738s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.3777s for 16384 events => throughput is 2.57E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0130s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/G5/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.3823s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9827s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.3864s for 16384 events => throughput is 2.57E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0133s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamguux/ajob1' in 71.4160 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamguux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamguux/G18/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 10.1935s + [COUNTERS] Fortran Overhead ( 0 ) : 3.8092s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.3711s for 16384 events => throughput is 2.57E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0132s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamguux/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.6160s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2307s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.3720s for 16384 events => throughput is 2.57E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0133s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamguux/G20/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.8908s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5015s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.3763s for 16384 events => throughput is 2.57E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0131s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamguux/G17/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 9.6270s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2348s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.3794s for 16384 events => throughput is 2.57E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0128s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamguux/G35/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.8273s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4296s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.3847s for 16384 events => throughput is 2.57E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0130s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamguux/G27/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 10.0223s + [COUNTERS] Fortran Overhead ( 0 ) : 3.6226s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.3866s for 16384 events => throughput is 2.57E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0131s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamguux/G34/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 9.9706s + [COUNTERS] Fortran Overhead ( 0 ) : 3.5850s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.3717s for 16384 events => throughput is 2.57E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0140s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamguux/G80/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.1589s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7790s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.3671s for 16384 events => throughput is 2.57E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0129s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamguux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_taptamgdsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_taptamgdsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_taptamgdsx/ajob1' in 3.0813 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_taptamgdsx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_taptamgdsx/G28/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.0647s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0809s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.9754s for 16384 events => throughput is 8.29E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0083s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_taptamgdsx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_taptamggg/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_taptamggg/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_taptamggg/ajob1' in 25.9863 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_taptamggg/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_taptamggg/G12/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.0498s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6683s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.3681s for 16384 events => throughput is 2.57E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0134s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_taptamggg/G9/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 9.8455s + [COUNTERS] Fortran Overhead ( 0 ) : 3.4716s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.3610s for 16384 events => throughput is 2.58E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0129s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_taptamggg/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.0476s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6808s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.3539s for 16384 events => throughput is 2.58E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0129s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_taptamggg/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxdx_epemguxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxdx_epemguxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxdx_epemguxdx/ajob1' in 17.3174 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxdx_epemguxdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxdx_epemguxdx/G14/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.1039s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1311s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.9642s for 16384 events => throughput is 1.70E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0086s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxdx_epemguxdx/G6/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.4768s + [COUNTERS] Fortran Overhead ( 0 ) : 4.5056s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.9627s for 16384 events => throughput is 1.70E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0084s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxdx_epemguxdx/G22/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.1932s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2183s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.9664s for 16384 events => throughput is 1.70E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0085s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxdx_epemguxdx/G8/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.1774s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2098s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.9594s for 16384 events => throughput is 1.71E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0081s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxdx_epemguxdx/G27/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.2969s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3145s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.9738s for 16384 events => throughput is 1.68E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0086s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxdx_epemguxdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_epemgdsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_epemgdsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_epemgdsx/ajob1' in 16.4424 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_epemgdsx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_epemgdsx/G31/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.8782s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8948s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.9747s for 16384 events => throughput is 1.68E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0086s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_epemgdsx/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.9234s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9295s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.9853s for 16384 events => throughput is 1.66E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0087s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_epemgdsx/G20/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.0870s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1054s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.9727s for 16384 events => throughput is 1.68E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0089s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_epemgdsx/G25/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.6611s + [COUNTERS] Fortran Overhead ( 0 ) : 3.6767s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.9759s for 16384 events => throughput is 1.68E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0085s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_epemgdsx/G28/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.0094s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0225s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.9784s for 16384 events => throughput is 1.67E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0085s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_epemgdsx/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.8004s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8153s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.9765s for 16384 events => throughput is 1.68E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0086s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_epemgdsx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamdddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamdddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamdddx/ajob1' in 7.9655 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamdddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamdddx/G17/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.9488s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2774s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.6589s for 16384 events => throughput is 3.52E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0125s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamdddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemggg/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemggg/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemggg/ajob1' in 28.9534 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemggg/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemggg/G56/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.8313s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6136s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.2048s for 16384 events => throughput is 5.11E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0128s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemggg/G12/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.9661s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7486s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.2045s for 16384 events => throughput is 5.11E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0130s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemggg/G11/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.9378s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7397s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.1846s for 16384 events => throughput is 5.14E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0134s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemggg/G40/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.7379s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5219s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.2028s for 16384 events => throughput is 5.12E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0133s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemggg/G8/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.5270s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3293s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.1849s for 16384 events => throughput is 5.14E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0128s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemggg/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.8696s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5890s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.2671s for 16384 events => throughput is 5.01E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0134s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemggg/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/ajob1' in 192.1205 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G14/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 9.4892s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1015s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.3738s for 16384 events => throughput is 2.57E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0138s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G74/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 10.1699s + [COUNTERS] Fortran Overhead ( 0 ) : 3.7534s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.4033s for 16384 events => throughput is 2.56E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0133s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G12/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.5810s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1819s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.3852s for 16384 events => throughput is 2.57E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0139s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G6/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.4235s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0323s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.3777s for 16384 events => throughput is 2.57E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0135s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G13/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 9.3835s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9697s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.4007s for 16384 events => throughput is 2.56E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0131s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G66/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.9366s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5602s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.3634s for 16384 events => throughput is 2.57E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0130s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G11/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.7446s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3696s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.3619s for 16384 events => throughput is 2.58E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0131s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G9/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 9.5229s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1506s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.3594s for 16384 events => throughput is 2.58E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0128s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G40/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.1336s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7279s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.3924s for 16384 events => throughput is 2.56E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0132s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G35/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.1288s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7460s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.3693s for 16384 events => throughput is 2.57E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0134s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.9752s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6030s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.3592s for 16384 events => throughput is 2.58E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0130s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G69/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.3858s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9935s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.3794s for 16384 events => throughput is 2.57E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0129s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G7/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.2569s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8503s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.3934s for 16384 events => throughput is 2.56E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0132s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G65/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 9.1191s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7002s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.4053s for 16384 events => throughput is 2.56E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0136s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G71/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.7466s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3316s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.4020s for 16384 events => throughput is 2.56E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0130s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G1/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 9.2961s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9139s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.3690s for 16384 events => throughput is 2.57E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0132s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.6194s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2318s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.3743s for 16384 events => throughput is 2.57E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0133s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G2/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.8220s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4255s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.3826s for 16384 events => throughput is 2.57E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0139s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G4/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.9995s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6174s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.3688s for 16384 events => throughput is 2.57E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0133s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G79/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.1928s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8259s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.3537s for 16384 events => throughput is 2.58E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0132s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.9479s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5775s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.3573s for 16384 events => throughput is 2.58E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0132s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G70/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.3567s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9829s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.3606s for 16384 events => throughput is 2.58E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0132s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G5/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.5756s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1225s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.4400s for 16384 events => throughput is 2.54E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0131s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/ajob1' in 93.5307 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/G46/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.3204s + [COUNTERS] Fortran Overhead ( 0 ) : 4.1169s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.1907s for 16384 events => throughput is 5.13E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0128s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/G18/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.5932s + [COUNTERS] Fortran Overhead ( 0 ) : 3.3985s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.1815s for 16384 events => throughput is 5.15E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0132s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.6603s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4682s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.1787s for 16384 events => throughput is 5.15E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0133s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/G33/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.6395s + [COUNTERS] Fortran Overhead ( 0 ) : 3.4448s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.1815s for 16384 events => throughput is 5.15E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0132s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/G20/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.8232s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6293s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.1808s for 16384 events => throughput is 5.15E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0130s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/G47/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.4592s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2700s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.1761s for 16384 events => throughput is 5.16E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0131s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/G17/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.4031s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2027s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.1874s for 16384 events => throughput is 5.14E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0130s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/G25/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.5396s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2664s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.2594s for 16384 events => throughput is 5.03E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0138s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/G35/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.5443s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3467s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.1844s for 16384 events => throughput is 5.15E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0132s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/G65/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.3943s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1875s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.1934s for 16384 events => throughput is 5.13E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0133s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/G27/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.9051s + [COUNTERS] Fortran Overhead ( 0 ) : 3.7126s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.1791s for 16384 events => throughput is 5.15E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0135s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/G34/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.6073s + [COUNTERS] Fortran Overhead ( 0 ) : 3.4116s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.1820s for 16384 events => throughput is 5.15E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0137s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/G80/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.1183s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9174s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.1874s for 16384 events => throughput is 5.14E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0136s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/G67/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.9083s + [COUNTERS] Fortran Overhead ( 0 ) : 3.7030s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.1918s for 16384 events => throughput is 5.13E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0136s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.6936s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4961s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.1846s for 16384 events => throughput is 5.14E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0130s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/G48/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.7038s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5042s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.1867s for 16384 events => throughput is 5.14E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0129s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_taptamgdux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_taptamgdux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_taptamgdux/ajob1' in 11.2177 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_taptamgdux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_taptamgdux/G14/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.4566s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4915s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.9567s for 16384 events => throughput is 8.37E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0084s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_taptamgdux/G40/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.8132s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8392s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.9655s for 16384 events => throughput is 8.34E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0084s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_taptamgdux/G44/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.9048s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9291s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.9674s for 16384 events => throughput is 8.33E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0083s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_taptamgdux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemgddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemgddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemgddx/ajob1' in 16.8208 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemgddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemgddx/G56/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.4590s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0615s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.3851s for 16384 events => throughput is 6.87E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0125s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemgddx/G12/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.2347s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9129s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.3090s for 16384 events => throughput is 7.10E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0128s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemgddx/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.2479s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9374s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.2984s for 16384 events => throughput is 7.13E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0121s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemgddx/G47/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.4771s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1360s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.3286s for 16384 events => throughput is 7.04E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0124s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemgddx/G42/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.3322s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0196s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.2998s for 16384 events => throughput is 7.12E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0127s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemgddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/ajob1' in 126.7675 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/G39/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.7946s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4062s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.3749s for 16384 events => throughput is 2.57E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0135s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/G12/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.5834s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2054s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.3651s for 16384 events => throughput is 2.57E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0128s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/G6/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.2906s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8868s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.3904s for 16384 events => throughput is 2.56E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0134s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/G66/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 9.0937s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7095s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.3709s for 16384 events => throughput is 2.57E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0133s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/G11/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.8731s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3733s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.4863s for 16384 events => throughput is 2.53E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0135s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/G9/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 9.9506s + [COUNTERS] Fortran Overhead ( 0 ) : 3.5358s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.4019s for 16384 events => throughput is 2.56E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0130s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.8449s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4342s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.3972s for 16384 events => throughput is 2.56E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0135s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/G69/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.5126s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1046s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.3947s for 16384 events => throughput is 2.56E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0134s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/G8/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.4143s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9563s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.4441s for 16384 events => throughput is 2.54E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0138s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/G71/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.7038s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3090s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.3816s for 16384 events => throughput is 2.57E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0133s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.5821s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1995s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.3695s for 16384 events => throughput is 2.57E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0130s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/G72/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.5441s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1709s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.3601s for 16384 events => throughput is 2.58E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0131s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/G79/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.1165s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7341s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.3695s for 16384 events => throughput is 2.57E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0129s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.2602s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8268s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.4201s for 16384 events => throughput is 2.55E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0133s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/G70/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.3899s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0071s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.3701s for 16384 events => throughput is 2.57E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0127s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/G5/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.5954s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2130s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.3690s for 16384 events => throughput is 2.57E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0135s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemdssx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemdssx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemdssx/ajob1' in 6.0841 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemdssx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemdssx/G46/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.0990s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1167s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.9739s for 16384 events => throughput is 1.68E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0084s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemdssx/G17/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.9554s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9718s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.9750s for 16384 events => throughput is 1.68E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0087s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemdssx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uc_taptamguc/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uc_taptamguc/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uc_taptamguc/ajob1' in 7.0778 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uc_taptamguc/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uc_taptamguc/G25/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.2266s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2688s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.9489s for 16384 events => throughput is 8.41E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0089s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uc_taptamguc/G37/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.8214s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8561s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.9562s for 16384 events => throughput is 8.38E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0092s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uc_taptamguc/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamgddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamgddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamgddx/ajob1' in 12.1285 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamgddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamgddx/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.2669s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6265s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.6280s for 16384 events => throughput is 3.54E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0124s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamgddx/G55/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.8319s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1909s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.6278s for 16384 events => throughput is 3.54E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0131s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamgddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemgddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemgddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemgddx/ajob1' in 2.4622 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemgddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemgddx/G8/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.4454s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4976s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.9394s for 16384 events => throughput is 1.74E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0083s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemgddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamgddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamgddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamgddx/ajob1' in 74.1114 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamgddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamgddx/G31/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.6039s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1734s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.4174s for 16384 events => throughput is 2.55E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0132s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamgddx/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.6351s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2309s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.3912s for 16384 events => throughput is 2.56E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0131s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamgddx/G33/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 9.6634s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1757s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.4741s for 16384 events => throughput is 2.53E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0136s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamgddx/G20/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.7648s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3872s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.3644s for 16384 events => throughput is 2.57E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0132s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamgddx/G35/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.6212s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2355s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.3725s for 16384 events => throughput is 2.57E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0132s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamgddx/G71/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.5027s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0985s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.3907s for 16384 events => throughput is 2.56E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0135s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamgddx/G34/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 10.1715s + [COUNTERS] Fortran Overhead ( 0 ) : 3.7856s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.3730s for 16384 events => throughput is 2.57E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0130s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamgddx/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.8962s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5036s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.3792s for 16384 events => throughput is 2.57E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0134s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamgddx/G48/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.1295s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6774s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.4385s for 16384 events => throughput is 2.54E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0136s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamgddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemddxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemddxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemddxdx/ajob1' in 7.5115 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemddxdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemddxdx/G11/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.9510s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5758s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.3621s for 16384 events => throughput is 6.94E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0132s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemddxdx/G71/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.5303s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2049s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.3125s for 16384 events => throughput is 7.09E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0130s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemddxdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dxsx_epemgdxsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dxsx_epemgdxsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dxsx_epemgdxsx/ajob1' in 2.2151 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dxsx_epemgdxsx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dxsx_epemgdxsx/G27/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.1985s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2279s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.9621s for 16384 events => throughput is 1.70E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0085s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dxsx_epemgdxsx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemduxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemduxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemduxdx/ajob1' in 5.7757 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemduxdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemduxdx/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.0357s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0577s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.9692s for 16384 events => throughput is 1.69E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0088s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemduxdx/G22/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.7106s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7304s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.9718s for 16384 events => throughput is 1.69E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0083s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemduxdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemguux/ajob1' in 29.9093 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemguux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemguux/G20/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.3592s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0463s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.3005s for 16384 events => throughput is 7.12E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0124s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemguux/G27/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.1452s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8173s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.3152s for 16384 events => throughput is 7.08E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0127s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemguux/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.7088s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3931s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.3031s for 16384 events => throughput is 7.11E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0126s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemguux/G55/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.5021s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1953s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.2945s for 16384 events => throughput is 7.14E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0123s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemguux/G37/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.3501s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0367s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.3003s for 16384 events => throughput is 7.12E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0131s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemguux/G41/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.2194s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8995s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.3077s for 16384 events => throughput is 7.10E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0122s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemguux/G57/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.9445s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6346s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.2973s for 16384 events => throughput is 7.13E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0126s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemguux/G48/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.5705s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2481s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.3099s for 16384 events => throughput is 7.09E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0125s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemguux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dd_taptamgdd/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dd_taptamgdd/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dd_taptamgdd/ajob1' in 12.4325 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dd_taptamgdd/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dd_taptamgdd/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.9321s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2219s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.6973s for 16384 events => throughput is 3.49E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0128s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dd_taptamgdd/G24/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.4707s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8084s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.6499s for 16384 events => throughput is 3.52E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0123s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dd_taptamgdd/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/ajob1' in 198.8199 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G14/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.5658s + [COUNTERS] Fortran Overhead ( 0 ) : 3.3719s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.1808s for 16384 events => throughput is 5.15E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0132s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G38/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.8880s + [COUNTERS] Fortran Overhead ( 0 ) : 3.6875s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.1876s for 16384 events => throughput is 5.14E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0129s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G39/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.6014s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3984s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.1899s for 16384 events => throughput is 5.14E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0131s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G12/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.3674s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1760s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.1786s for 16384 events => throughput is 5.15E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0128s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G6/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.0819s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8949s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.1740s for 16384 events => throughput is 5.16E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0130s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G3/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.9557s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7659s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.1767s for 16384 events => throughput is 5.16E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0131s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G13/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.5457s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2637s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.2681s for 16384 events => throughput is 5.01E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0139s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G66/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.1470s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9375s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.1963s for 16384 events => throughput is 5.13E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0132s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G33/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.1125s + [COUNTERS] Fortran Overhead ( 0 ) : 4.9212s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.1784s for 16384 events => throughput is 5.15E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0128s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G11/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.5413s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2913s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.2361s for 16384 events => throughput is 5.06E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0139s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G9/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 33.6886s + [COUNTERS] Fortran Overhead ( 0 ) : 17.7619s + [COUNTERS] CudaCpp MEs ( 2 ) : 15.9133s for 81920 events => throughput is 5.15E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0135s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G10/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.1161s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8919s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.2113s for 16384 events => throughput is 5.10E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0129s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G40/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.8093s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5865s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.2098s for 16384 events => throughput is 5.10E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0130s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G35/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.7419s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5579s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.1708s for 16384 events => throughput is 5.17E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0132s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.5726s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3784s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.1811s for 16384 events => throughput is 5.15E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0130s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G69/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.2344s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0408s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.1805s for 16384 events => throughput is 5.15E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0131s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G8/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.1384s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9343s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.1909s for 16384 events => throughput is 5.13E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0132s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G7/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.3083s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0380s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.2571s for 16384 events => throughput is 5.03E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0132s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G65/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.1836s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9934s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.1775s for 16384 events => throughput is 5.16E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0127s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G71/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.6421s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4561s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.1732s for 16384 events => throughput is 5.16E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0129s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G1/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.5663s + [COUNTERS] Fortran Overhead ( 0 ) : 3.3159s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.2372s for 16384 events => throughput is 5.06E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0132s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.4363s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2474s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.1754s for 16384 events => throughput is 5.16E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0135s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G2/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.3128s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1209s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.1789s for 16384 events => throughput is 5.15E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0130s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G4/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.0920s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9022s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.1767s for 16384 events => throughput is 5.16E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0131s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G80/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.3187s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1253s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.1804s for 16384 events => throughput is 5.15E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0130s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G41/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.7474s + [COUNTERS] Fortran Overhead ( 0 ) : 3.5430s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.1915s for 16384 events => throughput is 5.13E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0129s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G77/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.0522s + [COUNTERS] Fortran Overhead ( 0 ) : 3.8457s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.1933s for 16384 events => throughput is 5.13E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0131s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G76/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.4908s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3001s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.1780s for 16384 events => throughput is 5.16E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0128s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G79/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.0273s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8408s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.1734s for 16384 events => throughput is 5.16E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0130s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.6576s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4699s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.1744s for 16384 events => throughput is 5.16E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0133s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G70/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.1945s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9916s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.1899s for 16384 events => throughput is 5.14E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0130s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G5/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.2499s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0615s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.1751s for 16384 events => throughput is 5.16E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0133s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ds_taptamgds/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ds_taptamgds/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ds_taptamgds/ajob1' in 3.7832 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ds_taptamgds/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ds_taptamgds/G10/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.7667s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7877s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.9703s for 16384 events => throughput is 8.32E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0087s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ds_taptamgds/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_epemgudx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_epemgudx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_epemgudx/ajob1' in 13.9923 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_epemgudx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_epemgudx/G12/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.8109s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8255s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.9769s for 16384 events => throughput is 1.68E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0084s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_epemgudx/G11/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.7682s + [COUNTERS] Fortran Overhead ( 0 ) : 0.7821s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.9779s for 16384 events => throughput is 1.68E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0082s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_epemgudx/G17/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.3405s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3538s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.9781s for 16384 events => throughput is 1.68E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0086s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_epemgudx/G10/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.9100s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9235s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.9783s for 16384 events => throughput is 1.67E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0082s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_epemgudx/G27/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.0587s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0789s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.9713s for 16384 events => throughput is 1.69E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0084s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_epemgudx/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.0212s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0336s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.9785s for 16384 events => throughput is 1.67E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0090s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_epemgudx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_taptamgud/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_taptamgud/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_taptamgud/ajob1' in 29.2934 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_taptamgud/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_taptamgud/G12/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.7798s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8193s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.9520s for 16384 events => throughput is 8.39E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0085s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_taptamgud/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.0503s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0868s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.9551s for 16384 events => throughput is 8.38E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0083s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_taptamgud/G6/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.1142s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1406s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.9653s for 16384 events => throughput is 8.34E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0083s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_taptamgud/G23/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.8141s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8454s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.9602s for 16384 events => throughput is 8.36E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0085s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_taptamgud/G11/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.8205s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8580s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.9538s for 16384 events => throughput is 8.39E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0087s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_taptamgud/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.1636s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1934s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.9619s for 16384 events => throughput is 8.35E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0084s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_taptamgud/G22/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.3315s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3708s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.9525s for 16384 events => throughput is 8.39E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0082s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_taptamgud/G7/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.2332s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2511s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.9734s for 16384 events => throughput is 8.30E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0087s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_taptamgud/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.8628s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8937s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.9605s for 16384 events => throughput is 8.36E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0087s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_taptamgud/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemdddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemdddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemdddx/ajob1' in 7.7494 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemdddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemdddx/G31/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.0231s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6618s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.3491s for 16384 events => throughput is 6.97E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0122s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemdddx/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.6957s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3696s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.3135s for 16384 events => throughput is 7.08E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0126s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemdddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamdssx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamdssx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamdssx/ajob1' in 6.2871 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamdssx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamdssx/G45/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.0397s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0773s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.9537s for 16384 events => throughput is 8.39E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0088s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamdssx/G20/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.2175s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2460s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.9631s for 16384 events => throughput is 8.35E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0083s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamdssx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuddx/ajob1' in 9.2483 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuddx/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.0143s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0411s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.9647s for 16384 events => throughput is 8.34E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0085s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuddx/G20/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.0937s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1217s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.9636s for 16384 events => throughput is 8.34E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0085s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuddx/G47/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.0966s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1131s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.9749s for 16384 events => throughput is 8.30E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0086s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemuuxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemuuxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemuuxdx/ajob1' in 6.4186 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemuuxdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemuuxdx/G7/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.2980s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3116s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.9781s for 16384 events => throughput is 1.68E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0083s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemuuxdx/G41/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.0910s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1064s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.9761s for 16384 events => throughput is 1.68E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0084s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemuuxdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/ajob1' in 46.9751 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G14/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.0266s + [COUNTERS] Fortran Overhead ( 0 ) : 3.0414s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.9766s for 16384 events => throughput is 1.68E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0085s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G18/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.4890s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5124s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.9682s for 16384 events => throughput is 1.69E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0084s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G39/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.7881s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8092s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.9706s for 16384 events => throughput is 1.69E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0083s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G30/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.1057s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1170s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.9803s for 16384 events => throughput is 1.67E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0084s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G12/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.8082s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8141s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.9855s for 16384 events => throughput is 1.66E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0086s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.1568s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1778s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.9703s for 16384 events => throughput is 1.69E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0087s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G6/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.3624s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3793s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.9746s for 16384 events => throughput is 1.68E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0085s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G3/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.8255s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8418s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.9749s for 16384 events => throughput is 1.68E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0088s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G23/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.8450s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8626s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.9739s for 16384 events => throughput is 1.68E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0085s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G24/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.8518s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8507s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.9924s for 16384 events => throughput is 1.65E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0087s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G13/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.2689s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2869s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.9734s for 16384 events => throughput is 1.68E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0085s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G20/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.8654s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8835s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.9734s for 16384 events => throughput is 1.68E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0085s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G25/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.4122s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4313s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.9720s for 16384 events => throughput is 1.69E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0089s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.0978s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1188s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.9708s for 16384 events => throughput is 1.69E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0082s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G8/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.9769s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9702s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.9976s for 16384 events => throughput is 1.64E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0091s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G7/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.9897s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0133s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.9681s for 16384 events => throughput is 1.69E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0082s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G4/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.6520s + [COUNTERS] Fortran Overhead ( 0 ) : 0.6723s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.9715s for 16384 events => throughput is 1.69E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0083s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G41/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.3258s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3306s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.9866s for 16384 events => throughput is 1.66E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0086s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G32/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.8730s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8900s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.9745s for 16384 events => throughput is 1.68E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0085s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ucx_epemgucx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ucx_epemgucx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ucx_epemgucx/ajob1' in 6.0100 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ucx_epemgucx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ucx_epemgucx/G3/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.6400s + [COUNTERS] Fortran Overhead ( 0 ) : 0.6492s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.9824s for 16384 events => throughput is 1.67E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0085s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ucx_epemgucx/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.2560s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2711s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.9765s for 16384 events => throughput is 1.68E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0084s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ucx_epemgucx/G7/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.0708s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0902s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.9723s for 16384 events => throughput is 1.69E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0083s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ucx_epemgucx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_taptamgudx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_taptamgudx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_taptamgudx/ajob1' in 16.4854 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_taptamgudx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_taptamgudx/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.9257s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9453s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.9714s for 16384 events => throughput is 8.31E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0090s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_taptamgudx/G11/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.8459s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8693s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.9681s for 16384 events => throughput is 8.32E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0086s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_taptamgudx/G17/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.3599s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3903s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.9608s for 16384 events => throughput is 8.36E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0088s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_taptamgudx/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.2137s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2384s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.9670s for 16384 events => throughput is 8.33E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0083s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_taptamgudx/G27/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.0705s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0926s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.9689s for 16384 events => throughput is 8.32E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0091s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_taptamgudx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamuuxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamuuxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamuuxdx/ajob1' in 4.7426 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamuuxdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamuuxdx/G6/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.7259s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7574s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.9602s for 16384 events => throughput is 8.36E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0084s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamuuxdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_taptamguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_taptamguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_taptamguux/ajob1' in 5.8821 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_taptamguux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_taptamguux/G48/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.8656s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2187s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.6342s for 16384 events => throughput is 3.54E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0127s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_taptamguux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemudux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemudux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemudux/ajob1' in 10.9184 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemudux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemudux/G43/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.5404s + [COUNTERS] Fortran Overhead ( 0 ) : 3.5856s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.9465s for 16384 events => throughput is 1.73E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0083s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemudux/G8/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.1053s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1444s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.9526s for 16384 events => throughput is 1.72E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0083s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemudux/G5/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.2296s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2794s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.9416s for 16384 events => throughput is 1.74E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0085s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemudux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamsdxsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamsdxsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamsdxsx/ajob1' in 3.3104 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamsdxsx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamsdxsx/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.2936s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2819s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.0033s for 16384 events => throughput is 8.18E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0085s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamsdxsx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_epemguu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_epemguu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_epemguu/ajob1' in 36.0190 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_epemguu/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_epemguu/G30/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.3548s + [COUNTERS] Fortran Overhead ( 0 ) : 3.0273s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.3153s for 16384 events => throughput is 7.08E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0122s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_epemguu/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.2784s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9492s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.3166s for 16384 events => throughput is 7.07E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0126s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_epemguu/G9/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.6172s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2931s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.3114s for 16384 events => throughput is 7.09E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0127s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_epemguu/G49/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.2299s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9060s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.3111s for 16384 events => throughput is 7.09E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0128s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_epemguu/G10/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.5556s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2300s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.3134s for 16384 events => throughput is 7.08E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0122s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_epemguu/G81/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.3550s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0243s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.3183s for 16384 events => throughput is 7.07E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0124s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_epemguu/G62/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.1271s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8049s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.3097s for 16384 events => throughput is 7.09E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0124s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_epemguu/G51/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.2002s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8595s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.3282s for 16384 events => throughput is 7.04E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0125s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_epemguu/G28/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.1775s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8437s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.3214s for 16384 events => throughput is 7.06E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0123s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_epemguu/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uc_epemguc/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uc_epemguc/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uc_epemguc/ajob1' in 3.8140 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uc_epemguc/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uc_epemguc/G12/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.7572s + [COUNTERS] Fortran Overhead ( 0 ) : 0.7773s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.9715s for 16384 events => throughput is 1.69E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0084s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uc_epemguc/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.0272s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0496s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.9692s for 16384 events => throughput is 1.69E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0084s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uc_epemguc/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ds_epemgds/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ds_epemgds/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ds_epemgds/ajob1' in 2.0672 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ds_epemgds/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ds_epemgds/G27/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.0507s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0686s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.9739s for 16384 events => throughput is 1.68E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0082s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ds_epemgds/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_taptamguu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_taptamguu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_taptamguu/ajob1' in 17.6524 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_taptamguu/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_taptamguu/G52/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.6413s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9604s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.6685s for 16384 events => throughput is 3.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0124s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_taptamguu/G49/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.3338s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6395s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.6817s for 16384 events => throughput is 3.50E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0126s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_taptamguu/G32/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.6333s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9548s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.6656s for 16384 events => throughput is 3.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0129s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_taptamguu/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished in 1902.0475 seconds +__CUDACPP_DEBUG: gen_ximprove.launch finished in 1902.4828 seconds +__CUDACPP_DEBUG: GridPackCmd.refine4grid finished in 1902.4836 seconds +__CUDACPP_DEBUG: GridPackCmd.launch (3a) combine_events +combine_events +__CUDACPP_DEBUG: GridPackCmd.launch (3b) store_events +store_events + === Results Summary for run: GridRun_21 tag: tag_1 === + + Cross-section : 0 +- 0 pb + Nb of events : 500 + +INFO: No version of lhapdf. Can not run systematics computation +decay_events -from_cards +__CUDACPP_DEBUG: GridPackCmd.launch finished in 1920.0969 seconds +__CUDACPP_DEBUG: GridPackCmd.__init__ back from self.launch +DONE +write ./events.lhe.gz +END: Wed Sep 11 03:06:19 AM CEST 2024 +ELAPSED: 1923 seconds diff --git a/epochX/cudacpp/tlau/fromgridpacks/pp_dy3j.mad/cppsse4/output.txt b/epochX/cudacpp/tlau/fromgridpacks/pp_dy3j.mad/cppsse4/output.txt new file mode 100644 index 0000000000..62342ac6bb --- /dev/null +++ b/epochX/cudacpp/tlau/fromgridpacks/pp_dy3j.mad/cppsse4/output.txt @@ -0,0 +1,2336 @@ +START: Wed Sep 11 03:06:58 AM CEST 2024 +On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU (nproc=4)] [GPU: 1x Tesla V100S-PCIE-32GB]: +CUDACPP_RUNTIME_DISABLEFPE is set +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:06 madevent/SubProcesses/P0_dc_epemgdc/madevent -> build.sse4_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:06 madevent/SubProcesses/P0_dc_taptamgdc/madevent -> build.sse4_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:06 madevent/SubProcesses/P0_dd_epemgdd/madevent -> build.sse4_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:06 madevent/SubProcesses/P0_dd_taptamgdd/madevent -> build.sse4_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:06 madevent/SubProcesses/P0_ddx_epemgddx/madevent -> build.sse4_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:06 madevent/SubProcesses/P0_ddx_epemggg/madevent -> build.sse4_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:06 madevent/SubProcesses/P0_ddx_epemgssx/madevent -> build.sse4_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:06 madevent/SubProcesses/P0_ddx_epemguux/madevent -> build.sse4_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:06 madevent/SubProcesses/P0_ddx_taptamgddx/madevent -> build.sse4_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:06 madevent/SubProcesses/P0_ddx_taptamggg/madevent -> build.sse4_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:06 madevent/SubProcesses/P0_ddx_taptamgssx/madevent -> build.sse4_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:06 madevent/SubProcesses/P0_ddx_taptamguux/madevent -> build.sse4_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:06 madevent/SubProcesses/P0_ds_epemgds/madevent -> build.sse4_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:06 madevent/SubProcesses/P0_ds_taptamgds/madevent -> build.sse4_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:06 madevent/SubProcesses/P0_dsx_epemgdsx/madevent -> build.sse4_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:06 madevent/SubProcesses/P0_dsx_taptamgdsx/madevent -> build.sse4_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:06 madevent/SubProcesses/P0_dux_epemgdux/madevent -> build.sse4_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:06 madevent/SubProcesses/P0_dux_taptamgdux/madevent -> build.sse4_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:06 madevent/SubProcesses/P0_dxcx_epemgdxcx/madevent -> build.sse4_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:06 madevent/SubProcesses/P0_dxcx_taptamgdxcx/madevent -> build.sse4_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:06 madevent/SubProcesses/P0_dxdx_epemgdxdx/madevent -> build.sse4_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:06 madevent/SubProcesses/P0_dxdx_taptamgdxdx/madevent -> build.sse4_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:06 madevent/SubProcesses/P0_dxsx_epemgdxsx/madevent -> build.sse4_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:06 madevent/SubProcesses/P0_dxsx_taptamgdxsx/madevent -> build.sse4_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:06 madevent/SubProcesses/P0_gd_epemdddx/madevent -> build.sse4_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:06 madevent/SubProcesses/P0_gd_epemdssx/madevent -> build.sse4_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:06 madevent/SubProcesses/P0_gd_epemggd/madevent -> build.sse4_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:06 madevent/SubProcesses/P0_gd_epemudux/madevent -> build.sse4_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:06 madevent/SubProcesses/P0_gd_taptamdddx/madevent -> build.sse4_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:06 madevent/SubProcesses/P0_gd_taptamdssx/madevent -> build.sse4_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:06 madevent/SubProcesses/P0_gd_taptamggd/madevent -> build.sse4_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:06 madevent/SubProcesses/P0_gd_taptamudux/madevent -> build.sse4_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:06 madevent/SubProcesses/P0_gdx_epemddxdx/madevent -> build.sse4_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:06 madevent/SubProcesses/P0_gdx_epemggdx/madevent -> build.sse4_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:06 madevent/SubProcesses/P0_gdx_epemsdxsx/madevent -> build.sse4_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:06 madevent/SubProcesses/P0_gdx_epemuuxdx/madevent -> build.sse4_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:06 madevent/SubProcesses/P0_gdx_taptamddxdx/madevent -> build.sse4_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:06 madevent/SubProcesses/P0_gdx_taptamggdx/madevent -> build.sse4_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:06 madevent/SubProcesses/P0_gdx_taptamsdxsx/madevent -> build.sse4_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:06 madevent/SubProcesses/P0_gdx_taptamuuxdx/madevent -> build.sse4_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:06 madevent/SubProcesses/P0_gg_epemgddx/madevent -> build.sse4_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:06 madevent/SubProcesses/P0_gg_epemguux/madevent -> build.sse4_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:06 madevent/SubProcesses/P0_gg_taptamgddx/madevent -> build.sse4_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:06 madevent/SubProcesses/P0_gg_taptamguux/madevent -> build.sse4_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:06 madevent/SubProcesses/P0_gu_epemggu/madevent -> build.sse4_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:06 madevent/SubProcesses/P0_gu_epemuccx/madevent -> build.sse4_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:06 madevent/SubProcesses/P0_gu_epemuddx/madevent -> build.sse4_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:06 madevent/SubProcesses/P0_gu_epemuuux/madevent -> build.sse4_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:06 madevent/SubProcesses/P0_gu_taptamggu/madevent -> build.sse4_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:06 madevent/SubProcesses/P0_gu_taptamuccx/madevent -> build.sse4_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:06 madevent/SubProcesses/P0_gu_taptamuddx/madevent -> build.sse4_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:06 madevent/SubProcesses/P0_gu_taptamuuux/madevent -> build.sse4_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:06 madevent/SubProcesses/P0_gux_epemcuxcx/madevent -> build.sse4_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:06 madevent/SubProcesses/P0_gux_epemduxdx/madevent -> build.sse4_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:06 madevent/SubProcesses/P0_gux_epemggux/madevent -> build.sse4_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:06 madevent/SubProcesses/P0_gux_epemuuxux/madevent -> build.sse4_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:06 madevent/SubProcesses/P0_gux_taptamcuxcx/madevent -> build.sse4_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:06 madevent/SubProcesses/P0_gux_taptamduxdx/madevent -> build.sse4_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:06 madevent/SubProcesses/P0_gux_taptamggux/madevent -> build.sse4_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:06 madevent/SubProcesses/P0_gux_taptamuuxux/madevent -> build.sse4_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:06 madevent/SubProcesses/P0_uc_epemguc/madevent -> build.sse4_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:06 madevent/SubProcesses/P0_uc_taptamguc/madevent -> build.sse4_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:06 madevent/SubProcesses/P0_ucx_epemgucx/madevent -> build.sse4_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:06 madevent/SubProcesses/P0_ucx_taptamgucx/madevent -> build.sse4_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:06 madevent/SubProcesses/P0_ud_epemgud/madevent -> build.sse4_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:06 madevent/SubProcesses/P0_ud_taptamgud/madevent -> build.sse4_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:06 madevent/SubProcesses/P0_udx_epemgudx/madevent -> build.sse4_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:06 madevent/SubProcesses/P0_udx_taptamgudx/madevent -> build.sse4_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:06 madevent/SubProcesses/P0_uu_epemguu/madevent -> build.sse4_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:06 madevent/SubProcesses/P0_uu_taptamguu/madevent -> build.sse4_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:06 madevent/SubProcesses/P0_uux_epemgccx/madevent -> build.sse4_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:06 madevent/SubProcesses/P0_uux_epemgddx/madevent -> build.sse4_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:06 madevent/SubProcesses/P0_uux_epemggg/madevent -> build.sse4_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:06 madevent/SubProcesses/P0_uux_epemguux/madevent -> build.sse4_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:06 madevent/SubProcesses/P0_uux_taptamgccx/madevent -> build.sse4_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:06 madevent/SubProcesses/P0_uux_taptamgddx/madevent -> build.sse4_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:06 madevent/SubProcesses/P0_uux_taptamggg/madevent -> build.sse4_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:06 madevent/SubProcesses/P0_uux_taptamguux/madevent -> build.sse4_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:06 madevent/SubProcesses/P0_uxcx_epemguxcx/madevent -> build.sse4_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:06 madevent/SubProcesses/P0_uxcx_taptamguxcx/madevent -> build.sse4_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:06 madevent/SubProcesses/P0_uxdx_epemguxdx/madevent -> build.sse4_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:06 madevent/SubProcesses/P0_uxdx_taptamguxdx/madevent -> build.sse4_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:06 madevent/SubProcesses/P0_uxux_epemguxux/madevent -> build.sse4_m_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 11 03:06 madevent/SubProcesses/P0_uxux_taptamguxux/madevent -> build.sse4_m_inl0_hrd0/madevent_cpp +Now generating 500 events with random seed 21 and granularity 1 +************************************************************ +* * +* W E L C O M E to * +* M A D G R A P H 5 _ a M C @ N L O * +* M A D E V E N T * +* * +* * * * +* * * * * * +* * * * * 5 * * * * * +* * * * * * +* * * * +* * +* VERSION 3.5.3_lo_vect * +* * +* The MadGraph5_aMC@NLO Development Team - Find us at * +* https://server06.fynu.ucl.ac.be/projects/madgraph * +* * +* Type 'help' for in-line help. * +* * +************************************************************ +INFO: load configuration from /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/Cards/me5_configuration.txt +INFO: load configuration from /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/Cards/me5_configuration.txt +Using default text editor "vi". Set another one in ./input/mg5_configuration.txt +Using default eps viewer "evince". Set another one in ./input/mg5_configuration.txt +Using default web browser "firefox". Set another one in ./input/mg5_configuration.txt +WRITE GRIDCARD /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent +__CUDACPP_DEBUG: GridPackCmd.__init__ will call self.launch +__CUDACPP_DEBUG: GridPackCmd.launch starting +__CUDACPP_DEBUG: GridPackCmd.launch (1) restore_data +generate 500 events +P0_gg_epemguux +P0_gg_epemgddx +P0_gg_taptamguux +P0_gg_taptamgddx +P0_gu_epemggu +P0_gd_epemggd +P0_gux_epemggux +P0_gdx_epemggdx +P0_gu_taptamggu +P0_gd_taptamggd +P0_gux_taptamggux +P0_gdx_taptamggdx +P0_uux_epemggg +P0_ddx_epemggg +P0_uux_taptamggg +P0_ddx_taptamggg +P0_gu_epemuuux +P0_gd_epemdddx +P0_gux_epemuuxux +P0_gdx_epemddxdx +P0_gu_epemuddx +P0_gu_epemuccx +P0_gd_epemudux +P0_gd_epemdssx +P0_gux_epemduxdx +P0_gux_epemcuxcx +P0_gdx_epemuuxdx +P0_gdx_epemsdxsx +P0_gu_taptamuuux +P0_gd_taptamdddx +P0_gux_taptamuuxux +P0_gdx_taptamddxdx +P0_gu_taptamuddx +P0_gu_taptamuccx +P0_gd_taptamudux +P0_gd_taptamdssx +P0_gux_taptamduxdx +P0_gux_taptamcuxcx +P0_gdx_taptamuuxdx +P0_gdx_taptamsdxsx +P0_uu_epemguu +P0_uux_epemguux +P0_dd_epemgdd +P0_ddx_epemgddx +P0_uxux_epemguxux +P0_dxdx_epemgdxdx +P0_ud_epemgud +P0_uc_epemguc +P0_uux_epemgddx +P0_uux_epemgccx +P0_udx_epemgudx +P0_ucx_epemgucx +P0_dc_epemgdc +P0_ds_epemgds +P0_dux_epemgdux +P0_ddx_epemguux +P0_ddx_epemgssx +P0_dsx_epemgdsx +P0_uxdx_epemguxdx +P0_uxcx_epemguxcx +P0_dxcx_epemgdxcx +P0_dxsx_epemgdxsx +P0_uu_taptamguu +P0_uux_taptamguux +P0_dd_taptamgdd +P0_ddx_taptamgddx +P0_uxux_taptamguxux +P0_dxdx_taptamgdxdx +P0_ud_taptamgud +P0_uc_taptamguc +P0_uux_taptamgddx +P0_uux_taptamgccx +P0_udx_taptamgudx +P0_ucx_taptamgucx +P0_dc_taptamgdc +P0_ds_taptamgds +P0_dux_taptamgdux +P0_ddx_taptamguux +P0_ddx_taptamgssx +P0_dsx_taptamgdsx +P0_uxdx_taptamguxdx +P0_uxcx_taptamguxcx +P0_dxcx_taptamgdxcx +P0_dxsx_taptamgdxsx +__CUDACPP_DEBUG: GridPackCmd.launch (2) refine4grid +__CUDACPP_DEBUG: GridPackCmd.refine4grid starting +Using random number seed offset = 21 +INFO: Generating 500.0 unweighted events. +__CUDACPP_DEBUG: gen_ximprove.launch starting +sum of cpu time of last step: 5h02m29s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event starting +INFO: need to improve 362 channels +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuuux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuuux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuuux/ajob1' in 10.7348 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuuux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuuux/G12/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.3962s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1806s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.2089s for 16384 events => throughput is 1.36E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0067s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuuux/G20/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.4592s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2735s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.1793s for 16384 events => throughput is 1.39E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0064s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuuux/G29/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.8355s + [COUNTERS] Fortran Overhead ( 0 ) : 4.6464s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.1826s for 16384 events => throughput is 1.39E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0065s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuuux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemggg/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemggg/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemggg/ajob1' in 11.4593 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemggg/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemggg/G43/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.7992s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0908s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7015s for 16384 events => throughput is 9.63E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0069s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemggg/G10/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.0799s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2298s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.8427s for 16384 events => throughput is 8.89E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0074s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemggg/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.4000s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6956s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.6973s for 16384 events => throughput is 9.65E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0071s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemggg/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/ajob1' in 58.3614 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G39/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.9378s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2266s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7044s for 16384 events => throughput is 9.61E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0067s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G12/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.2873s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4968s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7831s for 16384 events => throughput is 9.19E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0074s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G6/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.8880s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1846s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.6964s for 16384 events => throughput is 9.66E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0069s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G13/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.1909s + [COUNTERS] Fortran Overhead ( 0 ) : 3.4869s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.6969s for 16384 events => throughput is 9.66E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0071s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G11/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.0412s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3140s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7202s for 16384 events => throughput is 9.52E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0069s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.9304s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2314s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.6920s for 16384 events => throughput is 9.68E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0069s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G69/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.7720s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0744s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.6906s for 16384 events => throughput is 9.69E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0070s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G8/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.7599s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0624s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.6906s for 16384 events => throughput is 9.69E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0069s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G7/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.7425s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0116s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7241s for 16384 events => throughput is 9.50E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0068s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G65/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.5784s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8658s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7059s for 16384 events => throughput is 9.60E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0067s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G1/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.1594s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4172s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7352s for 16384 events => throughput is 9.44E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0071s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.1420s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4389s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.6958s for 16384 events => throughput is 9.66E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0072s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G82/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.4432s + [COUNTERS] Fortran Overhead ( 0 ) : 3.7217s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7148s for 16384 events => throughput is 9.55E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0068s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G80/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.5265s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8276s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.6919s for 16384 events => throughput is 9.68E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0070s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G51/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.2907s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5801s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7038s for 16384 events => throughput is 9.62E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0068s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G70/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.8984s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0973s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7937s for 16384 events => throughput is 9.13E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0074s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G5/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.5432s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8446s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.6918s for 16384 events => throughput is 9.68E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0068s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamggux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamggux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamggux/ajob1' in 53.4130 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamggux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamggux/G59/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.3572s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0677s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.2824s for 16384 events => throughput is 4.99E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0071s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamggux/G12/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.7725s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4805s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.2847s for 16384 events => throughput is 4.99E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0073s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamggux/G13/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.7975s + [COUNTERS] Fortran Overhead ( 0 ) : 3.4981s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.2924s for 16384 events => throughput is 4.98E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0070s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamggux/G11/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.6685s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3419s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.3192s for 16384 events => throughput is 4.94E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0074s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamggux/G9/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.4651s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1419s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.3162s for 16384 events => throughput is 4.94E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0069s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamggux/G10/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.8064s + [COUNTERS] Fortran Overhead ( 0 ) : 3.4166s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.3823s for 16384 events => throughput is 4.84E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0074s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamggux/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.6762s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2556s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.4131s for 16384 events => throughput is 4.80E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0075s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamggux/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.9522s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6250s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.3196s for 16384 events => throughput is 4.94E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0076s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamggux/G80/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.1558s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8214s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.3274s for 16384 events => throughput is 4.92E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0070s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamggux/G70/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.6235s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1523s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.4638s for 16384 events => throughput is 4.73E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0073s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamggux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemsdxsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemsdxsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemsdxsx/ajob1' in 2.3520 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemsdxsx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemsdxsx/G8/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.3351s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8227s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5079s for 16384 events => throughput is 3.23E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0045s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemsdxsx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuccx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuccx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuccx/ajob1' in 2.7530 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuccx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuccx/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.7365s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6987s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.0330s for 16384 events => throughput is 1.59E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0048s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuccx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuuux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuuux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuuux/ajob1' in 9.2184 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuuux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuuux/G81/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.9483s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5620s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.3798s for 16384 events => throughput is 6.88E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0065s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuuux/G32/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.2401s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8662s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.3673s for 16384 events => throughput is 6.92E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0066s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuuux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamggg/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamggg/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamggg/ajob1' in 9.7182 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamggg/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamggg/G56/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.7754s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4754s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.2931s for 16384 events => throughput is 4.98E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0069s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamggg/G39/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.9124s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6134s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.2919s for 16384 events => throughput is 4.98E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0071s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamggg/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamduxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamduxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamduxdx/ajob1' in 6.9638 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamduxdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamduxdx/G45/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.1705s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1287s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.0371s for 16384 events => throughput is 1.58E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0047s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamduxdx/G17/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.7630s + [COUNTERS] Fortran Overhead ( 0 ) : 3.7253s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.0328s for 16384 events => throughput is 1.59E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0050s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamduxdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuccx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuccx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuccx/ajob1' in 4.3280 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuccx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuccx/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.1215s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6025s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5144s for 16384 events => throughput is 3.19E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0046s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuccx/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.1768s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6506s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5216s for 16384 events => throughput is 3.14E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0047s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuccx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuddx/ajob1' in 15.5625 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuddx/G46/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.6324s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1119s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5159s for 16384 events => throughput is 3.18E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0046s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuddx/G18/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.8412s + [COUNTERS] Fortran Overhead ( 0 ) : 3.3232s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5135s for 16384 events => throughput is 3.19E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0045s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuddx/G39/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.5114s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9808s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5256s for 16384 events => throughput is 3.12E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0050s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuddx/G45/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.6210s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0984s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5178s for 16384 events => throughput is 3.16E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0047s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuddx/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.5887s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0665s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5173s for 16384 events => throughput is 3.17E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0048s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuddx/G20/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.8127s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2952s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5127s for 16384 events => throughput is 3.20E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0048s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuddx/G47/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.7823s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2639s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5137s for 16384 events => throughput is 3.19E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0048s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuddx/G7/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.6627s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1179s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5397s for 16384 events => throughput is 3.04E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0050s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/ajob1' in 81.0131 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G14/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.5887s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8863s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.6951s for 16384 events => throughput is 9.67E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0073s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G74/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.6155s + [COUNTERS] Fortran Overhead ( 0 ) : 3.9189s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.6899s for 16384 events => throughput is 9.70E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0068s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G12/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.1084s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3945s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7070s for 16384 events => throughput is 9.60E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0069s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G6/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.9346s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2381s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.6897s for 16384 events => throughput is 9.70E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0069s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G13/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.7762s + [COUNTERS] Fortran Overhead ( 0 ) : 4.0686s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7003s for 16384 events => throughput is 9.64E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0072s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G83/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.1107s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4063s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.6973s for 16384 events => throughput is 9.65E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0070s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G33/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.0392s + [COUNTERS] Fortran Overhead ( 0 ) : 4.3366s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.6954s for 16384 events => throughput is 9.66E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0071s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G11/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.9435s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2391s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.6975s for 16384 events => throughput is 9.65E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0069s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G9/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.9339s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2134s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7135s for 16384 events => throughput is 9.56E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0070s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G10/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.9461s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2431s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.6959s for 16384 events => throughput is 9.66E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0071s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.9182s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2246s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.6865s for 16384 events => throughput is 9.72E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0071s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G69/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.7448s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0481s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.6896s for 16384 events => throughput is 9.70E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0070s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G65/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.1869s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4877s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.6925s for 16384 events => throughput is 9.68E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0067s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G1/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.1818s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4725s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7022s for 16384 events => throughput is 9.63E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0071s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.1678s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4195s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7409s for 16384 events => throughput is 9.41E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0074s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G72/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.9991s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3003s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.6915s for 16384 events => throughput is 9.69E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0073s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G2/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.6638s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9584s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.6983s for 16384 events => throughput is 9.65E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0070s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G80/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.4471s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7507s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.6894s for 16384 events => throughput is 9.70E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0069s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G79/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.9071s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2077s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.6926s for 16384 events => throughput is 9.68E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0067s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G70/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.8372s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1258s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7044s for 16384 events => throughput is 9.61E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0070s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G5/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.6764s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9618s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7073s for 16384 events => throughput is 9.60E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0073s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamguux/ajob1' in 2.5255 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamguux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamguux/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.5090s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4693s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.0349s for 16384 events => throughput is 1.58E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0047s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamguux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/ajob1' in 83.8769 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G14/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.5215s + [COUNTERS] Fortran Overhead ( 0 ) : 3.8199s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.6948s for 16384 events => throughput is 9.67E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0069s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G12/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.8452s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1407s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.6974s for 16384 events => throughput is 9.65E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0070s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G6/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.6745s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9695s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.6982s for 16384 events => throughput is 9.65E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0069s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G3/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.4362s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7284s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7009s for 16384 events => throughput is 9.63E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0069s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G13/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.9365s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2359s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.6933s for 16384 events => throughput is 9.68E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0073s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G11/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.0170s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2883s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7218s for 16384 events => throughput is 9.52E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0069s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G9/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.9803s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2746s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.6990s for 16384 events => throughput is 9.64E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0067s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G10/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.5541s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8299s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7174s for 16384 events => throughput is 9.54E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0068s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G40/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.2971s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5935s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.6967s for 16384 events => throughput is 9.66E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0070s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G78/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.6235s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9285s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.6879s for 16384 events => throughput is 9.71E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0071s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G35/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.4929s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7958s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.6901s for 16384 events => throughput is 9.69E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0069s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.0561s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3412s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7077s for 16384 events => throughput is 9.59E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0072s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G69/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.7588s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0513s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7005s for 16384 events => throughput is 9.63E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0070s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G7/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.5965s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8969s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.6928s for 16384 events => throughput is 9.68E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0069s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G71/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.0028s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2948s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7008s for 16384 events => throughput is 9.63E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0073s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G1/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.2758s + [COUNTERS] Fortran Overhead ( 0 ) : 3.4977s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7709s for 16384 events => throughput is 9.25E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0072s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.9518s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2515s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.6931s for 16384 events => throughput is 9.68E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0071s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G4/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.4255s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7233s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.6950s for 16384 events => throughput is 9.67E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0071s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G80/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.8518s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1483s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.6963s for 16384 events => throughput is 9.66E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0072s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G79/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.4031s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6962s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7000s for 16384 events => throughput is 9.64E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0070s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.3308s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6244s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.6997s for 16384 events => throughput is 9.64E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0068s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G70/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.7462s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0353s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7040s for 16384 events => throughput is 9.62E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0069s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G5/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.7867s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0806s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.6991s for 16384 events => throughput is 9.64E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0071s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxcx_epemguxcx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxcx_epemguxcx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxcx_epemguxcx/ajob1' in 5.5080 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxcx_epemguxcx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxcx_epemguxcx/G9/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.4370s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9223s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5100s for 16384 events => throughput is 3.21E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0046s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxcx_epemguxcx/G10/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.0413s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5251s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5115s for 16384 events => throughput is 3.20E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0047s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxcx_epemguxcx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/ajob1' in 54.8078 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/G46/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.1484s + [COUNTERS] Fortran Overhead ( 0 ) : 4.4632s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.6784s for 16384 events => throughput is 9.76E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0068s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/G18/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.6846s + [COUNTERS] Fortran Overhead ( 0 ) : 3.9878s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.6900s for 16384 events => throughput is 9.69E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0068s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/G39/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.5346s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8432s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.6844s for 16384 events => throughput is 9.73E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0070s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.9400s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2553s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.6776s for 16384 events => throughput is 9.77E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0070s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/G20/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.2568s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5678s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.6821s for 16384 events => throughput is 9.74E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0069s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/G17/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.9675s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2729s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.6876s for 16384 events => throughput is 9.71E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0070s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/G35/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.1818s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4973s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.6775s for 16384 events => throughput is 9.77E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0069s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/G71/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.1051s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4206s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.6777s for 16384 events => throughput is 9.77E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0068s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/G27/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.7617s + [COUNTERS] Fortran Overhead ( 0 ) : 4.0763s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.6782s for 16384 events => throughput is 9.76E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0072s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/G92/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.1550s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3997s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7485s for 16384 events => throughput is 9.37E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0069s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/G80/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.4648s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7760s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.6820s for 16384 events => throughput is 9.74E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0068s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.2836s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5930s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.6832s for 16384 events => throughput is 9.73E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0075s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/G29/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.6315s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9433s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.6814s for 16384 events => throughput is 9.74E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0069s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/G48/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.5016s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8163s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.6779s for 16384 events => throughput is 9.76E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0074s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_epemgdux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_epemgdux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_epemgdux/ajob1' in 12.0227 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_epemgdux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_epemgdux/G46/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.3026s + [COUNTERS] Fortran Overhead ( 0 ) : 0.7805s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5173s for 16384 events => throughput is 3.17E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0048s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_epemgdux/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.6254s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1019s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5186s for 16384 events => throughput is 3.16E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0050s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_epemgdux/G23/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.4399s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9199s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5151s for 16384 events => throughput is 3.18E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0049s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_epemgdux/G47/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.3750s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8524s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5178s for 16384 events => throughput is 3.16E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0048s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_epemgdux/G8/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.7357s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2150s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5160s for 16384 events => throughput is 3.18E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0047s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_epemgdux/G7/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.5201s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9997s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5156s for 16384 events => throughput is 3.18E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0048s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_epemgdux/G28/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.6770s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1511s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5212s for 16384 events => throughput is 3.14E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0046s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_epemgdux/G32/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.2374s + [COUNTERS] Fortran Overhead ( 0 ) : 0.7222s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5106s for 16384 events => throughput is 3.21E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0046s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_epemgdux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/ajob1' in 61.5010 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/G12/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.7506s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4654s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.2783s for 16384 events => throughput is 5.00E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0069s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/G6/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.4525s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1434s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.3021s for 16384 events => throughput is 4.96E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0071s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/G11/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.5322s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2171s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.3081s for 16384 events => throughput is 4.95E+03 events/s +INFO: The following Floating Point Exceptions have been reported: FE_UNDERFLOW + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0070s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.4401s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1732s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.2599s for 16384 events => throughput is 5.03E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0069s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/G69/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.3377s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0660s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.2643s for 16384 events => throughput is 5.02E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0073s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/G8/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.2491s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9433s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.2986s for 16384 events => throughput is 4.97E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0072s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/G84/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.6347s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3295s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.2980s for 16384 events => throughput is 4.97E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0072s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.9125s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5662s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.3395s for 16384 events => throughput is 4.91E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0068s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/G2/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.2560s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9316s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.3172s for 16384 events => throughput is 4.94E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0072s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/G80/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.0123s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6957s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.3097s for 16384 events => throughput is 4.95E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0069s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/G51/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.0903s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7593s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.3239s for 16384 events => throughput is 4.93E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0071s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/G70/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.3687s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0739s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.2877s for 16384 events => throughput is 4.98E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0071s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/G5/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.2864s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9840s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.2952s for 16384 events => throughput is 4.97E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0071s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamguux/ajob1' in 47.1277 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamguux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamguux/G18/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.1753s + [COUNTERS] Fortran Overhead ( 0 ) : 3.8196s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.3487s for 16384 events => throughput is 4.89E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0070s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamguux/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.5990s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2452s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.3468s for 16384 events => throughput is 4.90E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0069s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamguux/G20/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.8756s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5176s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.3506s for 16384 events => throughput is 4.89E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0074s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamguux/G17/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.5377s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2015s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.3293s for 16384 events => throughput is 4.92E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0069s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamguux/G35/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.7667s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4280s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.3316s for 16384 events => throughput is 4.92E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0071s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamguux/G27/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.9796s + [COUNTERS] Fortran Overhead ( 0 ) : 3.6281s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.3446s for 16384 events => throughput is 4.90E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0069s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamguux/G34/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.9011s + [COUNTERS] Fortran Overhead ( 0 ) : 3.5669s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.3273s for 16384 events => throughput is 4.92E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0069s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamguux/G80/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.1816s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7901s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.3841s for 16384 events => throughput is 4.84E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0074s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamguux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_taptamgdsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_taptamgdsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_taptamgdsx/ajob1' in 2.1393 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_taptamgdsx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_taptamgdsx/G28/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.1223s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0864s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.0314s for 16384 events => throughput is 1.59E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0046s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_taptamgdsx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_taptamggg/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_taptamggg/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_taptamggg/ajob1' in 16.9692 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_taptamggg/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_taptamggg/G12/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.0056s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6694s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.3290s for 16384 events => throughput is 4.92E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0072s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_taptamggg/G9/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.8865s + [COUNTERS] Fortran Overhead ( 0 ) : 3.4991s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.3800s for 16384 events => throughput is 4.85E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0073s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_taptamggg/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.0336s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6945s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.3320s for 16384 events => throughput is 4.92E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0071s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_taptamggg/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxdx_epemguxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxdx_epemguxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxdx_epemguxdx/ajob1' in 15.0051 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxdx_epemguxdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxdx_epemguxdx/G14/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.5876s + [COUNTERS] Fortran Overhead ( 0 ) : 3.0697s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5132s for 16384 events => throughput is 3.19E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0047s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxdx_epemguxdx/G6/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.0308s + [COUNTERS] Fortran Overhead ( 0 ) : 4.5130s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5131s for 16384 events => throughput is 3.19E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0048s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxdx_epemguxdx/G22/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.7450s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2280s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5123s for 16384 events => throughput is 3.20E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0047s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxdx_epemguxdx/G8/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.7306s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2131s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5130s for 16384 events => throughput is 3.19E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0045s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxdx_epemguxdx/G27/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.8421s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3207s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5164s for 16384 events => throughput is 3.17E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0050s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxdx_epemguxdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_epemgdsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_epemgdsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_epemgdsx/ajob1' in 13.6152 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_epemgdsx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_epemgdsx/G31/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.4137s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8989s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5104s for 16384 events => throughput is 3.21E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0045s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_epemgdsx/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.4509s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9319s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5143s for 16384 events => throughput is 3.19E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0046s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_epemgdsx/G20/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.6224s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1056s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5122s for 16384 events => throughput is 3.20E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0047s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_epemgdsx/G25/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.1853s + [COUNTERS] Fortran Overhead ( 0 ) : 3.6689s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5116s for 16384 events => throughput is 3.20E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0048s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_epemgdsx/G28/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.5257s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0096s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5112s for 16384 events => throughput is 3.20E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0048s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_epemgdsx/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.3345s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8170s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5129s for 16384 events => throughput is 3.19E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0046s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_epemgdsx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamdddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamdddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamdddx/ajob1' in 5.6585 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamdddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamdddx/G17/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.6422s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2730s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.3626s for 16384 events => throughput is 6.93E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0067s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamdddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemggg/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemggg/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemggg/ajob1' in 19.9307 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemggg/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemggg/G56/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.3473s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6378s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7027s for 16384 events => throughput is 9.62E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0069s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemggg/G12/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.4670s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7630s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.6969s for 16384 events => throughput is 9.66E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0071s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemggg/G11/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.4597s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7591s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.6935s for 16384 events => throughput is 9.67E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0071s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemggg/G40/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.2261s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5251s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.6941s for 16384 events => throughput is 9.67E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0069s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemggg/G8/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.0561s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3326s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7165s for 16384 events => throughput is 9.54E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0070s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemggg/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.2901s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5826s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7003s for 16384 events => throughput is 9.64E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0073s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemggg/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/ajob1' in 121.3207 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G14/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.4073s + [COUNTERS] Fortran Overhead ( 0 ) : 3.0836s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.3166s for 16384 events => throughput is 4.94E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0071s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G74/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.1676s + [COUNTERS] Fortran Overhead ( 0 ) : 3.7657s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.3949s for 16384 events => throughput is 4.83E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0070s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G12/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.4882s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1756s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.3053s for 16384 events => throughput is 4.96E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0073s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G6/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.3345s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0494s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.2781s for 16384 events => throughput is 5.00E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0070s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G13/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.2807s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9699s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.3036s for 16384 events => throughput is 4.96E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0072s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G66/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.8383s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5305s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.3005s for 16384 events => throughput is 4.96E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0073s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G11/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.6562s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3611s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.2882s for 16384 events => throughput is 4.98E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0069s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G9/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.4855s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1470s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.3313s for 16384 events => throughput is 4.92E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0072s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G40/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.0650s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7414s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.3163s for 16384 events => throughput is 4.94E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0073s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G35/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.0718s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7515s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.3130s for 16384 events => throughput is 4.95E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0073s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.9439s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6334s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.3033s for 16384 events => throughput is 4.96E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0073s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G69/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.3044s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9926s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.3048s for 16384 events => throughput is 4.96E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0070s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G7/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.1839s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8510s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.3259s for 16384 events => throughput is 4.93E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0070s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G65/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.9607s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6566s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.2972s for 16384 events => throughput is 4.97E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0070s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G71/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.6213s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3112s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.3032s for 16384 events => throughput is 4.96E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0069s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G1/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.3574s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9574s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.3931s for 16384 events => throughput is 4.83E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0069s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.5579s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2374s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.3135s for 16384 events => throughput is 4.94E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0071s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G2/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.6717s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3498s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.3145s for 16384 events => throughput is 4.94E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0073s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G4/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.8610s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6170s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.2369s for 16384 events => throughput is 5.06E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0072s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G79/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.1515s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8397s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.3045s for 16384 events => throughput is 4.96E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0073s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.8683s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5621s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.2991s for 16384 events => throughput is 4.97E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0071s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G70/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.3014s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9843s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.3102s for 16384 events => throughput is 4.95E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0069s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G5/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.4339s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1191s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.3077s for 16384 events => throughput is 4.95E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0070s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/ajob1' in 69.7391 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/G46/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.8436s + [COUNTERS] Fortran Overhead ( 0 ) : 4.1381s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.6987s for 16384 events => throughput is 9.64E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0067s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/G18/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.2042s + [COUNTERS] Fortran Overhead ( 0 ) : 3.4653s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7316s for 16384 events => throughput is 9.46E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0074s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.1638s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4672s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.6897s for 16384 events => throughput is 9.70E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0069s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/G33/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.2059s + [COUNTERS] Fortran Overhead ( 0 ) : 3.5086s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.6903s for 16384 events => throughput is 9.69E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0070s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/G20/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.3356s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6265s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7022s for 16384 events => throughput is 9.63E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0070s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/G47/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.0059s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2819s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7170s for 16384 events => throughput is 9.54E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0070s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/G17/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.8992s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1978s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.6943s for 16384 events => throughput is 9.67E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0070s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/G25/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.9256s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2279s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.6908s for 16384 events => throughput is 9.69E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0069s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/G35/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.0522s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3561s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.6892s for 16384 events => throughput is 9.70E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0069s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/G65/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.8930s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1914s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.6947s for 16384 events => throughput is 9.67E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0069s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/G27/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.4069s + [COUNTERS] Fortran Overhead ( 0 ) : 3.7105s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.6896s for 16384 events => throughput is 9.70E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0068s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/G34/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.1058s + [COUNTERS] Fortran Overhead ( 0 ) : 3.3968s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7022s for 16384 events => throughput is 9.63E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0068s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/G80/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.6657s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9205s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7382s for 16384 events => throughput is 9.43E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0070s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/G67/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.4025s + [COUNTERS] Fortran Overhead ( 0 ) : 3.6994s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.6959s for 16384 events => throughput is 9.66E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0072s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.2091s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5124s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.6897s for 16384 events => throughput is 9.70E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0070s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/G48/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.2026s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5040s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.6915s for 16384 events => throughput is 9.69E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0071s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_taptamgdux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_taptamgdux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_taptamgdux/ajob1' in 8.4283 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_taptamgdux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_taptamgdux/G14/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.5349s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4889s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.0414s for 16384 events => throughput is 1.57E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0046s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_taptamgdux/G40/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.8833s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8412s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.0374s for 16384 events => throughput is 1.58E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0046s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_taptamgdux/G44/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.9669s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9272s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.0350s for 16384 events => throughput is 1.58E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0047s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_taptamgdux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemgddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemgddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemgddx/ajob1' in 11.0874 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemgddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemgddx/G56/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.2507s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0529s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.1911s for 16384 events => throughput is 1.38E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0067s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemgddx/G12/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.1087s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9162s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.1852s for 16384 events => throughput is 1.38E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0073s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemgddx/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.1300s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9375s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.1857s for 16384 events => throughput is 1.38E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0068s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemgddx/G47/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.3145s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1201s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.1880s for 16384 events => throughput is 1.38E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0064s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemgddx/G42/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.2138s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0249s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.1823s for 16384 events => throughput is 1.39E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0066s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemgddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/ajob1' in 77.1901 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/G39/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.6997s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3970s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.2957s for 16384 events => throughput is 4.97E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0070s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/G12/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.5148s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2023s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.3054s for 16384 events => throughput is 4.96E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0071s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/G6/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.1862s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8958s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.2832s for 16384 events => throughput is 4.99E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0072s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/G66/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.0261s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6989s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.3202s for 16384 events => throughput is 4.93E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0071s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/G11/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.6780s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3817s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.2895s for 16384 events => throughput is 4.98E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0068s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/G9/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.8836s + [COUNTERS] Fortran Overhead ( 0 ) : 3.5989s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.2774s for 16384 events => throughput is 5.00E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0073s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.7589s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4361s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.3157s for 16384 events => throughput is 4.94E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0070s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/G69/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.3865s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1016s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.2780s for 16384 events => throughput is 5.00E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0069s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/G8/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.2573s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9589s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.2912s for 16384 events => throughput is 4.98E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0071s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/G71/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.6131s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3135s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.2926s for 16384 events => throughput is 4.98E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0070s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.5118s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2043s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.3005s for 16384 events => throughput is 4.96E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0070s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/G72/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.4598s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1717s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.2812s for 16384 events => throughput is 4.99E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0070s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/G79/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.0849s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7324s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.3454s for 16384 events => throughput is 4.90E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0072s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.1006s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8184s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.2754s for 16384 events => throughput is 5.00E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0068s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/G70/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.3151s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0219s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.2861s for 16384 events => throughput is 4.99E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0071s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/G5/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.4960s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2116s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.2776s for 16384 events => throughput is 5.00E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0068s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemdssx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemdssx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemdssx/ajob1' in 5.1265 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemdssx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemdssx/G46/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.6262s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1132s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5083s for 16384 events => throughput is 3.22E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0047s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemdssx/G17/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.4705s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9591s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5068s for 16384 events => throughput is 3.23E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0046s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemdssx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uc_taptamguc/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uc_taptamguc/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uc_taptamguc/ajob1' in 5.2306 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uc_taptamguc/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uc_taptamguc/G25/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.3069s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2648s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.0372s for 16384 events => throughput is 1.58E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0049s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uc_taptamguc/G37/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.8941s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8549s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.0345s for 16384 events => throughput is 1.58E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0048s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uc_taptamguc/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamgddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamgddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamgddx/ajob1' in 7.6388 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamgddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamgddx/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.0127s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6260s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.3801s for 16384 events => throughput is 6.88E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0066s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamgddx/G55/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.5961s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1947s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.3949s for 16384 events => throughput is 6.84E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0064s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamgddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemgddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemgddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemgddx/ajob1' in 2.0495 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemgddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemgddx/G8/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.0328s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5048s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5232s for 16384 events => throughput is 3.13E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0047s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemgddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamgddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamgddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamgddx/ajob1' in 46.3958 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamgddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamgddx/G31/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.4794s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1645s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.3077s for 16384 events => throughput is 4.95E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0071s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamgddx/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.5479s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2297s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.3108s for 16384 events => throughput is 4.95E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0074s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamgddx/G33/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.4587s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1385s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.3130s for 16384 events => throughput is 4.95E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0072s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamgddx/G20/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.7089s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3791s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.3226s for 16384 events => throughput is 4.93E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0072s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamgddx/G35/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.5650s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2404s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.3176s for 16384 events => throughput is 4.94E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0070s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamgddx/G71/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.4310s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0920s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.3319s for 16384 events => throughput is 4.92E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0072s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamgddx/G34/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.0933s + [COUNTERS] Fortran Overhead ( 0 ) : 3.7825s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.3036s for 16384 events => throughput is 4.96E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0073s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamgddx/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.8667s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5068s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.3528s for 16384 events => throughput is 4.89E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0071s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamgddx/G48/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.1213s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7483s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.3655s for 16384 events => throughput is 4.87E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0075s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamgddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemddxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemddxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemddxdx/ajob1' in 5.1486 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemddxdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemddxdx/G11/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.7249s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5318s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.1865s for 16384 events => throughput is 1.38E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0066s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemddxdx/G71/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.3941s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1994s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.1880s for 16384 events => throughput is 1.38E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0067s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemddxdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dxsx_epemgdxsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dxsx_epemgdxsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dxsx_epemgdxsx/ajob1' in 1.7561 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dxsx_epemgdxsx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dxsx_epemgdxsx/G27/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.7394s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2246s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5103s for 16384 events => throughput is 3.21E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0045s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dxsx_epemgdxsx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemduxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemduxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemduxdx/ajob1' in 4.8564 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemduxdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemduxdx/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.5743s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0542s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5155s for 16384 events => throughput is 3.18E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0046s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemduxdx/G22/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.2525s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7270s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5209s for 16384 events => throughput is 3.15E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0046s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemduxdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemguux/ajob1' in 21.0257 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemguux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemguux/G20/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.2371s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0387s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.1917s for 16384 events => throughput is 1.37E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0067s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemguux/G27/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.0495s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8180s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.2248s for 16384 events => throughput is 1.34E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0067s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemguux/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.6004s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3967s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.1969s for 16384 events => throughput is 1.37E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0068s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemguux/G55/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.3879s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1945s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.1868s for 16384 events => throughput is 1.38E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0066s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemguux/G37/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.2303s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0352s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.1887s for 16384 events => throughput is 1.38E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0064s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemguux/G41/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.0852s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8968s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.1814s for 16384 events => throughput is 1.39E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0070s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemguux/G57/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.8818s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6868s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.1883s for 16384 events => throughput is 1.38E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0066s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemguux/G48/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.4447s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2476s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.1901s for 16384 events => throughput is 1.38E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0070s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemguux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dd_taptamgdd/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dd_taptamgdd/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dd_taptamgdd/ajob1' in 7.8469 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dd_taptamgdd/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dd_taptamgdd/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.6288s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2321s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.3902s for 16384 events => throughput is 6.85E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0065s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dd_taptamgdd/G24/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.1886s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8027s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.3794s for 16384 events => throughput is 6.89E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0065s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dd_taptamgdd/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/ajob1' in 145.2937 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G14/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.0723s + [COUNTERS] Fortran Overhead ( 0 ) : 3.3773s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.6877s for 16384 events => throughput is 9.71E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0074s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G38/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.3743s + [COUNTERS] Fortran Overhead ( 0 ) : 3.6746s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.6928s for 16384 events => throughput is 9.68E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0068s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G39/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.1132s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4009s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7053s for 16384 events => throughput is 9.61E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0070s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G12/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.8914s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1927s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.6918s for 16384 events => throughput is 9.68E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0069s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G6/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.5980s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8940s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.6972s for 16384 events => throughput is 9.65E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0068s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G3/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.7564s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9458s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.8032s for 16384 events => throughput is 9.09E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0075s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G13/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.9520s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2556s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.6895s for 16384 events => throughput is 9.70E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0069s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G66/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.6140s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9153s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.6916s for 16384 events => throughput is 9.69E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0071s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G33/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.5931s + [COUNTERS] Fortran Overhead ( 0 ) : 4.8959s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.6902s for 16384 events => throughput is 9.69E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0069s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G11/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.9797s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2769s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.6956s for 16384 events => throughput is 9.66E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0072s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G9/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 26.3094s + [COUNTERS] Fortran Overhead ( 0 ) : 17.8291s + [COUNTERS] CudaCpp MEs ( 2 ) : 8.4731s for 81920 events => throughput is 9.67E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0071s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G10/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.5582s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8637s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.6873s for 16384 events => throughput is 9.71E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0072s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G40/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.2938s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5924s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.6946s for 16384 events => throughput is 9.67E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0068s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G35/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.2625s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5609s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.6945s for 16384 events => throughput is 9.67E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0071s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.1876s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3745s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.8057s for 16384 events => throughput is 9.07E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0074s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G69/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.7339s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0381s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.6887s for 16384 events => throughput is 9.70E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0070s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G8/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.6129s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9072s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.6985s for 16384 events => throughput is 9.65E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0072s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G7/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.7412s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0413s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.6929s for 16384 events => throughput is 9.68E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0070s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G65/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.7394s + [COUNTERS] Fortran Overhead ( 0 ) : 3.0409s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.6916s for 16384 events => throughput is 9.69E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0069s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G71/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.1710s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4693s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.6946s for 16384 events => throughput is 9.67E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0070s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G1/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.9949s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2994s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.6886s for 16384 events => throughput is 9.70E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0069s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.9434s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2471s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.6894s for 16384 events => throughput is 9.70E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0069s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G2/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.8012s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1052s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.6890s for 16384 events => throughput is 9.70E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0069s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G4/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.6274s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9204s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7000s for 16384 events => throughput is 9.64E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0070s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G80/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.8199s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1233s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.6896s for 16384 events => throughput is 9.70E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0069s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G41/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.2371s + [COUNTERS] Fortran Overhead ( 0 ) : 3.5336s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.6966s for 16384 events => throughput is 9.66E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0070s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G77/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.5367s + [COUNTERS] Fortran Overhead ( 0 ) : 3.8192s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7103s for 16384 events => throughput is 9.58E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0072s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G76/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.0820s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3204s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7546s for 16384 events => throughput is 9.34E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0070s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G79/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.5606s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8409s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7128s for 16384 events => throughput is 9.57E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0070s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.1803s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4676s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7055s for 16384 events => throughput is 9.61E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0072s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G70/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.6944s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9906s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.6967s for 16384 events => throughput is 9.66E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0070s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G5/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.8320s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0757s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.7489s for 16384 events => throughput is 9.37E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0074s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ds_taptamgds/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ds_taptamgds/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ds_taptamgds/ajob1' in 2.9042 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ds_taptamgds/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ds_taptamgds/G10/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.8876s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8282s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.0544s for 16384 events => throughput is 1.55E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0050s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ds_taptamgds/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_epemgudx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_epemgudx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_epemgudx/ajob1' in 11.3309 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_epemgudx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_epemgudx/G12/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.4128s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8565s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5518s for 16384 events => throughput is 2.97E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0045s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_epemgudx/G11/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.3054s + [COUNTERS] Fortran Overhead ( 0 ) : 0.7865s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5142s for 16384 events => throughput is 3.19E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0047s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_epemgudx/G17/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.9005s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3711s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5246s for 16384 events => throughput is 3.12E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0048s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_epemgudx/G10/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.4609s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9426s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5136s for 16384 events => throughput is 3.19E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0047s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_epemgudx/G27/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.6114s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0934s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5134s for 16384 events => throughput is 3.19E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0047s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_epemgudx/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.5574s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0335s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5193s for 16384 events => throughput is 3.15E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0046s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_epemgudx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_taptamgud/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_taptamgud/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_taptamgud/ajob1' in 20.9296 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_taptamgud/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_taptamgud/G12/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.8591s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8164s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.0380s for 16384 events => throughput is 1.58E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0048s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_taptamgud/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.1304s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0927s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.0330s for 16384 events => throughput is 1.59E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0048s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_taptamgud/G6/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.1901s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1463s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.0389s for 16384 events => throughput is 1.58E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0049s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_taptamgud/G23/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.8819s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8377s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.0394s for 16384 events => throughput is 1.58E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0048s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_taptamgud/G11/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.9024s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8601s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.0378s for 16384 events => throughput is 1.58E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0046s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_taptamgud/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.2261s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1867s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.0347s for 16384 events => throughput is 1.58E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0048s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_taptamgud/G22/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.4022s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3636s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.0340s for 16384 events => throughput is 1.58E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0046s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_taptamgud/G7/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.2763s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2354s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.0363s for 16384 events => throughput is 1.58E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0045s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_taptamgud/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.9386s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8970s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.0369s for 16384 events => throughput is 1.58E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0047s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_taptamgud/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemdddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemdddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemdddx/ajob1' in 5.4213 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemdddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemdddx/G31/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.8623s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6684s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.1873s for 16384 events => throughput is 1.38E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0066s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemdddx/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.5293s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3280s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.1948s for 16384 events => throughput is 1.37E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0065s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemdddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamdssx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamdssx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamdssx/ajob1' in 4.4694 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamdssx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamdssx/G45/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.1500s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0995s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.0456s for 16384 events => throughput is 1.57E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0049s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamdssx/G20/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.2895s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2439s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.0409s for 16384 events => throughput is 1.57E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0047s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamdssx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuddx/ajob1' in 6.4586 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuddx/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.0880s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0404s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.0428s for 16384 events => throughput is 1.57E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0047s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuddx/G20/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.1702s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1261s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.0393s for 16384 events => throughput is 1.58E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0047s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuddx/G47/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.1575s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1148s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.0382s for 16384 events => throughput is 1.58E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0045s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemuuxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemuuxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemuuxdx/ajob1' in 5.4651 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemuuxdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemuuxdx/G7/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.8315s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3070s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5197s for 16384 events => throughput is 3.15E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0048s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemuuxdx/G41/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.6039s + [COUNTERS] Fortran Overhead ( 0 ) : 3.0826s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5166s for 16384 events => throughput is 3.17E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0047s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemuuxdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/ajob1' in 38.1603 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G14/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.5507s + [COUNTERS] Fortran Overhead ( 0 ) : 3.0369s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5091s for 16384 events => throughput is 3.22E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0046s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G18/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.0485s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5323s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5112s for 16384 events => throughput is 3.20E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0049s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G39/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.3343s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8161s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5136s for 16384 events => throughput is 3.19E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0046s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G30/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.6525s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1370s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5110s for 16384 events => throughput is 3.21E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0046s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G12/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.3263s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8082s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5134s for 16384 events => throughput is 3.19E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0046s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.7036s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1777s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5211s for 16384 events => throughput is 3.14E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0049s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G6/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.8923s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3752s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5126s for 16384 events => throughput is 3.20E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0045s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G3/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.3557s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8403s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5105s for 16384 events => throughput is 3.21E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0048s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G23/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.4078s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8744s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5286s for 16384 events => throughput is 3.10E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0049s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G24/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.3689s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8540s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5100s for 16384 events => throughput is 3.21E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0049s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G13/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.7825s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2680s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5098s for 16384 events => throughput is 3.21E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0047s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G20/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.4010s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8839s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5125s for 16384 events => throughput is 3.20E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0046s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G25/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.9673s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4475s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5153s for 16384 events => throughput is 3.18E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0045s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.6434s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1271s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5115s for 16384 events => throughput is 3.20E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0047s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G8/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.4411s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9258s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5106s for 16384 events => throughput is 3.21E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0047s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G7/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.5244s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0089s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5107s for 16384 events => throughput is 3.21E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0048s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G4/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.2599s + [COUNTERS] Fortran Overhead ( 0 ) : 0.6987s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5563s for 16384 events => throughput is 2.95E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0049s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G41/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.8457s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3255s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5157s for 16384 events => throughput is 3.18E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0046s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G32/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.4011s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8853s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5109s for 16384 events => throughput is 3.21E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0049s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ucx_epemgucx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ucx_epemgucx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ucx_epemgucx/ajob1' in 4.5855 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ucx_epemgucx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ucx_epemgucx/G3/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.1616s + [COUNTERS] Fortran Overhead ( 0 ) : 0.6435s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5134s for 16384 events => throughput is 3.19E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0047s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ucx_epemgucx/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.7797s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2591s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5157s for 16384 events => throughput is 3.18E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0048s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ucx_epemgucx/G7/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.6014s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0870s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5099s for 16384 events => throughput is 3.21E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0045s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ucx_epemgucx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_taptamgudx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_taptamgudx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_taptamgudx/ajob1' in 11.8085 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_taptamgudx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_taptamgudx/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.9842s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9465s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.0331s for 16384 events => throughput is 1.59E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0045s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_taptamgudx/G11/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.9186s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8784s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.0356s for 16384 events => throughput is 1.58E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0046s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_taptamgudx/G17/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.4498s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3967s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.0485s for 16384 events => throughput is 1.56E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0046s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_taptamgudx/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.2628s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2242s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.0338s for 16384 events => throughput is 1.58E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0048s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_taptamgudx/G27/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.1233s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0834s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.0353s for 16384 events => throughput is 1.58E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0046s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_taptamgudx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamuuxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamuuxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamuuxdx/ajob1' in 3.8023 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamuuxdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamuuxdx/G6/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.7859s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7451s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.0361s for 16384 events => throughput is 1.58E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0046s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamuuxdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_taptamguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_taptamguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_taptamguux/ajob1' in 3.5986 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_taptamguux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_taptamguux/G48/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.5823s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2174s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.3583s for 16384 events => throughput is 6.95E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0067s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_taptamguux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemudux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemudux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemudux/ajob1' in 9.5724 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemudux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemudux/G43/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.0469s + [COUNTERS] Fortran Overhead ( 0 ) : 3.5277s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5144s for 16384 events => throughput is 3.19E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0049s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemudux/G8/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.6543s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1360s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5137s for 16384 events => throughput is 3.19E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0045s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemudux/G5/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.8283s + [COUNTERS] Fortran Overhead ( 0 ) : 3.3116s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5119s for 16384 events => throughput is 3.20E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0047s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemudux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamsdxsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamsdxsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamsdxsx/ajob1' in 2.3419 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamsdxsx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamsdxsx/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.3253s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2720s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.0486s for 16384 events => throughput is 1.56E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0047s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamsdxsx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_epemguu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_epemguu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_epemguu/ajob1' in 25.7935 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_epemguu/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_epemguu/G30/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.1838s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9931s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.1843s for 16384 events => throughput is 1.38E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0065s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_epemguu/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.1383s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9493s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.1827s for 16384 events => throughput is 1.39E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0063s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_epemguu/G9/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.4793s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2861s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.1865s for 16384 events => throughput is 1.38E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0066s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_epemguu/G49/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.0930s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9062s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.1801s for 16384 events => throughput is 1.39E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0067s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_epemguu/G10/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.4191s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2241s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.1885s for 16384 events => throughput is 1.38E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0065s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_epemguu/G81/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.2150s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0229s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.1856s for 16384 events => throughput is 1.38E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0065s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_epemguu/G62/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.9867s + [COUNTERS] Fortran Overhead ( 0 ) : 0.7983s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.1818s for 16384 events => throughput is 1.39E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0065s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_epemguu/G51/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.0496s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8591s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.1837s for 16384 events => throughput is 1.38E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0068s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_epemguu/G28/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.1066s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8369s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.2628s for 16384 events => throughput is 1.30E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0069s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_epemguu/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uc_epemguc/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uc_epemguc/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uc_epemguc/ajob1' in 2.8771 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uc_epemguc/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uc_epemguc/G12/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.2845s + [COUNTERS] Fortran Overhead ( 0 ) : 0.7679s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5119s for 16384 events => throughput is 3.20E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0047s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uc_epemguc/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.5629s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0474s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5108s for 16384 events => throughput is 3.21E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0047s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uc_epemguc/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ds_epemgds/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ds_epemgds/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ds_epemgds/ajob1' in 1.5976 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ds_epemgds/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ds_epemgds/G27/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.5810s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0669s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5096s for 16384 events => throughput is 3.21E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0045s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ds_epemgds/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_taptamguu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_taptamguu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_taptamguu/ajob1' in 10.6795 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_taptamguu/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_taptamguu/G52/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.3048s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9412s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.3572s for 16384 events => throughput is 6.95E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0064s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_taptamguu/G49/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.0232s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6416s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.3751s for 16384 events => throughput is 6.90E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0064s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_taptamguu/G32/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.3085s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9482s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.3537s for 16384 events => throughput is 6.96E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0067s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_taptamguu/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished in 1318.6667 seconds +__CUDACPP_DEBUG: gen_ximprove.launch finished in 1319.0954 seconds +__CUDACPP_DEBUG: GridPackCmd.refine4grid finished in 1319.0961 seconds +__CUDACPP_DEBUG: GridPackCmd.launch (3a) combine_events +combine_events +__CUDACPP_DEBUG: GridPackCmd.launch (3b) store_events +store_events + === Results Summary for run: GridRun_21 tag: tag_1 === + + Cross-section : 0 +- 0 pb + Nb of events : 500 + +INFO: No version of lhapdf. Can not run systematics computation +decay_events -from_cards +__CUDACPP_DEBUG: GridPackCmd.launch finished in 1336.0181 seconds +__CUDACPP_DEBUG: GridPackCmd.__init__ back from self.launch +DONE +write ./events.lhe.gz +END: Wed Sep 11 03:29:16 AM CEST 2024 +ELAPSED: 1338 seconds diff --git a/epochX/cudacpp/tlau/fromgridpacks/pp_dy3j.mad/cuda/output.txt b/epochX/cudacpp/tlau/fromgridpacks/pp_dy3j.mad/cuda/output.txt new file mode 100644 index 0000000000..30a7686bb3 --- /dev/null +++ b/epochX/cudacpp/tlau/fromgridpacks/pp_dy3j.mad/cuda/output.txt @@ -0,0 +1,2335 @@ +START: Wed Sep 11 02:17:25 AM CEST 2024 +On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU (nproc=4)] [GPU: 1x Tesla V100S-PCIE-32GB]: +CUDACPP_RUNTIME_DISABLEFPE is set +lrwxrwxrwx. 1 avalassi zg 36 Sep 11 02:17 madevent/SubProcesses/P0_dc_epemgdc/madevent -> build.cuda_m_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 11 02:17 madevent/SubProcesses/P0_dc_taptamgdc/madevent -> build.cuda_m_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 11 02:17 madevent/SubProcesses/P0_dd_epemgdd/madevent -> build.cuda_m_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 11 02:17 madevent/SubProcesses/P0_dd_taptamgdd/madevent -> build.cuda_m_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 11 02:17 madevent/SubProcesses/P0_ddx_epemgddx/madevent -> build.cuda_m_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 11 02:17 madevent/SubProcesses/P0_ddx_epemggg/madevent -> build.cuda_m_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 11 02:17 madevent/SubProcesses/P0_ddx_epemgssx/madevent -> build.cuda_m_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 11 02:17 madevent/SubProcesses/P0_ddx_epemguux/madevent -> build.cuda_m_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 11 02:17 madevent/SubProcesses/P0_ddx_taptamgddx/madevent -> build.cuda_m_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 11 02:17 madevent/SubProcesses/P0_ddx_taptamggg/madevent -> build.cuda_m_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 11 02:17 madevent/SubProcesses/P0_ddx_taptamgssx/madevent -> build.cuda_m_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 11 02:17 madevent/SubProcesses/P0_ddx_taptamguux/madevent -> build.cuda_m_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 11 02:17 madevent/SubProcesses/P0_ds_epemgds/madevent -> build.cuda_m_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 11 02:17 madevent/SubProcesses/P0_ds_taptamgds/madevent -> build.cuda_m_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 11 02:17 madevent/SubProcesses/P0_dsx_epemgdsx/madevent -> build.cuda_m_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 11 02:17 madevent/SubProcesses/P0_dsx_taptamgdsx/madevent -> build.cuda_m_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 11 02:17 madevent/SubProcesses/P0_dux_epemgdux/madevent -> build.cuda_m_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 11 02:17 madevent/SubProcesses/P0_dux_taptamgdux/madevent -> build.cuda_m_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 11 02:17 madevent/SubProcesses/P0_dxcx_epemgdxcx/madevent -> build.cuda_m_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 11 02:17 madevent/SubProcesses/P0_dxcx_taptamgdxcx/madevent -> build.cuda_m_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 11 02:17 madevent/SubProcesses/P0_dxdx_epemgdxdx/madevent -> build.cuda_m_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 11 02:17 madevent/SubProcesses/P0_dxdx_taptamgdxdx/madevent -> build.cuda_m_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 11 02:17 madevent/SubProcesses/P0_dxsx_epemgdxsx/madevent -> build.cuda_m_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 11 02:17 madevent/SubProcesses/P0_dxsx_taptamgdxsx/madevent -> build.cuda_m_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 11 02:17 madevent/SubProcesses/P0_gd_epemdddx/madevent -> build.cuda_m_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 11 02:17 madevent/SubProcesses/P0_gd_epemdssx/madevent -> build.cuda_m_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 11 02:17 madevent/SubProcesses/P0_gd_epemggd/madevent -> build.cuda_m_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 11 02:17 madevent/SubProcesses/P0_gd_epemudux/madevent -> build.cuda_m_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 11 02:17 madevent/SubProcesses/P0_gd_taptamdddx/madevent -> build.cuda_m_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 11 02:17 madevent/SubProcesses/P0_gd_taptamdssx/madevent -> build.cuda_m_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 11 02:17 madevent/SubProcesses/P0_gd_taptamggd/madevent -> build.cuda_m_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 11 02:17 madevent/SubProcesses/P0_gd_taptamudux/madevent -> build.cuda_m_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 11 02:17 madevent/SubProcesses/P0_gdx_epemddxdx/madevent -> build.cuda_m_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 11 02:17 madevent/SubProcesses/P0_gdx_epemggdx/madevent -> build.cuda_m_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 11 02:17 madevent/SubProcesses/P0_gdx_epemsdxsx/madevent -> build.cuda_m_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 11 02:17 madevent/SubProcesses/P0_gdx_epemuuxdx/madevent -> build.cuda_m_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 11 02:17 madevent/SubProcesses/P0_gdx_taptamddxdx/madevent -> build.cuda_m_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 11 02:17 madevent/SubProcesses/P0_gdx_taptamggdx/madevent -> build.cuda_m_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 11 02:17 madevent/SubProcesses/P0_gdx_taptamsdxsx/madevent -> build.cuda_m_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 11 02:17 madevent/SubProcesses/P0_gdx_taptamuuxdx/madevent -> build.cuda_m_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 11 02:17 madevent/SubProcesses/P0_gg_epemgddx/madevent -> build.cuda_m_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 11 02:17 madevent/SubProcesses/P0_gg_epemguux/madevent -> build.cuda_m_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 11 02:17 madevent/SubProcesses/P0_gg_taptamgddx/madevent -> build.cuda_m_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 11 02:17 madevent/SubProcesses/P0_gg_taptamguux/madevent -> build.cuda_m_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 11 02:17 madevent/SubProcesses/P0_gu_epemggu/madevent -> build.cuda_m_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 11 02:17 madevent/SubProcesses/P0_gu_epemuccx/madevent -> build.cuda_m_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 11 02:17 madevent/SubProcesses/P0_gu_epemuddx/madevent -> build.cuda_m_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 11 02:17 madevent/SubProcesses/P0_gu_epemuuux/madevent -> build.cuda_m_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 11 02:17 madevent/SubProcesses/P0_gu_taptamggu/madevent -> build.cuda_m_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 11 02:17 madevent/SubProcesses/P0_gu_taptamuccx/madevent -> build.cuda_m_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 11 02:17 madevent/SubProcesses/P0_gu_taptamuddx/madevent -> build.cuda_m_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 11 02:17 madevent/SubProcesses/P0_gu_taptamuuux/madevent -> build.cuda_m_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 11 02:17 madevent/SubProcesses/P0_gux_epemcuxcx/madevent -> build.cuda_m_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 11 02:17 madevent/SubProcesses/P0_gux_epemduxdx/madevent -> build.cuda_m_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 11 02:17 madevent/SubProcesses/P0_gux_epemggux/madevent -> build.cuda_m_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 11 02:17 madevent/SubProcesses/P0_gux_epemuuxux/madevent -> build.cuda_m_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 11 02:17 madevent/SubProcesses/P0_gux_taptamcuxcx/madevent -> build.cuda_m_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 11 02:17 madevent/SubProcesses/P0_gux_taptamduxdx/madevent -> build.cuda_m_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 11 02:17 madevent/SubProcesses/P0_gux_taptamggux/madevent -> build.cuda_m_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 11 02:17 madevent/SubProcesses/P0_gux_taptamuuxux/madevent -> build.cuda_m_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 11 02:17 madevent/SubProcesses/P0_uc_epemguc/madevent -> build.cuda_m_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 11 02:17 madevent/SubProcesses/P0_uc_taptamguc/madevent -> build.cuda_m_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 11 02:17 madevent/SubProcesses/P0_ucx_epemgucx/madevent -> build.cuda_m_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 11 02:17 madevent/SubProcesses/P0_ucx_taptamgucx/madevent -> build.cuda_m_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 11 02:17 madevent/SubProcesses/P0_ud_epemgud/madevent -> build.cuda_m_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 11 02:17 madevent/SubProcesses/P0_ud_taptamgud/madevent -> build.cuda_m_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 11 02:17 madevent/SubProcesses/P0_udx_epemgudx/madevent -> build.cuda_m_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 11 02:17 madevent/SubProcesses/P0_udx_taptamgudx/madevent -> build.cuda_m_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 11 02:17 madevent/SubProcesses/P0_uu_epemguu/madevent -> build.cuda_m_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 11 02:17 madevent/SubProcesses/P0_uu_taptamguu/madevent -> build.cuda_m_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 11 02:17 madevent/SubProcesses/P0_uux_epemgccx/madevent -> build.cuda_m_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 11 02:17 madevent/SubProcesses/P0_uux_epemgddx/madevent -> build.cuda_m_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 11 02:17 madevent/SubProcesses/P0_uux_epemggg/madevent -> build.cuda_m_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 11 02:17 madevent/SubProcesses/P0_uux_epemguux/madevent -> build.cuda_m_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 11 02:17 madevent/SubProcesses/P0_uux_taptamgccx/madevent -> build.cuda_m_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 11 02:17 madevent/SubProcesses/P0_uux_taptamgddx/madevent -> build.cuda_m_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 11 02:17 madevent/SubProcesses/P0_uux_taptamggg/madevent -> build.cuda_m_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 11 02:17 madevent/SubProcesses/P0_uux_taptamguux/madevent -> build.cuda_m_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 11 02:17 madevent/SubProcesses/P0_uxcx_epemguxcx/madevent -> build.cuda_m_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 11 02:17 madevent/SubProcesses/P0_uxcx_taptamguxcx/madevent -> build.cuda_m_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 11 02:17 madevent/SubProcesses/P0_uxdx_epemguxdx/madevent -> build.cuda_m_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 11 02:17 madevent/SubProcesses/P0_uxdx_taptamguxdx/madevent -> build.cuda_m_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 11 02:17 madevent/SubProcesses/P0_uxux_epemguxux/madevent -> build.cuda_m_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 11 02:17 madevent/SubProcesses/P0_uxux_taptamguxux/madevent -> build.cuda_m_inl0_hrd0/madevent_cuda +Now generating 500 events with random seed 21 and granularity 1 +************************************************************ +* * +* W E L C O M E to * +* M A D G R A P H 5 _ a M C @ N L O * +* M A D E V E N T * +* * +* * * * +* * * * * * +* * * * * 5 * * * * * +* * * * * * +* * * * +* * +* VERSION 3.5.3_lo_vect * +* * +* The MadGraph5_aMC@NLO Development Team - Find us at * +* https://server06.fynu.ucl.ac.be/projects/madgraph * +* * +* Type 'help' for in-line help. * +* * +************************************************************ +INFO: load configuration from /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/Cards/me5_configuration.txt +INFO: load configuration from /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/Cards/me5_configuration.txt +Using default text editor "vi". Set another one in ./input/mg5_configuration.txt +Using default eps viewer "evince". Set another one in ./input/mg5_configuration.txt +Using default web browser "firefox". Set another one in ./input/mg5_configuration.txt +WRITE GRIDCARD /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent +__CUDACPP_DEBUG: GridPackCmd.__init__ will call self.launch +__CUDACPP_DEBUG: GridPackCmd.launch starting +__CUDACPP_DEBUG: GridPackCmd.launch (1) restore_data +generate 500 events +P0_gg_epemguux +P0_gg_epemgddx +P0_gg_taptamguux +P0_gg_taptamgddx +P0_gu_epemggu +P0_gd_epemggd +P0_gux_epemggux +P0_gdx_epemggdx +P0_gu_taptamggu +P0_gd_taptamggd +P0_gux_taptamggux +P0_gdx_taptamggdx +P0_uux_epemggg +P0_ddx_epemggg +P0_uux_taptamggg +P0_ddx_taptamggg +P0_gu_epemuuux +P0_gd_epemdddx +P0_gux_epemuuxux +P0_gdx_epemddxdx +P0_gu_epemuddx +P0_gu_epemuccx +P0_gd_epemudux +P0_gd_epemdssx +P0_gux_epemduxdx +P0_gux_epemcuxcx +P0_gdx_epemuuxdx +P0_gdx_epemsdxsx +P0_gu_taptamuuux +P0_gd_taptamdddx +P0_gux_taptamuuxux +P0_gdx_taptamddxdx +P0_gu_taptamuddx +P0_gu_taptamuccx +P0_gd_taptamudux +P0_gd_taptamdssx +P0_gux_taptamduxdx +P0_gux_taptamcuxcx +P0_gdx_taptamuuxdx +P0_gdx_taptamsdxsx +P0_uu_epemguu +P0_uux_epemguux +P0_dd_epemgdd +P0_ddx_epemgddx +P0_uxux_epemguxux +P0_dxdx_epemgdxdx +P0_ud_epemgud +P0_uc_epemguc +P0_uux_epemgddx +P0_uux_epemgccx +P0_udx_epemgudx +P0_ucx_epemgucx +P0_dc_epemgdc +P0_ds_epemgds +P0_dux_epemgdux +P0_ddx_epemguux +P0_ddx_epemgssx +P0_dsx_epemgdsx +P0_uxdx_epemguxdx +P0_uxcx_epemguxcx +P0_dxcx_epemgdxcx +P0_dxsx_epemgdxsx +P0_uu_taptamguu +P0_uux_taptamguux +P0_dd_taptamgdd +P0_ddx_taptamgddx +P0_uxux_taptamguxux +P0_dxdx_taptamgdxdx +P0_ud_taptamgud +P0_uc_taptamguc +P0_uux_taptamgddx +P0_uux_taptamgccx +P0_udx_taptamgudx +P0_ucx_taptamgucx +P0_dc_taptamgdc +P0_ds_taptamgds +P0_dux_taptamgdux +P0_ddx_taptamguux +P0_ddx_taptamgssx +P0_dsx_taptamgdsx +P0_uxdx_taptamguxdx +P0_uxcx_taptamguxcx +P0_dxcx_taptamgdxcx +P0_dxsx_taptamgdxsx +__CUDACPP_DEBUG: GridPackCmd.launch (2) refine4grid +__CUDACPP_DEBUG: GridPackCmd.refine4grid starting +Using random number seed offset = 21 +INFO: Generating 500.0 unweighted events. +__CUDACPP_DEBUG: gen_ximprove.launch starting +sum of cpu time of last step: 5h02m29s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event starting +INFO: need to improve 362 channels +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuuux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuuux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuuux/ajob1' in 9.5725 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuuux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuuux/G12/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.7475s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6706s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0171s for 16384 events => throughput is 9.61E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0598s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuuux/G20/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.7893s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7130s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0166s for 16384 events => throughput is 9.85E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0596s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuuux/G29/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.2304s + [COUNTERS] Fortran Overhead ( 0 ) : 5.1543s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0165s for 16384 events => throughput is 9.92E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0596s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuuux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemggg/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemggg/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemggg/ajob1' in 8.3551 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemggg/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemggg/G43/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.6792s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5951s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0212s for 16384 events => throughput is 7.73E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0629s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemggg/G10/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.5815s + [COUNTERS] Fortran Overhead ( 0 ) : 3.4976s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0211s for 16384 events => throughput is 7.77E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0628s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemggg/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.1485s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0646s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0211s for 16384 events => throughput is 7.77E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0628s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemggg/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/ajob1' in 42.3519 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G39/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.7443s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6601s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0211s for 16384 events => throughput is 7.77E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0631s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G12/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.9611s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8766s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0213s for 16384 events => throughput is 7.70E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0632s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G6/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.6829s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5987s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0211s for 16384 events => throughput is 7.76E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0631s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G13/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.0620s + [COUNTERS] Fortran Overhead ( 0 ) : 3.9779s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0210s for 16384 events => throughput is 7.79E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0630s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G11/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.8338s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7496s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0210s for 16384 events => throughput is 7.81E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0633s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.7450s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6606s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0212s for 16384 events => throughput is 7.74E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0632s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G69/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.5726s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4884s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0210s for 16384 events => throughput is 7.81E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0632s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G8/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.5780s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4936s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0212s for 16384 events => throughput is 7.72E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0632s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G7/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.5109s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4269s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0211s for 16384 events => throughput is 7.76E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0629s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G65/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.3795s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2956s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0209s for 16384 events => throughput is 7.85E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0631s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G1/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.9343s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8500s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0209s for 16384 events => throughput is 7.85E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0634s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.9662s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8815s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0214s for 16384 events => throughput is 7.67E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0633s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G82/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.4361s + [COUNTERS] Fortran Overhead ( 0 ) : 4.3517s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0213s for 16384 events => throughput is 7.69E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0631s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G80/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.3392s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2551s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0210s for 16384 events => throughput is 7.80E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0631s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G51/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.0882s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0040s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0210s for 16384 events => throughput is 7.79E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0632s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G70/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.5944s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5100s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0212s for 16384 events => throughput is 7.74E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0631s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G5/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.3656s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2813s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0211s for 16384 events => throughput is 7.75E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0631s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamggux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamggux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamggux/ajob1' in 27.4850 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamggux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamggux/G59/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.6023s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5019s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0372s for 16384 events => throughput is 4.40E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0633s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamggux/G12/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.0077s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9071s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0373s for 16384 events => throughput is 4.40E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0634s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamggux/G13/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.0194s + [COUNTERS] Fortran Overhead ( 0 ) : 3.9189s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0373s for 16384 events => throughput is 4.39E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0632s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamggux/G11/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.8553s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7556s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0365s for 16384 events => throughput is 4.49E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0632s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamggux/G9/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.6781s + [COUNTERS] Fortran Overhead ( 0 ) : 3.5779s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0368s for 16384 events => throughput is 4.45E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0635s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamggux/G10/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.7934s + [COUNTERS] Fortran Overhead ( 0 ) : 3.6924s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0373s for 16384 events => throughput is 4.40E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0637s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamggux/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.7450s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6451s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0366s for 16384 events => throughput is 4.48E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0633s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamggux/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.1257s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0253s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0372s for 16384 events => throughput is 4.40E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0631s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamggux/G80/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.3356s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2359s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0365s for 16384 events => throughput is 4.49E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0632s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamggux/G70/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.6351s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5346s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0373s for 16384 events => throughput is 4.39E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0632s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamggux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemsdxsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemsdxsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemsdxsx/ajob1' in 2.5681 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemsdxsx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemsdxsx/G8/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.3071s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2654s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0087s for 16384 events => throughput is 1.89E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0331s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemsdxsx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuccx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuccx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuccx/ajob1' in 2.4454 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuccx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuccx/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.1724s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1261s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0126s for 16384 events => throughput is 1.30E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0336s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuccx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuuux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuuux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuuux/ajob1' in 6.0179 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuuux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuuux/G81/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.0910s + [COUNTERS] Fortran Overhead ( 0 ) : 3.0016s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0283s for 16384 events => throughput is 5.78E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0611s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuuux/G32/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.3828s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2934s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0283s for 16384 events => throughput is 5.79E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0611s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuuux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamggg/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamggg/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamggg/ajob1' in 4.6934 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamggg/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamggg/G56/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.0115s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9101s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0380s for 16384 events => throughput is 4.31E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0633s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamggg/G39/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.1485s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0482s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0367s for 16384 events => throughput is 4.46E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0636s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamggg/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamduxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamduxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamduxdx/ajob1' in 6.3657 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamduxdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamduxdx/G45/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.6106s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5642s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0128s for 16384 events => throughput is 1.28E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0336s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamduxdx/G17/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.2256s + [COUNTERS] Fortran Overhead ( 0 ) : 4.1794s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0127s for 16384 events => throughput is 1.29E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0336s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamduxdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuccx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuccx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuccx/ajob1' in 4.8335 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuccx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuccx/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.0851s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0437s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0088s for 16384 events => throughput is 1.87E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0327s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuccx/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.2305s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1885s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0092s for 16384 events => throughput is 1.79E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0328s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuccx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuddx/ajob1' in 17.0610 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuddx/G46/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.5854s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5438s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0086s for 16384 events => throughput is 1.90E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0330s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuddx/G18/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.8121s + [COUNTERS] Fortran Overhead ( 0 ) : 3.7702s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0088s for 16384 events => throughput is 1.87E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0331s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuddx/G39/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.4518s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4103s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0087s for 16384 events => throughput is 1.89E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0329s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuddx/G45/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.5690s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5273s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0086s for 16384 events => throughput is 1.90E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0331s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuddx/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.5390s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4973s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0086s for 16384 events => throughput is 1.90E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0331s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuddx/G20/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.7144s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6727s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0087s for 16384 events => throughput is 1.89E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0330s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuddx/G47/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.7316s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6897s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0088s for 16384 events => throughput is 1.85E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0330s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuddx/G7/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.5713s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5295s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0089s for 16384 events => throughput is 1.84E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0329s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/ajob1' in 61.4863 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G14/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.3983s + [COUNTERS] Fortran Overhead ( 0 ) : 3.3140s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0212s for 16384 events => throughput is 7.72E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0631s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G74/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.4882s + [COUNTERS] Fortran Overhead ( 0 ) : 4.4037s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0214s for 16384 events => throughput is 7.66E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0631s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G12/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.9339s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8495s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0213s for 16384 events => throughput is 7.70E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0631s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G6/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.7571s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6729s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0213s for 16384 events => throughput is 7.69E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0629s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G13/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.5699s + [COUNTERS] Fortran Overhead ( 0 ) : 4.4854s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0214s for 16384 events => throughput is 7.64E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0631s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G83/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.9227s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8384s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0211s for 16384 events => throughput is 7.75E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0631s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G33/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.8563s + [COUNTERS] Fortran Overhead ( 0 ) : 4.7718s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0213s for 16384 events => throughput is 7.70E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0632s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G11/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.7272s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6427s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0211s for 16384 events => throughput is 7.75E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0633s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G9/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.7617s + [COUNTERS] Fortran Overhead ( 0 ) : 3.6776s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0211s for 16384 events => throughput is 7.75E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0630s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G10/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.8020s + [COUNTERS] Fortran Overhead ( 0 ) : 3.7176s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0212s for 16384 events => throughput is 7.71E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0632s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.7462s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6617s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0214s for 16384 events => throughput is 7.66E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0631s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G69/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.5797s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4956s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0211s for 16384 events => throughput is 7.75E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0629s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G65/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.0070s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9226s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0213s for 16384 events => throughput is 7.69E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0631s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G1/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.0000s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9158s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0212s for 16384 events => throughput is 7.72E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0630s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.9294s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8448s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0215s for 16384 events => throughput is 7.61E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0630s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G72/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.8208s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7366s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0212s for 16384 events => throughput is 7.73E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0630s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G2/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.4612s + [COUNTERS] Fortran Overhead ( 0 ) : 3.3770s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0211s for 16384 events => throughput is 7.76E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0631s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G80/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.2698s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1856s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0210s for 16384 events => throughput is 7.82E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0632s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G79/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.7308s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6466s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0211s for 16384 events => throughput is 7.75E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0631s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G70/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.6331s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5488s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0212s for 16384 events => throughput is 7.73E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0632s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G5/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.4846s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4003s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0211s for 16384 events => throughput is 7.76E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0632s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamguux/ajob1' in 2.2112 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamguux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamguux/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.9498s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9028s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0129s for 16384 events => throughput is 1.27E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0341s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamguux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/ajob1' in 62.3123 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G14/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.3733s + [COUNTERS] Fortran Overhead ( 0 ) : 4.2889s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0215s for 16384 events => throughput is 7.62E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0629s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G12/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.6557s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5714s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0211s for 16384 events => throughput is 7.75E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0632s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G6/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.4893s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4050s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0214s for 16384 events => throughput is 7.65E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0629s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G3/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.2407s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1565s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0211s for 16384 events => throughput is 7.76E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0631s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G13/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.7478s + [COUNTERS] Fortran Overhead ( 0 ) : 3.6636s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0212s for 16384 events => throughput is 7.74E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0630s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G11/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.7962s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7121s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0213s for 16384 events => throughput is 7.69E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0629s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G9/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.7935s + [COUNTERS] Fortran Overhead ( 0 ) : 3.7091s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0213s for 16384 events => throughput is 7.69E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0631s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G10/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.3397s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2554s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0212s for 16384 events => throughput is 7.71E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0631s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G40/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.1297s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0454s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0213s for 16384 events => throughput is 7.68E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0629s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G78/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.4377s + [COUNTERS] Fortran Overhead ( 0 ) : 3.3533s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0212s for 16384 events => throughput is 7.73E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0631s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G35/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.3051s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2210s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0212s for 16384 events => throughput is 7.74E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0629s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.8551s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7708s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0213s for 16384 events => throughput is 7.70E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0630s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G69/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.5880s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5034s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0214s for 16384 events => throughput is 7.65E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0632s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G7/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.4099s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3256s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0214s for 16384 events => throughput is 7.65E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0628s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G71/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.8152s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7311s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0212s for 16384 events => throughput is 7.72E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0629s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G1/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.9855s + [COUNTERS] Fortran Overhead ( 0 ) : 3.9011s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0213s for 16384 events => throughput is 7.71E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0631s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.7694s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6853s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0213s for 16384 events => throughput is 7.69E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0629s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G4/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.2456s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1612s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0213s for 16384 events => throughput is 7.69E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0631s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G80/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.6641s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5799s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0212s for 16384 events => throughput is 7.75E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0631s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G79/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.2064s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1225s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0210s for 16384 events => throughput is 7.79E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0629s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.1444s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0604s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0211s for 16384 events => throughput is 7.75E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0629s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G70/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.5532s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4692s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0212s for 16384 events => throughput is 7.72E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0628s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G5/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.6107s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5268s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0213s for 16384 events => throughput is 7.71E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0627s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxcx_epemguxcx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxcx_epemguxcx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxcx_epemguxcx/ajob1' in 5.9518 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxcx_epemguxcx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxcx_epemguxcx/G9/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.4120s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3699s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0087s for 16384 events => throughput is 1.88E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0335s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxcx_epemguxcx/G10/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.0177s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9756s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0088s for 16384 events => throughput is 1.87E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0333s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxcx_epemguxcx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/ajob1' in 41.9002 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/G46/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.9757s + [COUNTERS] Fortran Overhead ( 0 ) : 4.8917s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0213s for 16384 events => throughput is 7.68E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0627s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/G18/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.5048s + [COUNTERS] Fortran Overhead ( 0 ) : 4.4209s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0212s for 16384 events => throughput is 7.73E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0627s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/G39/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.3668s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2830s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0211s for 16384 events => throughput is 7.75E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0627s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.7633s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6791s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0215s for 16384 events => throughput is 7.62E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0626s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/G20/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.0870s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0031s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0212s for 16384 events => throughput is 7.74E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0628s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/G17/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.8113s + [COUNTERS] Fortran Overhead ( 0 ) : 3.7272s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0212s for 16384 events => throughput is 7.72E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0628s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/G35/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.0300s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9460s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0212s for 16384 events => throughput is 7.72E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0628s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/G71/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.9539s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8698s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0214s for 16384 events => throughput is 7.64E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0626s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/G27/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.5911s + [COUNTERS] Fortran Overhead ( 0 ) : 4.5070s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0215s for 16384 events => throughput is 7.63E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0626s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/G92/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.8994s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8154s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0214s for 16384 events => throughput is 7.66E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0627s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/G80/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.2885s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2043s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0214s for 16384 events => throughput is 7.67E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0628s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.0876s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0035s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0214s for 16384 events => throughput is 7.67E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0627s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/G29/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.4693s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3854s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0212s for 16384 events => throughput is 7.72E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0627s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/G48/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.3244s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2403s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0214s for 16384 events => throughput is 7.67E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0628s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_epemgdux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_epemgdux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_epemgdux/ajob1' in 13.6069 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_epemgdux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_epemgdux/G46/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.2564s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2145s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0089s for 16384 events => throughput is 1.85E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0331s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_epemgdux/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.5744s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5325s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0087s for 16384 events => throughput is 1.88E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0333s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_epemgdux/G23/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.4056s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3639s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0087s for 16384 events => throughput is 1.89E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0331s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_epemgdux/G47/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.3256s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2836s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0089s for 16384 events => throughput is 1.85E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0331s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_epemgdux/G8/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.6983s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6566s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0087s for 16384 events => throughput is 1.89E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0331s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_epemgdux/G7/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.4749s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4330s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0087s for 16384 events => throughput is 1.88E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0332s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_epemgdux/G28/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.6206s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5787s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0086s for 16384 events => throughput is 1.90E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0333s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_epemgdux/G32/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.1850s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1430s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0089s for 16384 events => throughput is 1.85E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0332s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_epemgdux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/ajob1' in 28.7912 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/G12/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.9951s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8945s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0374s for 16384 events => throughput is 4.38E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0632s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/G6/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.6911s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5902s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0377s for 16384 events => throughput is 4.35E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0633s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/G11/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.7554s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6547s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0375s for 16384 events => throughput is 4.37E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0632s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.7101s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6094s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0373s for 16384 events => throughput is 4.39E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0633s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/G69/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.5928s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4920s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0374s for 16384 events => throughput is 4.38E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0634s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/G8/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.4740s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3733s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0374s for 16384 events => throughput is 4.38E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0633s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/G84/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.8644s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7641s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0370s for 16384 events => throughput is 4.43E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0634s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.1054s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0047s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0374s for 16384 events => throughput is 4.38E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0633s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/G2/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.4777s + [COUNTERS] Fortran Overhead ( 0 ) : 3.3771s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0372s for 16384 events => throughput is 4.41E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0634s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/G80/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.2312s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1306s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0374s for 16384 events => throughput is 4.38E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0633s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/G51/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.2844s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1836s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0375s for 16384 events => throughput is 4.37E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0632s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/G70/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.6089s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5083s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0375s for 16384 events => throughput is 4.37E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0632s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/G5/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.5239s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4234s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0373s for 16384 events => throughput is 4.39E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0632s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamguux/ajob1' in 26.6151 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamguux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamguux/G18/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.3529s + [COUNTERS] Fortran Overhead ( 0 ) : 4.2521s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0371s for 16384 events => throughput is 4.41E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0636s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamguux/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.7729s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6717s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0373s for 16384 events => throughput is 4.39E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0638s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamguux/G20/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.0363s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9357s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0370s for 16384 events => throughput is 4.42E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0636s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamguux/G17/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.7292s + [COUNTERS] Fortran Overhead ( 0 ) : 3.6292s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0366s for 16384 events => throughput is 4.48E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0635s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamguux/G35/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.0059s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9051s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0373s for 16384 events => throughput is 4.40E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0635s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamguux/G27/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.1598s + [COUNTERS] Fortran Overhead ( 0 ) : 4.0596s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0372s for 16384 events => throughput is 4.41E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0630s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamguux/G34/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.1018s + [COUNTERS] Fortran Overhead ( 0 ) : 4.0014s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0372s for 16384 events => throughput is 4.40E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0631s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamguux/G80/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.3194s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2189s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0370s for 16384 events => throughput is 4.42E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0635s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamguux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_taptamgdsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_taptamgdsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_taptamgdsx/ajob1' in 1.8212 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_taptamgdsx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_taptamgdsx/G28/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.5606s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5139s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0130s for 16384 events => throughput is 1.26E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0337s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_taptamgdsx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_taptamggg/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_taptamggg/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_taptamggg/ajob1' in 9.2633 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_taptamggg/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_taptamggg/G12/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.2133s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1125s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0375s for 16384 events => throughput is 4.37E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0633s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_taptamggg/G9/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.0137s + [COUNTERS] Fortran Overhead ( 0 ) : 3.9129s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0374s for 16384 events => throughput is 4.38E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0634s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_taptamggg/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.2388s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1379s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0373s for 16384 events => throughput is 4.40E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0636s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_taptamggg/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxdx_epemguxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxdx_epemguxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxdx_epemguxdx/ajob1' in 16.0569 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxdx_epemguxdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxdx_epemguxdx/G14/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.5589s + [COUNTERS] Fortran Overhead ( 0 ) : 3.5169s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0088s for 16384 events => throughput is 1.86E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0332s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxdx_epemguxdx/G6/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.9890s + [COUNTERS] Fortran Overhead ( 0 ) : 4.9469s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0087s for 16384 events => throughput is 1.88E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0334s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxdx_epemguxdx/G22/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.7157s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6736s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0088s for 16384 events => throughput is 1.86E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0333s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxdx_epemguxdx/G8/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.6929s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6505s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0089s for 16384 events => throughput is 1.84E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0334s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxdx_epemguxdx/G27/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.8032s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7610s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0089s for 16384 events => throughput is 1.84E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0332s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxdx_epemguxdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_epemgdsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_epemgdsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_epemgdsx/ajob1' in 14.8861 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_epemgdsx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_epemgdsx/G31/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.3796s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3376s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0087s for 16384 events => throughput is 1.89E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0334s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_epemgdsx/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.4023s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3602s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0088s for 16384 events => throughput is 1.86E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0333s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_epemgdsx/G20/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.5889s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5471s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0088s for 16384 events => throughput is 1.87E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0331s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_epemgdsx/G25/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.1660s + [COUNTERS] Fortran Overhead ( 0 ) : 4.1239s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0087s for 16384 events => throughput is 1.89E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0334s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_epemgdsx/G28/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.4941s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4519s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0088s for 16384 events => throughput is 1.87E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0334s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_epemgdsx/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.2963s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2538s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0091s for 16384 events => throughput is 1.81E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0335s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_epemgdsx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamdddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamdddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamdddx/ajob1' in 4.1319 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamdddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamdddx/G17/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.8624s + [COUNTERS] Fortran Overhead ( 0 ) : 3.7730s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0285s for 16384 events => throughput is 5.75E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0609s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamdddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemggg/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemggg/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemggg/ajob1' in 14.2682 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemggg/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemggg/G56/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.1353s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0507s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0213s for 16384 events => throughput is 7.70E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0632s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemggg/G12/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.2753s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1911s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0210s for 16384 events => throughput is 7.79E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0632s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemggg/G11/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.2559s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1716s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0212s for 16384 events => throughput is 7.73E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0632s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemggg/G40/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.0395s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9554s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0211s for 16384 events => throughput is 7.75E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0629s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemggg/G8/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.8439s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7595s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0215s for 16384 events => throughput is 7.62E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0629s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemggg/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.1071s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0227s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0214s for 16384 events => throughput is 7.65E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0630s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemggg/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/ajob1' in 63.4682 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G14/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.6120s + [COUNTERS] Fortran Overhead ( 0 ) : 3.5118s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0372s for 16384 events => throughput is 4.41E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0630s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G74/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.2874s + [COUNTERS] Fortran Overhead ( 0 ) : 4.1870s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0373s for 16384 events => throughput is 4.39E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0631s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G12/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.7092s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6086s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0375s for 16384 events => throughput is 4.37E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0631s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G6/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.5712s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4708s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0371s for 16384 events => throughput is 4.41E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0633s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G13/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.5128s + [COUNTERS] Fortran Overhead ( 0 ) : 3.4124s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0372s for 16384 events => throughput is 4.40E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0631s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G66/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.1185s + [COUNTERS] Fortran Overhead ( 0 ) : 3.0180s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0372s for 16384 events => throughput is 4.40E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0632s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G11/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.9093s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8088s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0373s for 16384 events => throughput is 4.39E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0632s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G9/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.6915s + [COUNTERS] Fortran Overhead ( 0 ) : 3.5919s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0367s for 16384 events => throughput is 4.46E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0629s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G40/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.3141s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2141s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0369s for 16384 events => throughput is 4.44E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0631s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G35/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.2623s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1617s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0374s for 16384 events => throughput is 4.38E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0631s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.1482s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0475s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0374s for 16384 events => throughput is 4.38E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0633s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G69/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.5979s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4972s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0373s for 16384 events => throughput is 4.39E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0633s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G7/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.3898s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2893s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0373s for 16384 events => throughput is 4.40E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0632s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G65/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.1885s + [COUNTERS] Fortran Overhead ( 0 ) : 3.0880s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0374s for 16384 events => throughput is 4.38E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0631s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G71/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.8473s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7471s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0370s for 16384 events => throughput is 4.43E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0632s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G1/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.6552s + [COUNTERS] Fortran Overhead ( 0 ) : 3.5544s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0373s for 16384 events => throughput is 4.39E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0634s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.7710s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6705s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0373s for 16384 events => throughput is 4.40E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0632s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G2/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.8879s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7873s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0375s for 16384 events => throughput is 4.37E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0631s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G4/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.2213s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1212s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0371s for 16384 events => throughput is 4.41E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0630s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G79/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.3576s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2574s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0368s for 16384 events => throughput is 4.45E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0633s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.1115s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0110s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0373s for 16384 events => throughput is 4.40E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0632s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G70/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.5190s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4185s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0373s for 16384 events => throughput is 4.39E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0632s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G5/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.6478s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5471s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0374s for 16384 events => throughput is 4.38E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0632s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/ajob1' in 54.7155 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/G46/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.6191s + [COUNTERS] Fortran Overhead ( 0 ) : 4.5351s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0213s for 16384 events => throughput is 7.67E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0627s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/G18/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.9042s + [COUNTERS] Fortran Overhead ( 0 ) : 3.8201s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0214s for 16384 events => throughput is 7.65E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0627s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.9822s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8980s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0213s for 16384 events => throughput is 7.68E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0628s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/G33/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.9653s + [COUNTERS] Fortran Overhead ( 0 ) : 3.8813s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0212s for 16384 events => throughput is 7.74E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0628s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/G20/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.1626s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0780s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0217s for 16384 events => throughput is 7.54E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0629s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/G47/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.7932s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7089s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0213s for 16384 events => throughput is 7.68E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0629s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/G17/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.6996s + [COUNTERS] Fortran Overhead ( 0 ) : 3.6157s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0213s for 16384 events => throughput is 7.70E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0627s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/G25/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.7628s + [COUNTERS] Fortran Overhead ( 0 ) : 3.6788s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0214s for 16384 events => throughput is 7.66E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0626s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/G35/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.8686s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7846s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0214s for 16384 events => throughput is 7.65E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0626s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/G65/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.7835s + [COUNTERS] Fortran Overhead ( 0 ) : 3.6990s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0217s for 16384 events => throughput is 7.56E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0628s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/G27/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.2305s + [COUNTERS] Fortran Overhead ( 0 ) : 4.1465s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0214s for 16384 events => throughput is 7.67E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0626s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/G34/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.9244s + [COUNTERS] Fortran Overhead ( 0 ) : 3.8404s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0215s for 16384 events => throughput is 7.62E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0625s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/G80/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.4393s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3553s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0214s for 16384 events => throughput is 7.66E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0626s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/G67/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.2496s + [COUNTERS] Fortran Overhead ( 0 ) : 4.1655s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0214s for 16384 events => throughput is 7.67E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0628s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.0171s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9330s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0215s for 16384 events => throughput is 7.63E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0627s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/G48/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.0190s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9350s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0212s for 16384 events => throughput is 7.74E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0629s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_taptamgdux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_taptamgdux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_taptamgdux/ajob1' in 7.4824 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_taptamgdux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_taptamgdux/G14/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.9662s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9188s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0134s for 16384 events => throughput is 1.22E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0339s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_taptamgdux/G40/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.3194s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2719s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0134s for 16384 events => throughput is 1.22E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0340s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_taptamgdux/G44/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.4205s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3731s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0133s for 16384 events => throughput is 1.24E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0342s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_taptamgdux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemgddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemgddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemgddx/ajob1' in 8.9447 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemgddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemgddx/G56/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.5607s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4838s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0169s for 16384 events => throughput is 9.72E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0601s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemgddx/G12/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.4207s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3442s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0168s for 16384 events => throughput is 9.78E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0598s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemgddx/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.4513s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3746s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0167s for 16384 events => throughput is 9.80E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0600s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemgddx/G47/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.6599s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5827s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0170s for 16384 events => throughput is 9.64E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0601s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemgddx/G42/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.5253s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4482s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0167s for 16384 events => throughput is 9.81E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0603s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemgddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/ajob1' in 36.9536 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/G39/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.9265s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8264s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0368s for 16384 events => throughput is 4.45E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0632s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/G12/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.7411s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6416s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0363s for 16384 events => throughput is 4.51E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0632s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/G6/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.4232s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3229s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0369s for 16384 events => throughput is 4.44E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0634s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/G66/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.3076s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2071s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0375s for 16384 events => throughput is 4.37E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0631s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/G11/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.9027s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8023s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0372s for 16384 events => throughput is 4.41E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0632s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/G9/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.0770s + [COUNTERS] Fortran Overhead ( 0 ) : 3.9768s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0370s for 16384 events => throughput is 4.42E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0631s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.9607s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8611s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0364s for 16384 events => throughput is 4.51E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0633s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/G69/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.6377s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5376s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0368s for 16384 events => throughput is 4.45E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0632s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/G8/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.4927s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3924s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0373s for 16384 events => throughput is 4.40E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0630s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/G71/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.8452s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7449s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0369s for 16384 events => throughput is 4.44E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0634s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.7384s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6380s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0370s for 16384 events => throughput is 4.42E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0634s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/G72/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.7124s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6119s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0373s for 16384 events => throughput is 4.39E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0632s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/G79/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.2680s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1677s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0373s for 16384 events => throughput is 4.40E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0631s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.3585s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2591s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0362s for 16384 events => throughput is 4.52E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0632s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/G70/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.5398s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4395s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0369s for 16384 events => throughput is 4.44E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0634s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/G5/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.7436s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6430s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0372s for 16384 events => throughput is 4.40E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0634s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemdssx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemdssx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemdssx/ajob1' in 5.5457 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemdssx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemdssx/G46/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.5826s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5406s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0087s for 16384 events => throughput is 1.89E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0333s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemdssx/G17/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.4461s + [COUNTERS] Fortran Overhead ( 0 ) : 3.4042s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0089s for 16384 events => throughput is 1.85E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0330s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemdssx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uc_taptamguc/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uc_taptamguc/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uc_taptamguc/ajob1' in 4.5923 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uc_taptamguc/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uc_taptamguc/G25/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.7430s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6961s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0129s for 16384 events => throughput is 1.27E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0340s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uc_taptamguc/G37/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.3272s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2805s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0129s for 16384 events => throughput is 1.28E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0339s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uc_taptamguc/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamgddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamgddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamgddx/ajob1' in 4.4293 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamgddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamgddx/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.1572s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0678s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0284s for 16384 events => throughput is 5.76E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0610s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamgddx/G55/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.7380s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6484s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0285s for 16384 events => throughput is 5.75E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0612s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamgddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemgddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemgddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemgddx/ajob1' in 2.2442 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemgddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemgddx/G8/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.9837s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9419s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0088s for 16384 events => throughput is 1.85E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0330s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemgddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamgddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamgddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamgddx/ajob1' in 23.4418 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamgddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamgddx/G31/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.7134s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6118s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0375s for 16384 events => throughput is 4.37E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0641s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamgddx/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.7577s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6575s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0365s for 16384 events => throughput is 4.49E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0637s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamgddx/G33/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.6727s + [COUNTERS] Fortran Overhead ( 0 ) : 3.5716s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0373s for 16384 events => throughput is 4.39E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0638s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamgddx/G20/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.9103s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8099s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0372s for 16384 events => throughput is 4.41E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0631s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamgddx/G35/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.7655s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6647s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0371s for 16384 events => throughput is 4.42E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0638s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamgddx/G71/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.6295s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5285s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0373s for 16384 events => throughput is 4.40E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0637s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamgddx/G34/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.3362s + [COUNTERS] Fortran Overhead ( 0 ) : 4.2354s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0372s for 16384 events => throughput is 4.41E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0636s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamgddx/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.0493s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9488s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0372s for 16384 events => throughput is 4.40E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0632s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamgddx/G48/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.2135s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1132s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0372s for 16384 events => throughput is 4.40E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0631s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamgddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemddxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemddxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemddxdx/ajob1' in 4.2971 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemddxdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemddxdx/G11/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.0480s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9711s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0168s for 16384 events => throughput is 9.78E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0601s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemddxdx/G71/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.7135s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6364s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0170s for 16384 events => throughput is 9.66E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0601s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemddxdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dxsx_epemgdxsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dxsx_epemgdxsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dxsx_epemgdxsx/ajob1' in 1.9678 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dxsx_epemgdxsx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dxsx_epemgdxsx/G27/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.7076s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6654s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0088s for 16384 events => throughput is 1.87E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0334s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dxsx_epemgdxsx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemduxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemduxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemduxdx/ajob1' in 5.2288 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemduxdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemduxdx/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.5254s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4833s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0088s for 16384 events => throughput is 1.87E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0333s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemduxdx/G22/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.1843s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1421s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0088s for 16384 events => throughput is 1.85E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0334s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemduxdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemguux/ajob1' in 17.5073 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemguux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemguux/G20/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.5507s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4739s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0168s for 16384 events => throughput is 9.78E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0601s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemguux/G27/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.3284s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2514s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0168s for 16384 events => throughput is 9.73E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0601s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemguux/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.8993s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8224s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0168s for 16384 events => throughput is 9.74E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0601s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemguux/G55/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.6996s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6227s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0169s for 16384 events => throughput is 9.69E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0601s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemguux/G37/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.5582s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4812s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0168s for 16384 events => throughput is 9.74E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0601s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemguux/G41/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.4029s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3260s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0170s for 16384 events => throughput is 9.65E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0599s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemguux/G57/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.1378s + [COUNTERS] Fortran Overhead ( 0 ) : 3.0605s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0172s for 16384 events => throughput is 9.53E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0601s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemguux/G48/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.7671s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6900s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0171s for 16384 events => throughput is 9.60E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0600s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemguux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dd_taptamgdd/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dd_taptamgdd/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dd_taptamgdd/ajob1' in 4.6205 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dd_taptamgdd/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dd_taptamgdd/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.7424s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6528s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0285s for 16384 events => throughput is 5.75E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0611s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dd_taptamgdd/G24/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.3318s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2418s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0286s for 16384 events => throughput is 5.73E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0613s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dd_taptamgdd/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/ajob1' in 108.4830 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G14/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.9409s + [COUNTERS] Fortran Overhead ( 0 ) : 3.8565s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0214s for 16384 events => throughput is 7.67E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0631s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G38/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.2446s + [COUNTERS] Fortran Overhead ( 0 ) : 4.1605s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0211s for 16384 events => throughput is 7.77E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0630s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G39/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.9141s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8295s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0213s for 16384 events => throughput is 7.70E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0634s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G12/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.7040s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6198s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0212s for 16384 events => throughput is 7.73E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0631s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G6/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.4192s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3352s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0211s for 16384 events => throughput is 7.76E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0629s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G3/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.2833s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1986s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0214s for 16384 events => throughput is 7.66E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0633s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G13/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.7619s + [COUNTERS] Fortran Overhead ( 0 ) : 3.6774s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0214s for 16384 events => throughput is 7.66E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0631s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G66/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.4403s + [COUNTERS] Fortran Overhead ( 0 ) : 3.3561s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0211s for 16384 events => throughput is 7.76E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0631s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G33/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.4299s + [COUNTERS] Fortran Overhead ( 0 ) : 5.3454s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0214s for 16384 events => throughput is 7.64E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0631s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G11/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.7995s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7154s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0213s for 16384 events => throughput is 7.70E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0629s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G9/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 18.4349s + [COUNTERS] Fortran Overhead ( 0 ) : 18.2850s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0866s for 81920 events => throughput is 9.46E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0633s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G10/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.4270s + [COUNTERS] Fortran Overhead ( 0 ) : 3.3428s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0213s for 16384 events => throughput is 7.70E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0630s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G40/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.1368s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0526s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0211s for 16384 events => throughput is 7.75E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0630s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G35/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.0914s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0072s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0212s for 16384 events => throughput is 7.73E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0631s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.8685s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7839s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0215s for 16384 events => throughput is 7.63E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0632s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G69/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.5519s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4672s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0217s for 16384 events => throughput is 7.54E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0630s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G8/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.4256s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3414s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0211s for 16384 events => throughput is 7.75E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0630s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G7/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.5631s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4789s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0213s for 16384 events => throughput is 7.68E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0629s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G65/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.5594s + [COUNTERS] Fortran Overhead ( 0 ) : 3.4747s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0215s for 16384 events => throughput is 7.62E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0632s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G71/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.9775s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8934s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0213s for 16384 events => throughput is 7.70E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0629s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G1/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.8048s + [COUNTERS] Fortran Overhead ( 0 ) : 3.7207s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0210s for 16384 events => throughput is 7.80E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0631s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.7764s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6922s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0211s for 16384 events => throughput is 7.77E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0631s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G2/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.6279s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5440s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0210s for 16384 events => throughput is 7.81E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0629s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G4/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.4245s + [COUNTERS] Fortran Overhead ( 0 ) : 3.3399s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0218s for 16384 events => throughput is 7.53E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0629s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G80/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.6375s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5532s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0213s for 16384 events => throughput is 7.69E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0630s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G41/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.0603s + [COUNTERS] Fortran Overhead ( 0 ) : 3.9762s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0211s for 16384 events => throughput is 7.77E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0630s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G77/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.3415s + [COUNTERS] Fortran Overhead ( 0 ) : 4.2570s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0215s for 16384 events => throughput is 7.64E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0630s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G76/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.8326s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7483s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0214s for 16384 events => throughput is 7.66E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0629s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G79/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.3604s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2761s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0213s for 16384 events => throughput is 7.70E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0630s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.9951s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9107s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0215s for 16384 events => throughput is 7.61E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0629s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G70/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.5081s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4235s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0214s for 16384 events => throughput is 7.65E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0633s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G5/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.5834s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4993s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0213s for 16384 events => throughput is 7.69E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0628s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ds_taptamgds/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ds_taptamgds/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ds_taptamgds/ajob1' in 2.5441 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ds_taptamgds/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ds_taptamgds/G10/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.2825s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2361s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0129s for 16384 events => throughput is 1.27E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0335s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ds_taptamgds/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_epemgudx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_epemgudx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_epemgudx/ajob1' in 12.4620 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_epemgudx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_epemgudx/G12/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.3173s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2757s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0086s for 16384 events => throughput is 1.92E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0331s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_epemgudx/G11/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.2581s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2159s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0090s for 16384 events => throughput is 1.82E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0332s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_epemgudx/G17/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.8396s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7976s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0089s for 16384 events => throughput is 1.84E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0331s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_epemgudx/G10/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.4062s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3643s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0088s for 16384 events => throughput is 1.87E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0331s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_epemgudx/G27/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.5766s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5351s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0087s for 16384 events => throughput is 1.89E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0328s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_epemgudx/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.5105s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4685s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0088s for 16384 events => throughput is 1.86E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0332s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_epemgudx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_taptamgud/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_taptamgud/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_taptamgud/ajob1' in 18.1226 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_taptamgud/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_taptamgud/G12/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.3015s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2547s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0129s for 16384 events => throughput is 1.27E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0339s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_taptamgud/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.5721s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5251s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0129s for 16384 events => throughput is 1.27E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0342s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_taptamgud/G6/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.6120s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5652s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0128s for 16384 events => throughput is 1.28E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0340s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_taptamgud/G23/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.3161s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2690s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0129s for 16384 events => throughput is 1.27E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0342s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_taptamgud/G11/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.3432s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2960s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0130s for 16384 events => throughput is 1.26E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0342s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_taptamgud/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.6726s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6258s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0129s for 16384 events => throughput is 1.27E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0339s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_taptamgud/G22/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.8564s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8095s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0130s for 16384 events => throughput is 1.26E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0339s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_taptamgud/G7/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.7112s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6642s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0130s for 16384 events => throughput is 1.26E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0341s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_taptamgud/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.3949s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3484s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0128s for 16384 events => throughput is 1.28E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0337s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_taptamgud/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemdddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemdddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemdddx/ajob1' in 4.5630 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemdddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemdddx/G31/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.1605s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0837s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0167s for 16384 events => throughput is 9.84E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0601s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemdddx/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.8550s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7783s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0167s for 16384 events => throughput is 9.80E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0600s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemdddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamdssx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamdssx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamdssx/ajob1' in 3.8277 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamdssx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamdssx/G45/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.5723s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5256s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0130s for 16384 events => throughput is 1.26E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0337s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamdssx/G20/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.7364s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6899s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0130s for 16384 events => throughput is 1.26E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0335s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamdssx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuddx/ajob1' in 5.4996 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuddx/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.5217s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4751s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0130s for 16384 events => throughput is 1.26E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0336s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuddx/G20/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.6050s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5587s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0128s for 16384 events => throughput is 1.28E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0335s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuddx/G47/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.5946s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5484s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0128s for 16384 events => throughput is 1.28E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0334s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemuuxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemuuxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemuuxdx/ajob1' in 5.8999 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemuuxdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemuuxdx/G7/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.7804s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7385s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0088s for 16384 events => throughput is 1.86E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0331s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemuuxdx/G41/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.6000s + [COUNTERS] Fortran Overhead ( 0 ) : 3.5576s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0091s for 16384 events => throughput is 1.81E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0333s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemuuxdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/ajob1' in 42.0210 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G14/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.5139s + [COUNTERS] Fortran Overhead ( 0 ) : 3.4721s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0087s for 16384 events => throughput is 1.88E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0332s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G18/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.0233s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9811s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0089s for 16384 events => throughput is 1.84E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0333s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G39/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.2886s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2463s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0092s for 16384 events => throughput is 1.78E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0330s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G30/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.6240s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5819s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0090s for 16384 events => throughput is 1.82E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0331s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G12/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.2863s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2443s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0087s for 16384 events => throughput is 1.87E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0333s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.6680s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6261s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0089s for 16384 events => throughput is 1.84E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0330s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G6/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.8491s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8072s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0088s for 16384 events => throughput is 1.86E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0332s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G3/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.3158s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2735s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0090s for 16384 events => throughput is 1.83E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0334s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G23/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.3359s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2939s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0089s for 16384 events => throughput is 1.83E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0330s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G24/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.3297s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2879s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0088s for 16384 events => throughput is 1.87E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0330s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G13/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.7528s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7110s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0088s for 16384 events => throughput is 1.85E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0330s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G20/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.3597s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3178s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0088s for 16384 events => throughput is 1.85E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0330s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G25/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.9032s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8612s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0089s for 16384 events => throughput is 1.85E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0331s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.5968s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5549s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0090s for 16384 events => throughput is 1.83E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0329s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G8/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.4033s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3613s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0088s for 16384 events => throughput is 1.86E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0332s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G7/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.4861s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4442s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0087s for 16384 events => throughput is 1.88E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0332s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G4/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.1546s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1125s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0089s for 16384 events => throughput is 1.83E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0332s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G41/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.8189s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7771s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0087s for 16384 events => throughput is 1.87E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0330s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G32/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.3652s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3234s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0087s for 16384 events => throughput is 1.88E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0331s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ucx_epemgucx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ucx_epemgucx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ucx_epemgucx/ajob1' in 5.2158 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ucx_epemgucx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ucx_epemgucx/G3/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.1222s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0803s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0088s for 16384 events => throughput is 1.85E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0330s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ucx_epemgucx/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.7343s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6927s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0087s for 16384 events => throughput is 1.88E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0329s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ucx_epemgucx/G7/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.5835s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5419s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0088s for 16384 events => throughput is 1.87E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0328s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ucx_epemgucx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_taptamgudx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_taptamgudx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_taptamgudx/ajob1' in 10.2633 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_taptamgudx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_taptamgudx/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.4342s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3871s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0133s for 16384 events => throughput is 1.23E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0337s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_taptamgudx/G11/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.3496s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3026s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0135s for 16384 events => throughput is 1.21E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0336s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_taptamgudx/G17/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.8716s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8245s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0135s for 16384 events => throughput is 1.22E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0336s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_taptamgudx/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.7094s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6623s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0133s for 16384 events => throughput is 1.23E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0338s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_taptamgudx/G27/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.5706s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5239s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0134s for 16384 events => throughput is 1.22E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0333s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_taptamgudx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamuuxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamuuxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamuuxdx/ajob1' in 3.4908 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamuuxdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamuuxdx/G6/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.2287s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1814s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0137s for 16384 events => throughput is 1.20E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0336s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamuuxdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_taptamguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_taptamguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_taptamguux/ajob1' in 2.0198 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_taptamguux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_taptamguux/G48/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.7441s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6546s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0285s for 16384 events => throughput is 5.75E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0611s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_taptamguux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemudux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemudux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemudux/ajob1' in 10.2011 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemudux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemudux/G43/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.0432s + [COUNTERS] Fortran Overhead ( 0 ) : 4.0012s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0088s for 16384 events => throughput is 1.87E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0333s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemudux/G8/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.6169s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5749s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0088s for 16384 events => throughput is 1.86E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0332s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemudux/G5/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.7641s + [COUNTERS] Fortran Overhead ( 0 ) : 3.7222s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0090s for 16384 events => throughput is 1.83E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0330s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemudux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamsdxsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamsdxsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamsdxsx/ajob1' in 2.0178 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamsdxsx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamsdxsx/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.7573s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7107s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0129s for 16384 events => throughput is 1.27E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0336s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamsdxsx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_epemguu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_epemguu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_epemguu/ajob1' in 21.9513 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_epemguu/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_epemguu/G30/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.5060s + [COUNTERS] Fortran Overhead ( 0 ) : 3.4290s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0170s for 16384 events => throughput is 9.65E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0600s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_epemguu/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.4670s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3900s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0167s for 16384 events => throughput is 9.78E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0603s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_epemguu/G9/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.8370s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7603s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0167s for 16384 events => throughput is 9.79E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0600s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_epemguu/G49/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.4207s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3438s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0169s for 16384 events => throughput is 9.71E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0600s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_epemguu/G10/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.7507s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6737s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0171s for 16384 events => throughput is 9.60E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0600s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_epemguu/G81/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.5502s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4732s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0170s for 16384 events => throughput is 9.66E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0601s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_epemguu/G62/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.3108s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2339s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0168s for 16384 events => throughput is 9.76E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0601s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_epemguu/G51/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.3747s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2977s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0170s for 16384 events => throughput is 9.65E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0600s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_epemguu/G28/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.3303s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2535s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0169s for 16384 events => throughput is 9.72E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0600s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_epemguu/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uc_epemguc/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uc_epemguc/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uc_epemguc/ajob1' in 3.3029 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uc_epemguc/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uc_epemguc/G12/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.2470s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2048s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0088s for 16384 events => throughput is 1.87E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0335s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uc_epemguc/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.5306s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4886s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0088s for 16384 events => throughput is 1.87E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0332s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uc_epemguc/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ds_epemgds/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ds_epemgds/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ds_epemgds/ajob1' in 1.8057 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ds_epemgds/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ds_epemgds/G27/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.5441s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5017s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0088s for 16384 events => throughput is 1.86E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0336s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ds_epemgds/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_taptamguu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_taptamguu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_taptamguu/ajob1' in 5.9200 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_taptamguu/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_taptamguu/G52/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.4673s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3771s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0290s for 16384 events => throughput is 5.66E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0612s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_taptamguu/G49/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.1595s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0698s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0284s for 16384 events => throughput is 5.76E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0612s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_taptamguu/G32/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.4822s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3925s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0286s for 16384 events => throughput is 5.74E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0612s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_taptamguu/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished in 950.2743 seconds +__CUDACPP_DEBUG: gen_ximprove.launch finished in 950.7177 seconds +__CUDACPP_DEBUG: GridPackCmd.refine4grid finished in 950.7184 seconds +__CUDACPP_DEBUG: GridPackCmd.launch (3a) combine_events +combine_events +__CUDACPP_DEBUG: GridPackCmd.launch (3b) store_events +store_events + === Results Summary for run: GridRun_21 tag: tag_1 === + + Cross-section : 0 +- 0 pb + Nb of events : 500 + +INFO: No version of lhapdf. Can not run systematics computation +decay_events -from_cards +__CUDACPP_DEBUG: GridPackCmd.launch finished in 969.4855 seconds +__CUDACPP_DEBUG: GridPackCmd.__init__ back from self.launch +DONE +write ./events.lhe.gz +END: Wed Sep 11 02:33:37 AM CEST 2024 +ELAPSED: 972 seconds diff --git a/epochX/cudacpp/tlau/fromgridpacks/pp_dy3j.mad/fortran/output.txt b/epochX/cudacpp/tlau/fromgridpacks/pp_dy3j.mad/fortran/output.txt new file mode 100644 index 0000000000..6b82127f97 --- /dev/null +++ b/epochX/cudacpp/tlau/fromgridpacks/pp_dy3j.mad/fortran/output.txt @@ -0,0 +1,1975 @@ +START: Wed Sep 11 01:44:17 AM CEST 2024 +On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU (nproc=4)] [GPU: 1x Tesla V100S-PCIE-32GB]: +CUDACPP_RUNTIME_DISABLEFPE is set +lrwxrwxrwx. 1 avalassi zg 16 Sep 11 01:44 madevent/SubProcesses/P0_dc_epemgdc/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 11 01:44 madevent/SubProcesses/P0_dc_taptamgdc/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 11 01:44 madevent/SubProcesses/P0_dd_epemgdd/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 11 01:44 madevent/SubProcesses/P0_dd_taptamgdd/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 11 01:44 madevent/SubProcesses/P0_ddx_epemgddx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 11 01:44 madevent/SubProcesses/P0_ddx_epemggg/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 11 01:44 madevent/SubProcesses/P0_ddx_epemgssx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 11 01:44 madevent/SubProcesses/P0_ddx_epemguux/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 11 01:44 madevent/SubProcesses/P0_ddx_taptamgddx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 11 01:44 madevent/SubProcesses/P0_ddx_taptamggg/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 11 01:44 madevent/SubProcesses/P0_ddx_taptamgssx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 11 01:44 madevent/SubProcesses/P0_ddx_taptamguux/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 11 01:44 madevent/SubProcesses/P0_ds_epemgds/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 11 01:44 madevent/SubProcesses/P0_ds_taptamgds/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 11 01:44 madevent/SubProcesses/P0_dsx_epemgdsx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 11 01:44 madevent/SubProcesses/P0_dsx_taptamgdsx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 11 01:44 madevent/SubProcesses/P0_dux_epemgdux/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 11 01:44 madevent/SubProcesses/P0_dux_taptamgdux/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 11 01:44 madevent/SubProcesses/P0_dxcx_epemgdxcx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 11 01:44 madevent/SubProcesses/P0_dxcx_taptamgdxcx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 11 01:44 madevent/SubProcesses/P0_dxdx_epemgdxdx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 11 01:44 madevent/SubProcesses/P0_dxdx_taptamgdxdx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 11 01:44 madevent/SubProcesses/P0_dxsx_epemgdxsx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 11 01:44 madevent/SubProcesses/P0_dxsx_taptamgdxsx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 11 01:44 madevent/SubProcesses/P0_gd_epemdddx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 11 01:44 madevent/SubProcesses/P0_gd_epemdssx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 11 01:44 madevent/SubProcesses/P0_gd_epemggd/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 11 01:44 madevent/SubProcesses/P0_gd_epemudux/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 11 01:44 madevent/SubProcesses/P0_gd_taptamdddx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 11 01:44 madevent/SubProcesses/P0_gd_taptamdssx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 11 01:44 madevent/SubProcesses/P0_gd_taptamggd/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 11 01:44 madevent/SubProcesses/P0_gd_taptamudux/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 11 01:44 madevent/SubProcesses/P0_gdx_epemddxdx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 11 01:44 madevent/SubProcesses/P0_gdx_epemggdx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 11 01:44 madevent/SubProcesses/P0_gdx_epemsdxsx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 11 01:44 madevent/SubProcesses/P0_gdx_epemuuxdx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 11 01:44 madevent/SubProcesses/P0_gdx_taptamddxdx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 11 01:44 madevent/SubProcesses/P0_gdx_taptamggdx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 11 01:44 madevent/SubProcesses/P0_gdx_taptamsdxsx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 11 01:44 madevent/SubProcesses/P0_gdx_taptamuuxdx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 11 01:44 madevent/SubProcesses/P0_gg_epemgddx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 11 01:44 madevent/SubProcesses/P0_gg_epemguux/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 11 01:44 madevent/SubProcesses/P0_gg_taptamgddx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 11 01:44 madevent/SubProcesses/P0_gg_taptamguux/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 11 01:44 madevent/SubProcesses/P0_gu_epemggu/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 11 01:44 madevent/SubProcesses/P0_gu_epemuccx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 11 01:44 madevent/SubProcesses/P0_gu_epemuddx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 11 01:44 madevent/SubProcesses/P0_gu_epemuuux/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 11 01:44 madevent/SubProcesses/P0_gu_taptamggu/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 11 01:44 madevent/SubProcesses/P0_gu_taptamuccx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 11 01:44 madevent/SubProcesses/P0_gu_taptamuddx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 11 01:44 madevent/SubProcesses/P0_gu_taptamuuux/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 11 01:44 madevent/SubProcesses/P0_gux_epemcuxcx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 11 01:44 madevent/SubProcesses/P0_gux_epemduxdx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 11 01:44 madevent/SubProcesses/P0_gux_epemggux/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 11 01:44 madevent/SubProcesses/P0_gux_epemuuxux/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 11 01:44 madevent/SubProcesses/P0_gux_taptamcuxcx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 11 01:44 madevent/SubProcesses/P0_gux_taptamduxdx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 11 01:44 madevent/SubProcesses/P0_gux_taptamggux/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 11 01:44 madevent/SubProcesses/P0_gux_taptamuuxux/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 11 01:44 madevent/SubProcesses/P0_uc_epemguc/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 11 01:44 madevent/SubProcesses/P0_uc_taptamguc/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 11 01:44 madevent/SubProcesses/P0_ucx_epemgucx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 11 01:44 madevent/SubProcesses/P0_ucx_taptamgucx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 11 01:44 madevent/SubProcesses/P0_ud_epemgud/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 11 01:44 madevent/SubProcesses/P0_ud_taptamgud/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 11 01:44 madevent/SubProcesses/P0_udx_epemgudx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 11 01:44 madevent/SubProcesses/P0_udx_taptamgudx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 11 01:44 madevent/SubProcesses/P0_uu_epemguu/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 11 01:44 madevent/SubProcesses/P0_uu_taptamguu/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 11 01:44 madevent/SubProcesses/P0_uux_epemgccx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 11 01:44 madevent/SubProcesses/P0_uux_epemgddx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 11 01:44 madevent/SubProcesses/P0_uux_epemggg/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 11 01:44 madevent/SubProcesses/P0_uux_epemguux/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 11 01:44 madevent/SubProcesses/P0_uux_taptamgccx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 11 01:44 madevent/SubProcesses/P0_uux_taptamgddx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 11 01:44 madevent/SubProcesses/P0_uux_taptamggg/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 11 01:44 madevent/SubProcesses/P0_uux_taptamguux/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 11 01:44 madevent/SubProcesses/P0_uxcx_epemguxcx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 11 01:44 madevent/SubProcesses/P0_uxcx_taptamguxcx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 11 01:44 madevent/SubProcesses/P0_uxdx_epemguxdx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 11 01:44 madevent/SubProcesses/P0_uxdx_taptamguxdx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 11 01:44 madevent/SubProcesses/P0_uxux_epemguxux/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 11 01:44 madevent/SubProcesses/P0_uxux_taptamguxux/madevent -> madevent_fortran +Now generating 500 events with random seed 21 and granularity 1 +************************************************************ +* * +* W E L C O M E to * +* M A D G R A P H 5 _ a M C @ N L O * +* M A D E V E N T * +* * +* * * * +* * * * * * +* * * * * 5 * * * * * +* * * * * * +* * * * +* * +* VERSION 3.5.3_lo_vect * +* * +* The MadGraph5_aMC@NLO Development Team - Find us at * +* https://server06.fynu.ucl.ac.be/projects/madgraph * +* * +* Type 'help' for in-line help. * +* * +************************************************************ +INFO: load configuration from /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/Cards/me5_configuration.txt +INFO: load configuration from /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/Cards/me5_configuration.txt +Using default text editor "vi". Set another one in ./input/mg5_configuration.txt +Using default eps viewer "evince". Set another one in ./input/mg5_configuration.txt +Using default web browser "firefox". Set another one in ./input/mg5_configuration.txt +WRITE GRIDCARD /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent +__CUDACPP_DEBUG: GridPackCmd.__init__ will call self.launch +__CUDACPP_DEBUG: GridPackCmd.launch starting +__CUDACPP_DEBUG: GridPackCmd.launch (1) restore_data +generate 500 events +P0_gg_epemguux +P0_gg_epemgddx +P0_gg_taptamguux +P0_gg_taptamgddx +P0_gu_epemggu +P0_gd_epemggd +P0_gux_epemggux +P0_gdx_epemggdx +P0_gu_taptamggu +P0_gd_taptamggd +P0_gux_taptamggux +P0_gdx_taptamggdx +P0_uux_epemggg +P0_ddx_epemggg +P0_uux_taptamggg +P0_ddx_taptamggg +P0_gu_epemuuux +P0_gd_epemdddx +P0_gux_epemuuxux +P0_gdx_epemddxdx +P0_gu_epemuddx +P0_gu_epemuccx +P0_gd_epemudux +P0_gd_epemdssx +P0_gux_epemduxdx +P0_gux_epemcuxcx +P0_gdx_epemuuxdx +P0_gdx_epemsdxsx +P0_gu_taptamuuux +P0_gd_taptamdddx +P0_gux_taptamuuxux +P0_gdx_taptamddxdx +P0_gu_taptamuddx +P0_gu_taptamuccx +P0_gd_taptamudux +P0_gd_taptamdssx +P0_gux_taptamduxdx +P0_gux_taptamcuxcx +P0_gdx_taptamuuxdx +P0_gdx_taptamsdxsx +P0_uu_epemguu +P0_uux_epemguux +P0_dd_epemgdd +P0_ddx_epemgddx +P0_uxux_epemguxux +P0_dxdx_epemgdxdx +P0_ud_epemgud +P0_uc_epemguc +P0_uux_epemgddx +P0_uux_epemgccx +P0_udx_epemgudx +P0_ucx_epemgucx +P0_dc_epemgdc +P0_ds_epemgds +P0_dux_epemgdux +P0_ddx_epemguux +P0_ddx_epemgssx +P0_dsx_epemgdsx +P0_uxdx_epemguxdx +P0_uxcx_epemguxcx +P0_dxcx_epemgdxcx +P0_dxsx_epemgdxsx +P0_uu_taptamguu +P0_uux_taptamguux +P0_dd_taptamgdd +P0_ddx_taptamgddx +P0_uxux_taptamguxux +P0_dxdx_taptamgdxdx +P0_ud_taptamgud +P0_uc_taptamguc +P0_uux_taptamgddx +P0_uux_taptamgccx +P0_udx_taptamgudx +P0_ucx_taptamgucx +P0_dc_taptamgdc +P0_ds_taptamgds +P0_dux_taptamgdux +P0_ddx_taptamguux +P0_ddx_taptamgssx +P0_dsx_taptamgdsx +P0_uxdx_taptamguxdx +P0_uxcx_taptamguxcx +P0_dxcx_taptamgdxcx +P0_dxsx_taptamgdxsx +__CUDACPP_DEBUG: GridPackCmd.launch (2) refine4grid +__CUDACPP_DEBUG: GridPackCmd.refine4grid starting +Using random number seed offset = 21 +INFO: Generating 500.0 unweighted events. +__CUDACPP_DEBUG: gen_ximprove.launch starting +sum of cpu time of last step: 5h02m29s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event starting +INFO: need to improve 362 channels +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuuux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuuux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuuux/ajob1' in 13.9901 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuuux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuuux/G12/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.4279s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2146s + [COUNTERS] Fortran MEs ( 1 ) : 2.2133s for 16384 events => throughput is 7.40E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuuux/G20/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.5313s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3253s + [COUNTERS] Fortran MEs ( 1 ) : 2.2060s for 16384 events => throughput is 7.43E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuuux/G29/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.8797s + [COUNTERS] Fortran Overhead ( 0 ) : 4.6662s + [COUNTERS] Fortran MEs ( 1 ) : 2.2135s for 16384 events => throughput is 7.40E+03 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuuux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemggg/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemggg/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemggg/ajob1' in 16.1904 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemggg/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemggg/G43/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.4857s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1334s + [COUNTERS] Fortran MEs ( 1 ) : 3.3523s for 16384 events => throughput is 4.89E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemggg/G10/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.4857s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1559s + [COUNTERS] Fortran MEs ( 1 ) : 3.3298s for 16384 events => throughput is 4.92E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemggg/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.1632s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6904s + [COUNTERS] Fortran MEs ( 1 ) : 3.4728s for 16384 events => throughput is 4.72E+03 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemggg/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/ajob1' in 86.0817 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G39/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.5942s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2772s + [COUNTERS] Fortran MEs ( 1 ) : 3.3170s for 16384 events => throughput is 4.94E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G12/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.7703s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4628s + [COUNTERS] Fortran MEs ( 1 ) : 3.3076s for 16384 events => throughput is 4.95E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G6/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.5136s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1993s + [COUNTERS] Fortran MEs ( 1 ) : 3.3144s for 16384 events => throughput is 4.94E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G13/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.8589s + [COUNTERS] Fortran Overhead ( 0 ) : 3.5511s + [COUNTERS] Fortran MEs ( 1 ) : 3.3079s for 16384 events => throughput is 4.95E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G11/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.6922s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3716s + [COUNTERS] Fortran MEs ( 1 ) : 3.3206s for 16384 events => throughput is 4.93E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.5802s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2565s + [COUNTERS] Fortran MEs ( 1 ) : 3.3237s for 16384 events => throughput is 4.93E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G69/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.3968s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0884s + [COUNTERS] Fortran MEs ( 1 ) : 3.3084s for 16384 events => throughput is 4.95E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G8/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.4049s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0888s + [COUNTERS] Fortran MEs ( 1 ) : 3.3162s for 16384 events => throughput is 4.94E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G7/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.3314s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0150s + [COUNTERS] Fortran MEs ( 1 ) : 3.3164s for 16384 events => throughput is 4.94E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G65/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.2294s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9095s + [COUNTERS] Fortran MEs ( 1 ) : 3.3199s for 16384 events => throughput is 4.94E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G1/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.7998s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4537s + [COUNTERS] Fortran MEs ( 1 ) : 3.3461s for 16384 events => throughput is 4.90E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.8155s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4808s + [COUNTERS] Fortran MEs ( 1 ) : 3.3347s for 16384 events => throughput is 4.91E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G82/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.0842s + [COUNTERS] Fortran Overhead ( 0 ) : 3.7742s + [COUNTERS] Fortran MEs ( 1 ) : 3.3100s for 16384 events => throughput is 4.95E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G80/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.1669s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8644s + [COUNTERS] Fortran MEs ( 1 ) : 3.3025s for 16384 events => throughput is 4.96E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G51/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.9186s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6062s + [COUNTERS] Fortran MEs ( 1 ) : 3.3124s for 16384 events => throughput is 4.95E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G70/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.4374s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0893s + [COUNTERS] Fortran MEs ( 1 ) : 3.3481s for 16384 events => throughput is 4.89E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/G5/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.1809s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8650s + [COUNTERS] Fortran MEs ( 1 ) : 3.3159s for 16384 events => throughput is 4.94E+03 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemggdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamggux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamggux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamggux/ajob1' in 86.3984 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamggux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamggux/G59/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.7298s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0976s + [COUNTERS] Fortran MEs ( 1 ) : 6.6323s for 16384 events => throughput is 2.47E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamggux/G12/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.1555s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5121s + [COUNTERS] Fortran MEs ( 1 ) : 6.6433s for 16384 events => throughput is 2.47E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamggux/G13/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 10.1820s + [COUNTERS] Fortran Overhead ( 0 ) : 3.5391s + [COUNTERS] Fortran MEs ( 1 ) : 6.6429s for 16384 events => throughput is 2.47E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamggux/G11/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.9717s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3467s + [COUNTERS] Fortran MEs ( 1 ) : 6.6250s for 16384 events => throughput is 2.47E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamggux/G9/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 9.8415s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1940s + [COUNTERS] Fortran MEs ( 1 ) : 6.6476s for 16384 events => throughput is 2.46E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamggux/G10/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 9.9274s + [COUNTERS] Fortran Overhead ( 0 ) : 3.3120s + [COUNTERS] Fortran MEs ( 1 ) : 6.6154s for 16384 events => throughput is 2.48E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamggux/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.8711s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2245s + [COUNTERS] Fortran MEs ( 1 ) : 6.6466s for 16384 events => throughput is 2.47E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamggux/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.2438s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6049s + [COUNTERS] Fortran MEs ( 1 ) : 6.6388s for 16384 events => throughput is 2.47E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamggux/G80/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.4901s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8431s + [COUNTERS] Fortran MEs ( 1 ) : 6.6470s for 16384 events => throughput is 2.46E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamggux/G70/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.8040s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1311s + [COUNTERS] Fortran MEs ( 1 ) : 6.6728s for 16384 events => throughput is 2.46E+03 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamggux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemsdxsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemsdxsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemsdxsx/ajob1' in 2.8765 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemsdxsx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemsdxsx/G8/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.8550s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8614s + [COUNTERS] Fortran MEs ( 1 ) : 0.9936s for 16384 events => throughput is 1.65E+04 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemsdxsx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuccx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuccx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuccx/ajob1' in 3.7369 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuccx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuccx/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.7155s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7337s + [COUNTERS] Fortran MEs ( 1 ) : 1.9819s for 16384 events => throughput is 8.27E+03 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuccx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuuux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuuux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuuux/ajob1' in 13.3629 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuuux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuuux/G81/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.0052s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5898s + [COUNTERS] Fortran MEs ( 1 ) : 4.4154s for 16384 events => throughput is 3.71E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuuux/G32/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.3185s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9005s + [COUNTERS] Fortran MEs ( 1 ) : 4.4180s for 16384 events => throughput is 3.71E+03 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuuux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamggg/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamggg/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamggg/ajob1' in 16.4267 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamggg/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamggg/G56/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.1408s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5452s + [COUNTERS] Fortran MEs ( 1 ) : 6.5956s for 16384 events => throughput is 2.48E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamggg/G39/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.2457s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6534s + [COUNTERS] Fortran MEs ( 1 ) : 6.5924s for 16384 events => throughput is 2.49E+03 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamggg/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamduxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamduxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamduxdx/ajob1' in 8.8804 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamduxdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamduxdx/G45/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.1230s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1536s + [COUNTERS] Fortran MEs ( 1 ) : 1.9695s for 16384 events => throughput is 8.32E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamduxdx/G17/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.7175s + [COUNTERS] Fortran Overhead ( 0 ) : 3.7454s + [COUNTERS] Fortran MEs ( 1 ) : 1.9721s for 16384 events => throughput is 8.31E+03 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_taptamduxdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuccx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuccx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuccx/ajob1' in 5.2993 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuccx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuccx/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.6051s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6240s + [COUNTERS] Fortran MEs ( 1 ) : 0.9812s for 16384 events => throughput is 1.67E+04 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuccx/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.6545s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6674s + [COUNTERS] Fortran MEs ( 1 ) : 0.9870s for 16384 events => throughput is 1.66E+04 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuccx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuddx/ajob1' in 19.5385 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuddx/G46/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.1245s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1324s + [COUNTERS] Fortran MEs ( 1 ) : 0.9921s for 16384 events => throughput is 1.65E+04 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuddx/G18/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.3557s + [COUNTERS] Fortran Overhead ( 0 ) : 3.3642s + [COUNTERS] Fortran MEs ( 1 ) : 0.9915s for 16384 events => throughput is 1.65E+04 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuddx/G39/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.0426s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0441s + [COUNTERS] Fortran MEs ( 1 ) : 0.9985s for 16384 events => throughput is 1.64E+04 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuddx/G45/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.1377s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1410s + [COUNTERS] Fortran MEs ( 1 ) : 0.9967s for 16384 events => throughput is 1.64E+04 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuddx/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.1233s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1047s + [COUNTERS] Fortran MEs ( 1 ) : 1.0186s for 16384 events => throughput is 1.61E+04 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuddx/G20/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.2465s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2597s + [COUNTERS] Fortran MEs ( 1 ) : 0.9869s for 16384 events => throughput is 1.66E+04 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuddx/G47/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.2655s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2834s + [COUNTERS] Fortran MEs ( 1 ) : 0.9821s for 16384 events => throughput is 1.67E+04 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuddx/G7/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.0961s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1126s + [COUNTERS] Fortran MEs ( 1 ) : 0.9834s for 16384 events => throughput is 1.67E+04 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemuddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/ajob1' in 116.4891 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G14/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.4606s + [COUNTERS] Fortran Overhead ( 0 ) : 3.0267s + [COUNTERS] Fortran MEs ( 1 ) : 3.4340s for 16384 events => throughput is 4.77E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G74/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.2927s + [COUNTERS] Fortran Overhead ( 0 ) : 3.9662s + [COUNTERS] Fortran MEs ( 1 ) : 3.3266s for 16384 events => throughput is 4.93E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G12/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.7433s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4230s + [COUNTERS] Fortran MEs ( 1 ) : 3.3202s for 16384 events => throughput is 4.93E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G6/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.6504s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2694s + [COUNTERS] Fortran MEs ( 1 ) : 3.3811s for 16384 events => throughput is 4.85E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G13/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.4590s + [COUNTERS] Fortran Overhead ( 0 ) : 4.1399s + [COUNTERS] Fortran MEs ( 1 ) : 3.3192s for 16384 events => throughput is 4.94E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G83/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.7467s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4280s + [COUNTERS] Fortran MEs ( 1 ) : 3.3187s for 16384 events => throughput is 4.94E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G33/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.7044s + [COUNTERS] Fortran Overhead ( 0 ) : 4.3854s + [COUNTERS] Fortran MEs ( 1 ) : 3.3189s for 16384 events => throughput is 4.94E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G11/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.5423s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2263s + [COUNTERS] Fortran MEs ( 1 ) : 3.3159s for 16384 events => throughput is 4.94E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G9/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.6254s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2967s + [COUNTERS] Fortran MEs ( 1 ) : 3.3287s for 16384 events => throughput is 4.92E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G10/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.6611s + [COUNTERS] Fortran Overhead ( 0 ) : 3.3415s + [COUNTERS] Fortran MEs ( 1 ) : 3.3196s for 16384 events => throughput is 4.94E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.6998s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2734s + [COUNTERS] Fortran MEs ( 1 ) : 3.4264s for 16384 events => throughput is 4.78E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G69/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.3956s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0780s + [COUNTERS] Fortran MEs ( 1 ) : 3.3176s for 16384 events => throughput is 4.94E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G65/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.8993s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5174s + [COUNTERS] Fortran MEs ( 1 ) : 3.3819s for 16384 events => throughput is 4.84E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G1/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.8375s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5186s + [COUNTERS] Fortran MEs ( 1 ) : 3.3189s for 16384 events => throughput is 4.94E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.7802s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4551s + [COUNTERS] Fortran MEs ( 1 ) : 3.3250s for 16384 events => throughput is 4.93E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G72/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.6690s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3254s + [COUNTERS] Fortran MEs ( 1 ) : 3.3437s for 16384 events => throughput is 4.90E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G2/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.3214s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9999s + [COUNTERS] Fortran MEs ( 1 ) : 3.3215s for 16384 events => throughput is 4.93E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G80/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.1837s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8049s + [COUNTERS] Fortran MEs ( 1 ) : 3.3788s for 16384 events => throughput is 4.85E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G79/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.5476s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2379s + [COUNTERS] Fortran MEs ( 1 ) : 3.3097s for 16384 events => throughput is 4.95E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G70/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.5727s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1833s + [COUNTERS] Fortran MEs ( 1 ) : 3.3894s for 16384 events => throughput is 4.83E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/G5/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.3105s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9708s + [COUNTERS] Fortran MEs ( 1 ) : 3.3397s for 16384 events => throughput is 4.91E+03 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemggux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamguux/ajob1' in 3.5056 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamguux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamguux/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.4837s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5115s + [COUNTERS] Fortran MEs ( 1 ) : 1.9722s for 16384 events => throughput is 8.31E+03 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamguux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/ajob1' in 121.2369 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G14/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.1870s + [COUNTERS] Fortran Overhead ( 0 ) : 3.8792s + [COUNTERS] Fortran MEs ( 1 ) : 3.3078s for 16384 events => throughput is 4.95E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G12/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.4640s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1677s + [COUNTERS] Fortran MEs ( 1 ) : 3.2963s for 16384 events => throughput is 4.97E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G6/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.2914s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0047s + [COUNTERS] Fortran MEs ( 1 ) : 3.2867s for 16384 events => throughput is 4.98E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G3/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.0440s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7476s + [COUNTERS] Fortran MEs ( 1 ) : 3.2964s for 16384 events => throughput is 4.97E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G13/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.5420s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2562s + [COUNTERS] Fortran MEs ( 1 ) : 3.2858s for 16384 events => throughput is 4.99E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G11/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.6333s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3361s + [COUNTERS] Fortran MEs ( 1 ) : 3.2971s for 16384 events => throughput is 4.97E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G9/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.5919s + [COUNTERS] Fortran Overhead ( 0 ) : 3.3057s + [COUNTERS] Fortran MEs ( 1 ) : 3.2862s for 16384 events => throughput is 4.99E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G10/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.1030s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8204s + [COUNTERS] Fortran MEs ( 1 ) : 3.2826s for 16384 events => throughput is 4.99E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G40/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.9289s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6208s + [COUNTERS] Fortran MEs ( 1 ) : 3.3081s for 16384 events => throughput is 4.95E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G78/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.2661s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9616s + [COUNTERS] Fortran MEs ( 1 ) : 3.3045s for 16384 events => throughput is 4.96E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G35/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.1166s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8191s + [COUNTERS] Fortran MEs ( 1 ) : 3.2975s for 16384 events => throughput is 4.97E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.7279s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3780s + [COUNTERS] Fortran MEs ( 1 ) : 3.3499s for 16384 events => throughput is 4.89E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G69/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.3845s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0895s + [COUNTERS] Fortran MEs ( 1 ) : 3.2950s for 16384 events => throughput is 4.97E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G7/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.2304s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9223s + [COUNTERS] Fortran MEs ( 1 ) : 3.3080s for 16384 events => throughput is 4.95E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G71/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.6090s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3147s + [COUNTERS] Fortran MEs ( 1 ) : 3.2943s for 16384 events => throughput is 4.97E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G1/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.8508s + [COUNTERS] Fortran Overhead ( 0 ) : 3.5090s + [COUNTERS] Fortran MEs ( 1 ) : 3.3418s for 16384 events => throughput is 4.90E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.5829s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2868s + [COUNTERS] Fortran MEs ( 1 ) : 3.2961s for 16384 events => throughput is 4.97E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G4/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.0295s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7366s + [COUNTERS] Fortran MEs ( 1 ) : 3.2929s for 16384 events => throughput is 4.98E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G80/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.4670s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1724s + [COUNTERS] Fortran MEs ( 1 ) : 3.2946s for 16384 events => throughput is 4.97E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G79/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.0355s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7275s + [COUNTERS] Fortran MEs ( 1 ) : 3.3080s for 16384 events => throughput is 4.95E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.9534s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6549s + [COUNTERS] Fortran MEs ( 1 ) : 3.2984s for 16384 events => throughput is 4.97E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G70/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.3616s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0562s + [COUNTERS] Fortran MEs ( 1 ) : 3.3053s for 16384 events => throughput is 4.96E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/G5/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.4129s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1180s + [COUNTERS] Fortran MEs ( 1 ) : 3.2949s for 16384 events => throughput is 4.97E+03 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemggd/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxcx_epemguxcx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxcx_epemguxcx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxcx_epemguxcx/ajob1' in 6.5272 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxcx_epemguxcx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxcx_epemguxcx/G9/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.9388s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9556s + [COUNTERS] Fortran MEs ( 1 ) : 0.9832s for 16384 events => throughput is 1.67E+04 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxcx_epemguxcx/G10/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.5485s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5694s + [COUNTERS] Fortran MEs ( 1 ) : 0.9791s for 16384 events => throughput is 1.67E+04 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxcx_epemguxcx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/ajob1' in 77.9162 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/G46/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.8344s + [COUNTERS] Fortran Overhead ( 0 ) : 4.5122s + [COUNTERS] Fortran MEs ( 1 ) : 3.3221s for 16384 events => throughput is 4.93E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/G18/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.3601s + [COUNTERS] Fortran Overhead ( 0 ) : 4.0488s + [COUNTERS] Fortran MEs ( 1 ) : 3.3112s for 16384 events => throughput is 4.95E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/G39/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.1738s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8620s + [COUNTERS] Fortran MEs ( 1 ) : 3.3118s for 16384 events => throughput is 4.95E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.5990s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2921s + [COUNTERS] Fortran MEs ( 1 ) : 3.3070s for 16384 events => throughput is 4.95E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/G20/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.8896s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5947s + [COUNTERS] Fortran MEs ( 1 ) : 3.2949s for 16384 events => throughput is 4.97E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/G17/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.6367s + [COUNTERS] Fortran Overhead ( 0 ) : 3.3154s + [COUNTERS] Fortran MEs ( 1 ) : 3.3213s for 16384 events => throughput is 4.93E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/G35/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.8434s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5209s + [COUNTERS] Fortran MEs ( 1 ) : 3.3225s for 16384 events => throughput is 4.93E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/G71/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.7504s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4405s + [COUNTERS] Fortran MEs ( 1 ) : 3.3099s for 16384 events => throughput is 4.95E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/G27/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.4133s + [COUNTERS] Fortran Overhead ( 0 ) : 4.1128s + [COUNTERS] Fortran MEs ( 1 ) : 3.3005s for 16384 events => throughput is 4.96E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/G92/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.7222s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4109s + [COUNTERS] Fortran MEs ( 1 ) : 3.3113s for 16384 events => throughput is 4.95E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/G80/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.1157s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8139s + [COUNTERS] Fortran MEs ( 1 ) : 3.3018s for 16384 events => throughput is 4.96E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.8996s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5943s + [COUNTERS] Fortran MEs ( 1 ) : 3.3053s for 16384 events => throughput is 4.96E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/G29/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.2898s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9689s + [COUNTERS] Fortran MEs ( 1 ) : 3.3209s for 16384 events => throughput is 4.93E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/G48/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.1318s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8212s + [COUNTERS] Fortran MEs ( 1 ) : 3.3106s for 16384 events => throughput is 4.95E+03 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemgddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_epemgdux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_epemgdux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_epemgdux/ajob1' in 15.9626 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_epemgdux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_epemgdux/G46/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.7853s + [COUNTERS] Fortran Overhead ( 0 ) : 0.7988s + [COUNTERS] Fortran MEs ( 1 ) : 0.9865s for 16384 events => throughput is 1.66E+04 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_epemgdux/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.1137s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1267s + [COUNTERS] Fortran MEs ( 1 ) : 0.9870s for 16384 events => throughput is 1.66E+04 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_epemgdux/G23/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.9343s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9472s + [COUNTERS] Fortran MEs ( 1 ) : 0.9872s for 16384 events => throughput is 1.66E+04 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_epemgdux/G47/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.8668s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8777s + [COUNTERS] Fortran MEs ( 1 ) : 0.9891s for 16384 events => throughput is 1.66E+04 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_epemgdux/G8/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.2227s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2412s + [COUNTERS] Fortran MEs ( 1 ) : 0.9815s for 16384 events => throughput is 1.67E+04 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_epemgdux/G7/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.0062s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0214s + [COUNTERS] Fortran MEs ( 1 ) : 0.9848s for 16384 events => throughput is 1.66E+04 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_epemgdux/G28/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.1649s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1728s + [COUNTERS] Fortran MEs ( 1 ) : 0.9921s for 16384 events => throughput is 1.65E+04 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_epemgdux/G32/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.7196s + [COUNTERS] Fortran Overhead ( 0 ) : 0.7377s + [COUNTERS] Fortran MEs ( 1 ) : 0.9819s for 16384 events => throughput is 1.67E+04 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_epemgdux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/ajob1' in 105.0271 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/G12/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.0753s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4763s + [COUNTERS] Fortran MEs ( 1 ) : 6.5991s for 16384 events => throughput is 2.48E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/G6/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.7878s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1740s + [COUNTERS] Fortran MEs ( 1 ) : 6.6139s for 16384 events => throughput is 2.48E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/G11/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.8672s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2508s + [COUNTERS] Fortran MEs ( 1 ) : 6.6164s for 16384 events => throughput is 2.48E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.8452s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2014s + [COUNTERS] Fortran MEs ( 1 ) : 6.6438s for 16384 events => throughput is 2.47E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/G69/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.7065s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0907s + [COUNTERS] Fortran MEs ( 1 ) : 6.6158s for 16384 events => throughput is 2.48E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/G8/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.5683s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9541s + [COUNTERS] Fortran MEs ( 1 ) : 6.6142s for 16384 events => throughput is 2.48E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/G84/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.0053s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3422s + [COUNTERS] Fortran MEs ( 1 ) : 6.6630s for 16384 events => throughput is 2.46E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.2668s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5962s + [COUNTERS] Fortran MEs ( 1 ) : 6.6706s for 16384 events => throughput is 2.46E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/G2/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 9.5875s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9770s + [COUNTERS] Fortran MEs ( 1 ) : 6.6105s for 16384 events => throughput is 2.48E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/G80/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.4015s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7363s + [COUNTERS] Fortran MEs ( 1 ) : 6.6652s for 16384 events => throughput is 2.46E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/G51/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.3590s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7534s + [COUNTERS] Fortran MEs ( 1 ) : 6.6056s for 16384 events => throughput is 2.48E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/G70/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.7035s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0901s + [COUNTERS] Fortran MEs ( 1 ) : 6.6134s for 16384 events => throughput is 2.48E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/G5/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.6115s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9991s + [COUNTERS] Fortran MEs ( 1 ) : 6.6124s for 16384 events => throughput is 2.48E+03 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamggdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamguux/ajob1' in 73.7733 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamguux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamguux/G18/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 10.5864s + [COUNTERS] Fortran Overhead ( 0 ) : 3.8959s + [COUNTERS] Fortran MEs ( 1 ) : 6.6906s for 16384 events => throughput is 2.45E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamguux/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.8792s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2564s + [COUNTERS] Fortran MEs ( 1 ) : 6.6228s for 16384 events => throughput is 2.47E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamguux/G20/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.1579s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5296s + [COUNTERS] Fortran MEs ( 1 ) : 6.6283s for 16384 events => throughput is 2.47E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamguux/G17/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 9.8852s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2356s + [COUNTERS] Fortran MEs ( 1 ) : 6.6496s for 16384 events => throughput is 2.46E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamguux/G35/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.0928s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4589s + [COUNTERS] Fortran MEs ( 1 ) : 6.6339s for 16384 events => throughput is 2.47E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamguux/G27/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 10.3401s + [COUNTERS] Fortran Overhead ( 0 ) : 3.6927s + [COUNTERS] Fortran MEs ( 1 ) : 6.6474s for 16384 events => throughput is 2.46E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamguux/G34/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 10.2561s + [COUNTERS] Fortran Overhead ( 0 ) : 3.6151s + [COUNTERS] Fortran MEs ( 1 ) : 6.6410s for 16384 events => throughput is 2.47E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamguux/G80/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.4257s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8039s + [COUNTERS] Fortran MEs ( 1 ) : 6.6218s for 16384 events => throughput is 2.47E+03 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamguux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_taptamgdsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_taptamgdsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_taptamgdsx/ajob1' in 3.0672 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_taptamgdsx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_taptamgdsx/G28/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.0453s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1034s + [COUNTERS] Fortran MEs ( 1 ) : 1.9419s for 16384 events => throughput is 8.44E+03 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_taptamgdsx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_taptamggg/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_taptamggg/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_taptamggg/ajob1' in 26.9408 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_taptamggg/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_taptamggg/G12/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.3341s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6914s + [COUNTERS] Fortran MEs ( 1 ) : 6.6427s for 16384 events => throughput is 2.47E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_taptamggg/G9/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 10.1615s + [COUNTERS] Fortran Overhead ( 0 ) : 3.5327s + [COUNTERS] Fortran MEs ( 1 ) : 6.6288s for 16384 events => throughput is 2.47E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_taptamggg/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.3870s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7225s + [COUNTERS] Fortran MEs ( 1 ) : 6.6645s for 16384 events => throughput is 2.46E+03 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_taptamggg/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxdx_epemguxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxdx_epemguxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxdx_epemguxdx/ajob1' in 17.5586 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxdx_epemguxdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxdx_epemguxdx/G14/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.1036s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1085s + [COUNTERS] Fortran MEs ( 1 ) : 0.9951s for 16384 events => throughput is 1.65E+04 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxdx_epemguxdx/G6/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.5441s + [COUNTERS] Fortran Overhead ( 0 ) : 4.5529s + [COUNTERS] Fortran MEs ( 1 ) : 0.9912s for 16384 events => throughput is 1.65E+04 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxdx_epemguxdx/G22/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.2440s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2431s + [COUNTERS] Fortran MEs ( 1 ) : 1.0009s for 16384 events => throughput is 1.64E+04 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxdx_epemguxdx/G8/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.2341s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2455s + [COUNTERS] Fortran MEs ( 1 ) : 0.9886s for 16384 events => throughput is 1.66E+04 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxdx_epemguxdx/G27/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.3381s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3455s + [COUNTERS] Fortran MEs ( 1 ) : 0.9925s for 16384 events => throughput is 1.65E+04 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uxdx_epemguxdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_epemgdsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_epemgdsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_epemgdsx/ajob1' in 16.5511 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_epemgdsx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_epemgdsx/G31/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.8958s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9196s + [COUNTERS] Fortran MEs ( 1 ) : 0.9762s for 16384 events => throughput is 1.68E+04 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_epemgdsx/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.9236s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9519s + [COUNTERS] Fortran MEs ( 1 ) : 0.9717s for 16384 events => throughput is 1.69E+04 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_epemgdsx/G20/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.1053s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1235s + [COUNTERS] Fortran MEs ( 1 ) : 0.9818s for 16384 events => throughput is 1.67E+04 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_epemgdsx/G25/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.6951s + [COUNTERS] Fortran Overhead ( 0 ) : 3.7167s + [COUNTERS] Fortran MEs ( 1 ) : 0.9783s for 16384 events => throughput is 1.67E+04 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_epemgdsx/G28/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.0085s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0321s + [COUNTERS] Fortran MEs ( 1 ) : 0.9764s for 16384 events => throughput is 1.68E+04 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_epemgdsx/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.8093s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8351s + [COUNTERS] Fortran MEs ( 1 ) : 0.9742s for 16384 events => throughput is 1.68E+04 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dsx_epemgdsx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamdddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamdddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamdddx/ajob1' in 7.7066 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamdddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamdddx/G17/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.6847s + [COUNTERS] Fortran Overhead ( 0 ) : 3.3102s + [COUNTERS] Fortran MEs ( 1 ) : 4.3744s for 16384 events => throughput is 3.75E+03 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamdddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemggg/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemggg/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemggg/ajob1' in 29.8185 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemggg/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemggg/G56/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.9559s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6361s + [COUNTERS] Fortran MEs ( 1 ) : 3.3198s for 16384 events => throughput is 4.94E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemggg/G12/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.1017s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7789s + [COUNTERS] Fortran MEs ( 1 ) : 3.3228s for 16384 events => throughput is 4.93E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemggg/G11/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.0926s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7698s + [COUNTERS] Fortran MEs ( 1 ) : 3.3228s for 16384 events => throughput is 4.93E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemggg/G40/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.9024s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5761s + [COUNTERS] Fortran MEs ( 1 ) : 3.3263s for 16384 events => throughput is 4.93E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemggg/G8/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.7013s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3534s + [COUNTERS] Fortran MEs ( 1 ) : 3.3478s for 16384 events => throughput is 4.89E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemggg/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.9524s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6199s + [COUNTERS] Fortran MEs ( 1 ) : 3.3324s for 16384 events => throughput is 4.92E+03 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemggg/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/ajob1' in 198.7301 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G14/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 9.7705s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1303s + [COUNTERS] Fortran MEs ( 1 ) : 6.6402s for 16384 events => throughput is 2.47E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G74/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 10.4177s + [COUNTERS] Fortran Overhead ( 0 ) : 3.8022s + [COUNTERS] Fortran MEs ( 1 ) : 6.6155s for 16384 events => throughput is 2.48E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G12/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.8539s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2083s + [COUNTERS] Fortran MEs ( 1 ) : 6.6456s for 16384 events => throughput is 2.47E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G6/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.7041s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0645s + [COUNTERS] Fortran MEs ( 1 ) : 6.6396s for 16384 events => throughput is 2.47E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G13/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 9.6400s + [COUNTERS] Fortran Overhead ( 0 ) : 3.0125s + [COUNTERS] Fortran MEs ( 1 ) : 6.6275s for 16384 events => throughput is 2.47E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G66/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 9.1981s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5621s + [COUNTERS] Fortran MEs ( 1 ) : 6.6359s for 16384 events => throughput is 2.47E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G11/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.0398s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3886s + [COUNTERS] Fortran MEs ( 1 ) : 6.6512s for 16384 events => throughput is 2.46E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G9/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 9.8840s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1951s + [COUNTERS] Fortran MEs ( 1 ) : 6.6889s for 16384 events => throughput is 2.45E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G40/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.3836s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7596s + [COUNTERS] Fortran MEs ( 1 ) : 6.6240s for 16384 events => throughput is 2.47E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G35/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.4791s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7699s + [COUNTERS] Fortran MEs ( 1 ) : 6.7091s for 16384 events => throughput is 2.44E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.2696s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6360s + [COUNTERS] Fortran MEs ( 1 ) : 6.6336s for 16384 events => throughput is 2.47E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G69/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.6948s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0186s + [COUNTERS] Fortran MEs ( 1 ) : 6.6762s for 16384 events => throughput is 2.45E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G7/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.5444s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8837s + [COUNTERS] Fortran MEs ( 1 ) : 6.6607s for 16384 events => throughput is 2.46E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G65/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 9.3334s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7085s + [COUNTERS] Fortran MEs ( 1 ) : 6.6249s for 16384 events => throughput is 2.47E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G71/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.9554s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3342s + [COUNTERS] Fortran MEs ( 1 ) : 6.6211s for 16384 events => throughput is 2.47E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G1/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 9.6258s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9951s + [COUNTERS] Fortran MEs ( 1 ) : 6.6308s for 16384 events => throughput is 2.47E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.9365s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2624s + [COUNTERS] Fortran MEs ( 1 ) : 6.6741s for 16384 events => throughput is 2.45E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G2/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 9.0425s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3924s + [COUNTERS] Fortran MEs ( 1 ) : 6.6500s for 16384 events => throughput is 2.46E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G4/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 9.3224s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6687s + [COUNTERS] Fortran MEs ( 1 ) : 6.6537s for 16384 events => throughput is 2.46E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G79/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.5333s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8523s + [COUNTERS] Fortran MEs ( 1 ) : 6.6810s for 16384 events => throughput is 2.45E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.2486s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6125s + [COUNTERS] Fortran MEs ( 1 ) : 6.6361s for 16384 events => throughput is 2.47E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G70/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.6508s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0156s + [COUNTERS] Fortran MEs ( 1 ) : 6.6351s for 16384 events => throughput is 2.47E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/G5/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.7798s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1361s + [COUNTERS] Fortran MEs ( 1 ) : 6.6437s for 16384 events => throughput is 2.47E+03 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamggu/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/ajob1' in 96.0997 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/G46/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.4919s + [COUNTERS] Fortran Overhead ( 0 ) : 4.1488s + [COUNTERS] Fortran MEs ( 1 ) : 3.3432s for 16384 events => throughput is 4.90E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/G18/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.7604s + [COUNTERS] Fortran Overhead ( 0 ) : 3.4393s + [COUNTERS] Fortran MEs ( 1 ) : 3.3211s for 16384 events => throughput is 4.93E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.8699s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5182s + [COUNTERS] Fortran MEs ( 1 ) : 3.3517s for 16384 events => throughput is 4.89E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/G33/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.8035s + [COUNTERS] Fortran Overhead ( 0 ) : 3.4852s + [COUNTERS] Fortran MEs ( 1 ) : 3.3183s for 16384 events => throughput is 4.94E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/G20/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.9955s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6666s + [COUNTERS] Fortran MEs ( 1 ) : 3.3289s for 16384 events => throughput is 4.92E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/G47/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.6207s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2945s + [COUNTERS] Fortran MEs ( 1 ) : 3.3261s for 16384 events => throughput is 4.93E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/G17/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.5563s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2359s + [COUNTERS] Fortran MEs ( 1 ) : 3.3204s for 16384 events => throughput is 4.93E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/G25/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.5880s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2725s + [COUNTERS] Fortran MEs ( 1 ) : 3.3155s for 16384 events => throughput is 4.94E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/G35/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.6974s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3836s + [COUNTERS] Fortran MEs ( 1 ) : 3.3138s for 16384 events => throughput is 4.94E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/G65/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.5466s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2186s + [COUNTERS] Fortran MEs ( 1 ) : 3.3280s for 16384 events => throughput is 4.92E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/G27/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.0588s + [COUNTERS] Fortran Overhead ( 0 ) : 3.7311s + [COUNTERS] Fortran MEs ( 1 ) : 3.3277s for 16384 events => throughput is 4.92E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/G34/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.7580s + [COUNTERS] Fortran Overhead ( 0 ) : 3.4353s + [COUNTERS] Fortran MEs ( 1 ) : 3.3226s for 16384 events => throughput is 4.93E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/G80/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.2563s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9438s + [COUNTERS] Fortran MEs ( 1 ) : 3.3125s for 16384 events => throughput is 4.95E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/G67/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.0626s + [COUNTERS] Fortran Overhead ( 0 ) : 3.7276s + [COUNTERS] Fortran MEs ( 1 ) : 3.3351s for 16384 events => throughput is 4.91E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.8907s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5660s + [COUNTERS] Fortran MEs ( 1 ) : 3.3248s for 16384 events => throughput is 4.93E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/G48/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.8493s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5344s + [COUNTERS] Fortran MEs ( 1 ) : 3.3148s for 16384 events => throughput is 4.94E+03 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_epemguux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_taptamgdux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_taptamgdux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_taptamgdux/ajob1' in 11.3936 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_taptamgdux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_taptamgdux/G14/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.5203s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5350s + [COUNTERS] Fortran MEs ( 1 ) : 1.9853s for 16384 events => throughput is 8.25E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_taptamgdux/G40/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.8659s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8728s + [COUNTERS] Fortran MEs ( 1 ) : 1.9931s for 16384 events => throughput is 8.22E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_taptamgdux/G44/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.9495s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9775s + [COUNTERS] Fortran MEs ( 1 ) : 1.9720s for 16384 events => throughput is 8.31E+03 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dux_taptamgdux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemgddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemgddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemgddx/ajob1' in 16.3782 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemgddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemgddx/G56/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.2945s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0920s + [COUNTERS] Fortran MEs ( 1 ) : 2.2025s for 16384 events => throughput is 7.44E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemgddx/G12/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.1455s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9427s + [COUNTERS] Fortran MEs ( 1 ) : 2.2028s for 16384 events => throughput is 7.44E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemgddx/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.2302s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9675s + [COUNTERS] Fortran MEs ( 1 ) : 2.2626s for 16384 events => throughput is 7.24E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemgddx/G47/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.3589s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1533s + [COUNTERS] Fortran MEs ( 1 ) : 2.2056s for 16384 events => throughput is 7.43E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemgddx/G42/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.2543s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0519s + [COUNTERS] Fortran MEs ( 1 ) : 2.2023s for 16384 events => throughput is 7.44E+03 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_epemgddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/ajob1' in 130.8680 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/G39/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.0966s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4225s + [COUNTERS] Fortran MEs ( 1 ) : 6.6741s for 16384 events => throughput is 2.45E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/G12/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.8550s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2306s + [COUNTERS] Fortran MEs ( 1 ) : 6.6244s for 16384 events => throughput is 2.47E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/G6/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.5442s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9337s + [COUNTERS] Fortran MEs ( 1 ) : 6.6105s for 16384 events => throughput is 2.48E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/G66/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 9.3635s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7601s + [COUNTERS] Fortran MEs ( 1 ) : 6.6034s for 16384 events => throughput is 2.48E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/G11/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.9884s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3918s + [COUNTERS] Fortran MEs ( 1 ) : 6.5967s for 16384 events => throughput is 2.48E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/G9/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 10.2424s + [COUNTERS] Fortran Overhead ( 0 ) : 3.5846s + [COUNTERS] Fortran MEs ( 1 ) : 6.6579s for 16384 events => throughput is 2.46E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.0657s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4493s + [COUNTERS] Fortran MEs ( 1 ) : 6.6164s for 16384 events => throughput is 2.48E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/G69/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.7422s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1228s + [COUNTERS] Fortran MEs ( 1 ) : 6.6194s for 16384 events => throughput is 2.48E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/G8/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.5925s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9832s + [COUNTERS] Fortran MEs ( 1 ) : 6.6094s for 16384 events => throughput is 2.48E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/G71/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.9440s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3397s + [COUNTERS] Fortran MEs ( 1 ) : 6.6043s for 16384 events => throughput is 2.48E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.8992s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2272s + [COUNTERS] Fortran MEs ( 1 ) : 6.6720s for 16384 events => throughput is 2.46E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/G72/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.8155s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2047s + [COUNTERS] Fortran MEs ( 1 ) : 6.6108s for 16384 events => throughput is 2.48E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/G79/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.4453s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7702s + [COUNTERS] Fortran MEs ( 1 ) : 6.6751s for 16384 events => throughput is 2.45E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.4804s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8660s + [COUNTERS] Fortran MEs ( 1 ) : 6.6144s for 16384 events => throughput is 2.48E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/G70/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.6562s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0287s + [COUNTERS] Fortran MEs ( 1 ) : 6.6275s for 16384 events => throughput is 2.47E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/G5/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.8409s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2385s + [COUNTERS] Fortran MEs ( 1 ) : 6.6023s for 16384 events => throughput is 2.48E+03 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamggd/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemdssx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemdssx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemdssx/ajob1' in 6.1467 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemdssx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemdssx/G46/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.1314s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1537s + [COUNTERS] Fortran MEs ( 1 ) : 0.9777s for 16384 events => throughput is 1.68E+04 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemdssx/G17/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.9755s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9952s + [COUNTERS] Fortran MEs ( 1 ) : 0.9803s for 16384 events => throughput is 1.67E+04 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemdssx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uc_taptamguc/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uc_taptamguc/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uc_taptamguc/ajob1' in 7.1241 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uc_taptamguc/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uc_taptamguc/G25/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.2587s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3018s + [COUNTERS] Fortran MEs ( 1 ) : 1.9569s for 16384 events => throughput is 8.37E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uc_taptamguc/G37/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.8257s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8746s + [COUNTERS] Fortran MEs ( 1 ) : 1.9511s for 16384 events => throughput is 8.40E+03 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uc_taptamguc/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamgddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamgddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamgddx/ajob1' in 11.7859 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamgddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamgddx/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.1343s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6828s + [COUNTERS] Fortran MEs ( 1 ) : 4.4515s for 16384 events => throughput is 3.68E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamgddx/G55/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.6102s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2229s + [COUNTERS] Fortran MEs ( 1 ) : 4.3873s for 16384 events => throughput is 3.73E+03 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ddx_taptamgddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemgddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemgddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemgddx/ajob1' in 2.5456 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemgddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemgddx/G8/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.5236s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5300s + [COUNTERS] Fortran MEs ( 1 ) : 0.9936s for 16384 events => throughput is 1.65E+04 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemgddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamgddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamgddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamgddx/ajob1' in 76.3456 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamgddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamgddx/G31/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.8222s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1982s + [COUNTERS] Fortran MEs ( 1 ) : 6.6239s for 16384 events => throughput is 2.47E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamgddx/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.8792s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2702s + [COUNTERS] Fortran MEs ( 1 ) : 6.6090s for 16384 events => throughput is 2.48E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamgddx/G33/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 9.8110s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1732s + [COUNTERS] Fortran MEs ( 1 ) : 6.6378s for 16384 events => throughput is 2.47E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamgddx/G20/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.0251s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4251s + [COUNTERS] Fortran MEs ( 1 ) : 6.6000s for 16384 events => throughput is 2.48E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamgddx/G35/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.9181s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2650s + [COUNTERS] Fortran MEs ( 1 ) : 6.6531s for 16384 events => throughput is 2.46E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamgddx/G71/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.7252s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1179s + [COUNTERS] Fortran MEs ( 1 ) : 6.6072s for 16384 events => throughput is 2.48E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamgddx/G34/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 10.5066s + [COUNTERS] Fortran Overhead ( 0 ) : 3.8127s + [COUNTERS] Fortran MEs ( 1 ) : 6.6939s for 16384 events => throughput is 2.45E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamgddx/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.1514s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5382s + [COUNTERS] Fortran MEs ( 1 ) : 6.6132s for 16384 events => throughput is 2.48E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamgddx/G48/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.3397s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7229s + [COUNTERS] Fortran MEs ( 1 ) : 6.6168s for 16384 events => throughput is 2.48E+03 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gg_taptamgddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemddxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemddxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemddxdx/ajob1' in 7.2744 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemddxdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemddxdx/G11/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.7836s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5782s + [COUNTERS] Fortran MEs ( 1 ) : 2.2053s for 16384 events => throughput is 7.43E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemddxdx/G71/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.4511s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2425s + [COUNTERS] Fortran MEs ( 1 ) : 2.2086s for 16384 events => throughput is 7.42E+03 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemddxdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dxsx_epemgdxsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dxsx_epemgdxsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dxsx_epemgdxsx/ajob1' in 2.2597 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dxsx_epemgdxsx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dxsx_epemgdxsx/G27/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.2380s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2585s + [COUNTERS] Fortran MEs ( 1 ) : 0.9795s for 16384 events => throughput is 1.67E+04 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dxsx_epemgdxsx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemduxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemduxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemduxdx/ajob1' in 5.8647 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemduxdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemduxdx/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.0757s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0775s + [COUNTERS] Fortran MEs ( 1 ) : 0.9983s for 16384 events => throughput is 1.64E+04 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemduxdx/G22/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.7492s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7560s + [COUNTERS] Fortran MEs ( 1 ) : 0.9932s for 16384 events => throughput is 1.65E+04 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gux_epemduxdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemguux/ajob1' in 29.4301 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemguux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemguux/G20/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.2724s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0602s + [COUNTERS] Fortran MEs ( 1 ) : 2.2123s for 16384 events => throughput is 7.41E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemguux/G27/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.0500s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8390s + [COUNTERS] Fortran MEs ( 1 ) : 2.2110s for 16384 events => throughput is 7.41E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemguux/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.6549s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4474s + [COUNTERS] Fortran MEs ( 1 ) : 2.2075s for 16384 events => throughput is 7.42E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemguux/G55/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.4380s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2256s + [COUNTERS] Fortran MEs ( 1 ) : 2.2124s for 16384 events => throughput is 7.41E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemguux/G37/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.2859s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0660s + [COUNTERS] Fortran MEs ( 1 ) : 2.2199s for 16384 events => throughput is 7.38E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemguux/G41/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.2049s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9430s + [COUNTERS] Fortran MEs ( 1 ) : 2.2619s for 16384 events => throughput is 7.24E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemguux/G57/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.8788s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6599s + [COUNTERS] Fortran MEs ( 1 ) : 2.2189s for 16384 events => throughput is 7.38E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemguux/G48/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.4959s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2746s + [COUNTERS] Fortran MEs ( 1 ) : 2.2213s for 16384 events => throughput is 7.38E+03 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_epemguux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dd_taptamgdd/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dd_taptamgdd/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dd_taptamgdd/ajob1' in 11.8540 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dd_taptamgdd/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dd_taptamgdd/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.6100s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2356s + [COUNTERS] Fortran MEs ( 1 ) : 4.3744s for 16384 events => throughput is 3.75E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dd_taptamgdd/G24/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.2040s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8403s + [COUNTERS] Fortran MEs ( 1 ) : 4.3637s for 16384 events => throughput is 3.75E+03 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_dd_taptamgdd/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/ajob1' in 177.2784 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G14/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.7666s + [COUNTERS] Fortran Overhead ( 0 ) : 3.4365s + [COUNTERS] Fortran MEs ( 1 ) : 3.3301s for 16384 events => throughput is 4.92E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G38/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.1147s + [COUNTERS] Fortran Overhead ( 0 ) : 3.7957s + [COUNTERS] Fortran MEs ( 1 ) : 3.3190s for 16384 events => throughput is 4.94E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G39/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.7580s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4337s + [COUNTERS] Fortran MEs ( 1 ) : 3.3243s for 16384 events => throughput is 4.93E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G12/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.5149s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2057s + [COUNTERS] Fortran MEs ( 1 ) : 3.3092s for 16384 events => throughput is 4.95E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G6/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.2446s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9332s + [COUNTERS] Fortran MEs ( 1 ) : 3.3115s for 16384 events => throughput is 4.95E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G3/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.1645s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8240s + [COUNTERS] Fortran MEs ( 1 ) : 3.3404s for 16384 events => throughput is 4.90E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G13/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.6034s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2819s + [COUNTERS] Fortran MEs ( 1 ) : 3.3215s for 16384 events => throughput is 4.93E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G66/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.2778s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9645s + [COUNTERS] Fortran MEs ( 1 ) : 3.3133s for 16384 events => throughput is 4.94E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G33/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.3257s + [COUNTERS] Fortran Overhead ( 0 ) : 4.9734s + [COUNTERS] Fortran MEs ( 1 ) : 3.3524s for 16384 events => throughput is 4.89E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G11/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.6895s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3234s + [COUNTERS] Fortran MEs ( 1 ) : 3.3661s for 16384 events => throughput is 4.87E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G9/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.0454s + [COUNTERS] Fortran Overhead ( 0 ) : 3.6725s + [COUNTERS] Fortran MEs ( 1 ) : 3.3729s for 16384 events => throughput is 4.86E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G10/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.2449s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9230s + [COUNTERS] Fortran MEs ( 1 ) : 3.3220s for 16384 events => throughput is 4.93E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G40/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.9503s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6400s + [COUNTERS] Fortran MEs ( 1 ) : 3.3103s for 16384 events => throughput is 4.95E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G35/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.9193s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5968s + [COUNTERS] Fortran MEs ( 1 ) : 3.3225s for 16384 events => throughput is 4.93E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.6756s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3636s + [COUNTERS] Fortran MEs ( 1 ) : 3.3120s for 16384 events => throughput is 4.95E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G69/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.3762s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0627s + [COUNTERS] Fortran MEs ( 1 ) : 3.3135s for 16384 events => throughput is 4.94E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G8/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.2809s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9399s + [COUNTERS] Fortran MEs ( 1 ) : 3.3410s for 16384 events => throughput is 4.90E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G7/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.3904s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0753s + [COUNTERS] Fortran MEs ( 1 ) : 3.3152s for 16384 events => throughput is 4.94E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G65/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.3482s + [COUNTERS] Fortran Overhead ( 0 ) : 3.0427s + [COUNTERS] Fortran MEs ( 1 ) : 3.3055s for 16384 events => throughput is 4.96E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G71/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.8021s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4894s + [COUNTERS] Fortran MEs ( 1 ) : 3.3127s for 16384 events => throughput is 4.95E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G1/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.6591s + [COUNTERS] Fortran Overhead ( 0 ) : 3.3353s + [COUNTERS] Fortran MEs ( 1 ) : 3.3238s for 16384 events => throughput is 4.93E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.5976s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2822s + [COUNTERS] Fortran MEs ( 1 ) : 3.3154s for 16384 events => throughput is 4.94E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G2/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.4640s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1464s + [COUNTERS] Fortran MEs ( 1 ) : 3.3177s for 16384 events => throughput is 4.94E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G4/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.3091s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9891s + [COUNTERS] Fortran MEs ( 1 ) : 3.3200s for 16384 events => throughput is 4.93E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G80/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.4912s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1633s + [COUNTERS] Fortran MEs ( 1 ) : 3.3278s for 16384 events => throughput is 4.92E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G41/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.9534s + [COUNTERS] Fortran Overhead ( 0 ) : 3.6401s + [COUNTERS] Fortran MEs ( 1 ) : 3.3133s for 16384 events => throughput is 4.94E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G77/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.2015s + [COUNTERS] Fortran Overhead ( 0 ) : 3.8953s + [COUNTERS] Fortran MEs ( 1 ) : 3.3062s for 16384 events => throughput is 4.96E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G76/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.6638s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3448s + [COUNTERS] Fortran MEs ( 1 ) : 3.3190s for 16384 events => throughput is 4.94E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G79/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.2748s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8924s + [COUNTERS] Fortran MEs ( 1 ) : 3.3825s for 16384 events => throughput is 4.84E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.8265s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5081s + [COUNTERS] Fortran MEs ( 1 ) : 3.3183s for 16384 events => throughput is 4.94E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G70/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.3344s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0182s + [COUNTERS] Fortran MEs ( 1 ) : 3.3161s for 16384 events => throughput is 4.94E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/G5/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.4192s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1035s + [COUNTERS] Fortran MEs ( 1 ) : 3.3157s for 16384 events => throughput is 4.94E+03 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ds_taptamgds/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ds_taptamgds/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ds_taptamgds/ajob1' in 3.8024 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ds_taptamgds/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ds_taptamgds/G10/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.7804s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8325s + [COUNTERS] Fortran MEs ( 1 ) : 1.9479s for 16384 events => throughput is 8.41E+03 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ds_taptamgds/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_epemgudx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_epemgudx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_epemgudx/ajob1' in 14.1898 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_epemgudx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_epemgudx/G12/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.8325s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8420s + [COUNTERS] Fortran MEs ( 1 ) : 0.9905s for 16384 events => throughput is 1.65E+04 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_epemgudx/G11/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.7936s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8068s + [COUNTERS] Fortran MEs ( 1 ) : 0.9868s for 16384 events => throughput is 1.66E+04 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_epemgudx/G17/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.3804s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3896s + [COUNTERS] Fortran MEs ( 1 ) : 0.9908s for 16384 events => throughput is 1.65E+04 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_epemgudx/G10/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.9450s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9569s + [COUNTERS] Fortran MEs ( 1 ) : 0.9881s for 16384 events => throughput is 1.66E+04 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_epemgudx/G27/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.0834s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0974s + [COUNTERS] Fortran MEs ( 1 ) : 0.9860s for 16384 events => throughput is 1.66E+04 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_epemgudx/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.0401s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0490s + [COUNTERS] Fortran MEs ( 1 ) : 0.9912s for 16384 events => throughput is 1.65E+04 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_epemgudx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_taptamgud/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_taptamgud/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_taptamgud/ajob1' in 29.5519 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_taptamgud/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_taptamgud/G12/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.8210s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8386s + [COUNTERS] Fortran MEs ( 1 ) : 1.9825s for 16384 events => throughput is 8.26E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_taptamgud/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.0676s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1076s + [COUNTERS] Fortran MEs ( 1 ) : 1.9600s for 16384 events => throughput is 8.36E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_taptamgud/G6/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.1345s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1771s + [COUNTERS] Fortran MEs ( 1 ) : 1.9574s for 16384 events => throughput is 8.37E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_taptamgud/G23/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.8284s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8667s + [COUNTERS] Fortran MEs ( 1 ) : 1.9617s for 16384 events => throughput is 8.35E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_taptamgud/G11/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.8544s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8870s + [COUNTERS] Fortran MEs ( 1 ) : 1.9674s for 16384 events => throughput is 8.33E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_taptamgud/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.1808s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2181s + [COUNTERS] Fortran MEs ( 1 ) : 1.9627s for 16384 events => throughput is 8.35E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_taptamgud/G22/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.3911s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4182s + [COUNTERS] Fortran MEs ( 1 ) : 1.9729s for 16384 events => throughput is 8.30E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_taptamgud/G7/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.2241s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2553s + [COUNTERS] Fortran MEs ( 1 ) : 1.9689s for 16384 events => throughput is 8.32E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_taptamgud/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.8813s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9230s + [COUNTERS] Fortran MEs ( 1 ) : 1.9582s for 16384 events => throughput is 8.37E+03 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_taptamgud/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemdddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemdddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemdddx/ajob1' in 7.5497 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemdddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemdddx/G31/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.9402s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7244s + [COUNTERS] Fortran MEs ( 1 ) : 2.2159s for 16384 events => throughput is 7.39E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemdddx/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.5690s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3640s + [COUNTERS] Fortran MEs ( 1 ) : 2.2049s for 16384 events => throughput is 7.43E+03 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemdddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamdssx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamdssx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamdssx/ajob1' in 6.3020 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamdssx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamdssx/G45/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.0509s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0976s + [COUNTERS] Fortran MEs ( 1 ) : 1.9533s for 16384 events => throughput is 8.39E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamdssx/G20/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.2107s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2646s + [COUNTERS] Fortran MEs ( 1 ) : 1.9460s for 16384 events => throughput is 8.42E+03 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_taptamdssx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuddx/ajob1' in 9.3047 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuddx/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.0294s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0729s + [COUNTERS] Fortran MEs ( 1 ) : 1.9565s for 16384 events => throughput is 8.37E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuddx/G20/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.1125s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1468s + [COUNTERS] Fortran MEs ( 1 ) : 1.9657s for 16384 events => throughput is 8.33E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuddx/G47/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.1038s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1431s + [COUNTERS] Fortran MEs ( 1 ) : 1.9606s for 16384 events => throughput is 8.36E+03 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_taptamuddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemuuxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemuuxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemuuxdx/ajob1' in 6.4968 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemuuxdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemuuxdx/G7/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.3381s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3515s + [COUNTERS] Fortran MEs ( 1 ) : 0.9867s for 16384 events => throughput is 1.66E+04 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemuuxdx/G41/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.1184s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1306s + [COUNTERS] Fortran MEs ( 1 ) : 0.9878s for 16384 events => throughput is 1.66E+04 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_epemuuxdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/ajob1' in 47.6712 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G14/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.0501s + [COUNTERS] Fortran Overhead ( 0 ) : 3.0664s + [COUNTERS] Fortran MEs ( 1 ) : 0.9837s for 16384 events => throughput is 1.67E+04 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G18/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.5533s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5591s + [COUNTERS] Fortran MEs ( 1 ) : 0.9942s for 16384 events => throughput is 1.65E+04 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G39/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.8510s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8371s + [COUNTERS] Fortran MEs ( 1 ) : 1.0140s for 16384 events => throughput is 1.62E+04 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G30/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.1454s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1563s + [COUNTERS] Fortran MEs ( 1 ) : 0.9891s for 16384 events => throughput is 1.66E+04 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G12/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.8242s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8373s + [COUNTERS] Fortran MEs ( 1 ) : 0.9869s for 16384 events => throughput is 1.66E+04 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.1916s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2068s + [COUNTERS] Fortran MEs ( 1 ) : 0.9848s for 16384 events => throughput is 1.66E+04 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G6/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.3791s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3937s + [COUNTERS] Fortran MEs ( 1 ) : 0.9854s for 16384 events => throughput is 1.66E+04 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G3/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.8971s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8810s + [COUNTERS] Fortran MEs ( 1 ) : 1.0161s for 16384 events => throughput is 1.61E+04 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G23/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.8727s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8855s + [COUNTERS] Fortran MEs ( 1 ) : 0.9872s for 16384 events => throughput is 1.66E+04 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G24/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.8665s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8803s + [COUNTERS] Fortran MEs ( 1 ) : 0.9862s for 16384 events => throughput is 1.66E+04 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G13/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.2918s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3072s + [COUNTERS] Fortran MEs ( 1 ) : 0.9846s for 16384 events => throughput is 1.66E+04 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G20/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.9068s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9210s + [COUNTERS] Fortran MEs ( 1 ) : 0.9858s for 16384 events => throughput is 1.66E+04 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G25/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.4560s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4739s + [COUNTERS] Fortran MEs ( 1 ) : 0.9821s for 16384 events => throughput is 1.67E+04 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.1703s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1833s + [COUNTERS] Fortran MEs ( 1 ) : 0.9870s for 16384 events => throughput is 1.66E+04 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G8/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.9427s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9599s + [COUNTERS] Fortran MEs ( 1 ) : 0.9828s for 16384 events => throughput is 1.67E+04 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G7/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.0217s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0359s + [COUNTERS] Fortran MEs ( 1 ) : 0.9858s for 16384 events => throughput is 1.66E+04 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G4/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.6748s + [COUNTERS] Fortran Overhead ( 0 ) : 0.6902s + [COUNTERS] Fortran MEs ( 1 ) : 0.9846s for 16384 events => throughput is 1.66E+04 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G41/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.3342s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3466s + [COUNTERS] Fortran MEs ( 1 ) : 0.9876s for 16384 events => throughput is 1.66E+04 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/G32/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.8919s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9091s + [COUNTERS] Fortran MEs ( 1 ) : 0.9827s for 16384 events => throughput is 1.67E+04 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ud_epemgud/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ucx_epemgucx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ucx_epemgucx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ucx_epemgucx/ajob1' in 6.0586 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ucx_epemgucx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ucx_epemgucx/G3/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.6523s + [COUNTERS] Fortran Overhead ( 0 ) : 0.6759s + [COUNTERS] Fortran MEs ( 1 ) : 0.9764s for 16384 events => throughput is 1.68E+04 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ucx_epemgucx/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.2563s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2784s + [COUNTERS] Fortran MEs ( 1 ) : 0.9779s for 16384 events => throughput is 1.68E+04 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ucx_epemgucx/G7/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.0915s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1123s + [COUNTERS] Fortran MEs ( 1 ) : 0.9792s for 16384 events => throughput is 1.67E+04 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ucx_epemgucx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_taptamgudx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_taptamgudx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_taptamgudx/ajob1' in 16.6930 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_taptamgudx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_taptamgudx/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.9799s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0081s + [COUNTERS] Fortran MEs ( 1 ) : 1.9718s for 16384 events => throughput is 8.31E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_taptamgudx/G11/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.8616s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9019s + [COUNTERS] Fortran MEs ( 1 ) : 1.9596s for 16384 events => throughput is 8.36E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_taptamgudx/G17/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.3901s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4255s + [COUNTERS] Fortran MEs ( 1 ) : 1.9646s for 16384 events => throughput is 8.34E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_taptamgudx/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.2400s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2731s + [COUNTERS] Fortran MEs ( 1 ) : 1.9669s for 16384 events => throughput is 8.33E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_taptamgudx/G27/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.1271s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1413s + [COUNTERS] Fortran MEs ( 1 ) : 1.9858s for 16384 events => throughput is 8.25E+03 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_udx_taptamgudx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamuuxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamuuxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamuuxdx/ajob1' in 4.7915 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamuuxdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamuuxdx/G6/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.7694s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7804s + [COUNTERS] Fortran MEs ( 1 ) : 1.9889s for 16384 events => throughput is 8.24E+03 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamuuxdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_taptamguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_taptamguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_taptamguux/ajob1' in 5.6787 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_taptamguux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_taptamguux/G48/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.6571s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2481s + [COUNTERS] Fortran MEs ( 1 ) : 4.4090s for 16384 events => throughput is 3.72E+03 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uux_taptamguux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemudux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemudux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemudux/ajob1' in 11.0997 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemudux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemudux/G43/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.5363s + [COUNTERS] Fortran Overhead ( 0 ) : 3.5479s + [COUNTERS] Fortran MEs ( 1 ) : 0.9884s for 16384 events => throughput is 1.66E+04 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemudux/G8/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.1693s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1794s + [COUNTERS] Fortran MEs ( 1 ) : 0.9899s for 16384 events => throughput is 1.66E+04 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemudux/G5/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.3348s + [COUNTERS] Fortran Overhead ( 0 ) : 3.3462s + [COUNTERS] Fortran MEs ( 1 ) : 0.9886s for 16384 events => throughput is 1.66E+04 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gd_epemudux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamsdxsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamsdxsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamsdxsx/ajob1' in 3.3235 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamsdxsx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamsdxsx/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.3017s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3105s + [COUNTERS] Fortran MEs ( 1 ) : 1.9912s for 16384 events => throughput is 8.23E+03 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gdx_taptamsdxsx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_epemguu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_epemguu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_epemguu/ajob1' in 35.3293 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_epemguu/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_epemguu/G30/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.2490s + [COUNTERS] Fortran Overhead ( 0 ) : 3.0327s + [COUNTERS] Fortran MEs ( 1 ) : 2.2163s for 16384 events => throughput is 7.39E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_epemguu/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.1975s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9884s + [COUNTERS] Fortran MEs ( 1 ) : 2.2091s for 16384 events => throughput is 7.42E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_epemguu/G9/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.5459s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3351s + [COUNTERS] Fortran MEs ( 1 ) : 2.2108s for 16384 events => throughput is 7.41E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_epemguu/G49/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.1957s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9799s + [COUNTERS] Fortran MEs ( 1 ) : 2.2158s for 16384 events => throughput is 7.39E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_epemguu/G10/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.4862s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2686s + [COUNTERS] Fortran MEs ( 1 ) : 2.2177s for 16384 events => throughput is 7.39E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_epemguu/G81/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.2746s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0644s + [COUNTERS] Fortran MEs ( 1 ) : 2.2102s for 16384 events => throughput is 7.41E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_epemguu/G62/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.0502s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8350s + [COUNTERS] Fortran MEs ( 1 ) : 2.2153s for 16384 events => throughput is 7.40E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_epemguu/G51/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.1108s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8938s + [COUNTERS] Fortran MEs ( 1 ) : 2.2170s for 16384 events => throughput is 7.39E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_epemguu/G28/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.0510s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8415s + [COUNTERS] Fortran MEs ( 1 ) : 2.2095s for 16384 events => throughput is 7.42E+03 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_epemguu/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uc_epemguc/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uc_epemguc/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uc_epemguc/ajob1' in 3.8762 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uc_epemguc/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uc_epemguc/G12/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.7807s + [COUNTERS] Fortran Overhead ( 0 ) : 0.7945s + [COUNTERS] Fortran MEs ( 1 ) : 0.9862s for 16384 events => throughput is 1.66E+04 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uc_epemguc/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.0555s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0717s + [COUNTERS] Fortran MEs ( 1 ) : 0.9838s for 16384 events => throughput is 1.67E+04 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uc_epemguc/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ds_epemgds/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ds_epemgds/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ds_epemgds/ajob1' in 2.1071 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ds_epemgds/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ds_epemgds/G27/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.0854s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1033s + [COUNTERS] Fortran MEs ( 1 ) : 0.9821s for 16384 events => throughput is 1.67E+04 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_ds_epemgds/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_taptamguu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_taptamguu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_taptamguu/ajob1' in 17.1294 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_taptamguu/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_taptamguu/G52/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.5215s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9695s + [COUNTERS] Fortran MEs ( 1 ) : 4.5519s for 16384 events => throughput is 3.60E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_taptamguu/G49/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.0812s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6658s + [COUNTERS] Fortran MEs ( 1 ) : 4.4154s for 16384 events => throughput is 3.71E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_taptamguu/G32/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.4679s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0236s + [COUNTERS] Fortran MEs ( 1 ) : 4.4443s for 16384 events => throughput is 3.69E+03 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_uu_taptamguu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/madevent/SubProcesses/P0_gu_epemggu/ajob1' in 6.9991 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished in 1924.3669 seconds +__CUDACPP_DEBUG: gen_ximprove.launch finished in 1924.7936 seconds +__CUDACPP_DEBUG: GridPackCmd.refine4grid finished in 1924.7943 seconds +__CUDACPP_DEBUG: GridPackCmd.launch (3a) combine_events +combine_events +__CUDACPP_DEBUG: GridPackCmd.launch (3b) store_events +store_events + === Results Summary for run: GridRun_21 tag: tag_1 === + + Cross-section : 0 +- 0 pb + Nb of events : 500 + +INFO: No version of lhapdf. Can not run systematics computation +decay_events -from_cards +__CUDACPP_DEBUG: GridPackCmd.launch finished in 1945.6279 seconds +__CUDACPP_DEBUG: GridPackCmd.__init__ back from self.launch +DONE +write ./events.lhe.gz +END: Wed Sep 11 02:16:45 AM CEST 2024 +ELAPSED: 1948 seconds diff --git a/epochX/cudacpp/tlau/fromgridpacks/pp_dy3j.mad/summary.txt b/epochX/cudacpp/tlau/fromgridpacks/pp_dy3j.mad/summary.txt new file mode 100644 index 0000000000..5393c884a3 --- /dev/null +++ b/epochX/cudacpp/tlau/fromgridpacks/pp_dy3j.mad/summary.txt @@ -0,0 +1,51 @@ +pp_dy3j.mad/fortran/output.txt +[GridPackCmd.launch] OVERALL TOTAL 1945.6279 seconds +[madevent COUNTERS] PROGRAM TOTAL 1910.3 +[madevent COUNTERS] Fortran Overhead 665.412 +[madevent COUNTERS] Fortran MEs 1244.89 +-------------------------------------------------------------------------------- +pp_dy3j.mad/cppnone/output.txt +[GridPackCmd.launch] OVERALL TOTAL 1920.0969 seconds +[madevent COUNTERS] PROGRAM TOTAL 1896.82 +[madevent COUNTERS] Fortran Overhead 668.916 +[madevent COUNTERS] CudaCpp MEs 1223.65 +[madevent COUNTERS] CudaCpp HEL 4.2527 +-------------------------------------------------------------------------------- +pp_dy3j.mad/cppsse4/output.txt +[GridPackCmd.launch] OVERALL TOTAL 1336.0181 seconds +[madevent COUNTERS] PROGRAM TOTAL 1313.34 +[madevent COUNTERS] Fortran Overhead 668.988 +[madevent COUNTERS] CudaCpp MEs 642.063 +[madevent COUNTERS] CudaCpp HEL 2.2873 +-------------------------------------------------------------------------------- +pp_dy3j.mad/cppavx2/output.txt +[GridPackCmd.launch] OVERALL TOTAL 960.2111 seconds +[madevent COUNTERS] PROGRAM TOTAL 937.127 +[madevent COUNTERS] Fortran Overhead 667.996 +[madevent COUNTERS] CudaCpp MEs 267.903 +[madevent COUNTERS] CudaCpp HEL 1.2269 +-------------------------------------------------------------------------------- +pp_dy3j.mad/cpp512y/output.txt +[GridPackCmd.launch] OVERALL TOTAL 940.0347 seconds +[madevent COUNTERS] PROGRAM TOTAL 917.336 +[madevent COUNTERS] Fortran Overhead 668.996 +[madevent COUNTERS] CudaCpp MEs 247.179 +[madevent COUNTERS] CudaCpp HEL 1.1605 +-------------------------------------------------------------------------------- +pp_dy3j.mad/cpp512z/output.txt +[GridPackCmd.launch] OVERALL TOTAL 1022.0703 seconds +[madevent COUNTERS] PROGRAM TOTAL 997.125 +[madevent COUNTERS] Fortran Overhead 669.147 +[madevent COUNTERS] CudaCpp MEs 326.476 +[madevent COUNTERS] CudaCpp HEL 1.503 +-------------------------------------------------------------------------------- +pp_dy3j.mad/cuda/output.txt +[GridPackCmd.launch] OVERALL TOTAL 969.4855 seconds +[madevent COUNTERS] PROGRAM TOTAL 853.823 +[madevent COUNTERS] Fortran Overhead 826.381 +[madevent COUNTERS] CudaCpp MEs 7.865 +[madevent COUNTERS] CudaCpp HEL 19.578 +-------------------------------------------------------------------------------- +pp_dy3j.mad/hip/output.txt +File not found: SKIP backend hip +-------------------------------------------------------------------------------- diff --git a/epochX/cudacpp/tlau/fromgridpacks/pp_dy4j.mad/cpp512y/output.txt b/epochX/cudacpp/tlau/fromgridpacks/pp_dy4j.mad/cpp512y/output.txt new file mode 100644 index 0000000000..29337c6e97 --- /dev/null +++ b/epochX/cudacpp/tlau/fromgridpacks/pp_dy4j.mad/cpp512y/output.txt @@ -0,0 +1,3327 @@ +START: Sun Sep 15 01:48:11 PM CEST 2024 +On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU (nproc=4)] [GPU: 1x Tesla V100S-PCIE-32GB]: +CUDACPP_RUNTIME_DISABLEFPE is set +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_bbx_epemdsdxsx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_bbx_taptamdsdxsx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_ccx_epemdcdxcx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_ccx_epemdsdxsx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_ccx_epemucuxcx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_ccx_epemuduxdx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_ccx_taptamdcdxcx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_ccx_taptamdsdxsx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_ccx_taptamucuxcx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_ccx_taptamuduxdx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_dc_epemdcccx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_dc_epemdcssx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_dc_epemddcdx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_dc_epemggdc/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_dc_epemudcux/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_dc_taptamdcccx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_dc_taptamdcssx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_dc_taptamddcdx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_dc_taptamggdc/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_dc_taptamudcux/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_dd_epemddddx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_dd_epemddssx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_dd_epemggdd/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_dd_epemuddux/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_dd_taptamddddx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_dd_taptamddssx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_dd_taptamggdd/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_dd_taptamuddux/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_ddx_epemdcdxcx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_ddx_epemdddxdx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_ddx_epemdsdxsx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_ddx_epemggddx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_ddx_epemgggg/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_ddx_epemggssx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_ddx_epemgguux/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_ddx_epemsbsxbx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_ddx_epemsssxsx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_ddx_epemucuxcx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_ddx_epemuduxdx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_ddx_epemusuxsx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_ddx_epemuuuxux/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_ddx_taptamdcdxcx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_ddx_taptamdddxdx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_ddx_taptamdsdxsx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_ddx_taptamggddx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_ddx_taptamgggg/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_ddx_taptamggssx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_ddx_taptamgguux/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_ddx_taptamsbsxbx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_ddx_taptamsssxsx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_ddx_taptamucuxcx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_ddx_taptamuduxdx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_ddx_taptamusuxsx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_ddx_taptamuuuxux/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_ds_epemddsdx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_ds_epemdsbbx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_ds_epemdsssx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_ds_epemggds/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_ds_epemudsux/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_ds_taptamddsdx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_ds_taptamdsbbx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_ds_taptamdsssx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_ds_taptamggds/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_ds_taptamudsux/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_dsx_epemdbsxbx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_dsx_epemdddxsx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_dsx_epemdssxsx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_dsx_epemggdsx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_dsx_epemuduxsx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_dsx_taptamdbsxbx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_dsx_taptamdddxsx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_dsx_taptamdssxsx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_dsx_taptamggdsx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_dsx_taptamuduxsx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_dux_epemdcuxcx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_dux_epemdduxdx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_dux_epemdsuxsx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_dux_epemggdux/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_dux_epemuduxux/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_dux_taptamdcuxcx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_dux_taptamdduxdx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_dux_taptamdsuxsx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_dux_taptamggdux/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_dux_taptamuduxux/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_dxcx_epemcdxcxcx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_dxcx_epemddxdxcx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_dxcx_epemggdxcx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_dxcx_epemsdxcxsx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_dxcx_epemuuxdxcx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_dxcx_taptamcdxcxcx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_dxcx_taptamddxdxcx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_dxcx_taptamggdxcx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_dxcx_taptamsdxcxsx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_dxcx_taptamuuxdxcx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_dxdx_epemddxdxdx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_dxdx_epemggdxdx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_dxdx_epemsdxdxsx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_dxdx_epemuuxdxdx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_dxdx_taptamddxdxdx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_dxdx_taptamggdxdx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_dxdx_taptamsdxdxsx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_dxdx_taptamuuxdxdx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_dxsx_epembdxsxbx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_dxsx_epemddxdxsx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_dxsx_epemggdxsx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_dxsx_epemsdxsxsx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_dxsx_epemuuxdxsx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_dxsx_taptambdxsxbx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_dxsx_taptamddxdxsx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_dxsx_taptamggdxsx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_dxsx_taptamsdxsxsx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_dxsx_taptamuuxdxsx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_gd_epemgdddx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_gd_epemgdssx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_gd_epemgggd/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_gd_epemgudux/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_gd_taptamgdddx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_gd_taptamgdssx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_gd_taptamgggd/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_gd_taptamgudux/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_gdx_epemgddxdx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_gdx_epemgggdx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_gdx_epemgsdxsx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_gdx_epemguuxdx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_gdx_taptamgddxdx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_gdx_taptamgggdx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_gdx_taptamgsdxsx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_gdx_taptamguuxdx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_gg_epemdddxdx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_gg_epemdsdxsx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_gg_epemggddx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_gg_epemgguux/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_gg_epemucuxcx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_gg_epemuduxdx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_gg_epemuuuxux/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_gg_taptamdddxdx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_gg_taptamdsdxsx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_gg_taptamggddx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_gg_taptamgguux/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_gg_taptamucuxcx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_gg_taptamuduxdx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_gg_taptamuuuxux/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_gu_epemgggu/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_gu_epemguccx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_gu_epemguddx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_gu_epemguuux/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_gu_taptamgggu/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_gu_taptamguccx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_gu_taptamguddx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_gu_taptamguuux/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_gux_epemgcuxcx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_gux_epemgduxdx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_gux_epemgggux/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_gux_epemguuxux/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_gux_taptamgcuxcx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_gux_taptamgduxdx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_gux_taptamgggux/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_gux_taptamguuxux/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_ssx_epemdbdxbx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_ssx_epemdcdxcx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_ssx_epemdsdxsx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_ssx_epemucuxcx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_ssx_epemuduxdx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_ssx_taptamdbdxbx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_ssx_taptamdcdxcx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_ssx_taptamdsdxsx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_ssx_taptamucuxcx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_ssx_taptamuduxdx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_uc_epemgguc/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_uc_epemucccx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_uc_epemudcdx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_uc_epemuucux/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_uc_taptamgguc/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_uc_taptamucccx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_uc_taptamudcdx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_uc_taptamuucux/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_ucx_epemggucx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_ucx_epemuccxcx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_ucx_epemuddxcx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_ucx_epemuuuxcx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_ucx_taptamggucx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_ucx_taptamuccxcx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_ucx_taptamuddxcx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_ucx_taptamuuuxcx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_ud_epemggud/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_ud_epemudccx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_ud_epemudddx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_ud_epemudssx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_ud_epemuudux/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_ud_taptamggud/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_ud_taptamudccx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_ud_taptamudddx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_ud_taptamudssx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_ud_taptamuudux/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_udx_epemggudx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_udx_epemucdxcx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_udx_epemuddxdx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_udx_epemusdxsx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_udx_epemuuuxdx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_udx_taptamggudx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_udx_taptamucdxcx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_udx_taptamuddxdx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_udx_taptamusdxsx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_udx_taptamuuuxdx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_uu_epemgguu/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_uu_epemuuccx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_uu_epemuuddx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_uu_epemuuuux/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_uu_taptamgguu/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_uu_taptamuuccx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_uu_taptamuuddx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_uu_taptamuuuux/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_uux_epemcccxcx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_uux_epemcscxsx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_uux_epemdcdxcx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_uux_epemdddxdx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_uux_epemdsdxsx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_uux_epemggccx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_uux_epemggddx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_uux_epemgggg/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_uux_epemgguux/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_uux_epemucuxcx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_uux_epemuduxdx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_uux_epemuuuxux/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_uux_taptamcccxcx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_uux_taptamcscxsx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_uux_taptamdcdxcx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_uux_taptamdddxdx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_uux_taptamdsdxsx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_uux_taptamggccx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_uux_taptamggddx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_uux_taptamgggg/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_uux_taptamgguux/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_uux_taptamucuxcx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_uux_taptamuduxdx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_uux_taptamuuuxux/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_uxcx_epemcuxcxcx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_uxcx_epemduxdxcx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_uxcx_epemgguxcx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_uxcx_epemuuxuxcx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_uxcx_taptamcuxcxcx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_uxcx_taptamduxdxcx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_uxcx_taptamgguxcx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_uxcx_taptamuuxuxcx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_uxdx_epemcuxdxcx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_uxdx_epemduxdxdx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_uxdx_epemgguxdx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_uxdx_epemsuxdxsx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_uxdx_epemuuxuxdx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_uxdx_taptamcuxdxcx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_uxdx_taptamduxdxdx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_uxdx_taptamgguxdx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_uxdx_taptamsuxdxsx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_uxdx_taptamuuxuxdx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_uxux_epemcuxuxcx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_uxux_epemduxuxdx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_uxux_epemgguxux/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_uxux_epemuuxuxux/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_uxux_taptamcuxuxcx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_uxux_taptamduxuxdx/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_uxux_taptamgguxux/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 13:48 madevent/SubProcesses/P0_uxux_taptamuuxuxux/madevent -> build.512y_d_inl0_hrd0/madevent_cpp +Now generating 500 events with random seed 21 and granularity 1 +************************************************************ +* * +* W E L C O M E to * +* M A D G R A P H 5 _ a M C @ N L O * +* M A D E V E N T * +* * +* * * * +* * * * * * +* * * * * 5 * * * * * +* * * * * * +* * * * +* * +* VERSION 3.5.3_lo_vect * +* * +* The MadGraph5_aMC@NLO Development Team - Find us at * +* https://server06.fynu.ucl.ac.be/projects/madgraph * +* * +* Type 'help' for in-line help. * +* * +************************************************************ +INFO: load configuration from /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/Cards/me5_configuration.txt +INFO: load configuration from /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/Cards/me5_configuration.txt +Using default text editor "vi". Set another one in ./input/mg5_configuration.txt +Using default eps viewer "evince". Set another one in ./input/mg5_configuration.txt +Using default web browser "firefox". Set another one in ./input/mg5_configuration.txt +WRITE GRIDCARD /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent +__CUDACPP_DEBUG: GridPackCmd.__init__ will call self.launch +__CUDACPP_DEBUG: GridPackCmd.launch starting +__CUDACPP_DEBUG: GridPackCmd.launch (1) restore_data +generate 500 events +P0_gg_epemgguux +P0_gg_epemggddx +P0_gg_taptamgguux +P0_gg_taptamggddx +P0_gu_epemgggu +P0_gd_epemgggd +P0_gux_epemgggux +P0_gdx_epemgggdx +P0_gu_taptamgggu +P0_gd_taptamgggd +P0_gux_taptamgggux +P0_gdx_taptamgggdx +P0_uux_epemgggg +P0_ddx_epemgggg +P0_uux_taptamgggg +P0_ddx_taptamgggg +P0_gg_epemuuuxux +P0_gg_epemdddxdx +P0_gg_epemuduxdx +P0_gg_epemucuxcx +P0_gg_epemdsdxsx +P0_gg_taptamuuuxux +P0_gg_taptamdddxdx +P0_gg_taptamuduxdx +P0_gg_taptamucuxcx +P0_gg_taptamdsdxsx +P0_gu_epemguuux +P0_gd_epemgdddx +P0_gux_epemguuxux +P0_gdx_epemgddxdx +P0_gu_epemguddx +P0_gu_epemguccx +P0_gd_epemgudux +P0_gd_epemgdssx +P0_gux_epemgduxdx +P0_gux_epemgcuxcx +P0_gdx_epemguuxdx +P0_gdx_epemgsdxsx +P0_gu_taptamguuux +P0_gd_taptamgdddx +P0_gux_taptamguuxux +P0_gdx_taptamgddxdx +P0_gu_taptamguddx +P0_gu_taptamguccx +P0_gd_taptamgudux +P0_gd_taptamgdssx +P0_gux_taptamgduxdx +P0_gux_taptamgcuxcx +P0_gdx_taptamguuxdx +P0_gdx_taptamgsdxsx +P0_uu_epemgguu +P0_uux_epemgguux +P0_dd_epemggdd +P0_ddx_epemggddx +P0_uxux_epemgguxux +P0_dxdx_epemggdxdx +P0_ud_epemggud +P0_uc_epemgguc +P0_uux_epemggddx +P0_uux_epemggccx +P0_udx_epemggudx +P0_ucx_epemggucx +P0_dc_epemggdc +P0_ds_epemggds +P0_dux_epemggdux +P0_ddx_epemgguux +P0_ddx_epemggssx +P0_dsx_epemggdsx +P0_uxdx_epemgguxdx +P0_uxcx_epemgguxcx +P0_dxcx_epemggdxcx +P0_dxsx_epemggdxsx +P0_uu_taptamgguu +P0_uux_taptamgguux +P0_dd_taptamggdd +P0_ddx_taptamggddx +P0_uxux_taptamgguxux +P0_dxdx_taptamggdxdx +P0_ud_taptamggud +P0_uc_taptamgguc +P0_uux_taptamggddx +P0_uux_taptamggccx +P0_udx_taptamggudx +P0_ucx_taptamggucx +P0_dc_taptamggdc +P0_ds_taptamggds +P0_dux_taptamggdux +P0_ddx_taptamgguux +P0_ddx_taptamggssx +P0_dsx_taptamggdsx +P0_uxdx_taptamgguxdx +P0_uxcx_taptamgguxcx +P0_dxcx_taptamggdxcx +P0_dxsx_taptamggdxsx +P0_uu_epemuuuux +P0_uux_epemuuuxux +P0_dd_epemddddx +P0_ddx_epemdddxdx +P0_uxux_epemuuxuxux +P0_dxdx_epemddxdxdx +P0_uu_epemuuddx +P0_uu_epemuuccx +P0_ud_epemuudux +P0_ud_epemudddx +P0_uc_epemuucux +P0_uc_epemucccx +P0_uux_epemuduxdx +P0_uux_epemucuxcx +P0_uux_epemdddxdx +P0_uux_epemcccxcx +P0_udx_epemuuuxdx +P0_udx_epemuddxdx +P0_ucx_epemuuuxcx +P0_ucx_epemuccxcx +P0_dd_epemuddux +P0_dd_epemddssx +P0_dc_epemddcdx +P0_dc_epemdcccx +P0_ds_epemddsdx +P0_ds_epemdsssx +P0_dux_epemuduxux +P0_dux_epemdduxdx +P0_ddx_epemuuuxux +P0_ddx_epemuduxdx +P0_ddx_epemdcdxcx +P0_ddx_epemdsdxsx +P0_ddx_epemsssxsx +P0_dsx_epemdddxsx +P0_dsx_epemdssxsx +P0_ccx_epemucuxcx +P0_ccx_epemdcdxcx +P0_ssx_epemdsdxsx +P0_uxux_epemduxuxdx +P0_uxux_epemcuxuxcx +P0_uxdx_epemuuxuxdx +P0_uxdx_epemduxdxdx +P0_uxcx_epemuuxuxcx +P0_uxcx_epemcuxcxcx +P0_dxdx_epemuuxdxdx +P0_dxdx_epemsdxdxsx +P0_dxcx_epemddxdxcx +P0_dxcx_epemcdxcxcx +P0_dxsx_epemddxdxsx +P0_dxsx_epemsdxsxsx +P0_ud_epemudccx +P0_ud_epemudssx +P0_uc_epemudcdx +P0_uux_epemdcdxcx +P0_uux_epemdsdxsx +P0_uux_epemcscxsx +P0_udx_epemucdxcx +P0_udx_epemusdxsx +P0_ucx_epemuddxcx +P0_dc_epemudcux +P0_dc_epemdcssx +P0_ds_epemudsux +P0_ds_epemdsbbx +P0_dux_epemdcuxcx +P0_dux_epemdsuxsx +P0_ddx_epemucuxcx +P0_ddx_epemusuxsx +P0_ddx_epemsbsxbx +P0_dsx_epemuduxsx +P0_dsx_epemdbsxbx +P0_ccx_epemuduxdx +P0_ccx_epemdsdxsx +P0_ssx_epemuduxdx +P0_ssx_epemucuxcx +P0_ssx_epemdcdxcx +P0_ssx_epemdbdxbx +P0_bbx_epemdsdxsx +P0_uxdx_epemcuxdxcx +P0_uxdx_epemsuxdxsx +P0_uxcx_epemduxdxcx +P0_dxcx_epemuuxdxcx +P0_dxcx_epemsdxcxsx +P0_dxsx_epemuuxdxsx +P0_dxsx_epembdxsxbx +P0_uu_taptamuuuux +P0_uux_taptamuuuxux +P0_dd_taptamddddx +P0_ddx_taptamdddxdx +P0_uxux_taptamuuxuxux +P0_dxdx_taptamddxdxdx +P0_uu_taptamuuddx +P0_uu_taptamuuccx +P0_ud_taptamuudux +P0_ud_taptamudddx +P0_uc_taptamuucux +P0_uc_taptamucccx +P0_uux_taptamuduxdx +P0_uux_taptamucuxcx +P0_uux_taptamdddxdx +P0_uux_taptamcccxcx +P0_udx_taptamuuuxdx +P0_udx_taptamuddxdx +P0_ucx_taptamuuuxcx +P0_ucx_taptamuccxcx +P0_dd_taptamuddux +P0_dd_taptamddssx +P0_dc_taptamddcdx +P0_dc_taptamdcccx +P0_ds_taptamddsdx +P0_ds_taptamdsssx +P0_dux_taptamuduxux +P0_dux_taptamdduxdx +P0_ddx_taptamuuuxux +P0_ddx_taptamuduxdx +P0_ddx_taptamdcdxcx +P0_ddx_taptamdsdxsx +P0_ddx_taptamsssxsx +P0_dsx_taptamdddxsx +P0_dsx_taptamdssxsx +P0_ccx_taptamucuxcx +P0_ccx_taptamdcdxcx +P0_ssx_taptamdsdxsx +P0_uxux_taptamduxuxdx +P0_uxux_taptamcuxuxcx +P0_uxdx_taptamuuxuxdx +P0_uxdx_taptamduxdxdx +P0_uxcx_taptamuuxuxcx +P0_uxcx_taptamcuxcxcx +P0_dxdx_taptamuuxdxdx +P0_dxdx_taptamsdxdxsx +P0_dxcx_taptamddxdxcx +P0_dxcx_taptamcdxcxcx +P0_dxsx_taptamddxdxsx +P0_dxsx_taptamsdxsxsx +P0_ud_taptamudccx +P0_ud_taptamudssx +P0_uc_taptamudcdx +P0_uux_taptamdcdxcx +P0_uux_taptamdsdxsx +P0_uux_taptamcscxsx +P0_udx_taptamucdxcx +P0_udx_taptamusdxsx +P0_ucx_taptamuddxcx +P0_dc_taptamudcux +P0_dc_taptamdcssx +P0_ds_taptamudsux +P0_ds_taptamdsbbx +P0_dux_taptamdcuxcx +P0_dux_taptamdsuxsx +P0_ddx_taptamucuxcx +P0_ddx_taptamusuxsx +P0_ddx_taptamsbsxbx +P0_dsx_taptamuduxsx +P0_dsx_taptamdbsxbx +P0_ccx_taptamuduxdx +P0_ccx_taptamdsdxsx +P0_ssx_taptamuduxdx +P0_ssx_taptamucuxcx +P0_ssx_taptamdcdxcx +P0_ssx_taptamdbdxbx +P0_bbx_taptamdsdxsx +P0_uxdx_taptamcuxdxcx +P0_uxdx_taptamsuxdxsx +P0_uxcx_taptamduxdxcx +P0_dxcx_taptamuuxdxcx +P0_dxcx_taptamsdxcxsx +P0_dxsx_taptamuuxdxsx +P0_dxsx_taptambdxsxbx +__CUDACPP_DEBUG: GridPackCmd.launch (2) refine4grid +__CUDACPP_DEBUG: GridPackCmd.refine4grid starting +Using random number seed offset = 21 +INFO: Generating 500.0 unweighted events. +__CUDACPP_DEBUG: gen_ximprove.launch starting +sum of cpu time of last step: 3 days,23h23m17s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event starting +INFO: need to improve 477 channels +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dxdx_taptamuuxdxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dxdx_taptamuuxdxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dxdx_taptamuuxdxdx/ajob1' in 5.3196 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dxdx_taptamuuxdxdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dxdx_taptamuuxdxdx/G21/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.2899s + [COUNTERS] Fortran Overhead ( 0 ) : 3.7972s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.4843s for 16384 events => throughput is 1.10E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0084s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dxdx_taptamuuxdxdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamguuxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamguuxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamguuxdx/ajob1' in 13.1604 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamguuxdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamguuxdx/G43/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.1815s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0859s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.0800s for 16384 events => throughput is 4.02E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0156s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamguuxdx/G335/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.8634s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7302s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.1175s for 16384 events => throughput is 3.98E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0158s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamguuxdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemuuuxux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemuuuxux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemuuuxux/ajob1' in 9.2672 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemuuuxux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemuuuxux/G689/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 9.2195s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6353s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.5574s for 16384 events => throughput is 2.50E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0268s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemuuuxux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamgguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamgguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamgguux/ajob1' in 240.5490 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamgguux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamgguux/G596/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 25.3285s + [COUNTERS] Fortran Overhead ( 0 ) : 3.5341s + [COUNTERS] CudaCpp MEs ( 2 ) : 21.7587s for 16384 events => throughput is 7.53E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0358s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamgguux/G304/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 24.8458s + [COUNTERS] Fortran Overhead ( 0 ) : 3.0411s + [COUNTERS] CudaCpp MEs ( 2 ) : 21.7693s for 16384 events => throughput is 7.53E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0354s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamgguux/G128/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 25.0609s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1576s + [COUNTERS] CudaCpp MEs ( 2 ) : 21.8678s for 16384 events => throughput is 7.49E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0355s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamgguux/G115/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 28.7043s + [COUNTERS] Fortran Overhead ( 0 ) : 6.9230s + [COUNTERS] CudaCpp MEs ( 2 ) : 21.7456s for 16384 events => throughput is 7.53E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0357s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamgguux/G599/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 25.8442s + [COUNTERS] Fortran Overhead ( 0 ) : 4.0161s + [COUNTERS] CudaCpp MEs ( 2 ) : 21.7921s for 16384 events => throughput is 7.52E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0359s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamgguux/G351/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 28.1636s + [COUNTERS] Fortran Overhead ( 0 ) : 6.3621s + [COUNTERS] CudaCpp MEs ( 2 ) : 21.7656s for 16384 events => throughput is 7.53E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0360s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamgguux/G111/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 24.3374s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4575s + [COUNTERS] CudaCpp MEs ( 2 ) : 21.8440s for 16384 events => throughput is 7.50E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0359s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamgguux/G354/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 28.7166s + [COUNTERS] Fortran Overhead ( 0 ) : 6.9244s + [COUNTERS] CudaCpp MEs ( 2 ) : 21.7565s for 16384 events => throughput is 7.53E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0356s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamgguux/G97/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 29.3562s + [COUNTERS] Fortran Overhead ( 0 ) : 7.5458s + [COUNTERS] CudaCpp MEs ( 2 ) : 21.7738s for 16384 events => throughput is 7.52E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0366s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamgguux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uxdx_epemsuxdxsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uxdx_epemsuxdxsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uxdx_epemsuxdxsx/ajob1' in 2.6095 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uxdx_epemsuxdxsx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uxdx_epemsuxdxsx/G74/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.5612s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2608s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2946s for 16384 events => throughput is 5.56E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0057s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uxdx_epemsuxdxsx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemdddxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemdddxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemdddxdx/ajob1' in 18.6877 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemdddxdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemdddxdx/G513/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 9.1741s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5567s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.5906s for 16384 events => throughput is 2.49E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0268s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemdddxdx/G180/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 9.4683s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8793s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.5620s for 16384 events => throughput is 2.50E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0270s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemdddxdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgddxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgddxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgddxdx/ajob1' in 9.9209 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgddxdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgddxdx/G515/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 9.8664s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2135s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.6262s for 16384 events => throughput is 2.47E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0267s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgddxdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguddx/ajob1' in 61.5030 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguddx/G104/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.7774s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6779s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.0838s for 16384 events => throughput is 4.01E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0157s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguddx/G120/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.3541s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2733s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.0648s for 16384 events => throughput is 4.03E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0160s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguddx/G305/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 10.2239s + [COUNTERS] Fortran Overhead ( 0 ) : 6.1279s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.0804s for 16384 events => throughput is 4.02E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0156s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguddx/G101/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.6325s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5134s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.1030s for 16384 events => throughput is 3.99E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0161s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguddx/G27/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.0477s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9511s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.0808s for 16384 events => throughput is 4.01E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0158s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguddx/G273/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.3242s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2335s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.0748s for 16384 events => throughput is 4.02E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0159s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguddx/G349/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.5176s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4407s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.0611s for 16384 events => throughput is 4.03E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0157s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguddx/G379/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.9442s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8452s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.0834s for 16384 events => throughput is 4.01E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0156s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguddx/G256/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 9.4927s + [COUNTERS] Fortran Overhead ( 0 ) : 5.3955s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.0814s for 16384 events => throughput is 4.01E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0158s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemguuxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemguuxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemguuxdx/ajob1' in 20.2524 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemguuxdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemguuxdx/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.3925s + [COUNTERS] Fortran Overhead ( 0 ) : 6.3036s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.0729s for 16384 events => throughput is 7.90E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0160s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemguuxdx/G68/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.5362s + [COUNTERS] Fortran Overhead ( 0 ) : 5.4740s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.0463s for 16384 events => throughput is 8.01E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0158s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemguuxdx/G355/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.2358s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1774s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.0423s for 16384 events => throughput is 8.02E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0161s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemguuxdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/ajob1' in 570.0340 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G14/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 23.8296s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1356s + [COUNTERS] CudaCpp MEs ( 2 ) : 21.6584s for 16384 events => throughput is 7.56E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0356s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G56/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 24.3043s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5896s + [COUNTERS] CudaCpp MEs ( 2 ) : 21.6796s for 16384 events => throughput is 7.56E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0352s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G517/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 28.1365s + [COUNTERS] Fortran Overhead ( 0 ) : 6.3590s + [COUNTERS] CudaCpp MEs ( 2 ) : 21.7421s for 16384 events => throughput is 7.54E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0353s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G52/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 24.5035s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8704s + [COUNTERS] CudaCpp MEs ( 2 ) : 21.5977s for 16384 events => throughput is 7.59E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0354s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G304/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 24.5464s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8181s + [COUNTERS] CudaCpp MEs ( 2 ) : 21.6936s for 16384 events => throughput is 7.55E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0347s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G3/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 26.7498s + [COUNTERS] Fortran Overhead ( 0 ) : 5.0553s + [COUNTERS] CudaCpp MEs ( 2 ) : 21.6587s for 16384 events => throughput is 7.56E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0358s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G23/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 23.5031s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8290s + [COUNTERS] CudaCpp MEs ( 2 ) : 21.6388s for 16384 events => throughput is 7.57E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0353s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G33/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 23.8905s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2397s + [COUNTERS] CudaCpp MEs ( 2 ) : 21.6152s for 16384 events => throughput is 7.58E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0356s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G53/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 28.6581s + [COUNTERS] Fortran Overhead ( 0 ) : 6.9762s + [COUNTERS] CudaCpp MEs ( 2 ) : 21.6464s for 16384 events => throughput is 7.57E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0354s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G495/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 23.3092s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6489s + [COUNTERS] CudaCpp MEs ( 2 ) : 21.6246s for 16384 events => throughput is 7.58E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0356s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G68/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 23.8614s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2758s + [COUNTERS] CudaCpp MEs ( 2 ) : 21.5498s for 16384 events => throughput is 7.60E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0359s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G49/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 23.8187s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1788s + [COUNTERS] CudaCpp MEs ( 2 ) : 21.6048s for 16384 events => throughput is 7.58E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0350s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G524/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 23.8743s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2500s + [COUNTERS] CudaCpp MEs ( 2 ) : 21.5882s for 16384 events => throughput is 7.59E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0361s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G65/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 23.7938s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1678s + [COUNTERS] CudaCpp MEs ( 2 ) : 21.5898s for 16384 events => throughput is 7.59E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0363s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G303/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 24.7482s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1171s + [COUNTERS] CudaCpp MEs ( 2 ) : 21.5960s for 16384 events => throughput is 7.59E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0352s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G55/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 24.4085s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8173s + [COUNTERS] CudaCpp MEs ( 2 ) : 21.5553s for 16384 events => throughput is 7.60E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0359s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G34/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 23.5806s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8936s + [COUNTERS] CudaCpp MEs ( 2 ) : 21.6513s for 16384 events => throughput is 7.57E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0357s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G509/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 29.6317s + [COUNTERS] Fortran Overhead ( 0 ) : 7.9416s + [COUNTERS] CudaCpp MEs ( 2 ) : 21.6546s for 16384 events => throughput is 7.57E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0355s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G512/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 23.5111s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8754s + [COUNTERS] CudaCpp MEs ( 2 ) : 21.6001s for 16384 events => throughput is 7.59E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0356s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G511/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 24.2856s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5185s + [COUNTERS] CudaCpp MEs ( 2 ) : 21.7318s for 16384 events => throughput is 7.54E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0353s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 24.6861s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9195s + [COUNTERS] CudaCpp MEs ( 2 ) : 21.7313s for 16384 events => throughput is 7.54E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0353s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G63/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 23.8011s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1882s + [COUNTERS] CudaCpp MEs ( 2 ) : 21.5776s for 16384 events => throughput is 7.59E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0354s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G48/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 24.1387s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4170s + [COUNTERS] CudaCpp MEs ( 2 ) : 21.6857s for 16384 events => throughput is 7.56E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0360s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/ajob1' in 375.0325 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G122/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.7285s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7959s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8963s for 16384 events => throughput is 1.50E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0362s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G108/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.3738s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4212s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.9145s for 16384 events => throughput is 1.50E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0381s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G105/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.4979s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5657s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8963s for 16384 events => throughput is 1.50E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0358s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G304/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.0821s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1673s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8787s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0361s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G301/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 18.7866s + [COUNTERS] Fortran Overhead ( 0 ) : 7.8924s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8580s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0362s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G128/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.1730s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2811s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8565s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0354s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G100/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 18.7335s + [COUNTERS] Fortran Overhead ( 0 ) : 7.8430s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8541s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0364s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G120/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.3979s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5275s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8348s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0357s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G302/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 22.0478s + [COUNTERS] Fortran Overhead ( 0 ) : 11.0973s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.9141s for 16384 events => throughput is 1.50E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0363s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G299/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.7408s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7794s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.9255s for 16384 events => throughput is 1.50E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0358s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G169/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 17.2026s + [COUNTERS] Fortran Overhead ( 0 ) : 6.2806s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8852s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0368s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G333/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 22.3500s + [COUNTERS] Fortran Overhead ( 0 ) : 11.3880s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.9265s for 16384 events => throughput is 1.50E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0355s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G594/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 20.6044s + [COUNTERS] Fortran Overhead ( 0 ) : 9.6165s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.9512s for 16384 events => throughput is 1.50E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0367s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G678/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 17.4927s + [COUNTERS] Fortran Overhead ( 0 ) : 6.5657s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8922s for 16384 events => throughput is 1.50E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0348s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G127/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.7164s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7896s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8912s for 16384 events => throughput is 1.50E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0356s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G126/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 20.7855s + [COUNTERS] Fortran Overhead ( 0 ) : 9.8796s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8680s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0380s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G598/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.1573s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2710s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8505s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0358s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G512/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.0507s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1644s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8506s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0357s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G103/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 20.2311s + [COUNTERS] Fortran Overhead ( 0 ) : 9.3484s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8467s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0360s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G107/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.3331s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4065s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8900s for 16384 events => throughput is 1.50E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0366s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G358/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.8261s + [COUNTERS] Fortran Overhead ( 0 ) : 3.8757s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.9142s for 16384 events => throughput is 1.50E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0362s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G97/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 18.5113s + [COUNTERS] Fortran Overhead ( 0 ) : 7.6453s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8305s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0356s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G179/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.7597s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8553s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8684s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0359s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_epemudcdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_epemudcdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_epemudcdx/ajob1' in 2.0865 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_epemudcdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_epemudcdx/G42/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.0584s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7613s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2912s for 16384 events => throughput is 5.63E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0059s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_epemudcdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ds_taptamggds/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ds_taptamggds/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ds_taptamggds/ajob1' in 14.6263 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ds_taptamggds/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ds_taptamggds/G17/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.8206s + [COUNTERS] Fortran Overhead ( 0 ) : 4.8205s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.9842s for 16384 events => throughput is 4.11E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0159s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ds_taptamggds/G110/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.7400s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7257s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.9982s for 16384 events => throughput is 4.10E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0161s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ds_taptamggds/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_epemggdux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_epemggdux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_epemggdux/ajob1' in 18.2525 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_epemggdux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_epemggdux/G208/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.1468s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0745s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.0566s for 16384 events => throughput is 7.97E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0157s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_epemggdux/G192/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.5193s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4479s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.0556s for 16384 events => throughput is 7.97E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0158s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_epemggdux/G380/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.7666s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6787s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.0723s for 16384 events => throughput is 7.91E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0157s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_epemggdux/G127/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.3669s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2989s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.0519s for 16384 events => throughput is 7.98E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0161s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_epemggdux/G223/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.3297s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2531s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.0607s for 16384 events => throughput is 7.95E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0159s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_epemggdux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_taptamgguc/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_taptamgguc/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_taptamgguc/ajob1' in 23.7588 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_taptamgguc/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_taptamgguc/G20/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.6027s + [COUNTERS] Fortran Overhead ( 0 ) : 4.5705s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.0164s for 16384 events => throughput is 4.08E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0159s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_taptamgguc/G110/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.4200s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3700s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.0339s for 16384 events => throughput is 4.06E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0161s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_taptamgguc/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 9.6679s + [COUNTERS] Fortran Overhead ( 0 ) : 5.6390s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.0131s for 16384 events => throughput is 4.08E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0158s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_taptamgguc/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uxcx_taptamgguxcx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uxcx_taptamgguxcx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uxcx_taptamgguxcx/ajob1' in 6.7336 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uxcx_taptamgguxcx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uxcx_taptamgguxcx/G187/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.6859s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6248s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.0454s for 16384 events => throughput is 4.05E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0157s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uxcx_taptamgguxcx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/ajob1' in 375.3845 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G499/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.1888s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2819s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8712s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0358s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G56/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.9669s + [COUNTERS] Fortran Overhead ( 0 ) : 3.0369s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8940s for 16384 events => throughput is 1.50E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0361s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G39/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.7510s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7973s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.9179s for 16384 events => throughput is 1.50E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0358s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G64/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.5554s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6293s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8901s for 16384 events => throughput is 1.50E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0361s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G13/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 12.6032s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6568s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.9103s for 16384 events => throughput is 1.50E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0360s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G33/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 12.7905s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8512s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.9030s for 16384 events => throughput is 1.50E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0363s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G43/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 18.2228s + [COUNTERS] Fortran Overhead ( 0 ) : 7.2992s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8880s for 16384 events => throughput is 1.50E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0356s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G68/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.0279s + [COUNTERS] Fortran Overhead ( 0 ) : 3.0330s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.9588s for 16384 events => throughput is 1.50E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0360s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G49/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.1434s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1964s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.9118s for 16384 events => throughput is 1.50E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0352s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G25/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 16.8868s + [COUNTERS] Fortran Overhead ( 0 ) : 5.9718s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8793s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0356s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G523/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.7609s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8285s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8959s for 16384 events => throughput is 1.50E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0365s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G395/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.4453s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4770s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.9326s for 16384 events => throughput is 1.50E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0357s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G524/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.1608s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0072s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.1173s for 16384 events => throughput is 1.47E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0363s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G65/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.0174s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0985s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8829s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0360s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G609/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.9483s + [COUNTERS] Fortran Overhead ( 0 ) : 3.0212s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8908s for 16384 events => throughput is 1.50E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0363s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G611/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.3467s + [COUNTERS] Fortran Overhead ( 0 ) : 3.3768s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.9340s for 16384 events => throughput is 1.50E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0358s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G1/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 15.6335s + [COUNTERS] Fortran Overhead ( 0 ) : 4.7380s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8601s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0353s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G55/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.5976s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6952s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8664s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0360s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G34/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.1554s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2388s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8809s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0357s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G26/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 17.4114s + [COUNTERS] Fortran Overhead ( 0 ) : 6.4938s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8823s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0353s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G58/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 16.9011s + [COUNTERS] Fortran Overhead ( 0 ) : 5.9046s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.9594s for 16384 events => throughput is 1.49E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0371s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G2/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 16.4908s + [COUNTERS] Fortran Overhead ( 0 ) : 5.3295s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.1249s for 16384 events => throughput is 1.47E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0363s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G512/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.2350s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1842s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.0152s for 16384 events => throughput is 1.49E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0356s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G51/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.7233s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7668s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.9206s for 16384 events => throughput is 1.50E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0359s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G29/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 17.8111s + [COUNTERS] Fortran Overhead ( 0 ) : 6.8963s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8794s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0354s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G48/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.0872s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1485s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.9021s for 16384 events => throughput is 1.50E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0366s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/ajob1' in 67.8541 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/G140/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.9989s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9140s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.0688s for 16384 events => throughput is 7.92E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0160s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/G114/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.7606s + [COUNTERS] Fortran Overhead ( 0 ) : 6.6938s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.0509s for 16384 events => throughput is 7.99E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0159s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/G49/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 9.8239s + [COUNTERS] Fortran Overhead ( 0 ) : 7.7279s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.0798s for 16384 events => throughput is 7.88E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0162s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/G120/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.1555s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0837s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.0561s for 16384 events => throughput is 7.97E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0156s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/G262/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.8913s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8191s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.0563s for 16384 events => throughput is 7.97E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0159s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/G27/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.0306s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9557s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.0591s for 16384 events => throughput is 7.96E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0158s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/G44/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.7055s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6346s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.0551s for 16384 events => throughput is 7.97E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0158s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/G96/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.7480s + [COUNTERS] Fortran Overhead ( 0 ) : 3.6373s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.0935s for 16384 events => throughput is 7.83E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0171s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/G261/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.2403s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1719s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.0526s for 16384 events => throughput is 7.98E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0158s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/G37/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.1892s + [COUNTERS] Fortran Overhead ( 0 ) : 5.1141s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.0591s for 16384 events => throughput is 7.96E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0160s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/G28/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.9210s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8579s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.0470s for 16384 events => throughput is 8.00E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0161s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.1310s + [COUNTERS] Fortran Overhead ( 0 ) : 6.0751s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.0399s for 16384 events => throughput is 8.03E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0161s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_epemuduxsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_epemuduxsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_epemuduxsx/ajob1' in 3.0139 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_epemuduxsx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_epemuduxsx/G42/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.9834s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6832s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2945s for 16384 events => throughput is 5.56E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0057s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_epemuduxsx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dd_epemggdd/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dd_epemggdd/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dd_epemggdd/ajob1' in 25.3876 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dd_epemggdd/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dd_epemggdd/G467/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.4962s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8913s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.5778s for 16384 events => throughput is 2.49E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0270s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dd_epemggdd/G351/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.8965s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3056s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.5637s for 16384 events => throughput is 2.50E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0272s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dd_epemggdd/G44/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.9064s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3021s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.5773s for 16384 events => throughput is 2.49E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0270s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dd_epemggdd/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamusdxsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamusdxsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamusdxsx/ajob1' in 1.9046 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamusdxsx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamusdxsx/G92/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.8597s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2590s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5949s for 16384 events => throughput is 2.75E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0058s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamusdxsx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemggddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemggddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemggddx/ajob1' in 58.0161 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemggddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemggddx/G175/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.7983s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1813s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.5904s for 16384 events => throughput is 2.49E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0267s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemggddx/G303/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 9.4162s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8163s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.5730s for 16384 events => throughput is 2.49E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0269s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemggddx/G111/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.4707s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8478s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.5960s for 16384 events => throughput is 2.48E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0269s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemggddx/G219/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 9.7667s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1556s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.5842s for 16384 events => throughput is 2.49E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0270s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemggddx/G318/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 12.7847s + [COUNTERS] Fortran Overhead ( 0 ) : 6.1902s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.5674s for 16384 events => throughput is 2.49E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0270s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemggddx/G143/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.6445s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0179s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.5997s for 16384 events => throughput is 2.48E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0269s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemggddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemguuxux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemguuxux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemguuxux/ajob1' in 13.9836 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemguuxux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemguuxux/G615/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.9284s + [COUNTERS] Fortran Overhead ( 0 ) : 7.3307s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.5706s for 16384 events => throughput is 2.49E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0271s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemguuxux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamgggg/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamgggg/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamgggg/ajob1' in 24.4557 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamgggg/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamgggg/G66/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 24.4065s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7520s + [COUNTERS] CudaCpp MEs ( 2 ) : 21.6183s for 16384 events => throughput is 7.58E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0362s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamgggg/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgduxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgduxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgduxdx/ajob1' in 6.3121 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgduxdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgduxdx/G44/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.2808s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1756s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.0893s for 16384 events => throughput is 4.01E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0159s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgduxdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgudux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgudux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgudux/ajob1' in 52.9725 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgudux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgudux/G39/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.4223s + [COUNTERS] Fortran Overhead ( 0 ) : 6.3627s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.0434s for 16384 events => throughput is 8.02E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0162s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgudux/G366/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.9109s + [COUNTERS] Fortran Overhead ( 0 ) : 4.8639s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.0312s for 16384 events => throughput is 8.07E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0158s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgudux/G259/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.4696s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4202s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.0337s for 16384 events => throughput is 8.06E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0157s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgudux/G28/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.2034s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1619s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.0260s for 16384 events => throughput is 8.09E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0155s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgudux/G32/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.5847s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5405s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.0285s for 16384 events => throughput is 8.08E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0157s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgudux/G61/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 10.7090s + [COUNTERS] Fortran Overhead ( 0 ) : 8.6625s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.0308s for 16384 events => throughput is 8.07E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0156s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgudux/G267/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.9616s + [COUNTERS] Fortran Overhead ( 0 ) : 5.9173s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.0286s for 16384 events => throughput is 8.08E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0157s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgudux/G144/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.5505s + [COUNTERS] Fortran Overhead ( 0 ) : 3.5031s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.0307s for 16384 events => throughput is 8.07E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0166s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgudux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemuudux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemuudux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemuudux/ajob1' in 24.3335 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemuudux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemuudux/G104/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 24.3050s + [COUNTERS] Fortran Overhead ( 0 ) : 20.4854s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.8108s for 81920 events => throughput is 2.15E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0088s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemuudux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamggddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamggddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamggddx/ajob1' in 199.9854 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamggddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamggddx/G108/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 24.0671s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2882s + [COUNTERS] CudaCpp MEs ( 2 ) : 21.7432s for 16384 events => throughput is 7.54E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0357s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamggddx/G304/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 24.6518s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8597s + [COUNTERS] CudaCpp MEs ( 2 ) : 21.7561s for 16384 events => throughput is 7.53E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0360s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamggddx/G128/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 24.8494s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9630s + [COUNTERS] CudaCpp MEs ( 2 ) : 21.8502s for 16384 events => throughput is 7.50E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0362s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamggddx/G360/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 24.2844s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4527s + [COUNTERS] CudaCpp MEs ( 2 ) : 21.7955s for 16384 events => throughput is 7.52E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0362s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamggddx/G169/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 28.3772s + [COUNTERS] Fortran Overhead ( 0 ) : 6.3573s + [COUNTERS] CudaCpp MEs ( 2 ) : 21.9847s for 16384 events => throughput is 7.45E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0352s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamggddx/G176/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 23.9024s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0745s + [COUNTERS] CudaCpp MEs ( 2 ) : 21.7924s for 16384 events => throughput is 7.52E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0356s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamggddx/G303/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 24.7670s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9475s + [COUNTERS] CudaCpp MEs ( 2 ) : 21.7836s for 16384 events => throughput is 7.52E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0359s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamggddx/G513/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 24.9241s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2106s + [COUNTERS] CudaCpp MEs ( 2 ) : 21.6779s for 16384 events => throughput is 7.56E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0355s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamggddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguccx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguccx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguccx/ajob1' in 24.6367 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguccx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguccx/G283/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.1224s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0869s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.0199s for 16384 events => throughput is 8.11E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0156s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguccx/G47/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.0567s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0179s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.0229s for 16384 events => throughput is 8.10E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0160s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguccx/G35/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.9086s + [COUNTERS] Fortran Overhead ( 0 ) : 6.8714s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.0215s for 16384 events => throughput is 8.10E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0157s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguccx/G373/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.7975s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7650s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.0166s for 16384 events => throughput is 8.12E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0159s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguccx/G249/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.6308s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5919s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.0229s for 16384 events => throughput is 8.10E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0160s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguccx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/ajob1' in 436.4202 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G14/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.0048s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0968s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8724s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0355s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G499/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 12.9135s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0221s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8554s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0361s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G56/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.7556s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8186s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.9013s for 16384 events => throughput is 1.50E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0357s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G39/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.2148s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3433s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8359s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0355s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G52/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.7797s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9077s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8365s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0354s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G304/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.6328s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7116s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8858s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0354s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G64/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.2749s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4120s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8268s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0360s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G3/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 16.0402s + [COUNTERS] Fortran Overhead ( 0 ) : 5.1501s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8541s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0360s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G13/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 12.5398s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7076s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.7965s for 16384 events => throughput is 1.52E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0357s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G33/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.0785s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2322s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8110s for 16384 events => throughput is 1.52E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0353s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G484/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.5398s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6630s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8409s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0359s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G68/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.1143s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2506s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8282s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0355s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G49/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 12.7143s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8477s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8308s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0358s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G40/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.7918s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9474s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8082s for 16384 events => throughput is 1.52E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0362s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G523/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.0277s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1836s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8087s for 16384 events => throughput is 1.52E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0354s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 12.4690s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6007s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8325s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0358s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G524/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.0137s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1703s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8079s for 16384 events => throughput is 1.52E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0355s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G65/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.1404s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2410s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8621s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0372s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G44/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 16.8130s + [COUNTERS] Fortran Overhead ( 0 ) : 5.9591s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8183s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0356s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 12.4082s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5457s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8263s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0361s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G303/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.9683s + [COUNTERS] Fortran Overhead ( 0 ) : 3.0695s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8632s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0356s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G528/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.2022s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2596s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.9072s for 16384 events => throughput is 1.50E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0354s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G55/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.9147s + [COUNTERS] Fortran Overhead ( 0 ) : 3.0447s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8346s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0353s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G298/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.0364s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1374s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8629s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0361s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G34/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 12.6621s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7554s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8711s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0356s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G54/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 18.2176s + [COUNTERS] Fortran Overhead ( 0 ) : 7.3360s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8458s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0358s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G67/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.2272s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3739s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8172s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0360s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G512/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 12.8926s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0518s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8055s for 16384 events => throughput is 1.52E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0354s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G511/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.6087s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7029s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8695s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0362s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.5681s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6166s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.9151s for 16384 events => throughput is 1.50E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0365s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G21/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 12.7329s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8403s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8574s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0351s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G48/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.5037s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6100s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8580s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0357s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_taptamggdux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_taptamggdux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_taptamggdux/ajob1' in 25.9645 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_taptamggdux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_taptamggdux/G68/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.2100s + [COUNTERS] Fortran Overhead ( 0 ) : 4.0508s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.1434s for 16384 events => throughput is 3.95E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0158s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_taptamggdux/G167/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.9976s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8488s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.1331s for 16384 events => throughput is 3.96E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0157s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_taptamggdux/G92/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.9787s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8155s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.1474s for 16384 events => throughput is 3.95E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0158s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_taptamggdux/G60/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.6931s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5363s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.1409s for 16384 events => throughput is 3.96E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0159s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_taptamggdux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/ajob1' in 315.2647 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G56/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.8380s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9329s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8696s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0356s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G39/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.1203s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1618s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.9223s for 16384 events => throughput is 1.50E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0361s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G517/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 19.3113s + [COUNTERS] Fortran Overhead ( 0 ) : 8.3992s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8762s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0359s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G497/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 12.9547s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0134s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.9051s for 16384 events => throughput is 1.50E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0362s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G13/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 12.6798s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7618s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8824s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0356s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G484/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 17.7658s + [COUNTERS] Fortran Overhead ( 0 ) : 6.8226s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.9076s for 16384 events => throughput is 1.50E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0356s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G68/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.1116s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1707s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.9058s for 16384 events => throughput is 1.50E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0352s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G49/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.4087s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4364s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.9368s for 16384 events => throughput is 1.50E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0356s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G523/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.5314s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6082s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8874s for 16384 events => throughput is 1.50E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0357s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G518/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 17.0402s + [COUNTERS] Fortran Overhead ( 0 ) : 6.0778s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.9268s for 16384 events => throughput is 1.50E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0357s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G42/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 19.2982s + [COUNTERS] Fortran Overhead ( 0 ) : 8.3550s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.9076s for 16384 events => throughput is 1.50E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0356s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G524/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 12.9749s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0083s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.9303s for 16384 events => throughput is 1.50E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0363s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G610/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.1464s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2074s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.9033s for 16384 events => throughput is 1.50E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0357s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G26/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 17.1028s + [COUNTERS] Fortran Overhead ( 0 ) : 6.1672s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8998s for 16384 events => throughput is 1.50E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0358s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G58/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 18.4234s + [COUNTERS] Fortran Overhead ( 0 ) : 7.4917s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8961s for 16384 events => throughput is 1.50E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0357s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G511/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.1532s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2068s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.9107s for 16384 events => throughput is 1.50E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0357s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.8915s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9587s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8959s for 16384 events => throughput is 1.50E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0368s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G32/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 19.2849s + [COUNTERS] Fortran Overhead ( 0 ) : 8.3048s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.9444s for 16384 events => throughput is 1.50E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0357s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G61/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 17.2697s + [COUNTERS] Fortran Overhead ( 0 ) : 6.3208s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.9133s for 16384 events => throughput is 1.50E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0357s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G390/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 20.5531s + [COUNTERS] Fortran Overhead ( 0 ) : 9.6069s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.9099s for 16384 events => throughput is 1.50E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0363s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_epemgguc/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_epemgguc/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_epemgguc/ajob1' in 18.1072 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_epemgguc/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_epemgguc/G164/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.5856s + [COUNTERS] Fortran Overhead ( 0 ) : 4.5715s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.9985s for 16384 events => throughput is 8.20E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0155s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_epemgguc/G24/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.1396s + [COUNTERS] Fortran Overhead ( 0 ) : 4.1152s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.0085s for 16384 events => throughput is 8.16E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0159s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_epemgguc/G162/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.2963s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2823s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.9986s for 16384 events => throughput is 8.20E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0155s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_epemgguc/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/ajob1' in 403.9237 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G336/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.5419s + [COUNTERS] Fortran Overhead ( 0 ) : 3.6761s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8300s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0358s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G108/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.4330s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5693s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8278s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0358s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G105/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.3820s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4309s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.9153s for 16384 events => throughput is 1.50E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0358s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G304/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.2213s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2411s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.9446s for 16384 events => throughput is 1.50E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0357s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G175/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.1842s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2730s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8754s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0358s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G128/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.5853s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6391s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.9105s for 16384 events => throughput is 1.50E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0357s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G360/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.5784s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5956s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.9462s for 16384 events => throughput is 1.50E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0366s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G120/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.2017s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3114s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8551s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0353s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G296/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.0488s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1366s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8769s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0353s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G299/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.7813s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8973s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8479s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0362s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G715/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.6090s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7032s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8704s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0354s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G300/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.3411s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4347s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8703s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0361s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G303/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.8243s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9210s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8676s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0357s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G127/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.8165s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9222s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8584s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0359s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G111/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.5705s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6106s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.9245s for 16384 events => throughput is 1.50E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0355s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G126/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 21.6920s + [COUNTERS] Fortran Overhead ( 0 ) : 10.7851s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8712s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0356s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G294/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 17.7584s + [COUNTERS] Fortran Overhead ( 0 ) : 6.8528s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8699s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0357s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G598/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.3177s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4675s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8151s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0351s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G610/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.2708s + [COUNTERS] Fortran Overhead ( 0 ) : 3.3851s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8508s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0349s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G178/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.5605s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6613s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8633s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0360s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G121/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.4680s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5165s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.9138s for 16384 events => throughput is 1.50E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0376s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G685/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.1420s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2460s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8602s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0357s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G690/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.0056s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1253s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8442s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0361s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G103/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 19.2644s + [COUNTERS] Fortran Overhead ( 0 ) : 8.3461s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8822s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0361s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G107/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.2684s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3774s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8547s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0363s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G180/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.3921s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5342s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8212s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0367s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G97/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 19.3365s + [COUNTERS] Fortran Overhead ( 0 ) : 8.4190s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8814s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0360s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G179/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.7702s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8228s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.9112s for 16384 events => throughput is 1.50E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0363s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamggddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamggddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamggddx/ajob1' in 6.6614 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamggddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamggddx/G189/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.6328s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5589s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.0577s for 16384 events => throughput is 4.04E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0162s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamggddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dd_taptamggdd/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dd_taptamggdd/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dd_taptamggdd/ajob1' in 15.4728 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dd_taptamggdd/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dd_taptamggdd/G463/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 15.4435s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1922s + [COUNTERS] CudaCpp MEs ( 2 ) : 13.2242s for 16384 events => throughput is 1.24E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0272s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dd_taptamggdd/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/ajob1' in 1015.2610 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G506/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 26.1401s + [COUNTERS] Fortran Overhead ( 0 ) : 4.3680s + [COUNTERS] CudaCpp MEs ( 2 ) : 21.7362s for 16384 events => throughput is 7.54E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0359s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G499/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 23.6503s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8947s + [COUNTERS] CudaCpp MEs ( 2 ) : 21.7198s for 16384 events => throughput is 7.54E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0358s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G46/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 29.4297s + [COUNTERS] Fortran Overhead ( 0 ) : 7.6685s + [COUNTERS] CudaCpp MEs ( 2 ) : 21.7253s for 16384 events => throughput is 7.54E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0359s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G56/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 24.4912s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5955s + [COUNTERS] CudaCpp MEs ( 2 ) : 21.8598s for 16384 events => throughput is 7.50E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0360s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G39/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 24.1467s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4440s + [COUNTERS] CudaCpp MEs ( 2 ) : 21.6663s for 16384 events => throughput is 7.56E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0363s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G517/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 27.0004s + [COUNTERS] Fortran Overhead ( 0 ) : 5.2389s + [COUNTERS] CudaCpp MEs ( 2 ) : 21.7257s for 16384 events => throughput is 7.54E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0357s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G52/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 24.7827s + [COUNTERS] Fortran Overhead ( 0 ) : 3.0328s + [COUNTERS] CudaCpp MEs ( 2 ) : 21.7138s for 16384 events => throughput is 7.55E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0361s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G31/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 28.6713s + [COUNTERS] Fortran Overhead ( 0 ) : 6.9102s + [COUNTERS] CudaCpp MEs ( 2 ) : 21.7254s for 16384 events => throughput is 7.54E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0357s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G324/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 28.1488s + [COUNTERS] Fortran Overhead ( 0 ) : 6.4893s + [COUNTERS] CudaCpp MEs ( 2 ) : 21.6237s for 16384 events => throughput is 7.58E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0358s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G64/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 23.6930s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8652s + [COUNTERS] CudaCpp MEs ( 2 ) : 21.7920s for 16384 events => throughput is 7.52E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0358s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G3/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 25.6955s + [COUNTERS] Fortran Overhead ( 0 ) : 4.0090s + [COUNTERS] CudaCpp MEs ( 2 ) : 21.6507s for 16384 events => throughput is 7.57E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0358s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G33/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 23.9913s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2380s + [COUNTERS] CudaCpp MEs ( 2 ) : 21.7175s for 16384 events => throughput is 7.54E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0358s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G53/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 28.3397s + [COUNTERS] Fortran Overhead ( 0 ) : 6.6248s + [COUNTERS] CudaCpp MEs ( 2 ) : 21.6790s for 16384 events => throughput is 7.56E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0358s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G43/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 27.2966s + [COUNTERS] Fortran Overhead ( 0 ) : 5.5248s + [COUNTERS] CudaCpp MEs ( 2 ) : 21.7361s for 16384 events => throughput is 7.54E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0357s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G49/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 23.5601s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8365s + [COUNTERS] CudaCpp MEs ( 2 ) : 21.6873s for 16384 events => throughput is 7.55E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0362s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G40/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 24.5964s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9221s + [COUNTERS] CudaCpp MEs ( 2 ) : 21.6384s for 16384 events => throughput is 7.57E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0358s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G523/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 23.9696s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2160s + [COUNTERS] CudaCpp MEs ( 2 ) : 21.7178s for 16384 events => throughput is 7.54E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0359s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G518/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 27.2692s + [COUNTERS] Fortran Overhead ( 0 ) : 5.5361s + [COUNTERS] CudaCpp MEs ( 2 ) : 21.6956s for 16384 events => throughput is 7.55E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0375s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G35/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 23.6591s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9639s + [COUNTERS] CudaCpp MEs ( 2 ) : 21.6581s for 16384 events => throughput is 7.56E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0371s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 23.3746s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6620s + [COUNTERS] CudaCpp MEs ( 2 ) : 21.6773s for 16384 events => throughput is 7.56E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0353s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G42/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 27.3077s + [COUNTERS] Fortran Overhead ( 0 ) : 5.5576s + [COUNTERS] CudaCpp MEs ( 2 ) : 21.7146s for 16384 events => throughput is 7.55E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0355s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G524/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 23.8924s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1891s + [COUNTERS] CudaCpp MEs ( 2 ) : 21.6680s for 16384 events => throughput is 7.56E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0354s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G65/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 23.7315s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9365s + [COUNTERS] CudaCpp MEs ( 2 ) : 21.7599s for 16384 events => throughput is 7.53E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0352s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G27/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 27.6887s + [COUNTERS] Fortran Overhead ( 0 ) : 5.8901s + [COUNTERS] CudaCpp MEs ( 2 ) : 21.7630s for 16384 events => throughput is 7.53E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0356s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G44/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 28.0453s + [COUNTERS] Fortran Overhead ( 0 ) : 6.3803s + [COUNTERS] CudaCpp MEs ( 2 ) : 21.6290s for 16384 events => throughput is 7.58E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0359s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 23.6603s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6840s + [COUNTERS] CudaCpp MEs ( 2 ) : 21.9400s for 16384 events => throughput is 7.47E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0363s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G303/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 24.9636s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2249s + [COUNTERS] CudaCpp MEs ( 2 ) : 21.7030s for 16384 events => throughput is 7.55E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0356s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G55/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 24.5599s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7740s + [COUNTERS] CudaCpp MEs ( 2 ) : 21.7502s for 16384 events => throughput is 7.53E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0357s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G34/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 23.6024s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8325s + [COUNTERS] CudaCpp MEs ( 2 ) : 21.7338s for 16384 events => throughput is 7.54E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0361s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G26/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 27.7132s + [COUNTERS] Fortran Overhead ( 0 ) : 6.0192s + [COUNTERS] CudaCpp MEs ( 2 ) : 21.6585s for 16384 events => throughput is 7.56E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0355s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G62/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 27.8106s + [COUNTERS] Fortran Overhead ( 0 ) : 6.0868s + [COUNTERS] CudaCpp MEs ( 2 ) : 21.6877s for 16384 events => throughput is 7.55E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0360s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G37/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 23.9562s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2091s + [COUNTERS] CudaCpp MEs ( 2 ) : 21.7115s for 16384 events => throughput is 7.55E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0356s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G485/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 27.1855s + [COUNTERS] Fortran Overhead ( 0 ) : 5.3962s + [COUNTERS] CudaCpp MEs ( 2 ) : 21.7533s for 16384 events => throughput is 7.53E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0360s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G67/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 24.0125s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2147s + [COUNTERS] CudaCpp MEs ( 2 ) : 21.7618s for 16384 events => throughput is 7.53E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0360s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G512/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 23.5686s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8628s + [COUNTERS] CudaCpp MEs ( 2 ) : 21.6697s for 16384 events => throughput is 7.56E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0361s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G511/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 24.2446s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5181s + [COUNTERS] CudaCpp MEs ( 2 ) : 21.6900s for 16384 events => throughput is 7.55E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0365s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G51/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 24.2529s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5357s + [COUNTERS] CudaCpp MEs ( 2 ) : 21.6811s for 16384 events => throughput is 7.56E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0360s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 24.1886s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4666s + [COUNTERS] CudaCpp MEs ( 2 ) : 21.6860s for 16384 events => throughput is 7.56E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0360s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G63/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 24.0820s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3521s + [COUNTERS] CudaCpp MEs ( 2 ) : 21.6940s for 16384 events => throughput is 7.55E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0360s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G48/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 24.1069s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4221s + [COUNTERS] CudaCpp MEs ( 2 ) : 21.6487s for 16384 events => throughput is 7.57E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0361s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgggux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgggux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgggux/ajob1' in 182.3961 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgggux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgggux/G33/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 23.6232s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8071s + [COUNTERS] CudaCpp MEs ( 2 ) : 21.7776s for 16384 events => throughput is 7.52E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0385s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgggux/G68/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 24.9978s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1817s + [COUNTERS] CudaCpp MEs ( 2 ) : 21.7801s for 16384 events => throughput is 7.52E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0359s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgggux/G25/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 28.4097s + [COUNTERS] Fortran Overhead ( 0 ) : 6.5294s + [COUNTERS] CudaCpp MEs ( 2 ) : 21.8440s for 16384 events => throughput is 7.50E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0362s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgggux/G494/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 23.4371s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6597s + [COUNTERS] CudaCpp MEs ( 2 ) : 21.7431s for 16384 events => throughput is 7.54E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0343s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgggux/G55/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 24.8515s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9576s + [COUNTERS] CudaCpp MEs ( 2 ) : 21.8580s for 16384 events => throughput is 7.50E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0359s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgggux/G509/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 28.9937s + [COUNTERS] Fortran Overhead ( 0 ) : 7.1570s + [COUNTERS] CudaCpp MEs ( 2 ) : 21.8012s for 16384 events => throughput is 7.52E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0356s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgggux/G29/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 27.9261s + [COUNTERS] Fortran Overhead ( 0 ) : 6.1186s + [COUNTERS] CudaCpp MEs ( 2 ) : 21.7722s for 16384 events => throughput is 7.53E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0353s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgggux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamgguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamgguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamgguux/ajob1' in 80.3063 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamgguux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamgguux/G260/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.8424s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5966s + [COUNTERS] CudaCpp MEs ( 2 ) : 13.2185s for 16384 events => throughput is 1.24E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0273s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamgguux/G447/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.8431s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6192s + [COUNTERS] CudaCpp MEs ( 2 ) : 13.1969s for 16384 events => throughput is 1.24E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0271s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamgguux/G229/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 17.3886s + [COUNTERS] Fortran Overhead ( 0 ) : 4.1190s + [COUNTERS] CudaCpp MEs ( 2 ) : 13.2424s for 16384 events => throughput is 1.24E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0272s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamgguux/G135/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 15.3528s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0286s + [COUNTERS] CudaCpp MEs ( 2 ) : 13.2963s for 16384 events => throughput is 1.23E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0278s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamgguux/G230/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 17.7581s + [COUNTERS] Fortran Overhead ( 0 ) : 4.4375s + [COUNTERS] CudaCpp MEs ( 2 ) : 13.2936s for 16384 events => throughput is 1.23E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0271s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamgguux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdddx/ajob1' in 17.5181 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdddx/G715/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.7455s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1237s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.5948s for 16384 events => throughput is 2.48E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0270s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdddx/G111/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.7032s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0826s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.5935s for 16384 events => throughput is 2.48E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0271s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemusdxsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemusdxsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemusdxsx/ajob1' in 6.6525 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemusdxsx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemusdxsx/G67/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.5996s + [COUNTERS] Fortran Overhead ( 0 ) : 6.3000s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2939s for 16384 events => throughput is 5.58E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0057s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemusdxsx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdssx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdssx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdssx/ajob1' in 24.6049 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdssx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdssx/G31/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.0062s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9673s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.0233s for 16384 events => throughput is 8.10E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0157s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdssx/G27/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.9971s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9683s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.0130s for 16384 events => throughput is 8.14E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0158s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdssx/G119/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.2410s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2154s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.0101s for 16384 events => throughput is 8.15E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0155s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdssx/G284/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.4828s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4372s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.0299s for 16384 events => throughput is 8.07E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0158s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdssx/G375/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.6538s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6152s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.0227s for 16384 events => throughput is 8.10E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0159s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdssx/G102/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.0811s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0580s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.0074s for 16384 events => throughput is 8.16E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0157s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdssx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_epemgguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_epemgguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_epemgguux/ajob1' in 40.1508 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_epemgguux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_epemgguux/G105/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.2264s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5794s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.6199s for 16384 events => throughput is 2.47E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0272s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_epemgguux/G223/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.9292s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2767s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.6245s for 16384 events => throughput is 2.47E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0280s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_epemgguux/G103/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 11.2485s + [COUNTERS] Fortran Overhead ( 0 ) : 4.5632s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.6579s for 16384 events => throughput is 2.46E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0275s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_epemgguux/G5/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 11.6409s + [COUNTERS] Fortran Overhead ( 0 ) : 4.9877s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.6254s for 16384 events => throughput is 2.47E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0278s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_epemgguux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/ajob1' in 921.4226 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G506/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 16.4825s + [COUNTERS] Fortran Overhead ( 0 ) : 5.5802s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8664s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0358s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G499/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 12.8212s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9230s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8626s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0356s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G38/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.7242s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8174s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8710s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0358s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G46/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 18.8856s + [COUNTERS] Fortran Overhead ( 0 ) : 7.9614s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8886s for 16384 events => throughput is 1.50E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0356s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G59/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 17.2814s + [COUNTERS] Fortran Overhead ( 0 ) : 6.3788s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8665s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0362s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G56/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.7155s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8429s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8369s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0357s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G525/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.3608s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4597s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8655s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0355s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G39/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.3095s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3777s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8953s for 16384 events => throughput is 1.50E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0365s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G517/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 16.6843s + [COUNTERS] Fortran Overhead ( 0 ) : 5.8102s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8383s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0357s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G487/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 15.8236s + [COUNTERS] Fortran Overhead ( 0 ) : 4.9187s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8692s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0357s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G52/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.1701s + [COUNTERS] Fortran Overhead ( 0 ) : 3.0322s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.1011s for 16384 events => throughput is 1.48E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0369s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G31/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 18.8422s + [COUNTERS] Fortran Overhead ( 0 ) : 7.9599s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8466s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0357s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G64/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 12.9561s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0433s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8771s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0357s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G3/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 15.9460s + [COUNTERS] Fortran Overhead ( 0 ) : 5.0449s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8655s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0356s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G497/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.2177s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3249s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8573s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0355s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G13/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 12.6286s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6959s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8973s for 16384 events => throughput is 1.50E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0355s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G66/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.3952s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4703s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8890s for 16384 events => throughput is 1.50E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0359s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G33/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 12.9262s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0016s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8894s for 16384 events => throughput is 1.50E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0351s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G53/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 19.1024s + [COUNTERS] Fortran Overhead ( 0 ) : 8.2112s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8558s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0354s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G43/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 15.8590s + [COUNTERS] Fortran Overhead ( 0 ) : 4.9553s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8681s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0357s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G68/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.1232s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2139s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8739s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0354s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G49/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 12.8333s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9567s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8417s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0349s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G302/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 19.3636s + [COUNTERS] Fortran Overhead ( 0 ) : 8.4679s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8591s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0366s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G25/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 17.2807s + [COUNTERS] Fortran Overhead ( 0 ) : 6.3539s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8911s for 16384 events => throughput is 1.50E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0357s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G40/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.0129s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1330s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8437s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0361s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G523/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 12.9576s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0674s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8544s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0359s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G518/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 16.2024s + [COUNTERS] Fortran Overhead ( 0 ) : 5.2639s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.9021s for 16384 events => throughput is 1.50E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0364s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G35/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.0145s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1540s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8248s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0357s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 12.5975s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6784s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8834s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0357s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G524/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.0063s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1381s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8288s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0394s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G65/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.0738s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1641s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8742s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0355s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G27/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 17.4456s + [COUNTERS] Fortran Overhead ( 0 ) : 6.5442s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8644s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0371s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G44/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 17.0648s + [COUNTERS] Fortran Overhead ( 0 ) : 6.1919s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8371s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0358s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G483/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 15.2070s + [COUNTERS] Fortran Overhead ( 0 ) : 4.3029s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8685s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0356s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G482/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 16.6342s + [COUNTERS] Fortran Overhead ( 0 ) : 5.7270s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8715s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0356s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G1/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 15.5763s + [COUNTERS] Fortran Overhead ( 0 ) : 4.6325s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.9077s for 16384 events => throughput is 1.50E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0362s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 12.4759s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5889s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8515s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0354s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G55/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.4142s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5226s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8557s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0359s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G298/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.9675s + [COUNTERS] Fortran Overhead ( 0 ) : 3.0850s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8461s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0364s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G513/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.8364s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9269s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8736s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0359s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G60/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 18.5741s + [COUNTERS] Fortran Overhead ( 0 ) : 7.6516s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8862s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0362s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G34/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.0021s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0359s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.9314s for 16384 events => throughput is 1.50E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0348s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G26/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 17.9526s + [COUNTERS] Fortran Overhead ( 0 ) : 7.0333s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8820s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0372s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G58/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 16.4779s + [COUNTERS] Fortran Overhead ( 0 ) : 5.6093s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8331s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0356s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G54/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 17.0124s + [COUNTERS] Fortran Overhead ( 0 ) : 6.1125s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8647s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0353s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G62/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 17.9711s + [COUNTERS] Fortran Overhead ( 0 ) : 7.0704s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8649s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0358s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G37/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 12.8212s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9160s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8697s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0354s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G493/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.0757s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1835s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8563s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0359s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G4/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 17.2715s + [COUNTERS] Fortran Overhead ( 0 ) : 6.3647s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8705s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0362s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G50/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.0411s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1510s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8549s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0351s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G485/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 16.7760s + [COUNTERS] Fortran Overhead ( 0 ) : 5.8837s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8567s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0355s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G41/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 18.1580s + [COUNTERS] Fortran Overhead ( 0 ) : 7.2321s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8887s for 16384 events => throughput is 1.50E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0372s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G67/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.0286s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1660s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8271s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0355s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G512/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 12.9901s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0744s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8801s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0356s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G511/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.5098s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5791s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8952s for 16384 events => throughput is 1.50E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0355s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G57/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 17.9202s + [COUNTERS] Fortran Overhead ( 0 ) : 7.0320s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8526s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0356s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.8033s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9424s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8257s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0352s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G63/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.3943s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4548s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.9039s for 16384 events => throughput is 1.50E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0356s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G32/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 18.7272s + [COUNTERS] Fortran Overhead ( 0 ) : 7.8133s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8782s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0357s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G29/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 17.2364s + [COUNTERS] Fortran Overhead ( 0 ) : 6.3294s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8713s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0357s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G48/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.2861s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3866s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8632s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0362s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_taptamggdsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_taptamggdsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_taptamggdsx/ajob1' in 19.0168 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_taptamggdsx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_taptamggdsx/G92/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.5435s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4781s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.0495s for 16384 events => throughput is 4.05E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0158s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_taptamggdsx/G144/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.2025s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1413s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.0456s for 16384 events => throughput is 4.05E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0157s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_taptamggdsx/G207/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.2021s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1052s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.0801s for 16384 events => throughput is 4.02E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0168s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_taptamggdsx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_epemggdsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_epemggdsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_epemggdsx/ajob1' in 8.9959 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_epemggdsx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_epemggdsx/G205/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.9672s + [COUNTERS] Fortran Overhead ( 0 ) : 6.9423s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.0093s for 16384 events => throughput is 8.15E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0155s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_epemggdsx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_taptamgggg/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_taptamgggg/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_taptamgggg/ajob1' in 24.5880 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_taptamgggg/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_taptamgggg/G67/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 24.5587s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7954s + [COUNTERS] CudaCpp MEs ( 2 ) : 21.7271s for 16384 events => throughput is 7.54E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0362s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_taptamgggg/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dc_epemdcssx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dc_epemdcssx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dc_epemdcssx/ajob1' in 2.5910 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dc_epemdcssx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dc_epemdcssx/G77/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.5479s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2465s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2958s for 16384 events => throughput is 5.54E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0057s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dc_epemdcssx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguccx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguccx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguccx/ajob1' in 6.6263 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguccx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguccx/G48/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.5919s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5162s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.0600s for 16384 events => throughput is 4.04E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0158s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguccx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ds_epemggds/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ds_epemggds/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ds_epemggds/ajob1' in 3.7426 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ds_epemggds/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ds_epemggds/G26/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.7145s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6747s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.0239s for 16384 events => throughput is 8.10E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0159s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ds_epemggds/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ucx_epemggucx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ucx_epemggucx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ucx_epemggucx/ajob1' in 9.8794 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ucx_epemggucx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ucx_epemggucx/G163/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.1697s + [COUNTERS] Fortran Overhead ( 0 ) : 4.1416s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.0123s for 16384 events => throughput is 8.14E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0157s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ucx_epemggucx/G60/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.6374s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6025s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.0191s for 16384 events => throughput is 8.11E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0158s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ucx_epemggucx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemggudx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemggudx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemggudx/ajob1' in 66.6323 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemggudx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemggudx/G109/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.5981s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5382s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.0441s for 16384 events => throughput is 8.02E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0158s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemggudx/G151/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.9352s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8798s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.0388s for 16384 events => throughput is 8.04E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0167s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemggudx/G106/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.3441s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2921s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.0364s for 16384 events => throughput is 8.05E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0155s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemggudx/G35/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 9.0808s + [COUNTERS] Fortran Overhead ( 0 ) : 7.0164s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.0485s for 16384 events => throughput is 8.00E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0159s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemggudx/G119/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.1338s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0546s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.0629s for 16384 events => throughput is 7.94E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0164s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemggudx/G187/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.9404s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8896s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.0350s for 16384 events => throughput is 8.05E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0159s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemggudx/G91/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.0886s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0391s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.0339s for 16384 events => throughput is 8.06E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0156s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemggudx/G363/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 25.7636s + [COUNTERS] Fortran Overhead ( 0 ) : 17.6048s + [COUNTERS] CudaCpp MEs ( 2 ) : 8.1429s for 65536 events => throughput is 8.05E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0159s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemggudx/G149/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.5703s + [COUNTERS] Fortran Overhead ( 0 ) : 4.5195s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.0353s for 16384 events => throughput is 8.05E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0155s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemggudx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgcuxcx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgcuxcx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgcuxcx/ajob1' in 4.5883 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgcuxcx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgcuxcx/G120/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.5401s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5052s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.0192s for 16384 events => throughput is 8.11E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0156s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgcuxcx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemgggg/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemgggg/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemgggg/ajob1' in 28.6820 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemgggg/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemgggg/G65/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.1657s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2921s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8379s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0357s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemgggg/G48/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.4452s + [COUNTERS] Fortran Overhead ( 0 ) : 3.5711s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8382s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0359s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemgggg/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_taptamggud/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_taptamggud/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_taptamggud/ajob1' in 58.5214 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_taptamggud/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_taptamggud/G46/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.3691s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2383s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.1151s for 16384 events => throughput is 3.98E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0158s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_taptamggud/G39/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.3654s + [COUNTERS] Fortran Overhead ( 0 ) : 4.2420s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.1072s for 16384 events => throughput is 3.99E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0162s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_taptamggud/G112/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.4056s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2838s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.1054s for 16384 events => throughput is 3.99E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0165s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_taptamggud/G115/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.6827s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5201s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.1467s for 16384 events => throughput is 3.95E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0159s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_taptamggud/G270/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.1274s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9787s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.1327s for 16384 events => throughput is 3.96E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0161s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_taptamggud/G96/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.7775s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6364s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.1253s for 16384 events => throughput is 3.97E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0158s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_taptamggud/G107/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.0940s + [COUNTERS] Fortran Overhead ( 0 ) : 3.9730s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.1051s for 16384 events => throughput is 3.99E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0159s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_taptamggud/G63/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.0560s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9221s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.1179s for 16384 events => throughput is 3.98E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0160s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_taptamggud/G375/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.4452s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3250s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.1043s for 16384 events => throughput is 3.99E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0160s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_taptamggud/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguuux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguuux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguuux/ajob1' in 42.5425 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguuux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguuux/G122/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.5153s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9165s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.5718s for 16384 events => throughput is 2.49E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0271s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguuux/G33/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 12.4563s + [COUNTERS] Fortran Overhead ( 0 ) : 5.8636s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.5661s for 16384 events => throughput is 2.50E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0265s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguuux/G127/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 9.6953s + [COUNTERS] Fortran Overhead ( 0 ) : 3.0962s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.5722s for 16384 events => throughput is 2.49E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0269s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguuux/G116/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 11.7896s + [COUNTERS] Fortran Overhead ( 0 ) : 5.1364s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.6263s for 16384 events => throughput is 2.47E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0268s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguuux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamgggdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamgggdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamgggdx/ajob1' in 216.0652 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamgggdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamgggdx/G56/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 24.4509s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6482s + [COUNTERS] CudaCpp MEs ( 2 ) : 21.7672s for 16384 events => throughput is 7.53E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0355s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamgggdx/G64/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 24.2409s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4498s + [COUNTERS] CudaCpp MEs ( 2 ) : 21.7559s for 16384 events => throughput is 7.53E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0351s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamgggdx/G13/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 23.4645s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6846s + [COUNTERS] CudaCpp MEs ( 2 ) : 21.7431s for 16384 events => throughput is 7.54E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0367s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamgggdx/G33/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 23.6331s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8727s + [COUNTERS] CudaCpp MEs ( 2 ) : 21.7249s for 16384 events => throughput is 7.54E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0356s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamgggdx/G49/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 24.0027s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1840s + [COUNTERS] CudaCpp MEs ( 2 ) : 21.7831s for 16384 events => throughput is 7.52E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0355s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamgggdx/G523/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 24.4145s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6795s + [COUNTERS] CudaCpp MEs ( 2 ) : 21.6994s for 16384 events => throughput is 7.55E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0357s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamgggdx/G65/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 23.8854s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0417s + [COUNTERS] CudaCpp MEs ( 2 ) : 21.8081s for 16384 events => throughput is 7.51E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0356s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamgggdx/G34/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 23.9042s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1464s + [COUNTERS] CudaCpp MEs ( 2 ) : 21.7226s for 16384 events => throughput is 7.54E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0353s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamgggdx/G48/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 23.8704s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0539s + [COUNTERS] CudaCpp MEs ( 2 ) : 21.7807s for 16384 events => throughput is 7.52E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0358s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamgggdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgudux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgudux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgudux/ajob1' in 16.0043 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgudux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgudux/G59/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.0311s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9130s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.1020s for 16384 events => throughput is 3.99E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0162s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgudux/G61/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 9.9081s + [COUNTERS] Fortran Overhead ( 0 ) : 5.7786s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.1136s for 16384 events => throughput is 3.98E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0160s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgudux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgduxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgduxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgduxdx/ajob1' in 4.0882 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgduxdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgduxdx/G104/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.0424s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9554s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.0712s for 16384 events => throughput is 7.91E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0158s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgduxdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgsdxsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgsdxsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgsdxsx/ajob1' in 3.8365 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgsdxsx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgsdxsx/G25/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.8080s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7930s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.9992s for 16384 events => throughput is 8.20E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0158s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgsdxsx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/ajob1' in 133.8352 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G208/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.0387s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9791s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.0438s for 16384 events => throughput is 8.02E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0159s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G235/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.7895s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7429s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.0306s for 16384 events => throughput is 8.07E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0160s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G140/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.0933s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0458s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.0317s for 16384 events => throughput is 8.06E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0158s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G109/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.8242s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7667s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.0415s for 16384 events => throughput is 8.03E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0160s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G33/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.9073s + [COUNTERS] Fortran Overhead ( 0 ) : 4.8340s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.0577s for 16384 events => throughput is 7.96E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0157s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G43/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.9120s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8075s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.0887s for 16384 events => throughput is 7.84E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0157s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G240/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.8680s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8151s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.0375s for 16384 events => throughput is 8.04E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0155s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G268/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.5739s + [COUNTERS] Fortran Overhead ( 0 ) : 4.5270s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.0310s for 16384 events => throughput is 8.07E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0159s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G139/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.1055s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0309s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.0585s for 16384 events => throughput is 7.96E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0162s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G120/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.8231s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7781s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.0291s for 16384 events => throughput is 8.07E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0159s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G236/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.8807s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8311s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.0339s for 16384 events => throughput is 8.06E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0157s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G181/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.5124s + [COUNTERS] Fortran Overhead ( 0 ) : 5.4659s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.0304s for 16384 events => throughput is 8.07E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0161s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G351/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.4714s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4153s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.0395s for 16384 events => throughput is 8.03E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0166s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G65/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.0161s + [COUNTERS] Fortran Overhead ( 0 ) : 3.9592s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.0411s for 16384 events => throughput is 8.03E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0159s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G71/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.6370s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5943s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.0273s for 16384 events => throughput is 8.08E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0154s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G119/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.0279s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9798s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.0325s for 16384 events => throughput is 8.06E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0157s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G127/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.6803s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6347s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.0298s for 16384 events => throughput is 8.07E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0158s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G96/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.7924s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7452s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.0314s for 16384 events => throughput is 8.07E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0158s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G60/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.3043s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2567s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.0319s for 16384 events => throughput is 8.06E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0157s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G26/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.5867s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5289s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.0419s for 16384 events => throughput is 8.02E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0158s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G37/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.9188s + [COUNTERS] Fortran Overhead ( 0 ) : 5.8752s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.0276s for 16384 events => throughput is 8.08E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0159s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G125/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.4436s + [COUNTERS] Fortran Overhead ( 0 ) : 3.3915s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.0363s for 16384 events => throughput is 8.05E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0158s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G91/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.5661s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5219s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.0282s for 16384 events => throughput is 8.08E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0159s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G143/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.4750s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4060s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.0525s for 16384 events => throughput is 7.98E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0165s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G116/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.3840s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3332s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.0352s for 16384 events => throughput is 8.05E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0157s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G61/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.7858s + [COUNTERS] Fortran Overhead ( 0 ) : 5.7256s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.0436s for 16384 events => throughput is 8.02E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0166s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G149/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.4693s + [COUNTERS] Fortran Overhead ( 0 ) : 4.4221s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.0315s for 16384 events => throughput is 8.07E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0157s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G144/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.5730s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5147s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.0423s for 16384 events => throughput is 8.02E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0159s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G207/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.8108s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7653s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.0299s for 16384 events => throughput is 8.07E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0156s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_epemgguu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_epemgguu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_epemgguu/ajob1' in 103.1014 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_epemgguu/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_epemgguu/G463/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.6026s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0362s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.5393s for 16384 events => throughput is 2.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0271s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_epemgguu/G100/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 9.7349s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1635s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.5440s for 16384 events => throughput is 2.50E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0274s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_epemgguu/G110/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.0767s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5104s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.5392s for 16384 events => throughput is 2.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0271s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_epemgguu/G124/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.1837s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5831s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.5738s for 16384 events => throughput is 2.49E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0268s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_epemgguu/G367/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.5429s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9642s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.5513s for 16384 events => throughput is 2.50E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0274s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_epemgguu/G126/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 10.7252s + [COUNTERS] Fortran Overhead ( 0 ) : 4.0981s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.6002s for 16384 events => throughput is 2.48E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0269s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_epemgguu/G361/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 11.1011s + [COUNTERS] Fortran Overhead ( 0 ) : 4.4947s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.5795s for 16384 events => throughput is 2.49E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0268s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_epemgguu/G116/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 9.2613s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7038s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.5307s for 16384 events => throughput is 2.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0268s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_epemgguu/G363/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 10.7272s + [COUNTERS] Fortran Overhead ( 0 ) : 4.1501s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.5499s for 16384 events => throughput is 2.50E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0272s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_epemgguu/G138/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.1922s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6205s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.5446s for 16384 events => throughput is 2.50E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0272s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_epemgguu/G102/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 9.7409s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1773s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.5365s for 16384 events => throughput is 2.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0272s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_epemgguu/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgdssx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgdssx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgdssx/ajob1' in 6.1627 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgdssx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgdssx/G28/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.1115s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0883s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.0075s for 16384 events => throughput is 4.09E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0157s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgdssx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_taptamgguu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_taptamgguu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_taptamgguu/ajob1' in 76.4429 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_taptamgguu/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_taptamgguu/G464/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 15.1961s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9737s + [COUNTERS] CudaCpp MEs ( 2 ) : 13.1932s for 16384 events => throughput is 1.24E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0291s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_taptamgguu/G59/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 15.2890s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1272s + [COUNTERS] CudaCpp MEs ( 2 ) : 13.1348s for 16384 events => throughput is 1.25E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0270s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_taptamgguu/G100/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 15.7370s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5906s + [COUNTERS] CudaCpp MEs ( 2 ) : 13.1192s for 16384 events => throughput is 1.25E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0272s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_taptamgguu/G106/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 15.0663s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8777s + [COUNTERS] CudaCpp MEs ( 2 ) : 13.1615s for 16384 events => throughput is 1.24E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0272s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_taptamgguu/G143/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 15.0345s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8381s + [COUNTERS] CudaCpp MEs ( 2 ) : 13.1698s for 16384 events => throughput is 1.24E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0266s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_taptamgguu/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dc_epemggdc/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dc_epemggdc/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dc_epemggdc/ajob1' in 7.3782 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dc_epemggdc/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dc_epemggdc/G18/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.3295s + [COUNTERS] Fortran Overhead ( 0 ) : 5.3077s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.0058s for 16384 events => throughput is 8.17E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0161s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dc_epemggdc/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ucx_taptamggucx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ucx_taptamggucx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ucx_taptamggucx/ajob1' in 8.8628 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ucx_taptamggucx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ucx_taptamggucx/G62/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.8121s + [COUNTERS] Fortran Overhead ( 0 ) : 4.7594s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.0359s for 16384 events => throughput is 4.06E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0168s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ucx_taptamggucx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamucdxcx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamucdxcx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamucdxcx/ajob1' in 2.4731 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamucdxcx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamucdxcx/G47/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.4369s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8348s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5962s for 16384 events => throughput is 2.75E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0059s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamucdxcx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_epemgggg/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_epemgggg/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_epemgggg/ajob1' in 14.3578 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_epemgggg/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_epemgggg/G49/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.3112s + [COUNTERS] Fortran Overhead ( 0 ) : 3.4459s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8298s for 16384 events => throughput is 1.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0356s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_epemgggg/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemudddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemudddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemudddx/ajob1' in 2.4260 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemudddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemudddx/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.3748s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6256s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.7403s for 16384 events => throughput is 2.21E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0089s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemudddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamggudx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamggudx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamggudx/ajob1' in 5.7497 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamggudx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamggudx/G60/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.7210s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6879s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.0168s for 16384 events => throughput is 4.08E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0163s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamggudx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_taptamuuddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_taptamuuddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_taptamuuddx/ajob1' in 3.2646 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_taptamuuddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_taptamuuddx/G11/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.2180s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6909s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.5181s for 16384 events => throughput is 1.08E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0090s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_taptamuuddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished in 6661.0695 seconds +__CUDACPP_DEBUG: gen_ximprove.launch finished in 6665.1490 seconds +__CUDACPP_DEBUG: GridPackCmd.refine4grid finished in 6665.1499 seconds +__CUDACPP_DEBUG: GridPackCmd.launch (3a) combine_events +combine_events +INFO: fail to reach target 500 +__CUDACPP_DEBUG: GridPackCmd.launch (3b) store_events +store_events + === Results Summary for run: GridRun_21 tag: tag_1 === + + Cross-section : 0 +- 0 pb + Nb of events : 195 + +INFO: No version of lhapdf. Can not run systematics computation +decay_events -from_cards +__CUDACPP_DEBUG: GridPackCmd.launch finished in 6831.8971 seconds +__CUDACPP_DEBUG: GridPackCmd.__init__ back from self.launch +DONE +write ./events.lhe.gz +END: Sun Sep 15 03:42:05 PM CEST 2024 +ELAPSED: 6834 seconds diff --git a/epochX/cudacpp/tlau/fromgridpacks/pp_dy4j.mad/cpp512z/output.txt b/epochX/cudacpp/tlau/fromgridpacks/pp_dy4j.mad/cpp512z/output.txt new file mode 100644 index 0000000000..1182fab1c5 --- /dev/null +++ b/epochX/cudacpp/tlau/fromgridpacks/pp_dy4j.mad/cpp512z/output.txt @@ -0,0 +1,3327 @@ +START: Sun Sep 15 03:45:18 PM CEST 2024 +On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU (nproc=4)] [GPU: 1x Tesla V100S-PCIE-32GB]: +CUDACPP_RUNTIME_DISABLEFPE is set +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_bbx_epemdsdxsx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_bbx_taptamdsdxsx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_ccx_epemdcdxcx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_ccx_epemdsdxsx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_ccx_epemucuxcx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_ccx_epemuduxdx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_ccx_taptamdcdxcx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_ccx_taptamdsdxsx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_ccx_taptamucuxcx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_ccx_taptamuduxdx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_dc_epemdcccx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_dc_epemdcssx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_dc_epemddcdx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_dc_epemggdc/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_dc_epemudcux/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_dc_taptamdcccx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_dc_taptamdcssx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_dc_taptamddcdx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_dc_taptamggdc/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_dc_taptamudcux/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_dd_epemddddx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_dd_epemddssx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_dd_epemggdd/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_dd_epemuddux/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_dd_taptamddddx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_dd_taptamddssx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_dd_taptamggdd/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_dd_taptamuddux/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_ddx_epemdcdxcx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_ddx_epemdddxdx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_ddx_epemdsdxsx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_ddx_epemggddx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_ddx_epemgggg/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_ddx_epemggssx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_ddx_epemgguux/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_ddx_epemsbsxbx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_ddx_epemsssxsx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_ddx_epemucuxcx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_ddx_epemuduxdx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_ddx_epemusuxsx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_ddx_epemuuuxux/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_ddx_taptamdcdxcx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_ddx_taptamdddxdx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_ddx_taptamdsdxsx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_ddx_taptamggddx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_ddx_taptamgggg/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_ddx_taptamggssx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_ddx_taptamgguux/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_ddx_taptamsbsxbx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_ddx_taptamsssxsx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_ddx_taptamucuxcx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_ddx_taptamuduxdx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_ddx_taptamusuxsx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_ddx_taptamuuuxux/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_ds_epemddsdx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_ds_epemdsbbx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_ds_epemdsssx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_ds_epemggds/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_ds_epemudsux/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_ds_taptamddsdx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_ds_taptamdsbbx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_ds_taptamdsssx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_ds_taptamggds/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_ds_taptamudsux/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_dsx_epemdbsxbx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_dsx_epemdddxsx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_dsx_epemdssxsx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_dsx_epemggdsx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_dsx_epemuduxsx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_dsx_taptamdbsxbx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_dsx_taptamdddxsx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_dsx_taptamdssxsx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_dsx_taptamggdsx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_dsx_taptamuduxsx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_dux_epemdcuxcx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_dux_epemdduxdx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_dux_epemdsuxsx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_dux_epemggdux/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_dux_epemuduxux/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_dux_taptamdcuxcx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_dux_taptamdduxdx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_dux_taptamdsuxsx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_dux_taptamggdux/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_dux_taptamuduxux/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_dxcx_epemcdxcxcx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_dxcx_epemddxdxcx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_dxcx_epemggdxcx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_dxcx_epemsdxcxsx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_dxcx_epemuuxdxcx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_dxcx_taptamcdxcxcx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_dxcx_taptamddxdxcx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_dxcx_taptamggdxcx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_dxcx_taptamsdxcxsx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_dxcx_taptamuuxdxcx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_dxdx_epemddxdxdx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_dxdx_epemggdxdx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_dxdx_epemsdxdxsx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_dxdx_epemuuxdxdx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_dxdx_taptamddxdxdx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_dxdx_taptamggdxdx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_dxdx_taptamsdxdxsx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_dxdx_taptamuuxdxdx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_dxsx_epembdxsxbx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_dxsx_epemddxdxsx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_dxsx_epemggdxsx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_dxsx_epemsdxsxsx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_dxsx_epemuuxdxsx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_dxsx_taptambdxsxbx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_dxsx_taptamddxdxsx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_dxsx_taptamggdxsx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_dxsx_taptamsdxsxsx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_dxsx_taptamuuxdxsx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_gd_epemgdddx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_gd_epemgdssx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_gd_epemgggd/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_gd_epemgudux/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_gd_taptamgdddx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_gd_taptamgdssx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_gd_taptamgggd/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_gd_taptamgudux/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_gdx_epemgddxdx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_gdx_epemgggdx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_gdx_epemgsdxsx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_gdx_epemguuxdx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_gdx_taptamgddxdx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_gdx_taptamgggdx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_gdx_taptamgsdxsx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_gdx_taptamguuxdx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_gg_epemdddxdx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_gg_epemdsdxsx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_gg_epemggddx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_gg_epemgguux/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_gg_epemucuxcx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_gg_epemuduxdx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_gg_epemuuuxux/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_gg_taptamdddxdx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_gg_taptamdsdxsx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_gg_taptamggddx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_gg_taptamgguux/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_gg_taptamucuxcx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_gg_taptamuduxdx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_gg_taptamuuuxux/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_gu_epemgggu/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_gu_epemguccx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_gu_epemguddx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_gu_epemguuux/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_gu_taptamgggu/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_gu_taptamguccx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_gu_taptamguddx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_gu_taptamguuux/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_gux_epemgcuxcx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_gux_epemgduxdx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_gux_epemgggux/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_gux_epemguuxux/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_gux_taptamgcuxcx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_gux_taptamgduxdx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_gux_taptamgggux/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_gux_taptamguuxux/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_ssx_epemdbdxbx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_ssx_epemdcdxcx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_ssx_epemdsdxsx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_ssx_epemucuxcx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_ssx_epemuduxdx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_ssx_taptamdbdxbx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_ssx_taptamdcdxcx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_ssx_taptamdsdxsx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_ssx_taptamucuxcx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_ssx_taptamuduxdx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_uc_epemgguc/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_uc_epemucccx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_uc_epemudcdx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_uc_epemuucux/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_uc_taptamgguc/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_uc_taptamucccx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_uc_taptamudcdx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_uc_taptamuucux/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_ucx_epemggucx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_ucx_epemuccxcx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_ucx_epemuddxcx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_ucx_epemuuuxcx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_ucx_taptamggucx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_ucx_taptamuccxcx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_ucx_taptamuddxcx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_ucx_taptamuuuxcx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_ud_epemggud/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_ud_epemudccx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_ud_epemudddx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_ud_epemudssx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_ud_epemuudux/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_ud_taptamggud/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_ud_taptamudccx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_ud_taptamudddx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_ud_taptamudssx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_ud_taptamuudux/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_udx_epemggudx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_udx_epemucdxcx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_udx_epemuddxdx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_udx_epemusdxsx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_udx_epemuuuxdx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_udx_taptamggudx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_udx_taptamucdxcx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_udx_taptamuddxdx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_udx_taptamusdxsx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_udx_taptamuuuxdx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_uu_epemgguu/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_uu_epemuuccx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_uu_epemuuddx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_uu_epemuuuux/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_uu_taptamgguu/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_uu_taptamuuccx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_uu_taptamuuddx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_uu_taptamuuuux/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_uux_epemcccxcx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_uux_epemcscxsx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_uux_epemdcdxcx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_uux_epemdddxdx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_uux_epemdsdxsx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_uux_epemggccx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_uux_epemggddx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_uux_epemgggg/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_uux_epemgguux/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_uux_epemucuxcx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_uux_epemuduxdx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_uux_epemuuuxux/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_uux_taptamcccxcx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_uux_taptamcscxsx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_uux_taptamdcdxcx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_uux_taptamdddxdx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_uux_taptamdsdxsx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_uux_taptamggccx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_uux_taptamggddx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_uux_taptamgggg/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_uux_taptamgguux/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_uux_taptamucuxcx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_uux_taptamuduxdx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_uux_taptamuuuxux/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_uxcx_epemcuxcxcx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_uxcx_epemduxdxcx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_uxcx_epemgguxcx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_uxcx_epemuuxuxcx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_uxcx_taptamcuxcxcx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_uxcx_taptamduxdxcx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_uxcx_taptamgguxcx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_uxcx_taptamuuxuxcx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_uxdx_epemcuxdxcx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_uxdx_epemduxdxdx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_uxdx_epemgguxdx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_uxdx_epemsuxdxsx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_uxdx_epemuuxuxdx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_uxdx_taptamcuxdxcx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_uxdx_taptamduxdxdx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_uxdx_taptamgguxdx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_uxdx_taptamsuxdxsx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_uxdx_taptamuuxuxdx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_uxux_epemcuxuxcx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_uxux_epemduxuxdx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_uxux_epemgguxux/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_uxux_epemuuxuxux/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_uxux_taptamcuxuxcx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_uxux_taptamduxuxdx/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_uxux_taptamgguxux/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 15:45 madevent/SubProcesses/P0_uxux_taptamuuxuxux/madevent -> build.512z_d_inl0_hrd0/madevent_cpp +Now generating 500 events with random seed 21 and granularity 1 +************************************************************ +* * +* W E L C O M E to * +* M A D G R A P H 5 _ a M C @ N L O * +* M A D E V E N T * +* * +* * * * +* * * * * * +* * * * * 5 * * * * * +* * * * * * +* * * * +* * +* VERSION 3.5.3_lo_vect * +* * +* The MadGraph5_aMC@NLO Development Team - Find us at * +* https://server06.fynu.ucl.ac.be/projects/madgraph * +* * +* Type 'help' for in-line help. * +* * +************************************************************ +INFO: load configuration from /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/Cards/me5_configuration.txt +INFO: load configuration from /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/Cards/me5_configuration.txt +Using default text editor "vi". Set another one in ./input/mg5_configuration.txt +Using default eps viewer "evince". Set another one in ./input/mg5_configuration.txt +Using default web browser "firefox". Set another one in ./input/mg5_configuration.txt +WRITE GRIDCARD /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent +__CUDACPP_DEBUG: GridPackCmd.__init__ will call self.launch +__CUDACPP_DEBUG: GridPackCmd.launch starting +__CUDACPP_DEBUG: GridPackCmd.launch (1) restore_data +generate 500 events +P0_gg_epemgguux +P0_gg_epemggddx +P0_gg_taptamgguux +P0_gg_taptamggddx +P0_gu_epemgggu +P0_gd_epemgggd +P0_gux_epemgggux +P0_gdx_epemgggdx +P0_gu_taptamgggu +P0_gd_taptamgggd +P0_gux_taptamgggux +P0_gdx_taptamgggdx +P0_uux_epemgggg +P0_ddx_epemgggg +P0_uux_taptamgggg +P0_ddx_taptamgggg +P0_gg_epemuuuxux +P0_gg_epemdddxdx +P0_gg_epemuduxdx +P0_gg_epemucuxcx +P0_gg_epemdsdxsx +P0_gg_taptamuuuxux +P0_gg_taptamdddxdx +P0_gg_taptamuduxdx +P0_gg_taptamucuxcx +P0_gg_taptamdsdxsx +P0_gu_epemguuux +P0_gd_epemgdddx +P0_gux_epemguuxux +P0_gdx_epemgddxdx +P0_gu_epemguddx +P0_gu_epemguccx +P0_gd_epemgudux +P0_gd_epemgdssx +P0_gux_epemgduxdx +P0_gux_epemgcuxcx +P0_gdx_epemguuxdx +P0_gdx_epemgsdxsx +P0_gu_taptamguuux +P0_gd_taptamgdddx +P0_gux_taptamguuxux +P0_gdx_taptamgddxdx +P0_gu_taptamguddx +P0_gu_taptamguccx +P0_gd_taptamgudux +P0_gd_taptamgdssx +P0_gux_taptamgduxdx +P0_gux_taptamgcuxcx +P0_gdx_taptamguuxdx +P0_gdx_taptamgsdxsx +P0_uu_epemgguu +P0_uux_epemgguux +P0_dd_epemggdd +P0_ddx_epemggddx +P0_uxux_epemgguxux +P0_dxdx_epemggdxdx +P0_ud_epemggud +P0_uc_epemgguc +P0_uux_epemggddx +P0_uux_epemggccx +P0_udx_epemggudx +P0_ucx_epemggucx +P0_dc_epemggdc +P0_ds_epemggds +P0_dux_epemggdux +P0_ddx_epemgguux +P0_ddx_epemggssx +P0_dsx_epemggdsx +P0_uxdx_epemgguxdx +P0_uxcx_epemgguxcx +P0_dxcx_epemggdxcx +P0_dxsx_epemggdxsx +P0_uu_taptamgguu +P0_uux_taptamgguux +P0_dd_taptamggdd +P0_ddx_taptamggddx +P0_uxux_taptamgguxux +P0_dxdx_taptamggdxdx +P0_ud_taptamggud +P0_uc_taptamgguc +P0_uux_taptamggddx +P0_uux_taptamggccx +P0_udx_taptamggudx +P0_ucx_taptamggucx +P0_dc_taptamggdc +P0_ds_taptamggds +P0_dux_taptamggdux +P0_ddx_taptamgguux +P0_ddx_taptamggssx +P0_dsx_taptamggdsx +P0_uxdx_taptamgguxdx +P0_uxcx_taptamgguxcx +P0_dxcx_taptamggdxcx +P0_dxsx_taptamggdxsx +P0_uu_epemuuuux +P0_uux_epemuuuxux +P0_dd_epemddddx +P0_ddx_epemdddxdx +P0_uxux_epemuuxuxux +P0_dxdx_epemddxdxdx +P0_uu_epemuuddx +P0_uu_epemuuccx +P0_ud_epemuudux +P0_ud_epemudddx +P0_uc_epemuucux +P0_uc_epemucccx +P0_uux_epemuduxdx +P0_uux_epemucuxcx +P0_uux_epemdddxdx +P0_uux_epemcccxcx +P0_udx_epemuuuxdx +P0_udx_epemuddxdx +P0_ucx_epemuuuxcx +P0_ucx_epemuccxcx +P0_dd_epemuddux +P0_dd_epemddssx +P0_dc_epemddcdx +P0_dc_epemdcccx +P0_ds_epemddsdx +P0_ds_epemdsssx +P0_dux_epemuduxux +P0_dux_epemdduxdx +P0_ddx_epemuuuxux +P0_ddx_epemuduxdx +P0_ddx_epemdcdxcx +P0_ddx_epemdsdxsx +P0_ddx_epemsssxsx +P0_dsx_epemdddxsx +P0_dsx_epemdssxsx +P0_ccx_epemucuxcx +P0_ccx_epemdcdxcx +P0_ssx_epemdsdxsx +P0_uxux_epemduxuxdx +P0_uxux_epemcuxuxcx +P0_uxdx_epemuuxuxdx +P0_uxdx_epemduxdxdx +P0_uxcx_epemuuxuxcx +P0_uxcx_epemcuxcxcx +P0_dxdx_epemuuxdxdx +P0_dxdx_epemsdxdxsx +P0_dxcx_epemddxdxcx +P0_dxcx_epemcdxcxcx +P0_dxsx_epemddxdxsx +P0_dxsx_epemsdxsxsx +P0_ud_epemudccx +P0_ud_epemudssx +P0_uc_epemudcdx +P0_uux_epemdcdxcx +P0_uux_epemdsdxsx +P0_uux_epemcscxsx +P0_udx_epemucdxcx +P0_udx_epemusdxsx +P0_ucx_epemuddxcx +P0_dc_epemudcux +P0_dc_epemdcssx +P0_ds_epemudsux +P0_ds_epemdsbbx +P0_dux_epemdcuxcx +P0_dux_epemdsuxsx +P0_ddx_epemucuxcx +P0_ddx_epemusuxsx +P0_ddx_epemsbsxbx +P0_dsx_epemuduxsx +P0_dsx_epemdbsxbx +P0_ccx_epemuduxdx +P0_ccx_epemdsdxsx +P0_ssx_epemuduxdx +P0_ssx_epemucuxcx +P0_ssx_epemdcdxcx +P0_ssx_epemdbdxbx +P0_bbx_epemdsdxsx +P0_uxdx_epemcuxdxcx +P0_uxdx_epemsuxdxsx +P0_uxcx_epemduxdxcx +P0_dxcx_epemuuxdxcx +P0_dxcx_epemsdxcxsx +P0_dxsx_epemuuxdxsx +P0_dxsx_epembdxsxbx +P0_uu_taptamuuuux +P0_uux_taptamuuuxux +P0_dd_taptamddddx +P0_ddx_taptamdddxdx +P0_uxux_taptamuuxuxux +P0_dxdx_taptamddxdxdx +P0_uu_taptamuuddx +P0_uu_taptamuuccx +P0_ud_taptamuudux +P0_ud_taptamudddx +P0_uc_taptamuucux +P0_uc_taptamucccx +P0_uux_taptamuduxdx +P0_uux_taptamucuxcx +P0_uux_taptamdddxdx +P0_uux_taptamcccxcx +P0_udx_taptamuuuxdx +P0_udx_taptamuddxdx +P0_ucx_taptamuuuxcx +P0_ucx_taptamuccxcx +P0_dd_taptamuddux +P0_dd_taptamddssx +P0_dc_taptamddcdx +P0_dc_taptamdcccx +P0_ds_taptamddsdx +P0_ds_taptamdsssx +P0_dux_taptamuduxux +P0_dux_taptamdduxdx +P0_ddx_taptamuuuxux +P0_ddx_taptamuduxdx +P0_ddx_taptamdcdxcx +P0_ddx_taptamdsdxsx +P0_ddx_taptamsssxsx +P0_dsx_taptamdddxsx +P0_dsx_taptamdssxsx +P0_ccx_taptamucuxcx +P0_ccx_taptamdcdxcx +P0_ssx_taptamdsdxsx +P0_uxux_taptamduxuxdx +P0_uxux_taptamcuxuxcx +P0_uxdx_taptamuuxuxdx +P0_uxdx_taptamduxdxdx +P0_uxcx_taptamuuxuxcx +P0_uxcx_taptamcuxcxcx +P0_dxdx_taptamuuxdxdx +P0_dxdx_taptamsdxdxsx +P0_dxcx_taptamddxdxcx +P0_dxcx_taptamcdxcxcx +P0_dxsx_taptamddxdxsx +P0_dxsx_taptamsdxsxsx +P0_ud_taptamudccx +P0_ud_taptamudssx +P0_uc_taptamudcdx +P0_uux_taptamdcdxcx +P0_uux_taptamdsdxsx +P0_uux_taptamcscxsx +P0_udx_taptamucdxcx +P0_udx_taptamusdxsx +P0_ucx_taptamuddxcx +P0_dc_taptamudcux +P0_dc_taptamdcssx +P0_ds_taptamudsux +P0_ds_taptamdsbbx +P0_dux_taptamdcuxcx +P0_dux_taptamdsuxsx +P0_ddx_taptamucuxcx +P0_ddx_taptamusuxsx +P0_ddx_taptamsbsxbx +P0_dsx_taptamuduxsx +P0_dsx_taptamdbsxbx +P0_ccx_taptamuduxdx +P0_ccx_taptamdsdxsx +P0_ssx_taptamuduxdx +P0_ssx_taptamucuxcx +P0_ssx_taptamdcdxcx +P0_ssx_taptamdbdxbx +P0_bbx_taptamdsdxsx +P0_uxdx_taptamcuxdxcx +P0_uxdx_taptamsuxdxsx +P0_uxcx_taptamduxdxcx +P0_dxcx_taptamuuxdxcx +P0_dxcx_taptamsdxcxsx +P0_dxsx_taptamuuxdxsx +P0_dxsx_taptambdxsxbx +__CUDACPP_DEBUG: GridPackCmd.launch (2) refine4grid +__CUDACPP_DEBUG: GridPackCmd.refine4grid starting +Using random number seed offset = 21 +INFO: Generating 500.0 unweighted events. +__CUDACPP_DEBUG: gen_ximprove.launch starting +sum of cpu time of last step: 3 days,23h23m17s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event starting +INFO: need to improve 477 channels +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dxdx_taptamuuxdxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dxdx_taptamuuxdxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dxdx_taptamuuxdxdx/ajob1' in 5.8994 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dxdx_taptamuuxdxdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dxdx_taptamuuxdxdx/G21/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.7769s + [COUNTERS] Fortran Overhead ( 0 ) : 3.8474s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.9181s for 16384 events => throughput is 8.54E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0114s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dxdx_taptamuuxdxdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamguuxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamguuxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamguuxdx/ajob1' in 15.0996 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamguuxdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamguuxdx/G43/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.2059s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0854s + [COUNTERS] CudaCpp MEs ( 2 ) : 5.0995s for 16384 events => throughput is 3.21E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0210s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamguuxdx/G335/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.8479s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7377s + [COUNTERS] CudaCpp MEs ( 2 ) : 5.0893s for 16384 events => throughput is 3.22E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0209s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamguuxdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemuuuxux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemuuuxux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemuuuxux/ajob1' in 9.6672 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemuuuxux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemuuuxux/G689/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 9.4109s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6256s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.7503s for 16384 events => throughput is 2.43E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0350s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemuuuxux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamgguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamgguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamgguux/ajob1' in 260.3814 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamgguux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamgguux/G596/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 27.5484s + [COUNTERS] Fortran Overhead ( 0 ) : 3.5163s + [COUNTERS] CudaCpp MEs ( 2 ) : 23.9844s for 16384 events => throughput is 6.83E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0477s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamgguux/G304/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 27.0795s + [COUNTERS] Fortran Overhead ( 0 ) : 3.0534s + [COUNTERS] CudaCpp MEs ( 2 ) : 23.9790s for 16384 events => throughput is 6.83E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0471s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamgguux/G128/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 27.1559s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1644s + [COUNTERS] CudaCpp MEs ( 2 ) : 23.9442s for 16384 events => throughput is 6.84E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0473s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamgguux/G115/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 31.1303s + [COUNTERS] Fortran Overhead ( 0 ) : 6.9417s + [COUNTERS] CudaCpp MEs ( 2 ) : 24.1412s for 16384 events => throughput is 6.79E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0474s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamgguux/G599/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 28.0336s + [COUNTERS] Fortran Overhead ( 0 ) : 3.9830s + [COUNTERS] CudaCpp MEs ( 2 ) : 24.0026s for 16384 events => throughput is 6.83E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0480s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamgguux/G351/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 30.3200s + [COUNTERS] Fortran Overhead ( 0 ) : 6.3565s + [COUNTERS] CudaCpp MEs ( 2 ) : 23.9155s for 16384 events => throughput is 6.85E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0480s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamgguux/G111/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 26.5117s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4832s + [COUNTERS] CudaCpp MEs ( 2 ) : 23.9807s for 16384 events => throughput is 6.83E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0478s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamgguux/G354/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 30.8784s + [COUNTERS] Fortran Overhead ( 0 ) : 6.8624s + [COUNTERS] CudaCpp MEs ( 2 ) : 23.9689s for 16384 events => throughput is 6.84E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0471s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamgguux/G97/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 31.5350s + [COUNTERS] Fortran Overhead ( 0 ) : 7.5483s + [COUNTERS] CudaCpp MEs ( 2 ) : 23.9397s for 16384 events => throughput is 6.84E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0471s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamgguux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uxdx_epemsuxdxsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uxdx_epemsuxdxsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uxdx_epemsuxdxsx/ajob1' in 2.6947 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uxdx_epemsuxdxsx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uxdx_epemsuxdxsx/G74/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.6471s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2470s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.3924s for 16384 events => throughput is 4.18E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0078s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uxdx_epemsuxdxsx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemdddxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemdddxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemdddxdx/ajob1' in 19.0071 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemdddxdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemdddxdx/G513/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 9.3265s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5862s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.7043s for 16384 events => throughput is 2.44E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0360s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemdddxdx/G180/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 9.6351s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8835s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.7150s for 16384 events => throughput is 2.44E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0366s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemdddxdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgddxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgddxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgddxdx/ajob1' in 9.9925 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgddxdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgddxdx/G515/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 9.9429s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2280s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.6798s for 16384 events => throughput is 2.45E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0351s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgddxdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguddx/ajob1' in 70.6658 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguddx/G104/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.7737s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6706s + [COUNTERS] CudaCpp MEs ( 2 ) : 5.0819s for 16384 events => throughput is 3.22E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0212s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguddx/G120/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.3584s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2581s + [COUNTERS] CudaCpp MEs ( 2 ) : 5.0797s for 16384 events => throughput is 3.23E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0206s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguddx/G305/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 11.2458s + [COUNTERS] Fortran Overhead ( 0 ) : 6.1320s + [COUNTERS] CudaCpp MEs ( 2 ) : 5.0925s for 16384 events => throughput is 3.22E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0213s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguddx/G101/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.6374s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5200s + [COUNTERS] CudaCpp MEs ( 2 ) : 5.0954s for 16384 events => throughput is 3.22E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0220s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguddx/G27/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.0615s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9547s + [COUNTERS] CudaCpp MEs ( 2 ) : 5.0857s for 16384 events => throughput is 3.22E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0210s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguddx/G273/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.3613s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2443s + [COUNTERS] CudaCpp MEs ( 2 ) : 5.0951s for 16384 events => throughput is 3.22E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0220s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguddx/G349/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.5738s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4443s + [COUNTERS] CudaCpp MEs ( 2 ) : 5.1092s for 16384 events => throughput is 3.21E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0203s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguddx/G379/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.9815s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8438s + [COUNTERS] CudaCpp MEs ( 2 ) : 5.1166s for 16384 events => throughput is 3.20E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0210s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguddx/G256/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 10.4863s + [COUNTERS] Fortran Overhead ( 0 ) : 5.3692s + [COUNTERS] CudaCpp MEs ( 2 ) : 5.0963s for 16384 events => throughput is 3.21E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0209s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemguuxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemguuxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemguuxdx/ajob1' in 21.6583 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemguuxdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemguuxdx/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.8316s + [COUNTERS] Fortran Overhead ( 0 ) : 6.2850s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.5257s for 16384 events => throughput is 6.49E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0209s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemguuxdx/G68/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.0440s + [COUNTERS] Fortran Overhead ( 0 ) : 5.5080s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.5157s for 16384 events => throughput is 6.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0204s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemguuxdx/G355/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.7018s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1661s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.5149s for 16384 events => throughput is 6.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0207s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemguuxdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/ajob1' in 625.6142 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G14/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 26.2298s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1435s + [COUNTERS] CudaCpp MEs ( 2 ) : 24.0391s for 16384 events => throughput is 6.82E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0471s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G56/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 26.6366s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5939s + [COUNTERS] CudaCpp MEs ( 2 ) : 23.9960s for 16384 events => throughput is 6.83E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0467s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G517/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 30.5129s + [COUNTERS] Fortran Overhead ( 0 ) : 6.4088s + [COUNTERS] CudaCpp MEs ( 2 ) : 24.0571s for 16384 events => throughput is 6.81E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0470s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G52/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 27.0026s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8703s + [COUNTERS] CudaCpp MEs ( 2 ) : 24.0845s for 16384 events => throughput is 6.80E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0478s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G304/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 26.8020s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8170s + [COUNTERS] CudaCpp MEs ( 2 ) : 23.9379s for 16384 events => throughput is 6.84E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0471s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G3/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 29.1087s + [COUNTERS] Fortran Overhead ( 0 ) : 5.0612s + [COUNTERS] CudaCpp MEs ( 2 ) : 23.9999s for 16384 events => throughput is 6.83E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0476s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G23/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 25.8846s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7905s + [COUNTERS] CudaCpp MEs ( 2 ) : 24.0460s for 16384 events => throughput is 6.81E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0481s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G33/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 26.3585s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2547s + [COUNTERS] CudaCpp MEs ( 2 ) : 24.0558s for 16384 events => throughput is 6.81E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0480s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G53/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 31.0283s + [COUNTERS] Fortran Overhead ( 0 ) : 6.9331s + [COUNTERS] CudaCpp MEs ( 2 ) : 24.0481s for 16384 events => throughput is 6.81E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0471s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G495/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 25.7272s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6545s + [COUNTERS] CudaCpp MEs ( 2 ) : 24.0252s for 16384 events => throughput is 6.82E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0475s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G68/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 26.5154s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2715s + [COUNTERS] CudaCpp MEs ( 2 ) : 24.1967s for 16384 events => throughput is 6.77E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0472s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G49/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 26.2615s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1590s + [COUNTERS] CudaCpp MEs ( 2 ) : 24.0553s for 16384 events => throughput is 6.81E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0472s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G524/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 26.2539s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2470s + [COUNTERS] CudaCpp MEs ( 2 ) : 23.9603s for 16384 events => throughput is 6.84E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0465s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G65/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 26.2720s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1612s + [COUNTERS] CudaCpp MEs ( 2 ) : 24.0629s for 16384 events => throughput is 6.81E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0479s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G303/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 27.2229s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1404s + [COUNTERS] CudaCpp MEs ( 2 ) : 24.0352s for 16384 events => throughput is 6.82E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0473s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G55/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 26.9260s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7997s + [COUNTERS] CudaCpp MEs ( 2 ) : 24.0784s for 16384 events => throughput is 6.80E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0479s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G34/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 25.9548s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8836s + [COUNTERS] CudaCpp MEs ( 2 ) : 24.0244s for 16384 events => throughput is 6.82E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0468s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G509/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 32.1061s + [COUNTERS] Fortran Overhead ( 0 ) : 8.0564s + [COUNTERS] CudaCpp MEs ( 2 ) : 24.0022s for 16384 events => throughput is 6.83E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0475s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G512/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 25.8799s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8713s + [COUNTERS] CudaCpp MEs ( 2 ) : 23.9609s for 16384 events => throughput is 6.84E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0477s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G511/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 26.6070s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5246s + [COUNTERS] CudaCpp MEs ( 2 ) : 24.0351s for 16384 events => throughput is 6.82E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0473s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 27.0505s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9185s + [COUNTERS] CudaCpp MEs ( 2 ) : 24.0850s for 16384 events => throughput is 6.80E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0471s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G63/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 26.2468s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1824s + [COUNTERS] CudaCpp MEs ( 2 ) : 24.0174s for 16384 events => throughput is 6.82E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0471s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G48/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 26.5729s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4320s + [COUNTERS] CudaCpp MEs ( 2 ) : 24.0935s for 16384 events => throughput is 6.80E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0474s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/ajob1' in 402.6262 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G122/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.8768s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8089s + [COUNTERS] CudaCpp MEs ( 2 ) : 12.0203s for 16384 events => throughput is 1.36E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0475s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G108/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.5171s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4200s + [COUNTERS] CudaCpp MEs ( 2 ) : 12.0495s for 16384 events => throughput is 1.36E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0475s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G105/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.6807s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5693s + [COUNTERS] CudaCpp MEs ( 2 ) : 12.0636s for 16384 events => throughput is 1.36E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0478s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G304/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 15.2627s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1741s + [COUNTERS] CudaCpp MEs ( 2 ) : 12.0407s for 16384 events => throughput is 1.36E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0480s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G301/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 20.0115s + [COUNTERS] Fortran Overhead ( 0 ) : 7.8978s + [COUNTERS] CudaCpp MEs ( 2 ) : 12.0659s for 16384 events => throughput is 1.36E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0478s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G128/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 15.4079s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2880s + [COUNTERS] CudaCpp MEs ( 2 ) : 12.0723s for 16384 events => throughput is 1.36E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0477s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G100/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 20.0233s + [COUNTERS] Fortran Overhead ( 0 ) : 7.8866s + [COUNTERS] CudaCpp MEs ( 2 ) : 12.0896s for 16384 events => throughput is 1.36E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0472s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G120/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.6546s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5248s + [COUNTERS] CudaCpp MEs ( 2 ) : 12.0830s for 16384 events => throughput is 1.36E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0467s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G302/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 23.1834s + [COUNTERS] Fortran Overhead ( 0 ) : 11.0768s + [COUNTERS] CudaCpp MEs ( 2 ) : 12.0582s for 16384 events => throughput is 1.36E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0483s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G299/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.8495s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7694s + [COUNTERS] CudaCpp MEs ( 2 ) : 12.0330s for 16384 events => throughput is 1.36E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0471s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G169/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 18.3779s + [COUNTERS] Fortran Overhead ( 0 ) : 6.2507s + [COUNTERS] CudaCpp MEs ( 2 ) : 12.0803s for 16384 events => throughput is 1.36E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0469s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G333/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 23.4724s + [COUNTERS] Fortran Overhead ( 0 ) : 11.3691s + [COUNTERS] CudaCpp MEs ( 2 ) : 12.0558s for 16384 events => throughput is 1.36E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0475s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G594/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 21.7316s + [COUNTERS] Fortran Overhead ( 0 ) : 9.6391s + [COUNTERS] CudaCpp MEs ( 2 ) : 12.0448s for 16384 events => throughput is 1.36E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0478s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G678/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 18.7191s + [COUNTERS] Fortran Overhead ( 0 ) : 6.5718s + [COUNTERS] CudaCpp MEs ( 2 ) : 12.1002s for 16384 events => throughput is 1.35E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0471s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G127/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.8954s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8104s + [COUNTERS] CudaCpp MEs ( 2 ) : 12.0379s for 16384 events => throughput is 1.36E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0471s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G126/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 21.9937s + [COUNTERS] Fortran Overhead ( 0 ) : 9.8491s + [COUNTERS] CudaCpp MEs ( 2 ) : 12.0970s for 16384 events => throughput is 1.35E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0477s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G598/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.3934s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2482s + [COUNTERS] CudaCpp MEs ( 2 ) : 12.0980s for 16384 events => throughput is 1.35E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0472s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G512/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.3081s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1778s + [COUNTERS] CudaCpp MEs ( 2 ) : 12.0844s for 16384 events => throughput is 1.36E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0458s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G103/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 21.4332s + [COUNTERS] Fortran Overhead ( 0 ) : 9.3311s + [COUNTERS] CudaCpp MEs ( 2 ) : 12.0557s for 16384 events => throughput is 1.36E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0463s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G107/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.5415s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4178s + [COUNTERS] CudaCpp MEs ( 2 ) : 12.0781s for 16384 events => throughput is 1.36E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0456s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G358/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 15.9847s + [COUNTERS] Fortran Overhead ( 0 ) : 3.8725s + [COUNTERS] CudaCpp MEs ( 2 ) : 12.0655s for 16384 events => throughput is 1.36E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0467s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G97/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 19.8604s + [COUNTERS] Fortran Overhead ( 0 ) : 7.7714s + [COUNTERS] CudaCpp MEs ( 2 ) : 12.0425s for 16384 events => throughput is 1.36E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0465s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G179/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 15.0078s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8508s + [COUNTERS] CudaCpp MEs ( 2 ) : 12.1093s for 16384 events => throughput is 1.35E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0477s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_epemudcdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_epemudcdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_epemudcdx/ajob1' in 2.1983 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_epemudcdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_epemudcdx/G42/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.1688s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7713s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.3899s for 16384 events => throughput is 4.20E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0076s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_epemudcdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ds_taptamggds/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ds_taptamggds/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ds_taptamggds/ajob1' in 16.8965 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ds_taptamggds/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ds_taptamggds/G17/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 9.9651s + [COUNTERS] Fortran Overhead ( 0 ) : 4.8472s + [COUNTERS] CudaCpp MEs ( 2 ) : 5.0966s for 16384 events => throughput is 3.21E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0214s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ds_taptamggds/G110/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.8677s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7393s + [COUNTERS] CudaCpp MEs ( 2 ) : 5.1075s for 16384 events => throughput is 3.21E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0209s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ds_taptamggds/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_epemggdux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_epemggdux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_epemggdux/ajob1' in 20.6899 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_epemggdux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_epemggdux/G208/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.6415s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0802s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.5406s for 16384 events => throughput is 6.45E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0207s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_epemggdux/G192/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.0137s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4494s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.5432s for 16384 events => throughput is 6.44E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0211s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_epemggdux/G380/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.2412s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6801s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.5398s for 16384 events => throughput is 6.45E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0213s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_epemggdux/G127/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.8597s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3038s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.5351s for 16384 events => throughput is 6.46E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0208s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_epemggdux/G223/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.8134s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2604s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.5313s for 16384 events => throughput is 6.47E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0217s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_epemggdux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_taptamgguc/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_taptamgguc/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_taptamgguc/ajob1' in 27.0174 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_taptamgguc/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_taptamgguc/G20/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 9.6887s + [COUNTERS] Fortran Overhead ( 0 ) : 4.5644s + [COUNTERS] CudaCpp MEs ( 2 ) : 5.1031s for 16384 events => throughput is 3.21E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0212s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_taptamgguc/G110/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.4790s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3644s + [COUNTERS] CudaCpp MEs ( 2 ) : 5.0938s for 16384 events => throughput is 3.22E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0208s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_taptamgguc/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 10.7814s + [COUNTERS] Fortran Overhead ( 0 ) : 5.6314s + [COUNTERS] CudaCpp MEs ( 2 ) : 5.1289s for 16384 events => throughput is 3.19E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0211s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_taptamgguc/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uxcx_taptamgguxcx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uxcx_taptamgguxcx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uxcx_taptamgguxcx/ajob1' in 7.7910 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uxcx_taptamgguxcx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uxcx_taptamgguxcx/G187/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.7431s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6342s + [COUNTERS] CudaCpp MEs ( 2 ) : 5.0868s for 16384 events => throughput is 3.22E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0221s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uxcx_taptamgguxcx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/ajob1' in 405.8713 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G499/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.3846s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2898s + [COUNTERS] CudaCpp MEs ( 2 ) : 12.0474s for 16384 events => throughput is 1.36E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0474s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G56/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 15.2002s + [COUNTERS] Fortran Overhead ( 0 ) : 3.0534s + [COUNTERS] CudaCpp MEs ( 2 ) : 12.0974s for 16384 events => throughput is 1.35E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0494s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G39/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.9152s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8046s + [COUNTERS] CudaCpp MEs ( 2 ) : 12.0634s for 16384 events => throughput is 1.36E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0472s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G64/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.7343s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6278s + [COUNTERS] CudaCpp MEs ( 2 ) : 12.0597s for 16384 events => throughput is 1.36E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0467s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G13/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.7790s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6546s + [COUNTERS] CudaCpp MEs ( 2 ) : 12.0783s for 16384 events => throughput is 1.36E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0460s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G33/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.9804s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8500s + [COUNTERS] CudaCpp MEs ( 2 ) : 12.0814s for 16384 events => throughput is 1.36E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0490s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G43/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 19.5560s + [COUNTERS] Fortran Overhead ( 0 ) : 7.2851s + [COUNTERS] CudaCpp MEs ( 2 ) : 12.2237s for 16384 events => throughput is 1.34E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0471s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G68/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 15.1951s + [COUNTERS] Fortran Overhead ( 0 ) : 3.0388s + [COUNTERS] CudaCpp MEs ( 2 ) : 12.1087s for 16384 events => throughput is 1.35E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0476s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G49/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.3114s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2027s + [COUNTERS] CudaCpp MEs ( 2 ) : 12.0613s for 16384 events => throughput is 1.36E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0475s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G25/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 18.0832s + [COUNTERS] Fortran Overhead ( 0 ) : 5.9663s + [COUNTERS] CudaCpp MEs ( 2 ) : 12.0687s for 16384 events => throughput is 1.36E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0482s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G523/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.9527s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8398s + [COUNTERS] CudaCpp MEs ( 2 ) : 12.0652s for 16384 events => throughput is 1.36E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0477s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G395/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.5869s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4887s + [COUNTERS] CudaCpp MEs ( 2 ) : 12.0510s for 16384 events => throughput is 1.36E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0473s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G524/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.1845s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9946s + [COUNTERS] CudaCpp MEs ( 2 ) : 12.1422s for 16384 events => throughput is 1.35E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0476s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G65/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.2213s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1034s + [COUNTERS] CudaCpp MEs ( 2 ) : 12.0707s for 16384 events => throughput is 1.36E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0471s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G609/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 15.1581s + [COUNTERS] Fortran Overhead ( 0 ) : 3.0525s + [COUNTERS] CudaCpp MEs ( 2 ) : 12.0581s for 16384 events => throughput is 1.36E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0475s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G611/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 15.5099s + [COUNTERS] Fortran Overhead ( 0 ) : 3.3836s + [COUNTERS] CudaCpp MEs ( 2 ) : 12.0786s for 16384 events => throughput is 1.36E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0477s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G1/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 16.8568s + [COUNTERS] Fortran Overhead ( 0 ) : 4.7708s + [COUNTERS] CudaCpp MEs ( 2 ) : 12.0391s for 16384 events => throughput is 1.36E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0469s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G55/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.8171s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7285s + [COUNTERS] CudaCpp MEs ( 2 ) : 12.0416s for 16384 events => throughput is 1.36E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0470s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G34/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.3107s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2380s + [COUNTERS] CudaCpp MEs ( 2 ) : 12.0246s for 16384 events => throughput is 1.36E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0482s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G26/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 18.6117s + [COUNTERS] Fortran Overhead ( 0 ) : 6.5183s + [COUNTERS] CudaCpp MEs ( 2 ) : 12.0461s for 16384 events => throughput is 1.36E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0473s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G58/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 17.9711s + [COUNTERS] Fortran Overhead ( 0 ) : 5.9087s + [COUNTERS] CudaCpp MEs ( 2 ) : 12.0150s for 16384 events => throughput is 1.36E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0474s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G2/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 17.5084s + [COUNTERS] Fortran Overhead ( 0 ) : 5.3404s + [COUNTERS] CudaCpp MEs ( 2 ) : 12.1213s for 16384 events => throughput is 1.35E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0466s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G512/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.4122s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2009s + [COUNTERS] CudaCpp MEs ( 2 ) : 12.1621s for 16384 events => throughput is 1.35E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0493s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G51/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.9156s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7890s + [COUNTERS] CudaCpp MEs ( 2 ) : 12.0791s for 16384 events => throughput is 1.36E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0475s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G29/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 18.9631s + [COUNTERS] Fortran Overhead ( 0 ) : 6.9048s + [COUNTERS] CudaCpp MEs ( 2 ) : 12.0115s for 16384 events => throughput is 1.36E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0468s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G48/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.2408s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1499s + [COUNTERS] CudaCpp MEs ( 2 ) : 12.0435s for 16384 events => throughput is 1.36E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0474s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/ajob1' in 73.6198 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/G140/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.4860s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9202s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.5453s for 16384 events => throughput is 6.44E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0206s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/G114/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 9.2732s + [COUNTERS] Fortran Overhead ( 0 ) : 6.7188s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.5336s for 16384 events => throughput is 6.47E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0208s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/G49/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 10.1620s + [COUNTERS] Fortran Overhead ( 0 ) : 7.6029s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.5380s for 16384 events => throughput is 6.46E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0211s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/G120/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.6696s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0951s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.5536s for 16384 events => throughput is 6.42E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0209s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/G262/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.3854s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8307s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.5342s for 16384 events => throughput is 6.47E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0205s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/G27/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.5195s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9517s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.5462s for 16384 events => throughput is 6.43E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0216s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/G44/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.2202s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6421s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.5576s for 16384 events => throughput is 6.41E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0205s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/G96/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.1445s + [COUNTERS] Fortran Overhead ( 0 ) : 3.5755s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.5479s for 16384 events => throughput is 6.43E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0211s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/G261/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.7486s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1817s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.5454s for 16384 events => throughput is 6.44E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0215s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/G37/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.6504s + [COUNTERS] Fortran Overhead ( 0 ) : 5.0773s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.5517s for 16384 events => throughput is 6.42E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0214s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/G28/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.4427s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8653s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.5564s for 16384 events => throughput is 6.41E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0210s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.6690s + [COUNTERS] Fortran Overhead ( 0 ) : 6.1187s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.5292s for 16384 events => throughput is 6.48E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0211s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_epemuduxsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_epemuduxsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_epemuduxsx/ajob1' in 3.1589 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_epemuduxsx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_epemuduxsx/G42/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.1300s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7262s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.3963s for 16384 events => throughput is 4.13E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0074s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_epemuduxsx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dd_epemggdd/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dd_epemggdd/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dd_epemggdd/ajob1' in 26.0621 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dd_epemggdd/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dd_epemggdd/G467/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.6816s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9061s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.7387s for 16384 events => throughput is 2.43E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0369s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dd_epemggdd/G351/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.1186s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3361s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.7462s for 16384 events => throughput is 2.43E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0363s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dd_epemggdd/G44/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 9.1734s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3214s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.8179s for 16384 events => throughput is 2.40E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0342s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dd_epemggdd/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamusdxsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamusdxsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamusdxsx/ajob1' in 2.1195 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamusdxsx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamusdxsx/G92/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.0740s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2641s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8022s for 16384 events => throughput is 2.04E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0076s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamusdxsx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemggddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemggddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemggddx/ajob1' in 59.1045 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemggddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemggddx/G175/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.9026s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1910s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.6763s for 16384 events => throughput is 2.45E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0352s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemggddx/G303/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 9.5147s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8053s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.6745s for 16384 events => throughput is 2.45E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0350s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemggddx/G111/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.5820s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8547s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.6922s for 16384 events => throughput is 2.45E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0351s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemggddx/G219/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 9.9883s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2072s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.7445s for 16384 events => throughput is 2.43E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0366s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemggddx/G318/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.2469s + [COUNTERS] Fortran Overhead ( 0 ) : 6.3238s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.8867s for 16384 events => throughput is 2.38E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0364s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemggddx/G143/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.7456s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0147s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.6952s for 16384 events => throughput is 2.45E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0357s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemggddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemguuxux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemguuxux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemguuxux/ajob1' in 14.0982 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemguuxux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemguuxux/G615/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.0506s + [COUNTERS] Fortran Overhead ( 0 ) : 7.3184s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.6965s for 16384 events => throughput is 2.45E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0358s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemguuxux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamgggg/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamgggg/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamgggg/ajob1' in 26.8387 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamgggg/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamgggg/G66/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 26.7869s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7442s + [COUNTERS] CudaCpp MEs ( 2 ) : 23.9953s for 16384 events => throughput is 6.83E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0474s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamgggg/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgduxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgduxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgduxdx/ajob1' in 7.3157 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgduxdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgduxdx/G44/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.2862s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1828s + [COUNTERS] CudaCpp MEs ( 2 ) : 5.0823s for 16384 events => throughput is 3.22E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0211s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgduxdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgudux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgudux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgudux/ajob1' in 57.2581 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgudux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgudux/G39/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 9.0005s + [COUNTERS] Fortran Overhead ( 0 ) : 6.4445s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.5352s for 16384 events => throughput is 6.46E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0207s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgudux/G366/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.4332s + [COUNTERS] Fortran Overhead ( 0 ) : 4.8797s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.5325s for 16384 events => throughput is 6.47E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0210s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgudux/G259/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.9541s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4121s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.5210s for 16384 events => throughput is 6.50E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0209s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgudux/G28/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.7603s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2009s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.5391s for 16384 events => throughput is 6.45E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0203s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgudux/G32/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.1865s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5734s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.5920s for 16384 events => throughput is 6.32E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0210s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgudux/G61/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 11.2488s + [COUNTERS] Fortran Overhead ( 0 ) : 8.7036s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.5243s for 16384 events => throughput is 6.49E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0209s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgudux/G267/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.4537s + [COUNTERS] Fortran Overhead ( 0 ) : 5.8985s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.5341s for 16384 events => throughput is 6.47E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0211s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgudux/G144/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.0614s + [COUNTERS] Fortran Overhead ( 0 ) : 3.5224s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.5183s for 16384 events => throughput is 6.51E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0207s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgudux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemuudux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemuudux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemuudux/ajob1' in 25.5120 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemuudux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemuudux/G104/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 25.4821s + [COUNTERS] Fortran Overhead ( 0 ) : 20.6082s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.8622s for 81920 events => throughput is 1.68E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0116s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemuudux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamggddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamggddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamggddx/ajob1' in 217.9367 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamggddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamggddx/G108/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 26.3489s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2970s + [COUNTERS] CudaCpp MEs ( 2 ) : 24.0048s for 16384 events => throughput is 6.83E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0471s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamggddx/G304/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 26.8104s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8511s + [COUNTERS] CudaCpp MEs ( 2 ) : 23.9107s for 16384 events => throughput is 6.85E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0486s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamggddx/G128/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 26.9550s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9671s + [COUNTERS] CudaCpp MEs ( 2 ) : 23.9416s for 16384 events => throughput is 6.84E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0462s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamggddx/G360/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 26.6203s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4622s + [COUNTERS] CudaCpp MEs ( 2 ) : 24.1095s for 16384 events => throughput is 6.80E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0485s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamggddx/G169/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 30.6891s + [COUNTERS] Fortran Overhead ( 0 ) : 6.4030s + [COUNTERS] CudaCpp MEs ( 2 ) : 24.2381s for 16384 events => throughput is 6.76E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0480s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamggddx/G176/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 26.1034s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0915s + [COUNTERS] CudaCpp MEs ( 2 ) : 23.9644s for 16384 events => throughput is 6.84E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0474s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamggddx/G303/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 26.9295s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9429s + [COUNTERS] CudaCpp MEs ( 2 ) : 23.9395s for 16384 events => throughput is 6.84E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0471s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamggddx/G513/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 27.3183s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2172s + [COUNTERS] CudaCpp MEs ( 2 ) : 24.0536s for 16384 events => throughput is 6.81E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0475s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamggddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguccx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguccx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguccx/ajob1' in 26.0894 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguccx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguccx/G283/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.4196s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0077s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.3926s for 16384 events => throughput is 6.85E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0193s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguccx/G47/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.3688s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9585s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.3904s for 16384 events => throughput is 6.85E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0199s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguccx/G35/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 9.1268s + [COUNTERS] Fortran Overhead ( 0 ) : 6.7142s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.3930s for 16384 events => throughput is 6.85E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0196s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguccx/G373/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.0908s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7059s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.3656s for 16384 events => throughput is 6.93E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0193s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguccx/G249/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.9563s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5413s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.3949s for 16384 events => throughput is 6.84E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0201s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguccx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/ajob1' in 449.4416 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G14/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.3812s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0219s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3162s for 16384 events => throughput is 1.45E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0431s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G499/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.2915s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9445s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3027s for 16384 events => throughput is 1.45E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0443s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G56/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.0716s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7046s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3236s for 16384 events => throughput is 1.45E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0434s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G39/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.6777s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3192s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3110s for 16384 events => throughput is 1.45E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0475s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G52/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.1480s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8125s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.2906s for 16384 events => throughput is 1.45E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0449s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G304/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.0203s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6319s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3413s for 16384 events => throughput is 1.44E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0471s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G64/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.7946s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3310s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.4183s for 16384 events => throughput is 1.43E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0452s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G3/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 16.4504s + [COUNTERS] Fortran Overhead ( 0 ) : 4.9559s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.4493s for 16384 events => throughput is 1.43E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0452s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G13/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.0030s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6515s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3078s for 16384 events => throughput is 1.45E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0437s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G33/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.4949s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1644s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.2876s for 16384 events => throughput is 1.45E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0429s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G484/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.9622s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5775s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3414s for 16384 events => throughput is 1.44E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0434s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G68/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.5187s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1878s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.2872s for 16384 events => throughput is 1.45E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0437s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G49/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.0729s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7707s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.2586s for 16384 events => throughput is 1.46E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0436s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G40/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.2559s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8363s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3707s for 16384 events => throughput is 1.44E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0489s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G523/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.4477s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1236s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.2788s for 16384 events => throughput is 1.45E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0453s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 12.9213s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5492s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3269s for 16384 events => throughput is 1.45E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0452s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G524/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.5126s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0866s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3810s for 16384 events => throughput is 1.44E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0450s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G65/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.4894s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1364s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3074s for 16384 events => throughput is 1.45E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0457s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G44/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 17.2651s + [COUNTERS] Fortran Overhead ( 0 ) : 5.7434s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.4776s for 16384 events => throughput is 1.43E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0442s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 12.8767s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5528s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.2806s for 16384 events => throughput is 1.45E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0433s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G303/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.3802s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9747s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3612s for 16384 events => throughput is 1.44E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0443s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G528/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.4964s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1397s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3134s for 16384 events => throughput is 1.45E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0433s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G55/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.3081s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9441s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3212s for 16384 events => throughput is 1.45E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0427s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G298/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.4572s + [COUNTERS] Fortran Overhead ( 0 ) : 3.0424s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3692s for 16384 events => throughput is 1.44E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0457s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G34/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.0438s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6915s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3084s for 16384 events => throughput is 1.45E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0439s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G54/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 18.4681s + [COUNTERS] Fortran Overhead ( 0 ) : 7.0649s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3578s for 16384 events => throughput is 1.44E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0454s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G67/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.7526s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2808s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.4276s for 16384 events => throughput is 1.43E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0442s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G512/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.2923s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9838s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.2650s for 16384 events => throughput is 1.45E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0435s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G511/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.9495s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6049s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3007s for 16384 events => throughput is 1.45E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0440s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.9048s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5062s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3552s for 16384 events => throughput is 1.44E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0435s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G21/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.2365s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7708s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.4219s for 16384 events => throughput is 1.43E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0438s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G48/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.8686s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5107s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3142s for 16384 events => throughput is 1.45E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0437s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_taptamggdux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_taptamggdux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_taptamggdux/ajob1' in 28.2410 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_taptamggdux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_taptamggdux/G68/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.6814s + [COUNTERS] Fortran Overhead ( 0 ) : 3.9013s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.7610s for 16384 events => throughput is 3.44E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0191s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_taptamggdux/G167/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.5698s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7803s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.7684s for 16384 events => throughput is 3.44E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0211s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_taptamggdux/G92/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.5527s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7617s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.7716s for 16384 events => throughput is 3.43E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0194s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_taptamggdux/G60/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.3465s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4820s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.8454s for 16384 events => throughput is 3.38E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0191s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_taptamggdux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/ajob1' in 321.5221 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G56/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.2227s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8447s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3351s for 16384 events => throughput is 1.45E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0429s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G39/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.5256s + [COUNTERS] Fortran Overhead ( 0 ) : 3.0727s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.4042s for 16384 events => throughput is 1.44E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0486s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G517/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 19.5144s + [COUNTERS] Fortran Overhead ( 0 ) : 8.1646s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3061s for 16384 events => throughput is 1.45E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0437s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G497/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.2886s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9386s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3066s for 16384 events => throughput is 1.45E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0434s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G13/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.0820s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7092s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3289s for 16384 events => throughput is 1.45E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0439s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G484/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 17.9824s + [COUNTERS] Fortran Overhead ( 0 ) : 6.6242s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3149s for 16384 events => throughput is 1.45E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0433s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G68/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.5320s + [COUNTERS] Fortran Overhead ( 0 ) : 3.0705s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.4173s for 16384 events => throughput is 1.44E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0442s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G49/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.7073s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3579s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3045s for 16384 events => throughput is 1.45E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0449s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G523/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.9042s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5307s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3296s for 16384 events => throughput is 1.45E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0439s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G518/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 17.3175s + [COUNTERS] Fortran Overhead ( 0 ) : 5.8798s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3935s for 16384 events => throughput is 1.44E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0443s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G42/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 19.6436s + [COUNTERS] Fortran Overhead ( 0 ) : 8.1805s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.4187s for 16384 events => throughput is 1.43E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0444s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G524/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.3140s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9103s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3599s for 16384 events => throughput is 1.44E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0439s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G610/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.5146s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1219s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3493s for 16384 events => throughput is 1.44E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0434s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G26/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 17.5285s + [COUNTERS] Fortran Overhead ( 0 ) : 6.0141s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.4666s for 16384 events => throughput is 1.43E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0478s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G58/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 18.6852s + [COUNTERS] Fortran Overhead ( 0 ) : 7.3011s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3365s for 16384 events => throughput is 1.45E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0477s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G511/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.5249s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1304s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3465s for 16384 events => throughput is 1.44E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0480s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.3052s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8509s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.4082s for 16384 events => throughput is 1.44E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0461s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G32/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 19.3126s + [COUNTERS] Fortran Overhead ( 0 ) : 7.9465s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3227s for 16384 events => throughput is 1.45E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0434s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G61/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 17.4502s + [COUNTERS] Fortran Overhead ( 0 ) : 6.0847s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3204s for 16384 events => throughput is 1.45E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0451s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G390/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 20.7495s + [COUNTERS] Fortran Overhead ( 0 ) : 9.2778s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.4281s for 16384 events => throughput is 1.43E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0435s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_epemgguc/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_epemgguc/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_epemgguc/ajob1' in 18.8640 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_epemgguc/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_epemgguc/G164/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.8781s + [COUNTERS] Fortran Overhead ( 0 ) : 4.4605s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.3979s for 16384 events => throughput is 6.83E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0197s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_epemgguc/G24/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.3592s + [COUNTERS] Fortran Overhead ( 0 ) : 3.9815s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.3583s for 16384 events => throughput is 6.95E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0193s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_epemgguc/G162/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.5428s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1607s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.3624s for 16384 events => throughput is 6.94E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0198s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_epemgguc/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/ajob1' in 413.5089 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G336/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 15.0277s + [COUNTERS] Fortran Overhead ( 0 ) : 3.5643s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.4193s for 16384 events => throughput is 1.43E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0441s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G108/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.8728s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4765s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3521s for 16384 events => throughput is 1.44E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0442s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G105/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.7277s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3347s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3493s for 16384 events => throughput is 1.44E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0437s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G304/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.4634s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1181s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3010s for 16384 events => throughput is 1.45E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0443s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G175/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.5229s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2042s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.2754s for 16384 events => throughput is 1.45E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0434s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G128/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.9473s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5457s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3578s for 16384 events => throughput is 1.44E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0438s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G360/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.8117s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4944s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.2723s for 16384 events => throughput is 1.45E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0450s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G120/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.5927s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2126s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3345s for 16384 events => throughput is 1.45E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0457s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G296/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.4397s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0780s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3183s for 16384 events => throughput is 1.45E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0434s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G299/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.1098s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7901s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.2751s for 16384 events => throughput is 1.45E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0446s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G715/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.9685s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6185s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3059s for 16384 events => throughput is 1.45E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0441s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G300/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.8366s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4003s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3928s for 16384 events => throughput is 1.44E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0436s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G303/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.2177s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8185s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3545s for 16384 events => throughput is 1.44E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0447s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G127/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.1821s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8250s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3124s for 16384 events => throughput is 1.45E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0446s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G111/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.9067s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5242s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3388s for 16384 events => throughput is 1.44E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0437s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G126/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 21.7584s + [COUNTERS] Fortran Overhead ( 0 ) : 10.3990s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3147s for 16384 events => throughput is 1.45E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0447s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G294/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 17.9625s + [COUNTERS] Fortran Overhead ( 0 ) : 6.6243s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.2932s for 16384 events => throughput is 1.45E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0450s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G598/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.7889s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3903s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3539s for 16384 events => throughput is 1.44E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0448s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G610/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.6151s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2982s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.2725s for 16384 events => throughput is 1.45E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0443s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G178/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.9018s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5643s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.2930s for 16384 events => throughput is 1.45E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0445s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G121/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.9076s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4380s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.4255s for 16384 events => throughput is 1.43E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0441s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G685/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.5523s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1577s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3501s for 16384 events => throughput is 1.44E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0445s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G690/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.5424s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0829s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.4140s for 16384 events => throughput is 1.44E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0454s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G103/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 19.3510s + [COUNTERS] Fortran Overhead ( 0 ) : 8.0460s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.2594s for 16384 events => throughput is 1.46E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0456s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G107/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.6519s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3016s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3061s for 16384 events => throughput is 1.45E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0441s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G180/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.7416s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4511s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.2465s for 16384 events => throughput is 1.46E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0440s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G97/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 19.4714s + [COUNTERS] Fortran Overhead ( 0 ) : 8.1509s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.2766s for 16384 events => throughput is 1.45E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0438s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G179/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.0675s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7217s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3009s for 16384 events => throughput is 1.45E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0449s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamggddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamggddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamggddx/ajob1' in 7.3094 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamggddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamggddx/G189/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.2788s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4763s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.7831s for 16384 events => throughput is 3.43E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0194s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamggddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dd_taptamggdd/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dd_taptamggdd/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dd_taptamggdd/ajob1' in 14.7691 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dd_taptamggdd/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dd_taptamggdd/G463/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.7379s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1320s + [COUNTERS] CudaCpp MEs ( 2 ) : 12.5721s for 16384 events => throughput is 1.30E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0337s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dd_taptamggdd/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/ajob1' in 1046.3658 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G506/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 26.9533s + [COUNTERS] Fortran Overhead ( 0 ) : 4.2076s + [COUNTERS] CudaCpp MEs ( 2 ) : 22.7004s for 16384 events => throughput is 7.22E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0453s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G499/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 24.6251s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8270s + [COUNTERS] CudaCpp MEs ( 2 ) : 22.7508s for 16384 events => throughput is 7.20E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0473s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G46/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 30.1239s + [COUNTERS] Fortran Overhead ( 0 ) : 7.4796s + [COUNTERS] CudaCpp MEs ( 2 ) : 22.5998s for 16384 events => throughput is 7.25E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0444s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G56/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 24.9943s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4958s + [COUNTERS] CudaCpp MEs ( 2 ) : 22.4544s for 16384 events => throughput is 7.30E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0440s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G39/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 25.0306s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3593s + [COUNTERS] CudaCpp MEs ( 2 ) : 22.6255s for 16384 events => throughput is 7.24E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0458s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G517/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 27.7420s + [COUNTERS] Fortran Overhead ( 0 ) : 5.0966s + [COUNTERS] CudaCpp MEs ( 2 ) : 22.6011s for 16384 events => throughput is 7.25E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0443s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G52/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 25.4693s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8646s + [COUNTERS] CudaCpp MEs ( 2 ) : 22.5602s for 16384 events => throughput is 7.26E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0445s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G31/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 29.2066s + [COUNTERS] Fortran Overhead ( 0 ) : 6.6147s + [COUNTERS] CudaCpp MEs ( 2 ) : 22.5468s for 16384 events => throughput is 7.27E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0451s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G324/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 28.9435s + [COUNTERS] Fortran Overhead ( 0 ) : 6.3280s + [COUNTERS] CudaCpp MEs ( 2 ) : 22.5716s for 16384 events => throughput is 7.26E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0439s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G64/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 24.5395s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8429s + [COUNTERS] CudaCpp MEs ( 2 ) : 22.6509s for 16384 events => throughput is 7.23E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0456s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G3/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 26.6272s + [COUNTERS] Fortran Overhead ( 0 ) : 3.8889s + [COUNTERS] CudaCpp MEs ( 2 ) : 22.6894s for 16384 events => throughput is 7.22E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0489s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G33/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 24.7616s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1549s + [COUNTERS] CudaCpp MEs ( 2 ) : 22.5631s for 16384 events => throughput is 7.26E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0436s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G53/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 29.1372s + [COUNTERS] Fortran Overhead ( 0 ) : 6.4026s + [COUNTERS] CudaCpp MEs ( 2 ) : 22.6914s for 16384 events => throughput is 7.22E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0432s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G43/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 27.8651s + [COUNTERS] Fortran Overhead ( 0 ) : 5.3042s + [COUNTERS] CudaCpp MEs ( 2 ) : 22.5155s for 16384 events => throughput is 7.28E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0454s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G49/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 24.7097s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7655s + [COUNTERS] CudaCpp MEs ( 2 ) : 22.9011s for 16384 events => throughput is 7.15E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0430s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G40/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 25.5217s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8500s + [COUNTERS] CudaCpp MEs ( 2 ) : 22.6248s for 16384 events => throughput is 7.24E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0470s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G523/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 24.6226s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1205s + [COUNTERS] CudaCpp MEs ( 2 ) : 22.4588s for 16384 events => throughput is 7.30E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0433s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G518/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 27.7830s + [COUNTERS] Fortran Overhead ( 0 ) : 5.2463s + [COUNTERS] CudaCpp MEs ( 2 ) : 22.4917s for 16384 events => throughput is 7.28E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0450s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G35/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 25.2833s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9056s + [COUNTERS] CudaCpp MEs ( 2 ) : 23.3327s for 16384 events => throughput is 7.02E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0450s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 24.1121s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6077s + [COUNTERS] CudaCpp MEs ( 2 ) : 22.4593s for 16384 events => throughput is 7.29E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0451s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G42/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 27.9576s + [COUNTERS] Fortran Overhead ( 0 ) : 5.3507s + [COUNTERS] CudaCpp MEs ( 2 ) : 22.5631s for 16384 events => throughput is 7.26E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0438s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G524/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 24.7522s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0860s + [COUNTERS] CudaCpp MEs ( 2 ) : 22.6215s for 16384 events => throughput is 7.24E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0447s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G65/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 24.5260s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8522s + [COUNTERS] CudaCpp MEs ( 2 ) : 22.6293s for 16384 events => throughput is 7.24E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0446s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G27/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 28.3607s + [COUNTERS] Fortran Overhead ( 0 ) : 5.6828s + [COUNTERS] CudaCpp MEs ( 2 ) : 22.6334s for 16384 events => throughput is 7.24E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0446s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G44/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 28.6923s + [COUNTERS] Fortran Overhead ( 0 ) : 6.1422s + [COUNTERS] CudaCpp MEs ( 2 ) : 22.5067s for 16384 events => throughput is 7.28E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0434s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 24.0669s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5977s + [COUNTERS] CudaCpp MEs ( 2 ) : 22.4247s for 16384 events => throughput is 7.31E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0446s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G303/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 25.7790s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1075s + [COUNTERS] CudaCpp MEs ( 2 ) : 22.6269s for 16384 events => throughput is 7.24E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0446s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G55/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 25.3909s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6853s + [COUNTERS] CudaCpp MEs ( 2 ) : 22.6607s for 16384 events => throughput is 7.23E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0450s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G34/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 24.3742s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7629s + [COUNTERS] CudaCpp MEs ( 2 ) : 22.5678s for 16384 events => throughput is 7.26E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0435s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G26/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 28.4585s + [COUNTERS] Fortran Overhead ( 0 ) : 5.7806s + [COUNTERS] CudaCpp MEs ( 2 ) : 22.6338s for 16384 events => throughput is 7.24E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0442s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G62/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 28.4989s + [COUNTERS] Fortran Overhead ( 0 ) : 5.8651s + [COUNTERS] CudaCpp MEs ( 2 ) : 22.5885s for 16384 events => throughput is 7.25E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0452s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G37/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 24.7203s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1202s + [COUNTERS] CudaCpp MEs ( 2 ) : 22.5560s for 16384 events => throughput is 7.26E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0440s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G485/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 27.8092s + [COUNTERS] Fortran Overhead ( 0 ) : 5.1927s + [COUNTERS] CudaCpp MEs ( 2 ) : 22.5710s for 16384 events => throughput is 7.26E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0455s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G67/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 24.7641s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1281s + [COUNTERS] CudaCpp MEs ( 2 ) : 22.5920s for 16384 events => throughput is 7.25E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0440s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G512/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 24.5548s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8022s + [COUNTERS] CudaCpp MEs ( 2 ) : 22.7087s for 16384 events => throughput is 7.21E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0439s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G511/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 24.9871s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4117s + [COUNTERS] CudaCpp MEs ( 2 ) : 22.5320s for 16384 events => throughput is 7.27E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0434s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G51/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 25.1759s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4424s + [COUNTERS] CudaCpp MEs ( 2 ) : 22.6875s for 16384 events => throughput is 7.22E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0460s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 24.9361s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3696s + [COUNTERS] CudaCpp MEs ( 2 ) : 22.5221s for 16384 events => throughput is 7.27E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0444s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G63/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 24.8173s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2620s + [COUNTERS] CudaCpp MEs ( 2 ) : 22.5099s for 16384 events => throughput is 7.28E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0454s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G48/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 24.9039s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3870s + [COUNTERS] CudaCpp MEs ( 2 ) : 22.4685s for 16384 events => throughput is 7.29E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0484s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgggux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgggux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgggux/ajob1' in 187.6255 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgggux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgggux/G33/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 24.5040s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7635s + [COUNTERS] CudaCpp MEs ( 2 ) : 22.6964s for 16384 events => throughput is 7.22E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0440s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgggux/G68/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 25.7538s + [COUNTERS] Fortran Overhead ( 0 ) : 3.0526s + [COUNTERS] CudaCpp MEs ( 2 ) : 22.6553s for 16384 events => throughput is 7.23E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0459s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgggux/G25/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 29.0403s + [COUNTERS] Fortran Overhead ( 0 ) : 6.2921s + [COUNTERS] CudaCpp MEs ( 2 ) : 22.7025s for 16384 events => throughput is 7.22E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0457s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgggux/G494/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 24.3615s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6238s + [COUNTERS] CudaCpp MEs ( 2 ) : 22.6928s for 16384 events => throughput is 7.22E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0449s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgggux/G55/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 25.5452s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8692s + [COUNTERS] CudaCpp MEs ( 2 ) : 22.6295s for 16384 events => throughput is 7.24E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0465s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgggux/G509/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 29.7040s + [COUNTERS] Fortran Overhead ( 0 ) : 6.9999s + [COUNTERS] CudaCpp MEs ( 2 ) : 22.6562s for 16384 events => throughput is 7.23E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0480s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgggux/G29/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 28.5512s + [COUNTERS] Fortran Overhead ( 0 ) : 5.9246s + [COUNTERS] CudaCpp MEs ( 2 ) : 22.5820s for 16384 events => throughput is 7.26E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0445s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgggux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamgguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamgguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamgguux/ajob1' in 76.9869 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamgguux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamgguux/G260/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.2061s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5437s + [COUNTERS] CudaCpp MEs ( 2 ) : 12.6282s for 16384 events => throughput is 1.30E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0342s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamgguux/G447/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.3654s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5704s + [COUNTERS] CudaCpp MEs ( 2 ) : 12.7609s for 16384 events => throughput is 1.28E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0341s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamgguux/G229/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 16.6338s + [COUNTERS] Fortran Overhead ( 0 ) : 4.0201s + [COUNTERS] CudaCpp MEs ( 2 ) : 12.5801s for 16384 events => throughput is 1.30E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0337s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamgguux/G135/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.6586s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9642s + [COUNTERS] CudaCpp MEs ( 2 ) : 12.6595s for 16384 events => throughput is 1.29E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0348s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamgguux/G230/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 16.9941s + [COUNTERS] Fortran Overhead ( 0 ) : 4.2693s + [COUNTERS] CudaCpp MEs ( 2 ) : 12.6920s for 16384 events => throughput is 1.29E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0329s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamgguux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdddx/ajob1' in 16.8638 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdddx/G715/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.3712s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0616s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.2753s for 16384 events => throughput is 2.61E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0343s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdddx/G111/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.4236s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0084s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.3784s for 16384 events => throughput is 2.57E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0368s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemusdxsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemusdxsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemusdxsx/ajob1' in 6.5177 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemusdxsx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemusdxsx/G67/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.4575s + [COUNTERS] Fortran Overhead ( 0 ) : 6.0830s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.3672s for 16384 events => throughput is 4.46E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0072s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemusdxsx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdssx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdssx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdssx/ajob1' in 26.4243 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdssx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdssx/G31/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.2928s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8995s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.3726s for 16384 events => throughput is 6.91E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0207s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdssx/G27/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.3282s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9084s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.4004s for 16384 events => throughput is 6.83E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0194s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdssx/G119/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.5498s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1451s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.3849s for 16384 events => throughput is 6.87E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0198s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdssx/G284/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.7444s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3502s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.3754s for 16384 events => throughput is 6.90E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0188s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdssx/G375/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.9742s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5565s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.3980s for 16384 events => throughput is 6.83E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0196s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdssx/G102/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.3910s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9962s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.3746s for 16384 events => throughput is 6.90E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0202s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdssx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_epemgguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_epemgguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_epemgguux/ajob1' in 38.3559 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_epemgguux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_epemgguux/G105/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.8324s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5279s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.2708s for 16384 events => throughput is 2.61E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0337s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_epemgguux/G223/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.5417s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2130s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.2953s for 16384 events => throughput is 2.60E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0334s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_epemgguux/G103/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 10.7442s + [COUNTERS] Fortran Overhead ( 0 ) : 4.4273s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.2849s for 16384 events => throughput is 2.61E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0319s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_epemgguux/G5/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 11.1282s + [COUNTERS] Fortran Overhead ( 0 ) : 4.8141s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.2811s for 16384 events => throughput is 2.61E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0330s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_epemgguux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/ajob1' in 945.5444 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G506/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 16.8150s + [COUNTERS] Fortran Overhead ( 0 ) : 5.3846s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3856s for 16384 events => throughput is 1.44E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0447s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G499/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.2677s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9199s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3039s for 16384 events => throughput is 1.45E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0440s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G38/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.1190s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7015s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3723s for 16384 events => throughput is 1.44E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0452s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G46/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 19.2335s + [COUNTERS] Fortran Overhead ( 0 ) : 7.7819s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.4081s for 16384 events => throughput is 1.44E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0435s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G59/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 17.5226s + [COUNTERS] Fortran Overhead ( 0 ) : 6.1589s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3180s for 16384 events => throughput is 1.45E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0457s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G56/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.0167s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7300s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.2428s for 16384 events => throughput is 1.46E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0439s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G525/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.7759s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3638s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3683s for 16384 events => throughput is 1.44E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0438s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G39/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.7014s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3053s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3524s for 16384 events => throughput is 1.44E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0437s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G517/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 16.9873s + [COUNTERS] Fortran Overhead ( 0 ) : 5.6265s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3182s for 16384 events => throughput is 1.45E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0426s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G487/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 16.0713s + [COUNTERS] Fortran Overhead ( 0 ) : 4.7373s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.2902s for 16384 events => throughput is 1.45E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0438s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G52/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.2480s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8884s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3140s for 16384 events => throughput is 1.45E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0456s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G31/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 19.4033s + [COUNTERS] Fortran Overhead ( 0 ) : 8.0868s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.2719s for 16384 events => throughput is 1.45E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0446s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G64/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.3632s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0070s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3128s for 16384 events => throughput is 1.45E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0434s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G3/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 16.2525s + [COUNTERS] Fortran Overhead ( 0 ) : 4.8929s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3159s for 16384 events => throughput is 1.45E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0437s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G497/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.5317s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3412s + [COUNTERS] CudaCpp MEs ( 2 ) : 12.1412s for 16384 events => throughput is 1.35E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0493s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G13/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.0079s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6403s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3231s for 16384 events => throughput is 1.45E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0444s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G66/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.7576s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3748s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3382s for 16384 events => throughput is 1.45E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0446s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G33/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.3411s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9453s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3521s for 16384 events => throughput is 1.44E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0437s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G53/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 19.3669s + [COUNTERS] Fortran Overhead ( 0 ) : 7.9172s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.4057s for 16384 events => throughput is 1.44E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0440s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G43/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 16.3373s + [COUNTERS] Fortran Overhead ( 0 ) : 4.8300s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.4636s for 16384 events => throughput is 1.43E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0437s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G68/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.6231s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1593s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.4157s for 16384 events => throughput is 1.44E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0481s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G49/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.2324s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8876s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3005s for 16384 events => throughput is 1.45E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0442s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G302/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 19.5407s + [COUNTERS] Fortran Overhead ( 0 ) : 8.1846s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3112s for 16384 events => throughput is 1.45E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0449s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G25/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 17.5572s + [COUNTERS] Fortran Overhead ( 0 ) : 6.1585s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3550s for 16384 events => throughput is 1.44E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0437s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G40/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.3987s + [COUNTERS] Fortran Overhead ( 0 ) : 3.0409s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3137s for 16384 events => throughput is 1.45E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0441s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G523/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.3441s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9912s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3090s for 16384 events => throughput is 1.45E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0439s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G518/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 16.5862s + [COUNTERS] Fortran Overhead ( 0 ) : 5.0455s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.4958s for 16384 events => throughput is 1.43E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0450s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G35/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.4908s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0965s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3498s for 16384 events => throughput is 1.44E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0445s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.2549s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6907s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.5172s for 16384 events => throughput is 1.42E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0470s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G524/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.5370s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0641s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.4280s for 16384 events => throughput is 1.43E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0449s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G65/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.4656s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1256s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.2947s for 16384 events => throughput is 1.45E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0454s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G27/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 17.6969s + [COUNTERS] Fortran Overhead ( 0 ) : 6.3947s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.2586s for 16384 events => throughput is 1.46E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0436s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G44/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 17.3693s + [COUNTERS] Fortran Overhead ( 0 ) : 6.0051s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3196s for 16384 events => throughput is 1.45E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0445s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G483/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 15.6374s + [COUNTERS] Fortran Overhead ( 0 ) : 4.1750s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.4175s for 16384 events => throughput is 1.43E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0450s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G482/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 16.8968s + [COUNTERS] Fortran Overhead ( 0 ) : 5.5306s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3230s for 16384 events => throughput is 1.45E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0431s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G1/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 15.9041s + [COUNTERS] Fortran Overhead ( 0 ) : 4.4522s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.4080s for 16384 events => throughput is 1.44E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0439s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.0780s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5854s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.4495s for 16384 events => throughput is 1.43E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0431s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G55/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.8282s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4308s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3504s for 16384 events => throughput is 1.44E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0471s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G298/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.3887s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9747s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3668s for 16384 events => throughput is 1.44E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0471s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G513/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.2290s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8443s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3411s for 16384 events => throughput is 1.44E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0437s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G60/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 18.7246s + [COUNTERS] Fortran Overhead ( 0 ) : 7.3851s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.2949s for 16384 events => throughput is 1.45E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0446s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G34/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.3224s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9651s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3108s for 16384 events => throughput is 1.45E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0464s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G26/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 18.1589s + [COUNTERS] Fortran Overhead ( 0 ) : 6.7894s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3255s for 16384 events => throughput is 1.45E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0441s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G58/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 16.9400s + [COUNTERS] Fortran Overhead ( 0 ) : 5.3359s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.5587s for 16384 events => throughput is 1.42E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0453s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G54/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 17.2818s + [COUNTERS] Fortran Overhead ( 0 ) : 5.8571s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3793s for 16384 events => throughput is 1.44E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0455s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G62/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 18.2197s + [COUNTERS] Fortran Overhead ( 0 ) : 6.8125s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3624s for 16384 events => throughput is 1.44E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0449s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G37/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.2250s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8565s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3236s for 16384 events => throughput is 1.45E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0450s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G493/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.5768s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0986s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.4348s for 16384 events => throughput is 1.43E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0434s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G4/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 17.4982s + [COUNTERS] Fortran Overhead ( 0 ) : 6.1374s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3171s for 16384 events => throughput is 1.45E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0436s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G50/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.4354s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0984s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.2937s for 16384 events => throughput is 1.45E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0433s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G485/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 17.1364s + [COUNTERS] Fortran Overhead ( 0 ) : 5.6764s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.4164s for 16384 events => throughput is 1.44E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0436s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G41/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 19.3928s + [COUNTERS] Fortran Overhead ( 0 ) : 7.2674s + [COUNTERS] CudaCpp MEs ( 2 ) : 12.0763s for 16384 events => throughput is 1.36E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0491s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G67/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.3814s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0746s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.2629s for 16384 events => throughput is 1.45E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0438s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G512/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.3851s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9956s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3451s for 16384 events => throughput is 1.44E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0444s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G511/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.8693s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4982s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3275s for 16384 events => throughput is 1.45E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0436s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G57/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 18.2236s + [COUNTERS] Fortran Overhead ( 0 ) : 6.8782s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3005s for 16384 events => throughput is 1.45E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0449s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.2798s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8559s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3791s for 16384 events => throughput is 1.44E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0447s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G63/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.7837s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3884s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3500s for 16384 events => throughput is 1.44E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0453s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G32/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 19.1109s + [COUNTERS] Fortran Overhead ( 0 ) : 7.6104s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.4542s for 16384 events => throughput is 1.43E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0464s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G29/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 17.4942s + [COUNTERS] Fortran Overhead ( 0 ) : 6.0922s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3582s for 16384 events => throughput is 1.44E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0437s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G48/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.7454s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3122s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3897s for 16384 events => throughput is 1.44E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0435s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_taptamggdsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_taptamggdsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_taptamggdsx/ajob1' in 21.0512 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_taptamggdsx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_taptamggdsx/G92/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.2448s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4000s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.8247s for 16384 events => throughput is 3.40E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0201s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_taptamggdsx/G144/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.8946s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0888s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.7857s for 16384 events => throughput is 3.42E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0200s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_taptamggdsx/G207/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.8447s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0280s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.7976s for 16384 events => throughput is 3.42E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0191s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_taptamggdsx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_epemggdsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_epemggdsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_epemggdsx/ajob1' in 9.0745 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_epemggdsx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_epemggdsx/G205/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 9.0408s + [COUNTERS] Fortran Overhead ( 0 ) : 6.6401s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.3810s for 16384 events => throughput is 6.88E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0197s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_epemggdsx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_taptamgggg/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_taptamgggg/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_taptamgggg/ajob1' in 25.4020 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_taptamgggg/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_taptamgggg/G67/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 25.3723s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6937s + [COUNTERS] CudaCpp MEs ( 2 ) : 22.6344s for 16384 events => throughput is 7.24E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0442s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_taptamgggg/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dc_epemdcssx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dc_epemdcssx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dc_epemdcssx/ajob1' in 2.6705 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dc_epemdcssx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dc_epemdcssx/G77/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.6241s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2321s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.3845s for 16384 events => throughput is 4.26E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0076s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dc_epemdcssx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguccx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguccx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguccx/ajob1' in 7.2589 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguccx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguccx/G48/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.2284s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4337s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.7735s for 16384 events => throughput is 3.43E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0212s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguccx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ds_epemggds/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ds_epemggds/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ds_epemggds/ajob1' in 4.0220 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ds_epemggds/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ds_epemggds/G26/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.9917s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5973s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.3745s for 16384 events => throughput is 6.90E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0199s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ds_epemggds/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ucx_epemggucx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ucx_epemggucx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ucx_epemggucx/ajob1' in 10.4195 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ucx_epemggucx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ucx_epemggucx/G163/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.4101s + [COUNTERS] Fortran Overhead ( 0 ) : 4.0056s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.3847s for 16384 events => throughput is 6.87E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0198s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ucx_epemggucx/G60/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.9430s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5448s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.3790s for 16384 events => throughput is 6.89E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0192s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ucx_epemggucx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemggudx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemggudx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemggudx/ajob1' in 69.5087 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemggudx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemggudx/G109/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.8673s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4778s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.3704s for 16384 events => throughput is 6.91E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0191s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemggudx/G151/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.2037s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8058s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.3785s for 16384 events => throughput is 6.89E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0194s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemggudx/G106/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.6221s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2220s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.3804s for 16384 events => throughput is 6.88E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0197s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemggudx/G35/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 9.2322s + [COUNTERS] Fortran Overhead ( 0 ) : 6.8234s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.3882s for 16384 events => throughput is 6.86E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0206s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemggudx/G119/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.4071s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9996s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.3861s for 16384 events => throughput is 6.87E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0213s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemggudx/G187/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.2323s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8250s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.3882s for 16384 events => throughput is 6.86E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0191s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemggudx/G91/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.3935s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9861s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.3868s for 16384 events => throughput is 6.86E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0206s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemggudx/G363/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 26.5804s + [COUNTERS] Fortran Overhead ( 0 ) : 17.0331s + [COUNTERS] CudaCpp MEs ( 2 ) : 9.5274s for 65536 events => throughput is 6.88E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0199s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemggudx/G149/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.7884s + [COUNTERS] Fortran Overhead ( 0 ) : 4.3734s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.3962s for 16384 events => throughput is 6.84E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0189s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemggudx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgcuxcx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgcuxcx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgcuxcx/ajob1' in 4.8915 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgcuxcx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgcuxcx/G120/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.8355s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4405s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.3758s for 16384 events => throughput is 6.90E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0192s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgcuxcx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemgggg/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemgggg/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemgggg/ajob1' in 29.7470 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemgggg/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemgggg/G65/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.8439s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1982s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.5992s for 16384 events => throughput is 1.41E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0465s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemgggg/G48/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.8328s + [COUNTERS] Fortran Overhead ( 0 ) : 3.4208s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3687s for 16384 events => throughput is 1.44E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0433s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemgggg/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_taptamggud/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_taptamggud/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_taptamggud/ajob1' in 63.9947 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_taptamggud/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_taptamggud/G46/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.9805s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1727s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.7882s for 16384 events => throughput is 3.42E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0196s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_taptamggud/G39/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.9122s + [COUNTERS] Fortran Overhead ( 0 ) : 4.1128s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.7782s for 16384 events => throughput is 3.43E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0212s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_taptamggud/G112/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.0286s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2331s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.7758s for 16384 events => throughput is 3.43E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0197s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_taptamggud/G115/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.2728s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4778s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.7750s for 16384 events => throughput is 3.43E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0200s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_taptamggud/G270/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.7135s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8743s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.8192s for 16384 events => throughput is 3.40E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0200s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_taptamggud/G96/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.4222s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5758s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.8255s for 16384 events => throughput is 3.40E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0209s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_taptamggud/G107/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.6098s + [COUNTERS] Fortran Overhead ( 0 ) : 3.7920s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.7986s for 16384 events => throughput is 3.41E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0193s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_taptamggud/G63/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.7070s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8536s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.8327s for 16384 events => throughput is 3.39E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0207s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_taptamggud/G375/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.1510s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2806s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.8490s for 16384 events => throughput is 3.38E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0215s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_taptamggud/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguuux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguuux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguuux/ajob1' in 41.9174 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguuux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguuux/G122/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.1852s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8383s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.3132s for 16384 events => throughput is 2.60E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0337s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguuux/G33/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 12.9571s + [COUNTERS] Fortran Overhead ( 0 ) : 5.9831s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.9371s for 16384 events => throughput is 2.36E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0368s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguuux/G127/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 9.2823s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9716s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.2772s for 16384 events => throughput is 2.61E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0336s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguuux/G116/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 11.4070s + [COUNTERS] Fortran Overhead ( 0 ) : 5.0092s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.3635s for 16384 events => throughput is 2.57E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0343s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguuux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamgggdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamgggdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamgggdx/ajob1' in 223.4761 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamgggdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamgggdx/G56/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 25.1219s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5477s + [COUNTERS] CudaCpp MEs ( 2 ) : 22.5294s for 16384 events => throughput is 7.27E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0448s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamgggdx/G64/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 25.1653s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3781s + [COUNTERS] CudaCpp MEs ( 2 ) : 22.7396s for 16384 events => throughput is 7.21E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0476s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamgggdx/G13/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 24.3234s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6418s + [COUNTERS] CudaCpp MEs ( 2 ) : 22.6360s for 16384 events => throughput is 7.24E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0455s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamgggdx/G33/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 24.4117s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8253s + [COUNTERS] CudaCpp MEs ( 2 ) : 22.5421s for 16384 events => throughput is 7.27E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0443s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamgggdx/G49/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 24.8897s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1082s + [COUNTERS] CudaCpp MEs ( 2 ) : 22.7373s for 16384 events => throughput is 7.21E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0442s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamgggdx/G523/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 25.2203s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5650s + [COUNTERS] CudaCpp MEs ( 2 ) : 22.6114s for 16384 events => throughput is 7.25E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0439s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamgggdx/G65/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 24.6296s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9696s + [COUNTERS] CudaCpp MEs ( 2 ) : 22.6145s for 16384 events => throughput is 7.24E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0456s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamgggdx/G34/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 24.8922s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0894s + [COUNTERS] CudaCpp MEs ( 2 ) : 22.7579s for 16384 events => throughput is 7.20E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0449s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamgggdx/G48/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 24.6251s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9840s + [COUNTERS] CudaCpp MEs ( 2 ) : 22.5977s for 16384 events => throughput is 7.25E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0434s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamgggdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgudux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgudux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgudux/ajob1' in 17.0784 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgudux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgudux/G59/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.6458s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8521s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.7741s for 16384 events => throughput is 3.43E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0196s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgudux/G61/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 10.3684s + [COUNTERS] Fortran Overhead ( 0 ) : 5.5799s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.7692s for 16384 events => throughput is 3.44E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0193s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgudux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgduxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgduxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgduxdx/ajob1' in 4.4549 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgduxdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgduxdx/G104/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.4070s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8917s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.4959s for 16384 events => throughput is 6.56E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0194s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgduxdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgsdxsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgsdxsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgsdxsx/ajob1' in 4.2117 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgsdxsx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgsdxsx/G25/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.1822s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7392s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.4239s for 16384 events => throughput is 6.76E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0191s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgsdxsx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/ajob1' in 142.5000 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G208/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.3515s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9320s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.3991s for 16384 events => throughput is 6.83E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0204s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G235/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.2690s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6985s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.5505s for 16384 events => throughput is 6.42E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0200s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G140/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.4272s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0041s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.4022s for 16384 events => throughput is 6.82E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0209s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G109/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.0748s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6714s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.3834s for 16384 events => throughput is 6.87E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0200s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G33/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.1183s + [COUNTERS] Fortran Overhead ( 0 ) : 4.6949s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.4036s for 16384 events => throughput is 6.82E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0199s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G43/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.2281s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7821s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.4266s for 16384 events => throughput is 6.75E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0194s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G240/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.1768s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7658s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.3915s for 16384 events => throughput is 6.85E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0195s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G268/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.8203s + [COUNTERS] Fortran Overhead ( 0 ) : 4.3923s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.4074s for 16384 events => throughput is 6.81E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0205s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G139/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.3909s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9642s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.4073s for 16384 events => throughput is 6.81E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0194s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G120/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.1343s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7233s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.3913s for 16384 events => throughput is 6.85E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0197s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G236/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.1924s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7803s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.3897s for 16384 events => throughput is 6.86E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0224s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G181/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.7283s + [COUNTERS] Fortran Overhead ( 0 ) : 5.3138s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.3951s for 16384 events => throughput is 6.84E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0194s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G351/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.7748s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3597s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.3954s for 16384 events => throughput is 6.84E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0197s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G65/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.2127s + [COUNTERS] Fortran Overhead ( 0 ) : 3.8116s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.3812s for 16384 events => throughput is 6.88E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0199s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G71/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.0699s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5629s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.4867s for 16384 events => throughput is 6.59E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0203s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G119/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.3324s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8998s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.4112s for 16384 events => throughput is 6.79E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0213s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G127/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.9972s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5751s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.4022s for 16384 events => throughput is 6.82E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0199s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G96/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.1200s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6913s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.4089s for 16384 events => throughput is 6.80E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0198s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G60/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.6370s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2216s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.3955s for 16384 events => throughput is 6.84E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0199s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G26/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.8765s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4617s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.3951s for 16384 events => throughput is 6.84E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0197s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G37/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.0965s + [COUNTERS] Fortran Overhead ( 0 ) : 5.6951s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.3800s for 16384 events => throughput is 6.88E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0214s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G125/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.6934s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2717s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.4009s for 16384 events => throughput is 6.82E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0207s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G91/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.8952s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4693s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.4064s for 16384 events => throughput is 6.81E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0195s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G143/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.7834s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3604s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.4022s for 16384 events => throughput is 6.82E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0208s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G116/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.7066s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2805s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.4062s for 16384 events => throughput is 6.81E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0200s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G61/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.9796s + [COUNTERS] Fortran Overhead ( 0 ) : 5.5548s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.4030s for 16384 events => throughput is 6.82E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0217s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G149/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.8681s + [COUNTERS] Fortran Overhead ( 0 ) : 4.3485s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.4979s for 16384 events => throughput is 6.56E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0216s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G144/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.8548s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4520s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.3829s for 16384 events => throughput is 6.88E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0199s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G207/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.1242s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7100s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.3944s for 16384 events => throughput is 6.84E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0198s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_epemgguu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_epemgguu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_epemgguu/ajob1' in 99.8100 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_epemgguu/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_epemgguu/G463/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.3182s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9743s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.3096s for 16384 events => throughput is 2.60E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0343s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_epemgguu/G100/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 9.5889s + [COUNTERS] Fortran Overhead ( 0 ) : 3.0494s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.5069s for 16384 events => throughput is 2.52E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0325s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_epemgguu/G110/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.9588s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4418s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.4842s for 16384 events => throughput is 2.53E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0328s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_epemgguu/G124/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.8857s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5190s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.3341s for 16384 events => throughput is 2.59E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0326s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_epemgguu/G367/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.2423s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9079s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.3010s for 16384 events => throughput is 2.60E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0334s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_epemgguu/G126/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 10.3237s + [COUNTERS] Fortran Overhead ( 0 ) : 3.9697s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.3214s for 16384 events => throughput is 2.59E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0326s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_epemgguu/G361/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 10.6447s + [COUNTERS] Fortran Overhead ( 0 ) : 4.3331s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.2754s for 16384 events => throughput is 2.61E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0361s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_epemgguu/G116/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.9465s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6275s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.2845s for 16384 events => throughput is 2.61E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0344s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_epemgguu/G363/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 10.3410s + [COUNTERS] Fortran Overhead ( 0 ) : 4.0231s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.2848s for 16384 events => throughput is 2.61E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0332s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_epemgguu/G138/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.9609s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5985s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.3279s for 16384 events => throughput is 2.59E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0345s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_epemgguu/G102/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 9.3839s + [COUNTERS] Fortran Overhead ( 0 ) : 3.0632s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.2884s for 16384 events => throughput is 2.61E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0324s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_epemgguu/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgdssx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgdssx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgdssx/ajob1' in 6.8706 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgdssx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgdssx/G28/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.8232s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0017s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.8008s for 16384 events => throughput is 3.41E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0208s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgdssx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_taptamgguu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_taptamgguu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_taptamgguu/ajob1' in 73.4956 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_taptamgguu/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_taptamgguu/G464/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.5352s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8773s + [COUNTERS] CudaCpp MEs ( 2 ) : 12.6255s for 16384 events => throughput is 1.30E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0324s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_taptamgguu/G59/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.7091s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0556s + [COUNTERS] CudaCpp MEs ( 2 ) : 12.6212s for 16384 events => throughput is 1.30E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0323s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_taptamgguu/G100/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 15.1344s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4867s + [COUNTERS] CudaCpp MEs ( 2 ) : 12.6155s for 16384 events => throughput is 1.30E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0321s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_taptamgguu/G106/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.4531s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7929s + [COUNTERS] CudaCpp MEs ( 2 ) : 12.6271s for 16384 events => throughput is 1.30E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0331s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_taptamgguu/G143/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.5418s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7777s + [COUNTERS] CudaCpp MEs ( 2 ) : 12.7312s for 16384 events => throughput is 1.29E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0329s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_taptamgguu/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dc_epemggdc/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dc_epemggdc/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dc_epemggdc/ajob1' in 7.5450 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dc_epemggdc/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dc_epemggdc/G18/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.4966s + [COUNTERS] Fortran Overhead ( 0 ) : 5.0914s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.3858s for 16384 events => throughput is 6.87E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0194s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dc_epemggdc/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ucx_taptamggucx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ucx_taptamggucx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ucx_taptamggucx/ajob1' in 9.4161 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ucx_taptamggucx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ucx_taptamggucx/G62/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 9.3698s + [COUNTERS] Fortran Overhead ( 0 ) : 4.5951s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.7553s for 16384 events => throughput is 3.45E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0194s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ucx_taptamggucx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamucdxcx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamucdxcx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamucdxcx/ajob1' in 2.5659 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamucdxcx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamucdxcx/G47/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.5248s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7682s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.7487s for 16384 events => throughput is 2.19E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0079s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamucdxcx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_epemgggg/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_epemgggg/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_epemgggg/ajob1' in 14.6972 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_epemgggg/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_epemgggg/G49/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.6490s + [COUNTERS] Fortran Overhead ( 0 ) : 3.3286s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.2764s for 16384 events => throughput is 1.45E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0441s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_epemgggg/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemudddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemudddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemudddx/ajob1' in 2.6349 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemudddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemudddx/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.5827s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5980s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.9742s for 16384 events => throughput is 1.68E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0105s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemudddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamggudx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamggudx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamggudx/ajob1' in 6.4096 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamggudx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamggudx/G60/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.3799s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5919s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.7681s for 16384 events => throughput is 3.44E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0199s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamggudx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_taptamuuddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_taptamuuddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_taptamuuddx/ajob1' in 3.5030 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_taptamuuddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_taptamuuddx/G11/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.4586s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6271s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.8207s for 16384 events => throughput is 9.00E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0108s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_taptamuuddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished in 6970.3120 seconds +__CUDACPP_DEBUG: gen_ximprove.launch finished in 6974.4503 seconds +__CUDACPP_DEBUG: GridPackCmd.refine4grid finished in 6974.4512 seconds +__CUDACPP_DEBUG: GridPackCmd.launch (3a) combine_events +combine_events +INFO: fail to reach target 500 +__CUDACPP_DEBUG: GridPackCmd.launch (3b) store_events +store_events + === Results Summary for run: GridRun_21 tag: tag_1 === + + Cross-section : 0 +- 0 pb + Nb of events : 195 + +INFO: No version of lhapdf. Can not run systematics computation +decay_events -from_cards +__CUDACPP_DEBUG: GridPackCmd.launch finished in 7136.2962 seconds +__CUDACPP_DEBUG: GridPackCmd.__init__ back from self.launch +DONE +write ./events.lhe.gz +END: Sun Sep 15 05:44:16 PM CEST 2024 +ELAPSED: 7138 seconds diff --git a/epochX/cudacpp/tlau/fromgridpacks/pp_dy4j.mad/cppavx2/output.txt b/epochX/cudacpp/tlau/fromgridpacks/pp_dy4j.mad/cppavx2/output.txt new file mode 100644 index 0000000000..8ffad4c20a --- /dev/null +++ b/epochX/cudacpp/tlau/fromgridpacks/pp_dy4j.mad/cppavx2/output.txt @@ -0,0 +1,3327 @@ +START: Sun Sep 15 11:42:44 AM CEST 2024 +On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU (nproc=4)] [GPU: 1x Tesla V100S-PCIE-32GB]: +CUDACPP_RUNTIME_DISABLEFPE is set +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_bbx_epemdsdxsx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_bbx_taptamdsdxsx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_ccx_epemdcdxcx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_ccx_epemdsdxsx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_ccx_epemucuxcx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_ccx_epemuduxdx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_ccx_taptamdcdxcx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_ccx_taptamdsdxsx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_ccx_taptamucuxcx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_ccx_taptamuduxdx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_dc_epemdcccx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_dc_epemdcssx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_dc_epemddcdx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_dc_epemggdc/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_dc_epemudcux/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_dc_taptamdcccx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_dc_taptamdcssx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_dc_taptamddcdx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_dc_taptamggdc/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_dc_taptamudcux/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_dd_epemddddx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_dd_epemddssx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_dd_epemggdd/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_dd_epemuddux/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_dd_taptamddddx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_dd_taptamddssx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_dd_taptamggdd/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_dd_taptamuddux/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_ddx_epemdcdxcx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_ddx_epemdddxdx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_ddx_epemdsdxsx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_ddx_epemggddx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_ddx_epemgggg/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_ddx_epemggssx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_ddx_epemgguux/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_ddx_epemsbsxbx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_ddx_epemsssxsx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_ddx_epemucuxcx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_ddx_epemuduxdx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_ddx_epemusuxsx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_ddx_epemuuuxux/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_ddx_taptamdcdxcx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_ddx_taptamdddxdx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_ddx_taptamdsdxsx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_ddx_taptamggddx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_ddx_taptamgggg/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_ddx_taptamggssx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_ddx_taptamgguux/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_ddx_taptamsbsxbx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_ddx_taptamsssxsx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_ddx_taptamucuxcx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_ddx_taptamuduxdx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_ddx_taptamusuxsx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_ddx_taptamuuuxux/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_ds_epemddsdx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_ds_epemdsbbx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_ds_epemdsssx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_ds_epemggds/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_ds_epemudsux/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_ds_taptamddsdx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_ds_taptamdsbbx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_ds_taptamdsssx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_ds_taptamggds/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_ds_taptamudsux/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_dsx_epemdbsxbx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_dsx_epemdddxsx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_dsx_epemdssxsx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_dsx_epemggdsx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_dsx_epemuduxsx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_dsx_taptamdbsxbx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_dsx_taptamdddxsx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_dsx_taptamdssxsx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_dsx_taptamggdsx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_dsx_taptamuduxsx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_dux_epemdcuxcx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_dux_epemdduxdx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_dux_epemdsuxsx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_dux_epemggdux/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_dux_epemuduxux/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_dux_taptamdcuxcx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_dux_taptamdduxdx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_dux_taptamdsuxsx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_dux_taptamggdux/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_dux_taptamuduxux/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_dxcx_epemcdxcxcx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_dxcx_epemddxdxcx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_dxcx_epemggdxcx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_dxcx_epemsdxcxsx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_dxcx_epemuuxdxcx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_dxcx_taptamcdxcxcx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_dxcx_taptamddxdxcx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_dxcx_taptamggdxcx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_dxcx_taptamsdxcxsx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_dxcx_taptamuuxdxcx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_dxdx_epemddxdxdx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_dxdx_epemggdxdx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_dxdx_epemsdxdxsx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_dxdx_epemuuxdxdx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_dxdx_taptamddxdxdx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_dxdx_taptamggdxdx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_dxdx_taptamsdxdxsx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_dxdx_taptamuuxdxdx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_dxsx_epembdxsxbx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_dxsx_epemddxdxsx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_dxsx_epemggdxsx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_dxsx_epemsdxsxsx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_dxsx_epemuuxdxsx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_dxsx_taptambdxsxbx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_dxsx_taptamddxdxsx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_dxsx_taptamggdxsx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_dxsx_taptamsdxsxsx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_dxsx_taptamuuxdxsx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_gd_epemgdddx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_gd_epemgdssx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_gd_epemgggd/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_gd_epemgudux/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_gd_taptamgdddx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_gd_taptamgdssx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_gd_taptamgggd/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_gd_taptamgudux/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_gdx_epemgddxdx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_gdx_epemgggdx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_gdx_epemgsdxsx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_gdx_epemguuxdx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_gdx_taptamgddxdx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_gdx_taptamgggdx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_gdx_taptamgsdxsx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_gdx_taptamguuxdx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_gg_epemdddxdx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_gg_epemdsdxsx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_gg_epemggddx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_gg_epemgguux/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_gg_epemucuxcx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_gg_epemuduxdx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_gg_epemuuuxux/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_gg_taptamdddxdx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_gg_taptamdsdxsx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_gg_taptamggddx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_gg_taptamgguux/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_gg_taptamucuxcx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_gg_taptamuduxdx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_gg_taptamuuuxux/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_gu_epemgggu/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_gu_epemguccx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_gu_epemguddx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_gu_epemguuux/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_gu_taptamgggu/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_gu_taptamguccx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_gu_taptamguddx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_gu_taptamguuux/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_gux_epemgcuxcx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_gux_epemgduxdx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_gux_epemgggux/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_gux_epemguuxux/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_gux_taptamgcuxcx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_gux_taptamgduxdx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_gux_taptamgggux/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_gux_taptamguuxux/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_ssx_epemdbdxbx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_ssx_epemdcdxcx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_ssx_epemdsdxsx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_ssx_epemucuxcx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_ssx_epemuduxdx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_ssx_taptamdbdxbx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_ssx_taptamdcdxcx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_ssx_taptamdsdxsx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_ssx_taptamucuxcx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_ssx_taptamuduxdx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_uc_epemgguc/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_uc_epemucccx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_uc_epemudcdx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_uc_epemuucux/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_uc_taptamgguc/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_uc_taptamucccx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_uc_taptamudcdx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_uc_taptamuucux/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_ucx_epemggucx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_ucx_epemuccxcx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_ucx_epemuddxcx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_ucx_epemuuuxcx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_ucx_taptamggucx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_ucx_taptamuccxcx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_ucx_taptamuddxcx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_ucx_taptamuuuxcx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_ud_epemggud/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_ud_epemudccx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_ud_epemudddx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_ud_epemudssx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_ud_epemuudux/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_ud_taptamggud/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_ud_taptamudccx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_ud_taptamudddx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_ud_taptamudssx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_ud_taptamuudux/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_udx_epemggudx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_udx_epemucdxcx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_udx_epemuddxdx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_udx_epemusdxsx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_udx_epemuuuxdx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_udx_taptamggudx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_udx_taptamucdxcx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_udx_taptamuddxdx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_udx_taptamusdxsx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_udx_taptamuuuxdx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_uu_epemgguu/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_uu_epemuuccx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_uu_epemuuddx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_uu_epemuuuux/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_uu_taptamgguu/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_uu_taptamuuccx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_uu_taptamuuddx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_uu_taptamuuuux/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_uux_epemcccxcx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_uux_epemcscxsx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_uux_epemdcdxcx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_uux_epemdddxdx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_uux_epemdsdxsx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_uux_epemggccx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_uux_epemggddx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_uux_epemgggg/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_uux_epemgguux/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_uux_epemucuxcx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_uux_epemuduxdx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_uux_epemuuuxux/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_uux_taptamcccxcx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_uux_taptamcscxsx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_uux_taptamdcdxcx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_uux_taptamdddxdx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_uux_taptamdsdxsx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_uux_taptamggccx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_uux_taptamggddx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_uux_taptamgggg/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_uux_taptamgguux/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_uux_taptamucuxcx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_uux_taptamuduxdx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_uux_taptamuuuxux/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_uxcx_epemcuxcxcx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_uxcx_epemduxdxcx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_uxcx_epemgguxcx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_uxcx_epemuuxuxcx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_uxcx_taptamcuxcxcx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_uxcx_taptamduxdxcx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_uxcx_taptamgguxcx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_uxcx_taptamuuxuxcx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_uxdx_epemcuxdxcx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_uxdx_epemduxdxdx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_uxdx_epemgguxdx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_uxdx_epemsuxdxsx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_uxdx_epemuuxuxdx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_uxdx_taptamcuxdxcx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_uxdx_taptamduxdxdx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_uxdx_taptamgguxdx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_uxdx_taptamsuxdxsx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_uxdx_taptamuuxuxdx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_uxux_epemcuxuxcx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_uxux_epemduxuxdx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_uxux_epemgguxux/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_uxux_epemuuxuxux/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_uxux_taptamcuxuxcx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_uxux_taptamduxuxdx/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_uxux_taptamgguxux/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 11:42 madevent/SubProcesses/P0_uxux_taptamuuxuxux/madevent -> build.avx2_d_inl0_hrd0/madevent_cpp +Now generating 500 events with random seed 21 and granularity 1 +************************************************************ +* * +* W E L C O M E to * +* M A D G R A P H 5 _ a M C @ N L O * +* M A D E V E N T * +* * +* * * * +* * * * * * +* * * * * 5 * * * * * +* * * * * * +* * * * +* * +* VERSION 3.5.3_lo_vect * +* * +* The MadGraph5_aMC@NLO Development Team - Find us at * +* https://server06.fynu.ucl.ac.be/projects/madgraph * +* * +* Type 'help' for in-line help. * +* * +************************************************************ +INFO: load configuration from /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/Cards/me5_configuration.txt +INFO: load configuration from /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/Cards/me5_configuration.txt +Using default text editor "vi". Set another one in ./input/mg5_configuration.txt +Using default eps viewer "evince". Set another one in ./input/mg5_configuration.txt +Using default web browser "firefox". Set another one in ./input/mg5_configuration.txt +WRITE GRIDCARD /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent +__CUDACPP_DEBUG: GridPackCmd.__init__ will call self.launch +__CUDACPP_DEBUG: GridPackCmd.launch starting +__CUDACPP_DEBUG: GridPackCmd.launch (1) restore_data +generate 500 events +P0_gg_epemgguux +P0_gg_epemggddx +P0_gg_taptamgguux +P0_gg_taptamggddx +P0_gu_epemgggu +P0_gd_epemgggd +P0_gux_epemgggux +P0_gdx_epemgggdx +P0_gu_taptamgggu +P0_gd_taptamgggd +P0_gux_taptamgggux +P0_gdx_taptamgggdx +P0_uux_epemgggg +P0_ddx_epemgggg +P0_uux_taptamgggg +P0_ddx_taptamgggg +P0_gg_epemuuuxux +P0_gg_epemdddxdx +P0_gg_epemuduxdx +P0_gg_epemucuxcx +P0_gg_epemdsdxsx +P0_gg_taptamuuuxux +P0_gg_taptamdddxdx +P0_gg_taptamuduxdx +P0_gg_taptamucuxcx +P0_gg_taptamdsdxsx +P0_gu_epemguuux +P0_gd_epemgdddx +P0_gux_epemguuxux +P0_gdx_epemgddxdx +P0_gu_epemguddx +P0_gu_epemguccx +P0_gd_epemgudux +P0_gd_epemgdssx +P0_gux_epemgduxdx +P0_gux_epemgcuxcx +P0_gdx_epemguuxdx +P0_gdx_epemgsdxsx +P0_gu_taptamguuux +P0_gd_taptamgdddx +P0_gux_taptamguuxux +P0_gdx_taptamgddxdx +P0_gu_taptamguddx +P0_gu_taptamguccx +P0_gd_taptamgudux +P0_gd_taptamgdssx +P0_gux_taptamgduxdx +P0_gux_taptamgcuxcx +P0_gdx_taptamguuxdx +P0_gdx_taptamgsdxsx +P0_uu_epemgguu +P0_uux_epemgguux +P0_dd_epemggdd +P0_ddx_epemggddx +P0_uxux_epemgguxux +P0_dxdx_epemggdxdx +P0_ud_epemggud +P0_uc_epemgguc +P0_uux_epemggddx +P0_uux_epemggccx +P0_udx_epemggudx +P0_ucx_epemggucx +P0_dc_epemggdc +P0_ds_epemggds +P0_dux_epemggdux +P0_ddx_epemgguux +P0_ddx_epemggssx +P0_dsx_epemggdsx +P0_uxdx_epemgguxdx +P0_uxcx_epemgguxcx +P0_dxcx_epemggdxcx +P0_dxsx_epemggdxsx +P0_uu_taptamgguu +P0_uux_taptamgguux +P0_dd_taptamggdd +P0_ddx_taptamggddx +P0_uxux_taptamgguxux +P0_dxdx_taptamggdxdx +P0_ud_taptamggud +P0_uc_taptamgguc +P0_uux_taptamggddx +P0_uux_taptamggccx +P0_udx_taptamggudx +P0_ucx_taptamggucx +P0_dc_taptamggdc +P0_ds_taptamggds +P0_dux_taptamggdux +P0_ddx_taptamgguux +P0_ddx_taptamggssx +P0_dsx_taptamggdsx +P0_uxdx_taptamgguxdx +P0_uxcx_taptamgguxcx +P0_dxcx_taptamggdxcx +P0_dxsx_taptamggdxsx +P0_uu_epemuuuux +P0_uux_epemuuuxux +P0_dd_epemddddx +P0_ddx_epemdddxdx +P0_uxux_epemuuxuxux +P0_dxdx_epemddxdxdx +P0_uu_epemuuddx +P0_uu_epemuuccx +P0_ud_epemuudux +P0_ud_epemudddx +P0_uc_epemuucux +P0_uc_epemucccx +P0_uux_epemuduxdx +P0_uux_epemucuxcx +P0_uux_epemdddxdx +P0_uux_epemcccxcx +P0_udx_epemuuuxdx +P0_udx_epemuddxdx +P0_ucx_epemuuuxcx +P0_ucx_epemuccxcx +P0_dd_epemuddux +P0_dd_epemddssx +P0_dc_epemddcdx +P0_dc_epemdcccx +P0_ds_epemddsdx +P0_ds_epemdsssx +P0_dux_epemuduxux +P0_dux_epemdduxdx +P0_ddx_epemuuuxux +P0_ddx_epemuduxdx +P0_ddx_epemdcdxcx +P0_ddx_epemdsdxsx +P0_ddx_epemsssxsx +P0_dsx_epemdddxsx +P0_dsx_epemdssxsx +P0_ccx_epemucuxcx +P0_ccx_epemdcdxcx +P0_ssx_epemdsdxsx +P0_uxux_epemduxuxdx +P0_uxux_epemcuxuxcx +P0_uxdx_epemuuxuxdx +P0_uxdx_epemduxdxdx +P0_uxcx_epemuuxuxcx +P0_uxcx_epemcuxcxcx +P0_dxdx_epemuuxdxdx +P0_dxdx_epemsdxdxsx +P0_dxcx_epemddxdxcx +P0_dxcx_epemcdxcxcx +P0_dxsx_epemddxdxsx +P0_dxsx_epemsdxsxsx +P0_ud_epemudccx +P0_ud_epemudssx +P0_uc_epemudcdx +P0_uux_epemdcdxcx +P0_uux_epemdsdxsx +P0_uux_epemcscxsx +P0_udx_epemucdxcx +P0_udx_epemusdxsx +P0_ucx_epemuddxcx +P0_dc_epemudcux +P0_dc_epemdcssx +P0_ds_epemudsux +P0_ds_epemdsbbx +P0_dux_epemdcuxcx +P0_dux_epemdsuxsx +P0_ddx_epemucuxcx +P0_ddx_epemusuxsx +P0_ddx_epemsbsxbx +P0_dsx_epemuduxsx +P0_dsx_epemdbsxbx +P0_ccx_epemuduxdx +P0_ccx_epemdsdxsx +P0_ssx_epemuduxdx +P0_ssx_epemucuxcx +P0_ssx_epemdcdxcx +P0_ssx_epemdbdxbx +P0_bbx_epemdsdxsx +P0_uxdx_epemcuxdxcx +P0_uxdx_epemsuxdxsx +P0_uxcx_epemduxdxcx +P0_dxcx_epemuuxdxcx +P0_dxcx_epemsdxcxsx +P0_dxsx_epemuuxdxsx +P0_dxsx_epembdxsxbx +P0_uu_taptamuuuux +P0_uux_taptamuuuxux +P0_dd_taptamddddx +P0_ddx_taptamdddxdx +P0_uxux_taptamuuxuxux +P0_dxdx_taptamddxdxdx +P0_uu_taptamuuddx +P0_uu_taptamuuccx +P0_ud_taptamuudux +P0_ud_taptamudddx +P0_uc_taptamuucux +P0_uc_taptamucccx +P0_uux_taptamuduxdx +P0_uux_taptamucuxcx +P0_uux_taptamdddxdx +P0_uux_taptamcccxcx +P0_udx_taptamuuuxdx +P0_udx_taptamuddxdx +P0_ucx_taptamuuuxcx +P0_ucx_taptamuccxcx +P0_dd_taptamuddux +P0_dd_taptamddssx +P0_dc_taptamddcdx +P0_dc_taptamdcccx +P0_ds_taptamddsdx +P0_ds_taptamdsssx +P0_dux_taptamuduxux +P0_dux_taptamdduxdx +P0_ddx_taptamuuuxux +P0_ddx_taptamuduxdx +P0_ddx_taptamdcdxcx +P0_ddx_taptamdsdxsx +P0_ddx_taptamsssxsx +P0_dsx_taptamdddxsx +P0_dsx_taptamdssxsx +P0_ccx_taptamucuxcx +P0_ccx_taptamdcdxcx +P0_ssx_taptamdsdxsx +P0_uxux_taptamduxuxdx +P0_uxux_taptamcuxuxcx +P0_uxdx_taptamuuxuxdx +P0_uxdx_taptamduxdxdx +P0_uxcx_taptamuuxuxcx +P0_uxcx_taptamcuxcxcx +P0_dxdx_taptamuuxdxdx +P0_dxdx_taptamsdxdxsx +P0_dxcx_taptamddxdxcx +P0_dxcx_taptamcdxcxcx +P0_dxsx_taptamddxdxsx +P0_dxsx_taptamsdxsxsx +P0_ud_taptamudccx +P0_ud_taptamudssx +P0_uc_taptamudcdx +P0_uux_taptamdcdxcx +P0_uux_taptamdsdxsx +P0_uux_taptamcscxsx +P0_udx_taptamucdxcx +P0_udx_taptamusdxsx +P0_ucx_taptamuddxcx +P0_dc_taptamudcux +P0_dc_taptamdcssx +P0_ds_taptamudsux +P0_ds_taptamdsbbx +P0_dux_taptamdcuxcx +P0_dux_taptamdsuxsx +P0_ddx_taptamucuxcx +P0_ddx_taptamusuxsx +P0_ddx_taptamsbsxbx +P0_dsx_taptamuduxsx +P0_dsx_taptamdbsxbx +P0_ccx_taptamuduxdx +P0_ccx_taptamdsdxsx +P0_ssx_taptamuduxdx +P0_ssx_taptamucuxcx +P0_ssx_taptamdcdxcx +P0_ssx_taptamdbdxbx +P0_bbx_taptamdsdxsx +P0_uxdx_taptamcuxdxcx +P0_uxdx_taptamsuxdxsx +P0_uxcx_taptamduxdxcx +P0_dxcx_taptamuuxdxcx +P0_dxcx_taptamsdxcxsx +P0_dxsx_taptamuuxdxsx +P0_dxsx_taptambdxsxbx +__CUDACPP_DEBUG: GridPackCmd.launch (2) refine4grid +__CUDACPP_DEBUG: GridPackCmd.refine4grid starting +Using random number seed offset = 21 +INFO: Generating 500.0 unweighted events. +__CUDACPP_DEBUG: gen_ximprove.launch starting +sum of cpu time of last step: 3 days,23h23m17s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event starting +INFO: need to improve 477 channels +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dxdx_taptamuuxdxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dxdx_taptamuuxdxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dxdx_taptamuuxdxdx/ajob1' in 5.5118 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dxdx_taptamuuxdxdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dxdx_taptamuuxdxdx/G21/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.4772s + [COUNTERS] Fortran Overhead ( 0 ) : 3.8179s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.6497s for 16384 events => throughput is 9.93E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0097s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dxdx_taptamuuxdxdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamguuxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamguuxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamguuxdx/ajob1' in 14.0824 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamguuxdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamguuxdx/G43/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.6576s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0906s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.5492s for 16384 events => throughput is 3.60E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0178s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamguuxdx/G335/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.3597s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7643s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.5780s for 16384 events => throughput is 3.58E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0174s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamguuxdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemuuuxux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemuuuxux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemuuuxux/ajob1' in 10.1234 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemuuuxux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemuuuxux/G689/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 10.0790s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6950s + [COUNTERS] CudaCpp MEs ( 2 ) : 7.3543s for 16384 events => throughput is 2.23E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0297s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemuuuxux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamgguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamgguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamgguux/ajob1' in 258.1842 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamgguux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamgguux/G596/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 27.2775s + [COUNTERS] Fortran Overhead ( 0 ) : 3.5357s + [COUNTERS] CudaCpp MEs ( 2 ) : 23.7019s for 16384 events => throughput is 6.91E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0399s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamgguux/G304/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 26.8837s + [COUNTERS] Fortran Overhead ( 0 ) : 3.0717s + [COUNTERS] CudaCpp MEs ( 2 ) : 23.7728s for 16384 events => throughput is 6.89E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0393s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamgguux/G128/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 26.9766s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1982s + [COUNTERS] CudaCpp MEs ( 2 ) : 23.7395s for 16384 events => throughput is 6.90E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0389s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamgguux/G115/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 30.7148s + [COUNTERS] Fortran Overhead ( 0 ) : 6.9924s + [COUNTERS] CudaCpp MEs ( 2 ) : 23.6829s for 16384 events => throughput is 6.92E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0395s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamgguux/G599/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 27.7505s + [COUNTERS] Fortran Overhead ( 0 ) : 3.9881s + [COUNTERS] CudaCpp MEs ( 2 ) : 23.7192s for 16384 events => throughput is 6.91E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0431s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamgguux/G351/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 30.1573s + [COUNTERS] Fortran Overhead ( 0 ) : 6.3891s + [COUNTERS] CudaCpp MEs ( 2 ) : 23.7283s for 16384 events => throughput is 6.90E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0399s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamgguux/G111/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 26.2778s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4710s + [COUNTERS] CudaCpp MEs ( 2 ) : 23.7674s for 16384 events => throughput is 6.89E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0393s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamgguux/G354/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 30.7011s + [COUNTERS] Fortran Overhead ( 0 ) : 6.8909s + [COUNTERS] CudaCpp MEs ( 2 ) : 23.7710s for 16384 events => throughput is 6.89E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0391s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamgguux/G97/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 31.2414s + [COUNTERS] Fortran Overhead ( 0 ) : 7.5640s + [COUNTERS] CudaCpp MEs ( 2 ) : 23.6382s for 16384 events => throughput is 6.93E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0392s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamgguux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uxdx_epemsuxdxsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uxdx_epemsuxdxsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uxdx_epemsuxdxsx/ajob1' in 2.6510 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uxdx_epemsuxdxsx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uxdx_epemsuxdxsx/G74/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.5975s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2691s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.3221s for 16384 events => throughput is 5.09E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0062s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uxdx_epemsuxdxsx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemdddxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemdddxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemdddxdx/ajob1' in 20.1730 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemdddxdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemdddxdx/G513/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 9.9024s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5791s + [COUNTERS] CudaCpp MEs ( 2 ) : 7.2927s for 16384 events => throughput is 2.25E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0306s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemdddxdx/G180/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 10.2177s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9092s + [COUNTERS] CudaCpp MEs ( 2 ) : 7.2785s for 16384 events => throughput is 2.25E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0300s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemdddxdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgddxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgddxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgddxdx/ajob1' in 10.8536 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgddxdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgddxdx/G515/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 10.5945s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2084s + [COUNTERS] CudaCpp MEs ( 2 ) : 7.3560s for 16384 events => throughput is 2.23E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0301s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgddxdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguddx/ajob1' in 66.2920 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguddx/G104/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.2478s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6713s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.5593s for 16384 events => throughput is 3.59E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0173s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguddx/G120/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.8836s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2576s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.6081s for 16384 events => throughput is 3.56E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0180s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguddx/G305/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 10.7647s + [COUNTERS] Fortran Overhead ( 0 ) : 6.1546s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.5926s for 16384 events => throughput is 3.57E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0175s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguddx/G101/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.1723s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5478s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.6065s for 16384 events => throughput is 3.56E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0179s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguddx/G27/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.5655s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9807s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.5673s for 16384 events => throughput is 3.59E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0175s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguddx/G273/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.8451s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2596s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.5680s for 16384 events => throughput is 3.59E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0175s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguddx/G349/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.1138s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4701s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.6264s for 16384 events => throughput is 3.54E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0173s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguddx/G379/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.4761s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8535s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.6046s for 16384 events => throughput is 3.56E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0181s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguddx/G256/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 10.0235s + [COUNTERS] Fortran Overhead ( 0 ) : 5.4046s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.6011s for 16384 events => throughput is 3.56E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0178s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemguuxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemguuxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemguuxdx/ajob1' in 21.0505 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemguuxdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemguuxdx/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.6006s + [COUNTERS] Fortran Overhead ( 0 ) : 6.3026s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.2807s for 16384 events => throughput is 7.18E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0173s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemguuxdx/G68/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.8670s + [COUNTERS] Fortran Overhead ( 0 ) : 5.5806s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.2690s for 16384 events => throughput is 7.22E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0174s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemguuxdx/G355/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.4961s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1882s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.2906s for 16384 events => throughput is 7.15E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0173s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemguuxdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/ajob1' in 621.5868 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G14/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 25.6136s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1445s + [COUNTERS] CudaCpp MEs ( 2 ) : 23.4298s for 16384 events => throughput is 6.99E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0393s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G56/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 27.0479s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6978s + [COUNTERS] CudaCpp MEs ( 2 ) : 24.3095s for 16384 events => throughput is 6.74E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0405s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G517/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 32.3996s + [COUNTERS] Fortran Overhead ( 0 ) : 6.9781s + [COUNTERS] CudaCpp MEs ( 2 ) : 25.3808s for 16384 events => throughput is 6.46E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0407s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G52/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 27.5002s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9545s + [COUNTERS] CudaCpp MEs ( 2 ) : 24.5056s for 16384 events => throughput is 6.69E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0402s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G304/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 26.3508s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8325s + [COUNTERS] CudaCpp MEs ( 2 ) : 23.4802s for 16384 events => throughput is 6.98E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0382s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G3/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 28.6572s + [COUNTERS] Fortran Overhead ( 0 ) : 5.0711s + [COUNTERS] CudaCpp MEs ( 2 ) : 23.5462s for 16384 events => throughput is 6.96E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0399s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G23/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 25.3890s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8086s + [COUNTERS] CudaCpp MEs ( 2 ) : 23.5411s for 16384 events => throughput is 6.96E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0393s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G33/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 25.7457s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2524s + [COUNTERS] CudaCpp MEs ( 2 ) : 23.4536s for 16384 events => throughput is 6.99E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0397s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G53/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 30.6900s + [COUNTERS] Fortran Overhead ( 0 ) : 7.0346s + [COUNTERS] CudaCpp MEs ( 2 ) : 23.6160s for 16384 events => throughput is 6.94E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0394s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G495/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 25.2040s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6677s + [COUNTERS] CudaCpp MEs ( 2 ) : 23.4972s for 16384 events => throughput is 6.97E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0391s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G68/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 25.8040s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2854s + [COUNTERS] CudaCpp MEs ( 2 ) : 23.4781s for 16384 events => throughput is 6.98E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0406s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G49/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 26.2208s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1714s + [COUNTERS] CudaCpp MEs ( 2 ) : 24.0102s for 16384 events => throughput is 6.82E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0392s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G524/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 26.5642s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3095s + [COUNTERS] CudaCpp MEs ( 2 ) : 24.2149s for 16384 events => throughput is 6.77E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0399s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G65/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 26.6466s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2320s + [COUNTERS] CudaCpp MEs ( 2 ) : 24.3745s for 16384 events => throughput is 6.72E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0401s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G303/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 26.6717s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1248s + [COUNTERS] CudaCpp MEs ( 2 ) : 23.5078s for 16384 events => throughput is 6.97E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0391s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G55/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 26.3883s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8314s + [COUNTERS] CudaCpp MEs ( 2 ) : 23.5182s for 16384 events => throughput is 6.97E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0388s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G34/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 25.3690s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8823s + [COUNTERS] CudaCpp MEs ( 2 ) : 23.4478s for 16384 events => throughput is 6.99E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0390s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G509/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 31.6666s + [COUNTERS] Fortran Overhead ( 0 ) : 8.0600s + [COUNTERS] CudaCpp MEs ( 2 ) : 23.5668s for 16384 events => throughput is 6.95E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0397s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G512/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 26.5906s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0115s + [COUNTERS] CudaCpp MEs ( 2 ) : 24.5387s for 16384 events => throughput is 6.68E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0405s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G511/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 26.1252s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5310s + [COUNTERS] CudaCpp MEs ( 2 ) : 23.5549s for 16384 events => throughput is 6.96E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0393s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 26.4851s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9562s + [COUNTERS] CudaCpp MEs ( 2 ) : 23.4880s for 16384 events => throughput is 6.98E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0410s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G63/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 25.7142s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1985s + [COUNTERS] CudaCpp MEs ( 2 ) : 23.4766s for 16384 events => throughput is 6.98E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0391s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G48/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 26.2383s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5432s + [COUNTERS] CudaCpp MEs ( 2 ) : 23.6560s for 16384 events => throughput is 6.93E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0390s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/ajob1' in 414.4499 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G122/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 15.3873s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8903s + [COUNTERS] CudaCpp MEs ( 2 ) : 12.4558s for 16384 events => throughput is 1.32E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0412s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G108/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.7170s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4866s + [COUNTERS] CudaCpp MEs ( 2 ) : 12.1899s for 16384 events => throughput is 1.34E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0405s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G105/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.9868s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6444s + [COUNTERS] CudaCpp MEs ( 2 ) : 12.3017s for 16384 events => throughput is 1.33E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0407s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G304/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 15.8780s + [COUNTERS] Fortran Overhead ( 0 ) : 3.3525s + [COUNTERS] CudaCpp MEs ( 2 ) : 12.4843s for 16384 events => throughput is 1.31E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0412s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G301/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 20.5571s + [COUNTERS] Fortran Overhead ( 0 ) : 8.1960s + [COUNTERS] CudaCpp MEs ( 2 ) : 12.3204s for 16384 events => throughput is 1.33E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0407s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G128/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 15.9415s + [COUNTERS] Fortran Overhead ( 0 ) : 3.3826s + [COUNTERS] CudaCpp MEs ( 2 ) : 12.5190s for 16384 events => throughput is 1.31E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0399s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G100/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 20.7395s + [COUNTERS] Fortran Overhead ( 0 ) : 8.1967s + [COUNTERS] CudaCpp MEs ( 2 ) : 12.5021s for 16384 events => throughput is 1.31E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0407s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G120/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 15.0901s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6531s + [COUNTERS] CudaCpp MEs ( 2 ) : 12.3963s for 16384 events => throughput is 1.32E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0406s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G302/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 23.8168s + [COUNTERS] Fortran Overhead ( 0 ) : 11.5187s + [COUNTERS] CudaCpp MEs ( 2 ) : 12.2571s for 16384 events => throughput is 1.34E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0410s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G299/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 15.3993s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9115s + [COUNTERS] CudaCpp MEs ( 2 ) : 12.4473s for 16384 events => throughput is 1.32E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0405s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G169/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 18.8174s + [COUNTERS] Fortran Overhead ( 0 ) : 6.4288s + [COUNTERS] CudaCpp MEs ( 2 ) : 12.3478s for 16384 events => throughput is 1.33E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0408s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G333/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 24.2630s + [COUNTERS] Fortran Overhead ( 0 ) : 11.8598s + [COUNTERS] CudaCpp MEs ( 2 ) : 12.3594s for 16384 events => throughput is 1.33E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0438s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G594/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 22.3081s + [COUNTERS] Fortran Overhead ( 0 ) : 9.9440s + [COUNTERS] CudaCpp MEs ( 2 ) : 12.3225s for 16384 events => throughput is 1.33E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0416s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G678/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 18.9833s + [COUNTERS] Fortran Overhead ( 0 ) : 6.7296s + [COUNTERS] CudaCpp MEs ( 2 ) : 12.2140s for 16384 events => throughput is 1.34E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0398s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G127/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 15.2274s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8872s + [COUNTERS] CudaCpp MEs ( 2 ) : 12.2998s for 16384 events => throughput is 1.33E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0403s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G126/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 22.5968s + [COUNTERS] Fortran Overhead ( 0 ) : 10.2639s + [COUNTERS] CudaCpp MEs ( 2 ) : 12.2891s for 16384 events => throughput is 1.33E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0437s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G598/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.6511s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3162s + [COUNTERS] CudaCpp MEs ( 2 ) : 12.2935s for 16384 events => throughput is 1.33E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0413s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G512/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.7955s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2855s + [COUNTERS] CudaCpp MEs ( 2 ) : 12.4681s for 16384 events => throughput is 1.31E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0419s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G103/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 22.0919s + [COUNTERS] Fortran Overhead ( 0 ) : 9.7727s + [COUNTERS] CudaCpp MEs ( 2 ) : 12.2789s for 16384 events => throughput is 1.33E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0403s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G107/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 15.0818s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5082s + [COUNTERS] CudaCpp MEs ( 2 ) : 12.5320s for 16384 events => throughput is 1.31E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0416s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G358/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 16.4160s + [COUNTERS] Fortran Overhead ( 0 ) : 4.0187s + [COUNTERS] CudaCpp MEs ( 2 ) : 12.3547s for 16384 events => throughput is 1.33E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0426s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G97/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 20.2542s + [COUNTERS] Fortran Overhead ( 0 ) : 7.8982s + [COUNTERS] CudaCpp MEs ( 2 ) : 12.3154s for 16384 events => throughput is 1.33E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0406s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G179/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 15.4336s + [COUNTERS] Fortran Overhead ( 0 ) : 3.0177s + [COUNTERS] CudaCpp MEs ( 2 ) : 12.3747s for 16384 events => throughput is 1.32E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0412s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_epemudcdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_epemudcdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_epemudcdx/ajob1' in 2.2580 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_epemudcdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_epemudcdx/G42/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.2235s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8895s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.3276s for 16384 events => throughput is 5.00E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0064s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_epemudcdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ds_taptamggds/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ds_taptamggds/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ds_taptamggds/ajob1' in 16.4640 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ds_taptamggds/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ds_taptamggds/G17/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 9.8594s + [COUNTERS] Fortran Overhead ( 0 ) : 5.0582s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.7834s for 16384 events => throughput is 3.43E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0178s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ds_taptamggds/G110/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.5310s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8169s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.6962s for 16384 events => throughput is 3.49E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0180s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ds_taptamggds/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_epemggdux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_epemggdux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_epemggdux/ajob1' in 20.1984 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_epemggdux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_epemggdux/G208/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.5492s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1834s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.3481s for 16384 events => throughput is 6.98E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0177s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_epemggdux/G192/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.8712s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5113s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.3420s for 16384 events => throughput is 7.00E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0179s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_epemggdux/G380/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.2259s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8200s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.3822s for 16384 events => throughput is 6.88E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0237s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_epemggdux/G127/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.7044s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3492s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.3378s for 16384 events => throughput is 7.01E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0173s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_epemggdux/G223/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.7172s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3332s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.3647s for 16384 events => throughput is 6.93E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0193s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_epemggdux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_taptamgguc/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_taptamgguc/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_taptamgguc/ajob1' in 26.5199 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_taptamgguc/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_taptamgguc/G20/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 9.4617s + [COUNTERS] Fortran Overhead ( 0 ) : 4.7324s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.7105s for 16384 events => throughput is 3.48E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0189s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_taptamgguc/G110/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.1922s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4661s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.7080s for 16384 events => throughput is 3.48E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0180s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_taptamgguc/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 10.5260s + [COUNTERS] Fortran Overhead ( 0 ) : 5.7887s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.7196s for 16384 events => throughput is 3.47E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0178s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_taptamgguc/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uxcx_taptamgguxcx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uxcx_taptamgguxcx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uxcx_taptamgguxcx/ajob1' in 7.4533 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uxcx_taptamgguxcx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uxcx_taptamgguxcx/G187/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.4068s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7502s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.6387s for 16384 events => throughput is 3.53E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0179s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uxcx_taptamgguxcx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/ajob1' in 410.3814 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G499/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.8514s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3876s + [COUNTERS] CudaCpp MEs ( 2 ) : 12.4231s for 16384 events => throughput is 1.32E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0407s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G56/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.9980s + [COUNTERS] Fortran Overhead ( 0 ) : 3.0306s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.9279s for 16384 events => throughput is 1.37E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0395s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G39/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.8515s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8410s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.9710s for 16384 events => throughput is 1.37E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0395s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G64/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 15.1489s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7551s + [COUNTERS] CudaCpp MEs ( 2 ) : 12.3535s for 16384 events => throughput is 1.33E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0403s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G13/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.6226s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6640s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.9181s for 16384 events => throughput is 1.37E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0404s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G33/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.8742s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8489s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.9862s for 16384 events => throughput is 1.37E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0391s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G43/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 19.3110s + [COUNTERS] Fortran Overhead ( 0 ) : 7.3304s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.9408s for 16384 events => throughput is 1.37E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0398s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G68/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 15.0581s + [COUNTERS] Fortran Overhead ( 0 ) : 3.0318s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.9871s for 16384 events => throughput is 1.37E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0392s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G49/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.1800s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1965s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.9445s for 16384 events => throughput is 1.37E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0391s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G25/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 18.7697s + [COUNTERS] Fortran Overhead ( 0 ) : 6.3501s + [COUNTERS] CudaCpp MEs ( 2 ) : 12.3793s for 16384 events => throughput is 1.32E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0403s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G523/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 15.4771s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9117s + [COUNTERS] CudaCpp MEs ( 2 ) : 12.5250s for 16384 events => throughput is 1.31E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0404s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G395/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.9777s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7049s + [COUNTERS] CudaCpp MEs ( 2 ) : 12.2247s for 16384 events => throughput is 1.34E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0481s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G524/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.0150s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0020s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.9736s for 16384 events => throughput is 1.37E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0394s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G65/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.1563s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1153s + [COUNTERS] CudaCpp MEs ( 2 ) : 12.0001s for 16384 events => throughput is 1.37E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0408s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G609/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 15.3923s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1391s + [COUNTERS] CudaCpp MEs ( 2 ) : 12.2131s for 16384 events => throughput is 1.34E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0400s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G611/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 16.0141s + [COUNTERS] Fortran Overhead ( 0 ) : 3.6033s + [COUNTERS] CudaCpp MEs ( 2 ) : 12.3702s for 16384 events => throughput is 1.32E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0406s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G1/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 17.2539s + [COUNTERS] Fortran Overhead ( 0 ) : 4.9486s + [COUNTERS] CudaCpp MEs ( 2 ) : 12.2648s for 16384 events => throughput is 1.34E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0405s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G55/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 15.2036s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8280s + [COUNTERS] CudaCpp MEs ( 2 ) : 12.3315s for 16384 events => throughput is 1.33E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0441s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G34/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.2473s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2483s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.9589s for 16384 events => throughput is 1.37E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0401s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G26/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 19.1179s + [COUNTERS] Fortran Overhead ( 0 ) : 6.6866s + [COUNTERS] CudaCpp MEs ( 2 ) : 12.3907s for 16384 events => throughput is 1.32E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0406s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G58/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 18.5392s + [COUNTERS] Fortran Overhead ( 0 ) : 6.0658s + [COUNTERS] CudaCpp MEs ( 2 ) : 12.4328s for 16384 events => throughput is 1.32E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0405s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G2/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 17.3809s + [COUNTERS] Fortran Overhead ( 0 ) : 5.3217s + [COUNTERS] CudaCpp MEs ( 2 ) : 12.0194s for 16384 events => throughput is 1.36E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0398s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G512/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.1625s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2083s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.9136s for 16384 events => throughput is 1.38E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0406s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G51/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 15.3954s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9137s + [COUNTERS] CudaCpp MEs ( 2 ) : 12.4389s for 16384 events => throughput is 1.32E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0427s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G29/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 19.5869s + [COUNTERS] Fortran Overhead ( 0 ) : 7.0969s + [COUNTERS] CudaCpp MEs ( 2 ) : 12.4498s for 16384 events => throughput is 1.32E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0403s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G48/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.2441s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2240s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.9799s for 16384 events => throughput is 1.37E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0401s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/ajob1' in 70.4326 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/G140/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.2310s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9370s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.2755s for 16384 events => throughput is 7.20E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0184s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/G114/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 9.0034s + [COUNTERS] Fortran Overhead ( 0 ) : 6.7491s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.2370s for 16384 events => throughput is 7.32E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0174s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/G49/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 9.8855s + [COUNTERS] Fortran Overhead ( 0 ) : 7.6252s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.2425s for 16384 events => throughput is 7.31E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0177s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/G120/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.4000s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1099s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.2727s for 16384 events => throughput is 7.21E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0175s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/G262/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.1170s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8450s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.2543s for 16384 events => throughput is 7.27E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0177s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/G27/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.2265s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9514s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.2570s for 16384 events => throughput is 7.26E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0182s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/G44/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.9350s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6571s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.2607s for 16384 events => throughput is 7.25E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0172s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/G96/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.8470s + [COUNTERS] Fortran Overhead ( 0 ) : 3.5761s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.2534s for 16384 events => throughput is 7.27E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0175s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/G261/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.4612s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1939s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.2497s for 16384 events => throughput is 7.28E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0177s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/G37/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.3922s + [COUNTERS] Fortran Overhead ( 0 ) : 5.1142s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.2602s for 16384 events => throughput is 7.25E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0177s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/G28/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.1301s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8655s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.2470s for 16384 events => throughput is 7.29E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0176s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.4342s + [COUNTERS] Fortran Overhead ( 0 ) : 6.1780s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.2388s for 16384 events => throughput is 7.32E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0173s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_epemuduxsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_epemuduxsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_epemuduxsx/ajob1' in 3.0604 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_epemuduxsx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_epemuduxsx/G42/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.0278s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6986s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.3229s for 16384 events => throughput is 5.07E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0063s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_epemuduxsx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dd_epemggdd/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dd_epemggdd/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dd_epemggdd/ajob1' in 27.8702 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dd_epemggdd/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dd_epemggdd/G467/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 9.2563s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9051s + [COUNTERS] CudaCpp MEs ( 2 ) : 7.3211s for 16384 events => throughput is 2.24E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0300s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dd_epemggdd/G351/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.7792s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3351s + [COUNTERS] CudaCpp MEs ( 2 ) : 7.4147s for 16384 events => throughput is 2.21E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0294s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dd_epemggdd/G44/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 9.7430s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3484s + [COUNTERS] CudaCpp MEs ( 2 ) : 7.3644s for 16384 events => throughput is 2.22E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0302s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dd_epemggdd/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamusdxsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamusdxsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamusdxsx/ajob1' in 1.9819 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamusdxsx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamusdxsx/G92/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.9323s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2749s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6510s for 16384 events => throughput is 2.52E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0063s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamusdxsx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemggddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemggddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemggddx/ajob1' in 62.3934 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemggddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemggddx/G175/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 9.6215s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3050s + [COUNTERS] CudaCpp MEs ( 2 ) : 7.2870s for 16384 events => throughput is 2.25E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0295s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemggddx/G303/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 10.1094s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8025s + [COUNTERS] CudaCpp MEs ( 2 ) : 7.2744s for 16384 events => throughput is 2.25E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0324s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemggddx/G111/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 9.1427s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8602s + [COUNTERS] CudaCpp MEs ( 2 ) : 7.2528s for 16384 events => throughput is 2.26E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0297s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemggddx/G219/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 10.5085s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1534s + [COUNTERS] CudaCpp MEs ( 2 ) : 7.3258s for 16384 events => throughput is 2.24E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0293s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemggddx/G318/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.5214s + [COUNTERS] Fortran Overhead ( 0 ) : 6.2227s + [COUNTERS] CudaCpp MEs ( 2 ) : 7.2687s for 16384 events => throughput is 2.25E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0299s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemggddx/G143/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 9.3416s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0146s + [COUNTERS] CudaCpp MEs ( 2 ) : 7.2975s for 16384 events => throughput is 2.25E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0296s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemggddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemguuxux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemguuxux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemguuxux/ajob1' in 14.7741 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemguuxux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemguuxux/G615/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.7248s + [COUNTERS] Fortran Overhead ( 0 ) : 7.3678s + [COUNTERS] CudaCpp MEs ( 2 ) : 7.3271s for 16384 events => throughput is 2.24E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0299s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemguuxux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamgggg/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamgggg/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamgggg/ajob1' in 26.5419 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamgggg/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamgggg/G66/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 26.4898s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7403s + [COUNTERS] CudaCpp MEs ( 2 ) : 23.7110s for 16384 events => throughput is 6.91E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0385s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamgggg/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgduxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgduxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgduxdx/ajob1' in 6.7287 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgduxdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgduxdx/G44/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.6788s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1829s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.4787s for 16384 events => throughput is 3.66E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0173s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgduxdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgudux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgudux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgudux/ajob1' in 55.7660 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgudux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgudux/G39/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 9.1719s + [COUNTERS] Fortran Overhead ( 0 ) : 6.7727s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.3809s for 16384 events => throughput is 6.88E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0183s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgudux/G366/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.1990s + [COUNTERS] Fortran Overhead ( 0 ) : 4.8839s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.2977s for 16384 events => throughput is 7.13E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0174s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgudux/G259/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.6693s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4162s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.2355s for 16384 events => throughput is 7.33E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0176s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgudux/G28/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.4064s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1626s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.2268s for 16384 events => throughput is 7.36E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0170s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgudux/G32/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.7940s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5670s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.2097s for 16384 events => throughput is 7.41E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0173s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgudux/G61/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 11.4781s + [COUNTERS] Fortran Overhead ( 0 ) : 9.1129s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.3441s for 16384 events => throughput is 6.99E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0211s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgudux/G267/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.1247s + [COUNTERS] Fortran Overhead ( 0 ) : 5.8751s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.2323s for 16384 events => throughput is 7.34E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0174s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgudux/G144/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.7579s + [COUNTERS] Fortran Overhead ( 0 ) : 3.5193s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.2211s for 16384 events => throughput is 7.38E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0176s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgudux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemuudux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemuudux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemuudux/ajob1' in 25.8162 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemuudux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemuudux/G104/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 25.7605s + [COUNTERS] Fortran Overhead ( 0 ) : 21.3843s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.3664s for 81920 events => throughput is 1.88E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0098s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemuudux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamggddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamggddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamggddx/ajob1' in 217.8197 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamggddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamggddx/G108/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 26.0416s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3241s + [COUNTERS] CudaCpp MEs ( 2 ) : 23.6786s for 16384 events => throughput is 6.92E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0390s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamggddx/G304/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 26.5229s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8597s + [COUNTERS] CudaCpp MEs ( 2 ) : 23.6240s for 16384 events => throughput is 6.94E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0391s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamggddx/G128/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 26.8787s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9747s + [COUNTERS] CudaCpp MEs ( 2 ) : 23.8648s for 16384 events => throughput is 6.87E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0393s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamggddx/G360/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 26.5727s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5284s + [COUNTERS] CudaCpp MEs ( 2 ) : 24.0038s for 16384 events => throughput is 6.83E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0405s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamggddx/G169/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 31.2568s + [COUNTERS] Fortran Overhead ( 0 ) : 6.7911s + [COUNTERS] CudaCpp MEs ( 2 ) : 24.4241s for 16384 events => throughput is 6.71E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0416s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamggddx/G176/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 25.7207s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1048s + [COUNTERS] CudaCpp MEs ( 2 ) : 23.5747s for 16384 events => throughput is 6.95E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0412s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamggddx/G303/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 26.6984s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9518s + [COUNTERS] CudaCpp MEs ( 2 ) : 23.7066s for 16384 events => throughput is 6.91E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0400s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamggddx/G513/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 27.9544s + [COUNTERS] Fortran Overhead ( 0 ) : 3.3148s + [COUNTERS] CudaCpp MEs ( 2 ) : 24.5986s for 16384 events => throughput is 6.66E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0411s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamggddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguccx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguccx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguccx/ajob1' in 25.7118 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguccx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguccx/G283/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.3461s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0901s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.2375s for 16384 events => throughput is 7.32E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0185s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguccx/G47/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.2694s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0244s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.2276s for 16384 events => throughput is 7.35E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0174s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguccx/G35/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 9.1179s + [COUNTERS] Fortran Overhead ( 0 ) : 6.8755s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.2253s for 16384 events => throughput is 7.36E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0171s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguccx/G373/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.0151s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7669s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.2307s for 16384 events => throughput is 7.34E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0175s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguccx/G249/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.8401s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5909s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.2316s for 16384 events => throughput is 7.34E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0176s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguccx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/ajob1' in 470.4092 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G14/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.0325s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0976s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.8937s for 16384 events => throughput is 1.38E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0412s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G499/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.9884s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0265s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.9228s for 16384 events => throughput is 1.37E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0391s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G56/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.7081s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7855s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.8834s for 16384 events => throughput is 1.38E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0392s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G39/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.2443s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3601s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.8453s for 16384 events => throughput is 1.38E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0389s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G52/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.8275s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9104s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.8777s for 16384 events => throughput is 1.38E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0394s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G304/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.7143s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7273s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.9478s for 16384 events => throughput is 1.37E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0391s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G64/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.3318s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4144s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.8780s for 16384 events => throughput is 1.38E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0394s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G3/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 17.1400s + [COUNTERS] Fortran Overhead ( 0 ) : 5.1769s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.9231s for 16384 events => throughput is 1.37E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0400s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G13/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.6183s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7190s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.8601s for 16384 events => throughput is 1.38E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0391s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G33/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.2229s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2398s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.9424s for 16384 events => throughput is 1.37E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0406s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G484/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.6308s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6796s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.9113s for 16384 events => throughput is 1.38E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0398s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G68/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.2000s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2782s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.8828s for 16384 events => throughput is 1.38E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0390s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G49/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.7729s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8578s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.8766s for 16384 events => throughput is 1.38E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0386s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G40/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.8625s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9446s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.8782s for 16384 events => throughput is 1.38E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0397s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G523/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.1351s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1924s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.9038s for 16384 events => throughput is 1.38E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0389s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.5569s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6046s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.9131s for 16384 events => throughput is 1.38E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0392s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G524/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.0948s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1734s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.8821s for 16384 events => throughput is 1.38E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0392s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G65/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.1350s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2166s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.8785s for 16384 events => throughput is 1.38E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0399s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G44/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 17.9562s + [COUNTERS] Fortran Overhead ( 0 ) : 6.0218s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.8934s for 16384 events => throughput is 1.38E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0410s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.4648s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5602s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.8656s for 16384 events => throughput is 1.38E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0390s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G303/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.9896s + [COUNTERS] Fortran Overhead ( 0 ) : 3.0806s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.8699s for 16384 events => throughput is 1.38E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0391s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G528/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 15.2985s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2664s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.9933s for 16384 events => throughput is 1.37E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0388s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G55/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 15.0462s + [COUNTERS] Fortran Overhead ( 0 ) : 3.0734s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.9336s for 16384 events => throughput is 1.37E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0393s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G298/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 15.1267s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1657s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.9207s for 16384 events => throughput is 1.37E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0403s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G34/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.7336s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7732s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.9211s for 16384 events => throughput is 1.37E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0392s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G54/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 19.3036s + [COUNTERS] Fortran Overhead ( 0 ) : 7.3714s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.8927s for 16384 events => throughput is 1.38E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0394s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G67/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.2390s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3646s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.8351s for 16384 events => throughput is 1.38E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0393s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G512/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.9731s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0459s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.8882s for 16384 events => throughput is 1.38E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0390s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G511/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.5940s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6963s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.8584s for 16384 events => throughput is 1.38E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0392s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.5785s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6054s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.9340s for 16384 events => throughput is 1.37E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0391s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G21/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.7364s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8310s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.8656s for 16384 events => throughput is 1.38E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0398s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G48/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.5125s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5974s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.8754s for 16384 events => throughput is 1.38E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0397s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_taptamggdux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_taptamggdux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_taptamggdux/ajob1' in 27.7613 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_taptamggdux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_taptamggdux/G68/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.6370s + [COUNTERS] Fortran Overhead ( 0 ) : 4.0362s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.5839s for 16384 events => throughput is 3.57E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0169s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_taptamggdux/G167/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.4589s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8865s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.5537s for 16384 events => throughput is 3.60E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0187s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_taptamggdux/G92/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.4015s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8290s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.5548s for 16384 events => throughput is 3.60E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0177s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_taptamggdux/G60/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.1566s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5615s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.5777s for 16384 events => throughput is 3.58E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0173s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_taptamggdux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/ajob1' in 328.9012 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G56/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.5221s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9299s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.5527s for 16384 events => throughput is 1.42E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0396s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G39/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.7280s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1812s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.5068s for 16384 events => throughput is 1.42E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0399s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G517/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 19.9281s + [COUNTERS] Fortran Overhead ( 0 ) : 8.4333s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.4558s for 16384 events => throughput is 1.43E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0390s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G497/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.6914s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0312s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.6212s for 16384 events => throughput is 1.41E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0390s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G13/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.3582s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7736s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.5446s for 16384 events => throughput is 1.42E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0400s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G484/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 18.6399s + [COUNTERS] Fortran Overhead ( 0 ) : 6.9098s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.6910s for 16384 events => throughput is 1.40E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0391s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G68/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.8746s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1860s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.6495s for 16384 events => throughput is 1.41E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0391s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G49/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.9976s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4429s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.5161s for 16384 events => throughput is 1.42E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0386s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G523/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.1670s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6125s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.5152s for 16384 events => throughput is 1.42E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0393s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G518/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 17.8110s + [COUNTERS] Fortran Overhead ( 0 ) : 6.1336s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.6382s for 16384 events => throughput is 1.41E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0393s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G42/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 19.9957s + [COUNTERS] Fortran Overhead ( 0 ) : 8.3840s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.5717s for 16384 events => throughput is 1.42E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0400s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G524/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.5456s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9772s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.5292s for 16384 events => throughput is 1.42E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0392s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G610/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.8870s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2311s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.6167s for 16384 events => throughput is 1.41E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0392s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G26/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 17.8225s + [COUNTERS] Fortran Overhead ( 0 ) : 6.1925s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.5884s for 16384 events => throughput is 1.41E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0416s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G58/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 19.1130s + [COUNTERS] Fortran Overhead ( 0 ) : 7.5380s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.5355s for 16384 events => throughput is 1.42E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0395s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G511/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.8304s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1987s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.5914s for 16384 events => throughput is 1.41E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0404s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.6053s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9564s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.6098s for 16384 events => throughput is 1.41E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0391s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G32/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 19.9047s + [COUNTERS] Fortran Overhead ( 0 ) : 8.2984s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.5642s for 16384 events => throughput is 1.42E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0421s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G61/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 17.9029s + [COUNTERS] Fortran Overhead ( 0 ) : 6.3431s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.5209s for 16384 events => throughput is 1.42E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0390s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G390/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 21.1676s + [COUNTERS] Fortran Overhead ( 0 ) : 9.6114s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.5171s for 16384 events => throughput is 1.42E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0392s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_epemgguc/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_epemgguc/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_epemgguc/ajob1' in 18.9292 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_epemgguc/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_epemgguc/G164/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.8660s + [COUNTERS] Fortran Overhead ( 0 ) : 4.6142s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.2345s for 16384 events => throughput is 7.33E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0173s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_epemgguc/G24/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.3963s + [COUNTERS] Fortran Overhead ( 0 ) : 4.1464s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.2322s for 16384 events => throughput is 7.34E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0176s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_epemgguc/G162/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.5808s + [COUNTERS] Fortran Overhead ( 0 ) : 3.3289s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.2343s for 16384 events => throughput is 7.33E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0175s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_epemgguc/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/ajob1' in 428.7536 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G336/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 15.4956s + [COUNTERS] Fortran Overhead ( 0 ) : 3.6669s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.7894s for 16384 events => throughput is 1.39E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0393s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G108/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.3697s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5685s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.7620s for 16384 events => throughput is 1.39E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0391s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G105/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.2053s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4307s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.7352s for 16384 events => throughput is 1.40E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0394s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G304/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 15.0122s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2340s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.7393s for 16384 events => throughput is 1.40E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0390s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G175/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.0713s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2805s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.7511s for 16384 events => throughput is 1.39E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0397s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G128/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.4328s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6382s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.7553s for 16384 events => throughput is 1.39E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0393s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G360/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.3878s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5985s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.7499s for 16384 events => throughput is 1.39E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0394s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G120/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.0872s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2847s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.7619s for 16384 events => throughput is 1.39E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0406s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G296/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.9032s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1365s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.7278s for 16384 events => throughput is 1.40E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0389s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G299/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.6766s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8734s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.7636s for 16384 events => throughput is 1.39E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0396s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G715/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.5094s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7131s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.7568s for 16384 events => throughput is 1.39E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0395s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G300/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.2515s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4499s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.7621s for 16384 events => throughput is 1.39E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0394s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G303/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.6422s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9193s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.6835s for 16384 events => throughput is 1.40E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0395s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G127/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.6875s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9301s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.7184s for 16384 events => throughput is 1.40E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0391s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G111/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.4457s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6170s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.7891s for 16384 events => throughput is 1.39E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0396s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G126/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 22.6006s + [COUNTERS] Fortran Overhead ( 0 ) : 10.8190s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.7419s for 16384 events => throughput is 1.40E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0397s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G294/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 18.6668s + [COUNTERS] Fortran Overhead ( 0 ) : 6.8404s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.7857s for 16384 events => throughput is 1.39E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0407s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G598/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.2856s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4862s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.7604s for 16384 events => throughput is 1.39E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0390s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G610/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 15.1649s + [COUNTERS] Fortran Overhead ( 0 ) : 3.4135s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.7123s for 16384 events => throughput is 1.40E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0392s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G178/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.4487s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6647s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.7449s for 16384 events => throughput is 1.39E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0392s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G121/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.3177s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5216s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.7566s for 16384 events => throughput is 1.39E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0395s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G685/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.0292s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2383s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.7512s for 16384 events => throughput is 1.39E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0397s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G690/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.9725s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1291s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.8041s for 16384 events => throughput is 1.39E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0393s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G103/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 20.1119s + [COUNTERS] Fortran Overhead ( 0 ) : 8.3172s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.7553s for 16384 events => throughput is 1.39E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0393s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G107/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.1618s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3798s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.7428s for 16384 events => throughput is 1.40E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0392s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G180/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.3213s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5577s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.7240s for 16384 events => throughput is 1.40E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0396s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G97/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 20.2236s + [COUNTERS] Fortran Overhead ( 0 ) : 8.4395s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.7447s for 16384 events => throughput is 1.40E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0394s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G179/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.7050s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8483s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.8171s for 16384 events => throughput is 1.39E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0396s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamggddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamggddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamggddx/ajob1' in 7.1590 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamggddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamggddx/G189/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.1108s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5823s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.5110s for 16384 events => throughput is 3.63E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0176s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamggddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dd_taptamggdd/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dd_taptamggdd/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dd_taptamggdd/ajob1' in 16.8695 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dd_taptamggdd/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dd_taptamggdd/G463/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 16.8219s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2078s + [COUNTERS] CudaCpp MEs ( 2 ) : 14.5842s for 16384 events => throughput is 1.12E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0299s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dd_taptamggdd/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/ajob1' in 1092.8475 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G506/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 27.9862s + [COUNTERS] Fortran Overhead ( 0 ) : 4.3444s + [COUNTERS] CudaCpp MEs ( 2 ) : 23.6023s for 16384 events => throughput is 6.94E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0394s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G499/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 25.5425s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8752s + [COUNTERS] CudaCpp MEs ( 2 ) : 23.6284s for 16384 events => throughput is 6.93E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0389s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G46/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 31.3937s + [COUNTERS] Fortran Overhead ( 0 ) : 7.6566s + [COUNTERS] CudaCpp MEs ( 2 ) : 23.6973s for 16384 events => throughput is 6.91E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0398s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G56/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 26.3149s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5900s + [COUNTERS] CudaCpp MEs ( 2 ) : 23.6856s for 16384 events => throughput is 6.92E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0393s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G39/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 26.0714s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4565s + [COUNTERS] CudaCpp MEs ( 2 ) : 23.5765s for 16384 events => throughput is 6.95E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0385s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G517/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 29.4359s + [COUNTERS] Fortran Overhead ( 0 ) : 5.2204s + [COUNTERS] CudaCpp MEs ( 2 ) : 24.1761s for 16384 events => throughput is 6.78E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0394s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G52/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 26.7511s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9810s + [COUNTERS] CudaCpp MEs ( 2 ) : 23.7311s for 16384 events => throughput is 6.90E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0390s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G31/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 30.6752s + [COUNTERS] Fortran Overhead ( 0 ) : 6.9473s + [COUNTERS] CudaCpp MEs ( 2 ) : 23.6876s for 16384 events => throughput is 6.92E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0403s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G324/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 30.2270s + [COUNTERS] Fortran Overhead ( 0 ) : 6.4897s + [COUNTERS] CudaCpp MEs ( 2 ) : 23.6963s for 16384 events => throughput is 6.91E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0410s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G64/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 25.5456s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8623s + [COUNTERS] CudaCpp MEs ( 2 ) : 23.6436s for 16384 events => throughput is 6.93E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0397s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G3/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 27.7620s + [COUNTERS] Fortran Overhead ( 0 ) : 4.0227s + [COUNTERS] CudaCpp MEs ( 2 ) : 23.6995s for 16384 events => throughput is 6.91E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0398s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G33/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 25.9335s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2395s + [COUNTERS] CudaCpp MEs ( 2 ) : 23.6549s for 16384 events => throughput is 6.93E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0391s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G53/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 30.2293s + [COUNTERS] Fortran Overhead ( 0 ) : 6.6165s + [COUNTERS] CudaCpp MEs ( 2 ) : 23.5737s for 16384 events => throughput is 6.95E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0390s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G43/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 29.2169s + [COUNTERS] Fortran Overhead ( 0 ) : 5.4926s + [COUNTERS] CudaCpp MEs ( 2 ) : 23.6858s for 16384 events => throughput is 6.92E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0385s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G49/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 25.4824s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8187s + [COUNTERS] CudaCpp MEs ( 2 ) : 23.6247s for 16384 events => throughput is 6.94E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0391s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G40/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 26.5612s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9088s + [COUNTERS] CudaCpp MEs ( 2 ) : 23.6133s for 16384 events => throughput is 6.94E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0392s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G523/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 25.8949s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2078s + [COUNTERS] CudaCpp MEs ( 2 ) : 23.6478s for 16384 events => throughput is 6.93E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0393s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G518/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 29.0998s + [COUNTERS] Fortran Overhead ( 0 ) : 5.4533s + [COUNTERS] CudaCpp MEs ( 2 ) : 23.6076s for 16384 events => throughput is 6.94E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0389s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G35/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 25.5720s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9606s + [COUNTERS] CudaCpp MEs ( 2 ) : 23.5722s for 16384 events => throughput is 6.95E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0392s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 25.2703s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6658s + [COUNTERS] CudaCpp MEs ( 2 ) : 23.5653s for 16384 events => throughput is 6.95E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0393s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G42/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 29.1528s + [COUNTERS] Fortran Overhead ( 0 ) : 5.5339s + [COUNTERS] CudaCpp MEs ( 2 ) : 23.5798s for 16384 events => throughput is 6.95E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0390s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G524/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 25.9189s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1595s + [COUNTERS] CudaCpp MEs ( 2 ) : 23.7205s for 16384 events => throughput is 6.91E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0390s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G65/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 25.5771s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9205s + [COUNTERS] CudaCpp MEs ( 2 ) : 23.6176s for 16384 events => throughput is 6.94E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0390s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G27/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 29.5174s + [COUNTERS] Fortran Overhead ( 0 ) : 5.8503s + [COUNTERS] CudaCpp MEs ( 2 ) : 23.6272s for 16384 events => throughput is 6.93E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0398s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G44/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 30.0884s + [COUNTERS] Fortran Overhead ( 0 ) : 6.3509s + [COUNTERS] CudaCpp MEs ( 2 ) : 23.6979s for 16384 events => throughput is 6.91E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0396s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 25.4312s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6611s + [COUNTERS] CudaCpp MEs ( 2 ) : 23.7312s for 16384 events => throughput is 6.90E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0390s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G303/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 26.8717s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2096s + [COUNTERS] CudaCpp MEs ( 2 ) : 23.6231s for 16384 events => throughput is 6.94E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0390s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G55/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 26.3604s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7363s + [COUNTERS] CudaCpp MEs ( 2 ) : 23.5848s for 16384 events => throughput is 6.95E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0392s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G34/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 25.4591s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8270s + [COUNTERS] CudaCpp MEs ( 2 ) : 23.5930s for 16384 events => throughput is 6.94E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0392s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G26/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 29.6616s + [COUNTERS] Fortran Overhead ( 0 ) : 5.9877s + [COUNTERS] CudaCpp MEs ( 2 ) : 23.6349s for 16384 events => throughput is 6.93E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0389s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G62/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 29.7183s + [COUNTERS] Fortran Overhead ( 0 ) : 6.0505s + [COUNTERS] CudaCpp MEs ( 2 ) : 23.6284s for 16384 events => throughput is 6.93E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0395s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G37/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 25.9620s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2000s + [COUNTERS] CudaCpp MEs ( 2 ) : 23.7227s for 16384 events => throughput is 6.91E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0393s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G485/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 29.0660s + [COUNTERS] Fortran Overhead ( 0 ) : 5.3786s + [COUNTERS] CudaCpp MEs ( 2 ) : 23.6479s for 16384 events => throughput is 6.93E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0395s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G67/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 26.0393s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2168s + [COUNTERS] CudaCpp MEs ( 2 ) : 23.7835s for 16384 events => throughput is 6.89E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0390s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G512/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 25.5038s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8860s + [COUNTERS] CudaCpp MEs ( 2 ) : 23.5785s for 16384 events => throughput is 6.95E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0393s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G511/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 26.2185s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5135s + [COUNTERS] CudaCpp MEs ( 2 ) : 23.6663s for 16384 events => throughput is 6.92E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0388s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G51/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 26.2843s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5263s + [COUNTERS] CudaCpp MEs ( 2 ) : 23.7188s for 16384 events => throughput is 6.91E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0393s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 26.1598s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4579s + [COUNTERS] CudaCpp MEs ( 2 ) : 23.6625s for 16384 events => throughput is 6.92E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0394s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G63/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 26.0401s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3431s + [COUNTERS] CudaCpp MEs ( 2 ) : 23.6581s for 16384 events => throughput is 6.93E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0388s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G48/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 26.0721s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4128s + [COUNTERS] CudaCpp MEs ( 2 ) : 23.6200s for 16384 events => throughput is 6.94E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0392s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgggux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgggux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgggux/ajob1' in 195.3040 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgggux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgggux/G33/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 25.4674s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8177s + [COUNTERS] CudaCpp MEs ( 2 ) : 23.6102s for 16384 events => throughput is 6.94E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0395s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgggux/G68/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 26.8532s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2007s + [COUNTERS] CudaCpp MEs ( 2 ) : 23.6135s for 16384 events => throughput is 6.94E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0391s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgggux/G25/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 30.2223s + [COUNTERS] Fortran Overhead ( 0 ) : 6.5353s + [COUNTERS] CudaCpp MEs ( 2 ) : 23.6476s for 16384 events => throughput is 6.93E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0393s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgggux/G494/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 25.3193s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6604s + [COUNTERS] CudaCpp MEs ( 2 ) : 23.6200s for 16384 events => throughput is 6.94E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0390s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgggux/G55/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 26.6065s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9576s + [COUNTERS] CudaCpp MEs ( 2 ) : 23.6093s for 16384 events => throughput is 6.94E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0396s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgggux/G509/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 30.8607s + [COUNTERS] Fortran Overhead ( 0 ) : 7.1452s + [COUNTERS] CudaCpp MEs ( 2 ) : 23.6757s for 16384 events => throughput is 6.92E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0397s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgggux/G29/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 29.8060s + [COUNTERS] Fortran Overhead ( 0 ) : 6.1293s + [COUNTERS] CudaCpp MEs ( 2 ) : 23.6373s for 16384 events => throughput is 6.93E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0395s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgggux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamgguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamgguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamgguux/ajob1' in 86.9603 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamgguux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamgguux/G260/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 16.1754s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6129s + [COUNTERS] CudaCpp MEs ( 2 ) : 14.5331s for 16384 events => throughput is 1.13E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0294s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamgguux/G447/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 16.2398s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6303s + [COUNTERS] CudaCpp MEs ( 2 ) : 14.5801s for 16384 events => throughput is 1.12E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0294s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamgguux/G229/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 18.7381s + [COUNTERS] Fortran Overhead ( 0 ) : 4.1271s + [COUNTERS] CudaCpp MEs ( 2 ) : 14.5814s for 16384 events => throughput is 1.12E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0295s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamgguux/G135/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 16.6345s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0303s + [COUNTERS] CudaCpp MEs ( 2 ) : 14.5746s for 16384 events => throughput is 1.12E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0296s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamgguux/G230/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 19.0396s + [COUNTERS] Fortran Overhead ( 0 ) : 4.4314s + [COUNTERS] CudaCpp MEs ( 2 ) : 14.5784s for 16384 events => throughput is 1.12E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0298s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamgguux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdddx/ajob1' in 18.8034 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdddx/G715/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 9.4033s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1195s + [COUNTERS] CudaCpp MEs ( 2 ) : 7.2543s for 16384 events => throughput is 2.26E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0295s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdddx/G111/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 9.3256s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0775s + [COUNTERS] CudaCpp MEs ( 2 ) : 7.2184s for 16384 events => throughput is 2.27E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0297s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemusdxsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemusdxsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemusdxsx/ajob1' in 6.6537 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemusdxsx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemusdxsx/G67/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.6037s + [COUNTERS] Fortran Overhead ( 0 ) : 6.2806s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.3170s for 16384 events => throughput is 5.17E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0061s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemusdxsx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdssx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdssx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdssx/ajob1' in 25.9555 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdssx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdssx/G31/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.2283s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9638s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.2473s for 16384 events => throughput is 7.29E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0172s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdssx/G27/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.2310s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9726s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.2412s for 16384 events => throughput is 7.31E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0172s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdssx/G119/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.4948s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2348s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.2426s for 16384 events => throughput is 7.31E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0174s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdssx/G284/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.6692s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4131s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.2389s for 16384 events => throughput is 7.32E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0172s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdssx/G375/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.8673s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6112s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.2394s for 16384 events => throughput is 7.32E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0167s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdssx/G102/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.3183s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0635s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.2374s for 16384 events => throughput is 7.32E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0175s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdssx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_epemgguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_epemgguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_epemgguux/ajob1' in 42.7349 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_epemgguux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_epemgguux/G105/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.8778s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5737s + [COUNTERS] CudaCpp MEs ( 2 ) : 7.2749s for 16384 events => throughput is 2.25E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0292s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_epemgguux/G223/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 9.5511s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2649s + [COUNTERS] CudaCpp MEs ( 2 ) : 7.2565s for 16384 events => throughput is 2.26E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0298s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_epemgguux/G103/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 11.9008s + [COUNTERS] Fortran Overhead ( 0 ) : 4.5762s + [COUNTERS] CudaCpp MEs ( 2 ) : 7.2941s for 16384 events => throughput is 2.25E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0304s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_epemgguux/G5/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 12.2969s + [COUNTERS] Fortran Overhead ( 0 ) : 4.9778s + [COUNTERS] CudaCpp MEs ( 2 ) : 7.2896s for 16384 events => throughput is 2.25E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0295s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_epemgguux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/ajob1' in 953.8304 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G506/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 16.9614s + [COUNTERS] Fortran Overhead ( 0 ) : 5.5665s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3557s for 16384 events => throughput is 1.44E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0391s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G499/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.4164s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9068s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.4714s for 16384 events => throughput is 1.43E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0382s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G38/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.2268s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8119s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3756s for 16384 events => throughput is 1.44E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0393s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G46/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 19.5076s + [COUNTERS] Fortran Overhead ( 0 ) : 7.9617s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.5069s for 16384 events => throughput is 1.42E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0390s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G59/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 17.8750s + [COUNTERS] Fortran Overhead ( 0 ) : 6.4191s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.4171s for 16384 events => throughput is 1.44E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0389s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G56/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.3410s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8439s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.4577s for 16384 events => throughput is 1.43E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0394s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G525/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.8980s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4551s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.4036s for 16384 events => throughput is 1.44E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0392s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G39/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.7911s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3901s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3619s for 16384 events => throughput is 1.44E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0390s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G517/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 17.2437s + [COUNTERS] Fortran Overhead ( 0 ) : 5.8423s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3621s for 16384 events => throughput is 1.44E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0393s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G487/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 16.3377s + [COUNTERS] Fortran Overhead ( 0 ) : 4.8851s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.4122s for 16384 events => throughput is 1.44E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0404s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G52/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.4249s + [COUNTERS] Fortran Overhead ( 0 ) : 3.0101s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3732s for 16384 events => throughput is 1.44E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0416s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G31/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 19.3173s + [COUNTERS] Fortran Overhead ( 0 ) : 7.9471s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3305s for 16384 events => throughput is 1.45E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0397s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G64/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.4439s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0301s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3749s for 16384 events => throughput is 1.44E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0389s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G3/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 16.4385s + [COUNTERS] Fortran Overhead ( 0 ) : 5.0404s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3588s for 16384 events => throughput is 1.44E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0393s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G497/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.7469s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3197s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3881s for 16384 events => throughput is 1.44E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0392s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G13/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.1568s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6900s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.4287s for 16384 events => throughput is 1.43E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0381s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G66/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.0282s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4744s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.5149s for 16384 events => throughput is 1.42E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0389s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G33/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.4122s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9948s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3783s for 16384 events => throughput is 1.44E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0391s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G53/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 19.5899s + [COUNTERS] Fortran Overhead ( 0 ) : 8.1937s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3569s for 16384 events => throughput is 1.44E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0393s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G43/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 16.3733s + [COUNTERS] Fortran Overhead ( 0 ) : 4.9502s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3840s for 16384 events => throughput is 1.44E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0391s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G68/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.7961s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2193s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.5375s for 16384 events => throughput is 1.42E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0393s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G49/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.3814s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9612s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3812s for 16384 events => throughput is 1.44E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0390s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G302/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 19.9169s + [COUNTERS] Fortran Overhead ( 0 ) : 8.5062s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3716s for 16384 events => throughput is 1.44E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0392s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G25/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 17.8177s + [COUNTERS] Fortran Overhead ( 0 ) : 6.3652s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.4134s for 16384 events => throughput is 1.44E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0391s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G40/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.7765s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1765s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.5593s for 16384 events => throughput is 1.42E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0407s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G523/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.4925s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0639s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3899s for 16384 events => throughput is 1.44E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0388s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G518/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 16.6518s + [COUNTERS] Fortran Overhead ( 0 ) : 5.2286s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3842s for 16384 events => throughput is 1.44E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0390s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G35/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.6582s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1730s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.4458s for 16384 events => throughput is 1.43E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0394s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.0683s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6695s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3596s for 16384 events => throughput is 1.44E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0393s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G524/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.5618s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1071s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.4153s for 16384 events => throughput is 1.44E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0393s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G65/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.6881s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1805s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.4687s for 16384 events => throughput is 1.43E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0388s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G27/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 17.9326s + [COUNTERS] Fortran Overhead ( 0 ) : 6.5214s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3724s for 16384 events => throughput is 1.44E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0388s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G44/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 17.6689s + [COUNTERS] Fortran Overhead ( 0 ) : 6.2259s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.4036s for 16384 events => throughput is 1.44E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0394s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G483/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 15.7424s + [COUNTERS] Fortran Overhead ( 0 ) : 4.3237s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3798s for 16384 events => throughput is 1.44E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0390s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G482/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 17.1596s + [COUNTERS] Fortran Overhead ( 0 ) : 5.7493s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3712s for 16384 events => throughput is 1.44E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0391s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G1/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 16.0490s + [COUNTERS] Fortran Overhead ( 0 ) : 4.6242s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3851s for 16384 events => throughput is 1.44E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0397s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.0270s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5713s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.4164s for 16384 events => throughput is 1.44E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0393s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G55/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.1646s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5637s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.5611s for 16384 events => throughput is 1.42E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0398s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G298/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.5379s + [COUNTERS] Fortran Overhead ( 0 ) : 3.0666s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.4320s for 16384 events => throughput is 1.43E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0392s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G513/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.2752s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9275s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3085s for 16384 events => throughput is 1.45E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0393s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G60/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 18.9887s + [COUNTERS] Fortran Overhead ( 0 ) : 7.5696s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3796s for 16384 events => throughput is 1.44E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0394s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G34/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.4555s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0359s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3806s for 16384 events => throughput is 1.44E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0389s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G26/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 18.4677s + [COUNTERS] Fortran Overhead ( 0 ) : 7.0407s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3874s for 16384 events => throughput is 1.44E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0396s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G58/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 17.1210s + [COUNTERS] Fortran Overhead ( 0 ) : 5.6152s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.4665s for 16384 events => throughput is 1.43E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0393s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G54/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 17.5118s + [COUNTERS] Fortran Overhead ( 0 ) : 6.0810s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3918s for 16384 events => throughput is 1.44E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0390s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G62/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 18.4515s + [COUNTERS] Fortran Overhead ( 0 ) : 7.0586s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3541s for 16384 events => throughput is 1.44E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0387s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G37/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.3081s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9190s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3500s for 16384 events => throughput is 1.44E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0391s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G493/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.5906s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1802s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3716s for 16384 events => throughput is 1.44E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0387s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G4/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 17.7629s + [COUNTERS] Fortran Overhead ( 0 ) : 6.3469s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3768s for 16384 events => throughput is 1.44E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0391s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G50/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.5663s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1441s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3836s for 16384 events => throughput is 1.44E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0386s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G485/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 17.3256s + [COUNTERS] Fortran Overhead ( 0 ) : 5.8913s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3949s for 16384 events => throughput is 1.44E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0394s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G41/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 18.6662s + [COUNTERS] Fortran Overhead ( 0 ) : 7.2241s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.4011s for 16384 events => throughput is 1.44E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0410s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G67/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.6098s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1626s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.4078s for 16384 events => throughput is 1.44E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0394s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G512/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.4747s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0580s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3775s for 16384 events => throughput is 1.44E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0392s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G511/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.0024s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5733s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3901s for 16384 events => throughput is 1.44E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0389s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G57/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 18.4354s + [COUNTERS] Fortran Overhead ( 0 ) : 7.0313s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3651s for 16384 events => throughput is 1.44E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0390s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.4050s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9445s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.4218s for 16384 events => throughput is 1.43E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0387s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G63/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.9335s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4595s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.4348s for 16384 events => throughput is 1.43E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0392s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G32/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 19.2367s + [COUNTERS] Fortran Overhead ( 0 ) : 7.8417s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3559s for 16384 events => throughput is 1.44E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0391s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G29/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 17.6837s + [COUNTERS] Fortran Overhead ( 0 ) : 6.3083s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3363s for 16384 events => throughput is 1.45E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0391s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G48/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.7573s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3868s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.3315s for 16384 events => throughput is 1.45E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0390s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_taptamggdsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_taptamggdsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_taptamggdsx/ajob1' in 20.2707 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_taptamggdsx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_taptamggdsx/G92/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.9598s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4743s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.4685s for 16384 events => throughput is 3.67E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0171s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_taptamggdsx/G144/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.6207s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1317s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.4719s for 16384 events => throughput is 3.66E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0172s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_taptamggdsx/G207/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.5990s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1010s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.4805s for 16384 events => throughput is 3.66E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0176s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_taptamggdsx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_epemggdsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_epemggdsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_epemggdsx/ajob1' in 9.1433 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_epemggdsx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_epemggdsx/G205/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 9.0901s + [COUNTERS] Fortran Overhead ( 0 ) : 6.8358s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.2369s for 16384 events => throughput is 7.32E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0174s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_epemggdsx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_taptamgggg/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_taptamgggg/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_taptamgggg/ajob1' in 26.3533 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_taptamgggg/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_taptamgggg/G67/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 26.3161s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7975s + [COUNTERS] CudaCpp MEs ( 2 ) : 23.4788s for 16384 events => throughput is 6.98E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0399s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_taptamgggg/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dc_epemdcssx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dc_epemdcssx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dc_epemdcssx/ajob1' in 2.6410 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dc_epemdcssx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dc_epemdcssx/G77/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.5883s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2622s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.3200s for 16384 events => throughput is 5.12E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0061s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dc_epemdcssx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguccx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguccx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguccx/ajob1' in 7.0304 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguccx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguccx/G48/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.9799s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5434s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.4193s for 16384 events => throughput is 3.71E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0172s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguccx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ds_epemggds/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ds_epemggds/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ds_epemggds/ajob1' in 3.9237 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ds_epemggds/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ds_epemggds/G26/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.8923s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6560s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.2190s for 16384 events => throughput is 7.38E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0173s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ds_epemggds/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ucx_epemggucx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ucx_epemggucx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ucx_epemggucx/ajob1' in 10.3463 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ucx_epemggucx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ucx_epemggucx/G163/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.4009s + [COUNTERS] Fortran Overhead ( 0 ) : 4.1411s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.2427s for 16384 events => throughput is 7.31E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0172s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ucx_epemggucx/G60/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.8740s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6061s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.2506s for 16384 events => throughput is 7.28E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0173s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ucx_epemggucx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemggudx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemggudx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemggudx/ajob1' in 69.1174 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemggudx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemggudx/G109/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.7755s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5282s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.2301s for 16384 events => throughput is 7.35E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0172s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemggudx/G151/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.1251s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8751s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.2328s for 16384 events => throughput is 7.34E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0171s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemggudx/G106/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.5434s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2866s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.2395s for 16384 events => throughput is 7.32E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0173s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemggudx/G35/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 9.2627s + [COUNTERS] Fortran Overhead ( 0 ) : 7.0158s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.2296s for 16384 events => throughput is 7.35E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0173s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemggudx/G119/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.2851s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0452s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.2227s for 16384 events => throughput is 7.37E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0173s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemggudx/G187/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.1356s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8931s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.2255s for 16384 events => throughput is 7.36E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0170s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemggudx/G91/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.3057s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0520s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.2367s for 16384 events => throughput is 7.33E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0170s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemggudx/G363/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 26.6161s + [COUNTERS] Fortran Overhead ( 0 ) : 17.6731s + [COUNTERS] CudaCpp MEs ( 2 ) : 8.9261s for 65536 events => throughput is 7.34E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0169s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemggudx/G149/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.8671s + [COUNTERS] Fortran Overhead ( 0 ) : 4.5849s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.2648s for 16384 events => throughput is 7.23E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0173s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemggudx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgcuxcx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgcuxcx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgcuxcx/ajob1' in 4.8345 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgcuxcx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgcuxcx/G120/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.7843s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5235s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.2433s for 16384 events => throughput is 7.30E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0175s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgcuxcx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemgggg/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemgggg/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemgggg/ajob1' in 30.6656 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemgggg/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemgggg/G65/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 15.2169s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2999s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.8772s for 16384 events => throughput is 1.38E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0398s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemgggg/G48/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 15.3800s + [COUNTERS] Fortran Overhead ( 0 ) : 3.5423s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.7988s for 16384 events => throughput is 1.39E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0389s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemgggg/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_taptamggud/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_taptamggud/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_taptamggud/ajob1' in 61.6396 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_taptamggud/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_taptamggud/G46/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.7637s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2778s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.4684s for 16384 events => throughput is 3.67E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0175s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_taptamggud/G39/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.7206s + [COUNTERS] Fortran Overhead ( 0 ) : 4.2285s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.4747s for 16384 events => throughput is 3.66E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0174s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_taptamggud/G112/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.7981s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2799s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.5005s for 16384 events => throughput is 3.64E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0177s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_taptamggud/G115/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.0327s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5273s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.4880s for 16384 events => throughput is 3.65E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0174s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_taptamggud/G270/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.4381s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9645s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.4562s for 16384 events => throughput is 3.68E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0175s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_taptamggud/G96/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.0878s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6223s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.4484s for 16384 events => throughput is 3.68E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0171s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_taptamggud/G107/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.4243s + [COUNTERS] Fortran Overhead ( 0 ) : 3.9388s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.4679s for 16384 events => throughput is 3.67E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0176s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_taptamggud/G63/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.3946s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9080s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.4693s for 16384 events => throughput is 3.67E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0173s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_taptamggud/G375/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.7859s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3163s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.4521s for 16384 events => throughput is 3.68E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0175s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_taptamggud/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguuux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguuux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguuux/ajob1' in 45.5901 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguuux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguuux/G122/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 9.2547s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9036s + [COUNTERS] CudaCpp MEs ( 2 ) : 7.3220s for 16384 events => throughput is 2.24E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0291s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguuux/G33/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.2529s + [COUNTERS] Fortran Overhead ( 0 ) : 5.9109s + [COUNTERS] CudaCpp MEs ( 2 ) : 7.3124s for 16384 events => throughput is 2.24E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0296s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguuux/G127/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 10.4843s + [COUNTERS] Fortran Overhead ( 0 ) : 3.0922s + [COUNTERS] CudaCpp MEs ( 2 ) : 7.3624s for 16384 events => throughput is 2.23E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0297s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguuux/G116/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 12.4959s + [COUNTERS] Fortran Overhead ( 0 ) : 5.1382s + [COUNTERS] CudaCpp MEs ( 2 ) : 7.3279s for 16384 events => throughput is 2.24E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0297s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguuux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamgggdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamgggdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamgggdx/ajob1' in 231.9228 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamgggdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamgggdx/G56/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 26.2521s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6355s + [COUNTERS] CudaCpp MEs ( 2 ) : 23.5768s for 16384 events => throughput is 6.95E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0398s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamgggdx/G64/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 25.9616s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4538s + [COUNTERS] CudaCpp MEs ( 2 ) : 23.4688s for 16384 events => throughput is 6.98E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0390s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamgggdx/G13/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 25.2206s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6808s + [COUNTERS] CudaCpp MEs ( 2 ) : 23.5006s for 16384 events => throughput is 6.97E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0392s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamgggdx/G33/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 25.4629s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8636s + [COUNTERS] CudaCpp MEs ( 2 ) : 23.5602s for 16384 events => throughput is 6.95E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0391s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamgggdx/G49/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 25.7550s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1749s + [COUNTERS] CudaCpp MEs ( 2 ) : 23.5407s for 16384 events => throughput is 6.96E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0393s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamgggdx/G523/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 26.2587s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6451s + [COUNTERS] CudaCpp MEs ( 2 ) : 23.5742s for 16384 events => throughput is 6.95E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0394s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamgggdx/G65/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 25.5308s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0296s + [COUNTERS] CudaCpp MEs ( 2 ) : 23.4622s for 16384 events => throughput is 6.98E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0390s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamgggdx/G34/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 25.7067s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1546s + [COUNTERS] CudaCpp MEs ( 2 ) : 23.5126s for 16384 events => throughput is 6.97E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0395s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamgggdx/G48/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 25.5750s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0477s + [COUNTERS] CudaCpp MEs ( 2 ) : 23.4879s for 16384 events => throughput is 6.98E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0393s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamgggdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgudux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgudux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgudux/ajob1' in 16.7520 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgudux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgudux/G59/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.3930s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9119s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.4638s for 16384 events => throughput is 3.67E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0172s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgudux/G61/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 10.2921s + [COUNTERS] Fortran Overhead ( 0 ) : 5.7918s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.4828s for 16384 events => throughput is 3.65E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0175s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgudux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgduxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgduxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgduxdx/ajob1' in 4.3330 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgduxdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgduxdx/G104/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.2834s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9810s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.2851s for 16384 events => throughput is 7.17E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0173s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgduxdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgsdxsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgsdxsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgsdxsx/ajob1' in 4.0852 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgsdxsx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgsdxsx/G25/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.0464s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7943s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.2347s for 16384 events => throughput is 7.33E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0174s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgsdxsx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/ajob1' in 139.5528 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G208/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.2413s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9710s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.2528s for 16384 events => throughput is 7.27E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0175s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G235/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.0079s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7548s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.2359s for 16384 events => throughput is 7.33E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0171s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G140/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.2925s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0498s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.2257s for 16384 events => throughput is 7.36E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0170s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G109/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.9884s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7325s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.2388s for 16384 events => throughput is 7.32E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0170s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G33/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.1005s + [COUNTERS] Fortran Overhead ( 0 ) : 4.8498s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.2338s for 16384 events => throughput is 7.33E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0170s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G43/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.0787s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8216s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.2399s for 16384 events => throughput is 7.31E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0172s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G240/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.0740s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8208s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.2361s for 16384 events => throughput is 7.33E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0171s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G268/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.8023s + [COUNTERS] Fortran Overhead ( 0 ) : 4.5509s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.2342s for 16384 events => throughput is 7.33E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0172s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G139/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.2855s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0281s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.2404s for 16384 events => throughput is 7.31E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0170s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G120/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.0337s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7816s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.2349s for 16384 events => throughput is 7.33E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0172s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G236/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.1036s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8541s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.2324s for 16384 events => throughput is 7.34E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0171s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G181/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.6595s + [COUNTERS] Fortran Overhead ( 0 ) : 5.4064s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.2357s for 16384 events => throughput is 7.33E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0173s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G351/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.6603s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4012s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.2418s for 16384 events => throughput is 7.31E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0173s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G65/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.2204s + [COUNTERS] Fortran Overhead ( 0 ) : 3.9649s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.2385s for 16384 events => throughput is 7.32E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0170s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G71/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.8544s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5992s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.2381s for 16384 events => throughput is 7.32E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0172s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G119/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.2163s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9630s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.2360s for 16384 events => throughput is 7.33E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0172s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G127/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.8571s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6065s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.2331s for 16384 events => throughput is 7.34E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0175s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G96/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.0119s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7331s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.2603s for 16384 events => throughput is 7.25E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0185s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G60/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.5264s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2728s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.2363s for 16384 events => throughput is 7.33E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0173s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G26/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.7491s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4976s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.2341s for 16384 events => throughput is 7.33E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0174s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G37/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.1598s + [COUNTERS] Fortran Overhead ( 0 ) : 5.9071s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.2352s for 16384 events => throughput is 7.33E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0175s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G125/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.6424s + [COUNTERS] Fortran Overhead ( 0 ) : 3.3964s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.2289s for 16384 events => throughput is 7.35E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0172s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G91/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.7558s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5150s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.2237s for 16384 events => throughput is 7.37E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0171s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G143/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.6287s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3850s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.2265s for 16384 events => throughput is 7.36E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0172s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G116/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.5834s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3281s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.2381s for 16384 events => throughput is 7.32E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0172s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G61/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.9857s + [COUNTERS] Fortran Overhead ( 0 ) : 5.7362s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.2325s for 16384 events => throughput is 7.34E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0170s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G149/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.6937s + [COUNTERS] Fortran Overhead ( 0 ) : 4.4525s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.2240s for 16384 events => throughput is 7.37E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0172s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G144/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.7615s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4944s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.2500s for 16384 events => throughput is 7.28E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0171s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G207/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.0030s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7588s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.2270s for 16384 events => throughput is 7.36E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0172s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_epemgguu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_epemgguu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_epemgguu/ajob1' in 112.0588 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_epemgguu/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_epemgguu/G463/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 9.3536s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0743s + [COUNTERS] CudaCpp MEs ( 2 ) : 7.2501s for 16384 events => throughput is 2.26E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0292s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_epemgguu/G100/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 10.4746s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1704s + [COUNTERS] CudaCpp MEs ( 2 ) : 7.2745s for 16384 events => throughput is 2.25E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0297s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_epemgguu/G110/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.7851s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4794s + [COUNTERS] CudaCpp MEs ( 2 ) : 7.2756s for 16384 events => throughput is 2.25E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0302s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_epemgguu/G124/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.8753s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5833s + [COUNTERS] CudaCpp MEs ( 2 ) : 7.2625s for 16384 events => throughput is 2.26E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0295s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_epemgguu/G367/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 9.2767s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9661s + [COUNTERS] CudaCpp MEs ( 2 ) : 7.2810s for 16384 events => throughput is 2.25E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0296s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_epemgguu/G126/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 12.3425s + [COUNTERS] Fortran Overhead ( 0 ) : 4.1011s + [COUNTERS] CudaCpp MEs ( 2 ) : 8.2109s for 16384 events => throughput is 2.00E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0306s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_epemgguu/G361/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 11.7822s + [COUNTERS] Fortran Overhead ( 0 ) : 4.4789s + [COUNTERS] CudaCpp MEs ( 2 ) : 7.2739s for 16384 events => throughput is 2.25E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0294s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_epemgguu/G116/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 10.0407s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7056s + [COUNTERS] CudaCpp MEs ( 2 ) : 7.3043s for 16384 events => throughput is 2.24E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0307s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_epemgguu/G363/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 11.4810s + [COUNTERS] Fortran Overhead ( 0 ) : 4.1757s + [COUNTERS] CudaCpp MEs ( 2 ) : 7.2736s for 16384 events => throughput is 2.25E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0317s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_epemgguu/G138/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.9408s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6265s + [COUNTERS] CudaCpp MEs ( 2 ) : 7.2843s for 16384 events => throughput is 2.25E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0300s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_epemgguu/G102/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 10.4724s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1880s + [COUNTERS] CudaCpp MEs ( 2 ) : 7.2551s for 16384 events => throughput is 2.26E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0293s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_epemgguu/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgdssx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgdssx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgdssx/ajob1' in 6.6698 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgdssx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgdssx/G28/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.6182s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0840s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.5168s for 16384 events => throughput is 3.63E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0174s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgdssx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_taptamgguu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_taptamgguu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_taptamgguu/ajob1' in 84.2981 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_taptamgguu/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_taptamgguu/G464/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 16.7034s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9567s + [COUNTERS] CudaCpp MEs ( 2 ) : 14.7170s for 16384 events => throughput is 1.11E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0297s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_taptamgguu/G59/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 16.8291s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1223s + [COUNTERS] CudaCpp MEs ( 2 ) : 14.6767s for 16384 events => throughput is 1.12E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0302s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_taptamgguu/G100/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 17.3201s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5788s + [COUNTERS] CudaCpp MEs ( 2 ) : 14.7117s for 16384 events => throughput is 1.11E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0296s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_taptamgguu/G106/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 16.6307s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8694s + [COUNTERS] CudaCpp MEs ( 2 ) : 14.7315s for 16384 events => throughput is 1.11E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0297s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_taptamgguu/G143/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 16.6897s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8456s + [COUNTERS] CudaCpp MEs ( 2 ) : 14.8141s for 16384 events => throughput is 1.11E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0300s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_taptamgguu/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dc_epemggdc/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dc_epemggdc/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dc_epemggdc/ajob1' in 7.5511 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dc_epemggdc/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dc_epemggdc/G18/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.5040s + [COUNTERS] Fortran Overhead ( 0 ) : 5.2813s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.2053s for 16384 events => throughput is 7.43E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0174s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dc_epemggdc/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ucx_taptamggucx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ucx_taptamggucx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ucx_taptamggucx/ajob1' in 9.2745 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ucx_taptamggucx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ucx_taptamggucx/G62/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 9.2236s + [COUNTERS] Fortran Overhead ( 0 ) : 4.7541s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.4522s for 16384 events => throughput is 3.68E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0173s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ucx_taptamggucx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamucdxcx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamucdxcx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamucdxcx/ajob1' in 2.5472 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamucdxcx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamucdxcx/G47/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.4978s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8416s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6500s for 16384 events => throughput is 2.52E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0062s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamucdxcx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_epemgggg/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_epemgggg/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_epemgggg/ajob1' in 14.9875 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_epemgggg/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_epemgggg/G49/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.9358s + [COUNTERS] Fortran Overhead ( 0 ) : 3.4567s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.4401s for 16384 events => throughput is 1.43E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0390s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_epemgggg/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemudddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemudddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemudddx/ajob1' in 2.5037 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemudddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemudddx/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.4527s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6298s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8136s for 16384 events => throughput is 2.01E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0093s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemudddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamggudx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamggudx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamggudx/ajob1' in 6.2082 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamggudx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamggudx/G60/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.1779s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6774s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.4826s for 16384 events => throughput is 3.66E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0180s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamggudx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_taptamuuddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_taptamuuddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_taptamuuddx/ajob1' in 3.4014 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_taptamuuddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_taptamuuddx/G11/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.3534s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6858s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.6582s for 16384 events => throughput is 9.88E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0094s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_taptamuuddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished in 7127.5099 seconds +__CUDACPP_DEBUG: gen_ximprove.launch finished in 7131.7128 seconds +__CUDACPP_DEBUG: GridPackCmd.refine4grid finished in 7131.7138 seconds +__CUDACPP_DEBUG: GridPackCmd.launch (3a) combine_events +combine_events +INFO: fail to reach target 500 +__CUDACPP_DEBUG: GridPackCmd.launch (3b) store_events +store_events + === Results Summary for run: GridRun_21 tag: tag_1 === + + Cross-section : 0 +- 0 pb + Nb of events : 195 + +INFO: No version of lhapdf. Can not run systematics computation +decay_events -from_cards +__CUDACPP_DEBUG: GridPackCmd.launch finished in 7335.2356 seconds +__CUDACPP_DEBUG: GridPackCmd.__init__ back from self.launch +DONE +write ./events.lhe.gz +END: Sun Sep 15 01:45:02 PM CEST 2024 +ELAPSED: 7338 seconds diff --git a/epochX/cudacpp/tlau/fromgridpacks/pp_dy4j.mad/cppnone/output.txt b/epochX/cudacpp/tlau/fromgridpacks/pp_dy4j.mad/cppnone/output.txt new file mode 100644 index 0000000000..1815809970 --- /dev/null +++ b/epochX/cudacpp/tlau/fromgridpacks/pp_dy4j.mad/cppnone/output.txt @@ -0,0 +1,3327 @@ +START: Sun Sep 15 12:10:19 AM CEST 2024 +On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU (nproc=4)] [GPU: 1x Tesla V100S-PCIE-32GB]: +CUDACPP_RUNTIME_DISABLEFPE is set +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_bbx_epemdsdxsx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_bbx_taptamdsdxsx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_ccx_epemdcdxcx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_ccx_epemdsdxsx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_ccx_epemucuxcx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_ccx_epemuduxdx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_ccx_taptamdcdxcx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_ccx_taptamdsdxsx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_ccx_taptamucuxcx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_ccx_taptamuduxdx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_dc_epemdcccx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_dc_epemdcssx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_dc_epemddcdx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_dc_epemggdc/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_dc_epemudcux/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_dc_taptamdcccx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_dc_taptamdcssx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_dc_taptamddcdx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_dc_taptamggdc/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_dc_taptamudcux/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_dd_epemddddx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_dd_epemddssx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_dd_epemggdd/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_dd_epemuddux/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_dd_taptamddddx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_dd_taptamddssx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_dd_taptamggdd/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_dd_taptamuddux/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_ddx_epemdcdxcx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_ddx_epemdddxdx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_ddx_epemdsdxsx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_ddx_epemggddx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_ddx_epemgggg/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_ddx_epemggssx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_ddx_epemgguux/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_ddx_epemsbsxbx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_ddx_epemsssxsx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_ddx_epemucuxcx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_ddx_epemuduxdx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_ddx_epemusuxsx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_ddx_epemuuuxux/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_ddx_taptamdcdxcx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_ddx_taptamdddxdx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_ddx_taptamdsdxsx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_ddx_taptamggddx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_ddx_taptamgggg/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_ddx_taptamggssx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_ddx_taptamgguux/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_ddx_taptamsbsxbx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_ddx_taptamsssxsx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_ddx_taptamucuxcx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_ddx_taptamuduxdx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_ddx_taptamusuxsx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_ddx_taptamuuuxux/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_ds_epemddsdx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_ds_epemdsbbx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_ds_epemdsssx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_ds_epemggds/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_ds_epemudsux/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_ds_taptamddsdx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_ds_taptamdsbbx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_ds_taptamdsssx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_ds_taptamggds/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_ds_taptamudsux/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_dsx_epemdbsxbx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_dsx_epemdddxsx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_dsx_epemdssxsx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_dsx_epemggdsx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_dsx_epemuduxsx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_dsx_taptamdbsxbx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_dsx_taptamdddxsx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_dsx_taptamdssxsx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_dsx_taptamggdsx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_dsx_taptamuduxsx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_dux_epemdcuxcx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_dux_epemdduxdx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_dux_epemdsuxsx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_dux_epemggdux/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_dux_epemuduxux/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_dux_taptamdcuxcx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_dux_taptamdduxdx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_dux_taptamdsuxsx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_dux_taptamggdux/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_dux_taptamuduxux/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_dxcx_epemcdxcxcx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_dxcx_epemddxdxcx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_dxcx_epemggdxcx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_dxcx_epemsdxcxsx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_dxcx_epemuuxdxcx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_dxcx_taptamcdxcxcx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_dxcx_taptamddxdxcx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_dxcx_taptamggdxcx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_dxcx_taptamsdxcxsx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_dxcx_taptamuuxdxcx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_dxdx_epemddxdxdx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_dxdx_epemggdxdx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_dxdx_epemsdxdxsx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_dxdx_epemuuxdxdx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_dxdx_taptamddxdxdx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_dxdx_taptamggdxdx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_dxdx_taptamsdxdxsx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_dxdx_taptamuuxdxdx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_dxsx_epembdxsxbx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_dxsx_epemddxdxsx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_dxsx_epemggdxsx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_dxsx_epemsdxsxsx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_dxsx_epemuuxdxsx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_dxsx_taptambdxsxbx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_dxsx_taptamddxdxsx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_dxsx_taptamggdxsx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_dxsx_taptamsdxsxsx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_dxsx_taptamuuxdxsx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_gd_epemgdddx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_gd_epemgdssx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_gd_epemgggd/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_gd_epemgudux/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_gd_taptamgdddx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_gd_taptamgdssx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_gd_taptamgggd/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_gd_taptamgudux/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_gdx_epemgddxdx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_gdx_epemgggdx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_gdx_epemgsdxsx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_gdx_epemguuxdx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_gdx_taptamgddxdx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_gdx_taptamgggdx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_gdx_taptamgsdxsx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_gdx_taptamguuxdx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_gg_epemdddxdx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_gg_epemdsdxsx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_gg_epemggddx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_gg_epemgguux/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_gg_epemucuxcx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_gg_epemuduxdx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_gg_epemuuuxux/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_gg_taptamdddxdx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_gg_taptamdsdxsx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_gg_taptamggddx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_gg_taptamgguux/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_gg_taptamucuxcx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_gg_taptamuduxdx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_gg_taptamuuuxux/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_gu_epemgggu/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_gu_epemguccx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_gu_epemguddx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_gu_epemguuux/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_gu_taptamgggu/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_gu_taptamguccx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_gu_taptamguddx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_gu_taptamguuux/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_gux_epemgcuxcx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_gux_epemgduxdx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_gux_epemgggux/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_gux_epemguuxux/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_gux_taptamgcuxcx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_gux_taptamgduxdx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_gux_taptamgggux/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_gux_taptamguuxux/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_ssx_epemdbdxbx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_ssx_epemdcdxcx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_ssx_epemdsdxsx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_ssx_epemucuxcx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_ssx_epemuduxdx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_ssx_taptamdbdxbx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_ssx_taptamdcdxcx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_ssx_taptamdsdxsx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_ssx_taptamucuxcx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_ssx_taptamuduxdx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_uc_epemgguc/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_uc_epemucccx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_uc_epemudcdx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_uc_epemuucux/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_uc_taptamgguc/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_uc_taptamucccx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_uc_taptamudcdx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_uc_taptamuucux/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_ucx_epemggucx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_ucx_epemuccxcx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_ucx_epemuddxcx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_ucx_epemuuuxcx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_ucx_taptamggucx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_ucx_taptamuccxcx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_ucx_taptamuddxcx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_ucx_taptamuuuxcx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_ud_epemggud/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_ud_epemudccx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_ud_epemudddx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_ud_epemudssx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_ud_epemuudux/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_ud_taptamggud/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_ud_taptamudccx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_ud_taptamudddx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_ud_taptamudssx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_ud_taptamuudux/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_udx_epemggudx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_udx_epemucdxcx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_udx_epemuddxdx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_udx_epemusdxsx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_udx_epemuuuxdx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_udx_taptamggudx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_udx_taptamucdxcx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_udx_taptamuddxdx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_udx_taptamusdxsx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_udx_taptamuuuxdx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_uu_epemgguu/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_uu_epemuuccx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_uu_epemuuddx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_uu_epemuuuux/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_uu_taptamgguu/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_uu_taptamuuccx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_uu_taptamuuddx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_uu_taptamuuuux/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_uux_epemcccxcx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_uux_epemcscxsx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_uux_epemdcdxcx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_uux_epemdddxdx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_uux_epemdsdxsx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_uux_epemggccx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_uux_epemggddx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_uux_epemgggg/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_uux_epemgguux/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_uux_epemucuxcx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_uux_epemuduxdx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_uux_epemuuuxux/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_uux_taptamcccxcx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_uux_taptamcscxsx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_uux_taptamdcdxcx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_uux_taptamdddxdx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_uux_taptamdsdxsx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_uux_taptamggccx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_uux_taptamggddx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_uux_taptamgggg/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_uux_taptamgguux/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_uux_taptamucuxcx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_uux_taptamuduxdx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_uux_taptamuuuxux/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_uxcx_epemcuxcxcx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_uxcx_epemduxdxcx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_uxcx_epemgguxcx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_uxcx_epemuuxuxcx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_uxcx_taptamcuxcxcx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_uxcx_taptamduxdxcx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_uxcx_taptamgguxcx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_uxcx_taptamuuxuxcx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_uxdx_epemcuxdxcx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_uxdx_epemduxdxdx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_uxdx_epemgguxdx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_uxdx_epemsuxdxsx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_uxdx_epemuuxuxdx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_uxdx_taptamcuxdxcx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_uxdx_taptamduxdxdx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_uxdx_taptamgguxdx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_uxdx_taptamsuxdxsx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_uxdx_taptamuuxuxdx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_uxux_epemcuxuxcx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_uxux_epemduxuxdx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_uxux_epemgguxux/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_uxux_epemuuxuxux/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_uxux_taptamcuxuxcx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_uxux_taptamduxuxdx/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_uxux_taptamgguxux/madevent -> build.none_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 00:10 madevent/SubProcesses/P0_uxux_taptamuuxuxux/madevent -> build.none_d_inl0_hrd0/madevent_cpp +Now generating 500 events with random seed 21 and granularity 1 +************************************************************ +* * +* W E L C O M E to * +* M A D G R A P H 5 _ a M C @ N L O * +* M A D E V E N T * +* * +* * * * +* * * * * * +* * * * * 5 * * * * * +* * * * * * +* * * * +* * +* VERSION 3.5.3_lo_vect * +* * +* The MadGraph5_aMC@NLO Development Team - Find us at * +* https://server06.fynu.ucl.ac.be/projects/madgraph * +* * +* Type 'help' for in-line help. * +* * +************************************************************ +INFO: load configuration from /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/Cards/me5_configuration.txt +INFO: load configuration from /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/Cards/me5_configuration.txt +Using default text editor "vi". Set another one in ./input/mg5_configuration.txt +Using default eps viewer "evince". Set another one in ./input/mg5_configuration.txt +Using default web browser "firefox". Set another one in ./input/mg5_configuration.txt +WRITE GRIDCARD /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent +__CUDACPP_DEBUG: GridPackCmd.__init__ will call self.launch +__CUDACPP_DEBUG: GridPackCmd.launch starting +__CUDACPP_DEBUG: GridPackCmd.launch (1) restore_data +generate 500 events +P0_gg_epemgguux +P0_gg_epemggddx +P0_gg_taptamgguux +P0_gg_taptamggddx +P0_gu_epemgggu +P0_gd_epemgggd +P0_gux_epemgggux +P0_gdx_epemgggdx +P0_gu_taptamgggu +P0_gd_taptamgggd +P0_gux_taptamgggux +P0_gdx_taptamgggdx +P0_uux_epemgggg +P0_ddx_epemgggg +P0_uux_taptamgggg +P0_ddx_taptamgggg +P0_gg_epemuuuxux +P0_gg_epemdddxdx +P0_gg_epemuduxdx +P0_gg_epemucuxcx +P0_gg_epemdsdxsx +P0_gg_taptamuuuxux +P0_gg_taptamdddxdx +P0_gg_taptamuduxdx +P0_gg_taptamucuxcx +P0_gg_taptamdsdxsx +P0_gu_epemguuux +P0_gd_epemgdddx +P0_gux_epemguuxux +P0_gdx_epemgddxdx +P0_gu_epemguddx +P0_gu_epemguccx +P0_gd_epemgudux +P0_gd_epemgdssx +P0_gux_epemgduxdx +P0_gux_epemgcuxcx +P0_gdx_epemguuxdx +P0_gdx_epemgsdxsx +P0_gu_taptamguuux +P0_gd_taptamgdddx +P0_gux_taptamguuxux +P0_gdx_taptamgddxdx +P0_gu_taptamguddx +P0_gu_taptamguccx +P0_gd_taptamgudux +P0_gd_taptamgdssx +P0_gux_taptamgduxdx +P0_gux_taptamgcuxcx +P0_gdx_taptamguuxdx +P0_gdx_taptamgsdxsx +P0_uu_epemgguu +P0_uux_epemgguux +P0_dd_epemggdd +P0_ddx_epemggddx +P0_uxux_epemgguxux +P0_dxdx_epemggdxdx +P0_ud_epemggud +P0_uc_epemgguc +P0_uux_epemggddx +P0_uux_epemggccx +P0_udx_epemggudx +P0_ucx_epemggucx +P0_dc_epemggdc +P0_ds_epemggds +P0_dux_epemggdux +P0_ddx_epemgguux +P0_ddx_epemggssx +P0_dsx_epemggdsx +P0_uxdx_epemgguxdx +P0_uxcx_epemgguxcx +P0_dxcx_epemggdxcx +P0_dxsx_epemggdxsx +P0_uu_taptamgguu +P0_uux_taptamgguux +P0_dd_taptamggdd +P0_ddx_taptamggddx +P0_uxux_taptamgguxux +P0_dxdx_taptamggdxdx +P0_ud_taptamggud +P0_uc_taptamgguc +P0_uux_taptamggddx +P0_uux_taptamggccx +P0_udx_taptamggudx +P0_ucx_taptamggucx +P0_dc_taptamggdc +P0_ds_taptamggds +P0_dux_taptamggdux +P0_ddx_taptamgguux +P0_ddx_taptamggssx +P0_dsx_taptamggdsx +P0_uxdx_taptamgguxdx +P0_uxcx_taptamgguxcx +P0_dxcx_taptamggdxcx +P0_dxsx_taptamggdxsx +P0_uu_epemuuuux +P0_uux_epemuuuxux +P0_dd_epemddddx +P0_ddx_epemdddxdx +P0_uxux_epemuuxuxux +P0_dxdx_epemddxdxdx +P0_uu_epemuuddx +P0_uu_epemuuccx +P0_ud_epemuudux +P0_ud_epemudddx +P0_uc_epemuucux +P0_uc_epemucccx +P0_uux_epemuduxdx +P0_uux_epemucuxcx +P0_uux_epemdddxdx +P0_uux_epemcccxcx +P0_udx_epemuuuxdx +P0_udx_epemuddxdx +P0_ucx_epemuuuxcx +P0_ucx_epemuccxcx +P0_dd_epemuddux +P0_dd_epemddssx +P0_dc_epemddcdx +P0_dc_epemdcccx +P0_ds_epemddsdx +P0_ds_epemdsssx +P0_dux_epemuduxux +P0_dux_epemdduxdx +P0_ddx_epemuuuxux +P0_ddx_epemuduxdx +P0_ddx_epemdcdxcx +P0_ddx_epemdsdxsx +P0_ddx_epemsssxsx +P0_dsx_epemdddxsx +P0_dsx_epemdssxsx +P0_ccx_epemucuxcx +P0_ccx_epemdcdxcx +P0_ssx_epemdsdxsx +P0_uxux_epemduxuxdx +P0_uxux_epemcuxuxcx +P0_uxdx_epemuuxuxdx +P0_uxdx_epemduxdxdx +P0_uxcx_epemuuxuxcx +P0_uxcx_epemcuxcxcx +P0_dxdx_epemuuxdxdx +P0_dxdx_epemsdxdxsx +P0_dxcx_epemddxdxcx +P0_dxcx_epemcdxcxcx +P0_dxsx_epemddxdxsx +P0_dxsx_epemsdxsxsx +P0_ud_epemudccx +P0_ud_epemudssx +P0_uc_epemudcdx +P0_uux_epemdcdxcx +P0_uux_epemdsdxsx +P0_uux_epemcscxsx +P0_udx_epemucdxcx +P0_udx_epemusdxsx +P0_ucx_epemuddxcx +P0_dc_epemudcux +P0_dc_epemdcssx +P0_ds_epemudsux +P0_ds_epemdsbbx +P0_dux_epemdcuxcx +P0_dux_epemdsuxsx +P0_ddx_epemucuxcx +P0_ddx_epemusuxsx +P0_ddx_epemsbsxbx +P0_dsx_epemuduxsx +P0_dsx_epemdbsxbx +P0_ccx_epemuduxdx +P0_ccx_epemdsdxsx +P0_ssx_epemuduxdx +P0_ssx_epemucuxcx +P0_ssx_epemdcdxcx +P0_ssx_epemdbdxbx +P0_bbx_epemdsdxsx +P0_uxdx_epemcuxdxcx +P0_uxdx_epemsuxdxsx +P0_uxcx_epemduxdxcx +P0_dxcx_epemuuxdxcx +P0_dxcx_epemsdxcxsx +P0_dxsx_epemuuxdxsx +P0_dxsx_epembdxsxbx +P0_uu_taptamuuuux +P0_uux_taptamuuuxux +P0_dd_taptamddddx +P0_ddx_taptamdddxdx +P0_uxux_taptamuuxuxux +P0_dxdx_taptamddxdxdx +P0_uu_taptamuuddx +P0_uu_taptamuuccx +P0_ud_taptamuudux +P0_ud_taptamudddx +P0_uc_taptamuucux +P0_uc_taptamucccx +P0_uux_taptamuduxdx +P0_uux_taptamucuxcx +P0_uux_taptamdddxdx +P0_uux_taptamcccxcx +P0_udx_taptamuuuxdx +P0_udx_taptamuddxdx +P0_ucx_taptamuuuxcx +P0_ucx_taptamuccxcx +P0_dd_taptamuddux +P0_dd_taptamddssx +P0_dc_taptamddcdx +P0_dc_taptamdcccx +P0_ds_taptamddsdx +P0_ds_taptamdsssx +P0_dux_taptamuduxux +P0_dux_taptamdduxdx +P0_ddx_taptamuuuxux +P0_ddx_taptamuduxdx +P0_ddx_taptamdcdxcx +P0_ddx_taptamdsdxsx +P0_ddx_taptamsssxsx +P0_dsx_taptamdddxsx +P0_dsx_taptamdssxsx +P0_ccx_taptamucuxcx +P0_ccx_taptamdcdxcx +P0_ssx_taptamdsdxsx +P0_uxux_taptamduxuxdx +P0_uxux_taptamcuxuxcx +P0_uxdx_taptamuuxuxdx +P0_uxdx_taptamduxdxdx +P0_uxcx_taptamuuxuxcx +P0_uxcx_taptamcuxcxcx +P0_dxdx_taptamuuxdxdx +P0_dxdx_taptamsdxdxsx +P0_dxcx_taptamddxdxcx +P0_dxcx_taptamcdxcxcx +P0_dxsx_taptamddxdxsx +P0_dxsx_taptamsdxsxsx +P0_ud_taptamudccx +P0_ud_taptamudssx +P0_uc_taptamudcdx +P0_uux_taptamdcdxcx +P0_uux_taptamdsdxsx +P0_uux_taptamcscxsx +P0_udx_taptamucdxcx +P0_udx_taptamusdxsx +P0_ucx_taptamuddxcx +P0_dc_taptamudcux +P0_dc_taptamdcssx +P0_ds_taptamudsux +P0_ds_taptamdsbbx +P0_dux_taptamdcuxcx +P0_dux_taptamdsuxsx +P0_ddx_taptamucuxcx +P0_ddx_taptamusuxsx +P0_ddx_taptamsbsxbx +P0_dsx_taptamuduxsx +P0_dsx_taptamdbsxbx +P0_ccx_taptamuduxdx +P0_ccx_taptamdsdxsx +P0_ssx_taptamuduxdx +P0_ssx_taptamucuxcx +P0_ssx_taptamdcdxcx +P0_ssx_taptamdbdxbx +P0_bbx_taptamdsdxsx +P0_uxdx_taptamcuxdxcx +P0_uxdx_taptamsuxdxsx +P0_uxcx_taptamduxdxcx +P0_dxcx_taptamuuxdxcx +P0_dxcx_taptamsdxcxsx +P0_dxsx_taptamuuxdxsx +P0_dxsx_taptambdxsxbx +__CUDACPP_DEBUG: GridPackCmd.launch (2) refine4grid +__CUDACPP_DEBUG: GridPackCmd.refine4grid starting +Using random number seed offset = 21 +INFO: Generating 500.0 unweighted events. +__CUDACPP_DEBUG: gen_ximprove.launch starting +sum of cpu time of last step: 3 days,23h23m17s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event starting +INFO: need to improve 477 channels +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dxdx_taptamuuxdxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dxdx_taptamuuxdxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dxdx_taptamuuxdxdx/ajob1' in 10.9704 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dxdx_taptamuuxdxdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dxdx_taptamuuxdxdx/G21/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 10.9324s + [COUNTERS] Fortran Overhead ( 0 ) : 3.8328s + [COUNTERS] CudaCpp MEs ( 2 ) : 7.0634s for 16384 events => throughput is 2.32E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0362s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dxdx_taptamuuxdxdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamguuxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamguuxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamguuxdx/ajob1' in 44.6916 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamguuxdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamguuxdx/G43/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 21.9451s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0129s + [COUNTERS] CudaCpp MEs ( 2 ) : 19.8610s for 16384 events => throughput is 8.25E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0712s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamguuxdx/G335/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 22.6968s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6433s + [COUNTERS] CudaCpp MEs ( 2 ) : 19.9775s for 16384 events => throughput is 8.20E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0759s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamguuxdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemuuuxux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemuuuxux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemuuuxux/ajob1' in 34.9613 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemuuuxux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemuuuxux/G689/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 34.9046s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5129s + [COUNTERS] CudaCpp MEs ( 2 ) : 32.2627s for 16384 events => throughput is 5.08E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1290s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemuuuxux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamgguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamgguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamgguux/ajob1' in 1029.3499 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamgguux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamgguux/G596/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 113.1870s + [COUNTERS] Fortran Overhead ( 0 ) : 3.3811s + [COUNTERS] CudaCpp MEs ( 2 ) : 109.6334s for 16384 events => throughput is 1.49E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1725s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamgguux/G304/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 112.9484s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9365s + [COUNTERS] CudaCpp MEs ( 2 ) : 109.8373s for 16384 events => throughput is 1.49E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1746s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamgguux/G128/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 112.6810s + [COUNTERS] Fortran Overhead ( 0 ) : 3.0365s + [COUNTERS] CudaCpp MEs ( 2 ) : 109.4711s for 16384 events => throughput is 1.50E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1734s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamgguux/G115/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 116.5243s + [COUNTERS] Fortran Overhead ( 0 ) : 6.7203s + [COUNTERS] CudaCpp MEs ( 2 ) : 109.6280s for 16384 events => throughput is 1.49E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1760s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamgguux/G599/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 113.3586s + [COUNTERS] Fortran Overhead ( 0 ) : 3.8187s + [COUNTERS] CudaCpp MEs ( 2 ) : 109.3644s for 16384 events => throughput is 1.50E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1755s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamgguux/G351/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 115.5941s + [COUNTERS] Fortran Overhead ( 0 ) : 6.1227s + [COUNTERS] CudaCpp MEs ( 2 ) : 109.2979s for 16384 events => throughput is 1.50E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1735s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamgguux/G111/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 111.9865s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4103s + [COUNTERS] CudaCpp MEs ( 2 ) : 109.4004s for 16384 events => throughput is 1.50E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1759s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamgguux/G354/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 115.9697s + [COUNTERS] Fortran Overhead ( 0 ) : 6.5853s + [COUNTERS] CudaCpp MEs ( 2 ) : 109.2131s for 16384 events => throughput is 1.50E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1713s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamgguux/G97/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 116.8922s + [COUNTERS] Fortran Overhead ( 0 ) : 7.2624s + [COUNTERS] CudaCpp MEs ( 2 ) : 109.4567s for 16384 events => throughput is 1.50E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1730s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamgguux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uxdx_epemsuxdxsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uxdx_epemsuxdxsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uxdx_epemsuxdxsx/ajob1' in 3.6317 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uxdx_epemsuxdxsx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uxdx_epemsuxdxsx/G74/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.5859s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2062s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.3578s for 16384 events => throughput is 1.21E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0219s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uxdx_epemsuxdxsx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemdddxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemdddxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemdddxdx/ajob1' in 70.2598 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemdddxdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemdddxdx/G513/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 34.9374s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4640s + [COUNTERS] CudaCpp MEs ( 2 ) : 32.3435s for 16384 events => throughput is 5.07E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1299s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemdddxdx/G180/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 35.2697s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7986s + [COUNTERS] CudaCpp MEs ( 2 ) : 32.3419s for 16384 events => throughput is 5.07E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1292s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemdddxdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgddxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgddxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgddxdx/ajob1' in 35.4352 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgddxdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgddxdx/G515/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 35.3816s + [COUNTERS] Fortran Overhead ( 0 ) : 3.0966s + [COUNTERS] CudaCpp MEs ( 2 ) : 32.1540s for 16384 events => throughput is 5.10E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1310s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgddxdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguddx/ajob1' in 205.6575 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguddx/G104/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 21.8349s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6110s + [COUNTERS] CudaCpp MEs ( 2 ) : 20.1521s for 16384 events => throughput is 8.13E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0718s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguddx/G120/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 22.3934s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1539s + [COUNTERS] CudaCpp MEs ( 2 ) : 20.1683s for 16384 events => throughput is 8.12E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0712s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguddx/G305/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 26.1143s + [COUNTERS] Fortran Overhead ( 0 ) : 5.8830s + [COUNTERS] CudaCpp MEs ( 2 ) : 20.1591s for 16384 events => throughput is 8.13E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0721s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguddx/G101/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 21.6326s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4621s + [COUNTERS] CudaCpp MEs ( 2 ) : 20.0993s for 16384 events => throughput is 8.15E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0712s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguddx/G27/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 22.0642s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9075s + [COUNTERS] CudaCpp MEs ( 2 ) : 20.0851s for 16384 events => throughput is 8.16E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0716s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguddx/G273/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 22.3543s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1592s + [COUNTERS] CudaCpp MEs ( 2 ) : 20.1238s for 16384 events => throughput is 8.14E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0714s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguddx/G349/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 21.5962s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3914s + [COUNTERS] CudaCpp MEs ( 2 ) : 20.1332s for 16384 events => throughput is 8.14E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0717s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguddx/G379/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 22.0984s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7746s + [COUNTERS] CudaCpp MEs ( 2 ) : 20.2534s for 16384 events => throughput is 8.09E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0704s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguddx/G256/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 25.3693s + [COUNTERS] Fortran Overhead ( 0 ) : 5.2186s + [COUNTERS] CudaCpp MEs ( 2 ) : 20.0789s for 16384 events => throughput is 8.16E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0719s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemguuxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemguuxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemguuxdx/ajob1' in 43.1843 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemguuxdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemguuxdx/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 15.8978s + [COUNTERS] Fortran Overhead ( 0 ) : 6.0416s + [COUNTERS] CudaCpp MEs ( 2 ) : 9.7863s for 16384 events => throughput is 1.67E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0698s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemguuxdx/G68/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 15.1807s + [COUNTERS] Fortran Overhead ( 0 ) : 5.3180s + [COUNTERS] CudaCpp MEs ( 2 ) : 9.7907s for 16384 events => throughput is 1.67E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0721s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemguuxdx/G355/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 12.0169s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0991s + [COUNTERS] CudaCpp MEs ( 2 ) : 9.8479s for 16384 events => throughput is 1.66E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0699s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemguuxdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/ajob1' in 2585.0681 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G14/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 111.5237s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0469s + [COUNTERS] CudaCpp MEs ( 2 ) : 109.3040s for 16384 events => throughput is 1.50E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1728s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G56/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 111.7440s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4978s + [COUNTERS] CudaCpp MEs ( 2 ) : 109.0749s for 16384 events => throughput is 1.50E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1713s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G517/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 115.8780s + [COUNTERS] Fortran Overhead ( 0 ) : 6.1154s + [COUNTERS] CudaCpp MEs ( 2 ) : 109.5886s for 16384 events => throughput is 1.50E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1740s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G52/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 112.2301s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7598s + [COUNTERS] CudaCpp MEs ( 2 ) : 109.3004s for 16384 events => throughput is 1.50E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1698s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G304/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 112.2316s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7189s + [COUNTERS] CudaCpp MEs ( 2 ) : 109.3405s for 16384 events => throughput is 1.50E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1722s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G3/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 115.1895s + [COUNTERS] Fortran Overhead ( 0 ) : 4.8730s + [COUNTERS] CudaCpp MEs ( 2 ) : 110.1448s for 16384 events => throughput is 1.49E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1718s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G23/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 111.0644s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7259s + [COUNTERS] CudaCpp MEs ( 2 ) : 109.1649s for 16384 events => throughput is 1.50E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1736s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G33/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 111.1881s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1604s + [COUNTERS] CudaCpp MEs ( 2 ) : 108.8538s for 16384 events => throughput is 1.51E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1738s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G53/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 116.1894s + [COUNTERS] Fortran Overhead ( 0 ) : 6.7278s + [COUNTERS] CudaCpp MEs ( 2 ) : 109.2823s for 16384 events => throughput is 1.50E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1793s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G495/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 110.7916s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6049s + [COUNTERS] CudaCpp MEs ( 2 ) : 109.0122s for 16384 events => throughput is 1.50E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1744s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G68/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 111.2536s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1843s + [COUNTERS] CudaCpp MEs ( 2 ) : 108.8973s for 16384 events => throughput is 1.50E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1720s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G49/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 111.1795s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0956s + [COUNTERS] CudaCpp MEs ( 2 ) : 108.9102s for 16384 events => throughput is 1.50E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1737s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G524/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 111.3965s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1680s + [COUNTERS] CudaCpp MEs ( 2 ) : 109.0564s for 16384 events => throughput is 1.50E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1721s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G65/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 111.3139s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0771s + [COUNTERS] CudaCpp MEs ( 2 ) : 109.0630s for 16384 events => throughput is 1.50E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1738s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G303/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 112.7862s + [COUNTERS] Fortran Overhead ( 0 ) : 3.0207s + [COUNTERS] CudaCpp MEs ( 2 ) : 109.5937s for 16384 events => throughput is 1.49E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1718s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G55/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 111.8401s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6944s + [COUNTERS] CudaCpp MEs ( 2 ) : 108.9722s for 16384 events => throughput is 1.50E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1735s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G34/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 111.4118s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8621s + [COUNTERS] CudaCpp MEs ( 2 ) : 109.3757s for 16384 events => throughput is 1.50E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1740s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G509/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 116.9382s + [COUNTERS] Fortran Overhead ( 0 ) : 7.6687s + [COUNTERS] CudaCpp MEs ( 2 ) : 109.0968s for 16384 events => throughput is 1.50E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1727s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G512/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 111.2290s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8236s + [COUNTERS] CudaCpp MEs ( 2 ) : 109.2314s for 16384 events => throughput is 1.50E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1740s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G511/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 111.9249s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4237s + [COUNTERS] CudaCpp MEs ( 2 ) : 109.3298s for 16384 events => throughput is 1.50E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1714s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 112.2374s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8648s + [COUNTERS] CudaCpp MEs ( 2 ) : 109.1992s for 16384 events => throughput is 1.50E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1734s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G63/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 111.3877s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1174s + [COUNTERS] CudaCpp MEs ( 2 ) : 109.0935s for 16384 events => throughput is 1.50E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1768s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G48/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 111.6629s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3526s + [COUNTERS] CudaCpp MEs ( 2 ) : 109.1382s for 16384 events => throughput is 1.50E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1721s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/ajob1' in 1378.9537 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G122/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.2993s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6762s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.4516s for 16384 events => throughput is 3.01E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1715s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G108/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.1101s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3351s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.6039s for 16384 events => throughput is 3.00E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1711s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G105/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.1607s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4635s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.5257s for 16384 events => throughput is 3.00E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1715s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G304/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.8742s + [COUNTERS] Fortran Overhead ( 0 ) : 3.0434s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.6596s for 16384 events => throughput is 3.00E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1712s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G301/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 62.5129s + [COUNTERS] Fortran Overhead ( 0 ) : 7.6249s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.7168s for 16384 events => throughput is 2.99E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1712s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G128/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.7577s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1403s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.4465s for 16384 events => throughput is 3.01E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1709s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G100/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 62.2320s + [COUNTERS] Fortran Overhead ( 0 ) : 7.6029s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.4528s for 16384 events => throughput is 3.01E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1764s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G120/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.0720s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4401s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.4582s for 16384 events => throughput is 3.01E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1737s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G302/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 65.4565s + [COUNTERS] Fortran Overhead ( 0 ) : 10.6457s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.6368s for 16384 events => throughput is 3.00E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1740s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G299/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.6633s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6647s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.8267s for 16384 events => throughput is 2.99E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1719s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G169/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 60.6858s + [COUNTERS] Fortran Overhead ( 0 ) : 6.0142s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.4971s for 16384 events => throughput is 3.01E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1745s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G333/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 65.6470s + [COUNTERS] Fortran Overhead ( 0 ) : 10.9535s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.5217s for 16384 events => throughput is 3.01E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1717s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G594/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 63.9530s + [COUNTERS] Fortran Overhead ( 0 ) : 9.2130s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.5659s for 16384 events => throughput is 3.00E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1740s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G678/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 61.0122s + [COUNTERS] Fortran Overhead ( 0 ) : 6.3100s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.5314s for 16384 events => throughput is 3.00E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1708s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G127/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.4677s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7038s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.5908s for 16384 events => throughput is 3.00E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1731s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G126/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 64.2416s + [COUNTERS] Fortran Overhead ( 0 ) : 9.4666s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.6020s for 16384 events => throughput is 3.00E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1730s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G598/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.1698s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1677s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.8296s for 16384 events => throughput is 2.99E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1725s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G512/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.0510s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2256s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.6436s for 16384 events => throughput is 3.00E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1818s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G103/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 63.7746s + [COUNTERS] Fortran Overhead ( 0 ) : 9.0038s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.5979s for 16384 events => throughput is 3.00E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1729s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G107/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.0146s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3433s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.4983s for 16384 events => throughput is 3.01E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1730s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G358/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 58.4863s + [COUNTERS] Fortran Overhead ( 0 ) : 3.7019s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.6078s for 16384 events => throughput is 3.00E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1765s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G97/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 62.1097s + [COUNTERS] Fortran Overhead ( 0 ) : 7.4334s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.5061s for 16384 events => throughput is 3.01E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1702s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G179/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.7468s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7505s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.8243s for 16384 events => throughput is 2.99E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1720s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_epemudcdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_epemudcdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_epemudcdx/ajob1' in 3.1684 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_epemudcdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_epemudcdx/G42/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.1196s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7488s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.3495s for 16384 events => throughput is 1.21E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0213s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_epemudcdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ds_taptamggds/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ds_taptamggds/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ds_taptamggds/ajob1' in 46.5075 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ds_taptamggds/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ds_taptamggds/G17/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 24.7066s + [COUNTERS] Fortran Overhead ( 0 ) : 4.6582s + [COUNTERS] CudaCpp MEs ( 2 ) : 19.9762s for 16384 events => throughput is 8.20E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0722s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ds_taptamggds/G110/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 21.7312s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6724s + [COUNTERS] CudaCpp MEs ( 2 ) : 19.9870s for 16384 events => throughput is 8.20E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0718s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ds_taptamggds/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_epemggdux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_epemggdux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_epemggdux/ajob1' in 59.1208 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_epemggdux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_epemggdux/G208/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 12.3042s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0073s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.2249s for 16384 events => throughput is 1.60E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0719s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_epemggdux/G192/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 11.7212s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3891s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.2581s for 16384 events => throughput is 1.60E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0740s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_epemggdux/G380/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 11.9755s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6186s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.2843s for 16384 events => throughput is 1.59E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0727s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_epemggdux/G127/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 11.5181s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2553s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.1896s for 16384 events => throughput is 1.61E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0732s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_epemggdux/G223/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 11.4750s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2046s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.1988s for 16384 events => throughput is 1.61E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0716s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_epemggdux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_taptamgguc/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_taptamgguc/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_taptamgguc/ajob1' in 71.5676 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_taptamgguc/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_taptamgguc/G20/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 24.5258s + [COUNTERS] Fortran Overhead ( 0 ) : 4.4029s + [COUNTERS] CudaCpp MEs ( 2 ) : 20.0509s for 16384 events => throughput is 8.17E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0721s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_taptamgguc/G110/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 21.4166s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3238s + [COUNTERS] CudaCpp MEs ( 2 ) : 20.0212s for 16384 events => throughput is 8.18E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0716s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_taptamgguc/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 25.5578s + [COUNTERS] Fortran Overhead ( 0 ) : 5.4458s + [COUNTERS] CudaCpp MEs ( 2 ) : 20.0403s for 16384 events => throughput is 8.18E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0717s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_taptamgguc/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uxcx_taptamgguxcx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uxcx_taptamgguxcx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uxcx_taptamgguxcx/ajob1' in 23.3966 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uxcx_taptamgguxcx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uxcx_taptamgguxcx/G187/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 23.3411s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5713s + [COUNTERS] CudaCpp MEs ( 2 ) : 20.6919s for 16384 events => throughput is 7.92E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0779s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uxcx_taptamgguxcx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/ajob1' in 1515.7356 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G499/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.0294s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2024s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.6558s for 16384 events => throughput is 3.00E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1712s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G56/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.7800s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9250s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.6820s for 16384 events => throughput is 3.00E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1731s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G39/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.7258s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7029s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.8504s for 16384 events => throughput is 2.99E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1725s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G64/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.4297s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5422s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.7108s for 16384 events => throughput is 2.99E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1768s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G13/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 56.5343s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6018s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.7596s for 16384 events => throughput is 2.99E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1729s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G33/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 56.6019s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7694s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.6591s for 16384 events => throughput is 3.00E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1734s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G43/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 61.9257s + [COUNTERS] Fortran Overhead ( 0 ) : 6.9821s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.7718s for 16384 events => throughput is 2.99E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1718s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G68/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.9766s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9177s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.8864s for 16384 events => throughput is 2.99E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1725s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G49/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 56.8554s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1106s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.5660s for 16384 events => throughput is 3.00E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1787s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G25/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 60.5356s + [COUNTERS] Fortran Overhead ( 0 ) : 5.7358s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.6257s for 16384 events => throughput is 3.00E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1741s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G523/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.6685s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7325s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.7605s for 16384 events => throughput is 2.99E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1755s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G395/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.0896s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3828s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.5348s for 16384 events => throughput is 3.00E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1720s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G524/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.1808s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0121s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.9854s for 16384 events => throughput is 2.98E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1833s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G65/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 56.9661s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0220s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.7723s for 16384 events => throughput is 2.99E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1717s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G609/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.6360s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9131s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.5523s for 16384 events => throughput is 3.00E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1706s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G611/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 58.0362s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2401s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.6246s for 16384 events => throughput is 3.00E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1715s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G1/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 59.8218s + [COUNTERS] Fortran Overhead ( 0 ) : 4.6169s + [COUNTERS] CudaCpp MEs ( 2 ) : 55.0330s for 16384 events => throughput is 2.98E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1719s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G55/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.5315s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5801s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.7786s for 16384 events => throughput is 2.99E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1728s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G34/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.0161s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1490s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.6932s for 16384 events => throughput is 3.00E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1738s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G26/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 61.1637s + [COUNTERS] Fortran Overhead ( 0 ) : 6.2764s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.7136s for 16384 events => throughput is 2.99E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1737s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G58/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 60.7726s + [COUNTERS] Fortran Overhead ( 0 ) : 5.6730s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.9241s for 16384 events => throughput is 2.98E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1755s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G2/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 60.4493s + [COUNTERS] Fortran Overhead ( 0 ) : 5.1082s + [COUNTERS] CudaCpp MEs ( 2 ) : 55.1664s for 16384 events => throughput is 2.97E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1746s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G512/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.0362s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0981s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.7655s for 16384 events => throughput is 2.99E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1725s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G51/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.7238s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6861s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.8618s for 16384 events => throughput is 2.99E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1759s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G29/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 61.7863s + [COUNTERS] Fortran Overhead ( 0 ) : 6.6388s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.9749s for 16384 events => throughput is 2.98E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1726s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G48/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 56.9302s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0483s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.7071s for 16384 events => throughput is 2.99E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1747s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/ajob1' in 162.1701 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/G140/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 11.9136s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8326s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.0075s for 16384 events => throughput is 1.64E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0735s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/G114/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 16.5633s + [COUNTERS] Fortran Overhead ( 0 ) : 6.4491s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.0415s for 16384 events => throughput is 1.63E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0727s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/G49/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 17.3760s + [COUNTERS] Fortran Overhead ( 0 ) : 7.2756s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.0290s for 16384 events => throughput is 1.63E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0715s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/G120/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 12.0820s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0012s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.0096s for 16384 events => throughput is 1.64E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0712s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/G262/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 11.8634s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7582s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.0343s for 16384 events => throughput is 1.63E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0709s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/G27/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 11.9412s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8688s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.0008s for 16384 events => throughput is 1.64E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0717s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/G44/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 11.6981s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5761s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.0507s for 16384 events => throughput is 1.63E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0714s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/G96/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.4801s + [COUNTERS] Fortran Overhead ( 0 ) : 3.4075s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.0005s for 16384 events => throughput is 1.64E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0721s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/G261/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 12.1483s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0877s + [COUNTERS] CudaCpp MEs ( 2 ) : 9.9887s for 16384 events => throughput is 1.64E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0719s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/G37/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.9745s + [COUNTERS] Fortran Overhead ( 0 ) : 4.8858s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.0176s for 16384 events => throughput is 1.64E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0710s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/G28/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 11.8592s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7859s + [COUNTERS] CudaCpp MEs ( 2 ) : 9.9999s for 16384 events => throughput is 1.64E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0734s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 15.9954s + [COUNTERS] Fortran Overhead ( 0 ) : 5.9256s + [COUNTERS] CudaCpp MEs ( 2 ) : 9.9968s for 16384 events => throughput is 1.64E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0730s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_epemuduxsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_epemuduxsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_epemuduxsx/ajob1' in 3.9712 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_epemuduxsx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_epemuduxsx/G42/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.9404s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5878s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.3315s for 16384 events => throughput is 1.23E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0211s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_epemuduxsx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dd_epemggdd/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dd_epemggdd/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dd_epemggdd/ajob1' in 101.5867 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dd_epemggdd/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dd_epemggdd/G467/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 33.8974s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8060s + [COUNTERS] CudaCpp MEs ( 2 ) : 31.9635s for 16384 events => throughput is 5.13E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1279s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dd_epemggdd/G351/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 33.3270s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2796s + [COUNTERS] CudaCpp MEs ( 2 ) : 31.9189s for 16384 events => throughput is 5.13E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1285s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dd_epemggdd/G44/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 34.2719s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2207s + [COUNTERS] CudaCpp MEs ( 2 ) : 31.9230s for 16384 events => throughput is 5.13E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1282s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dd_epemggdd/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamusdxsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamusdxsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamusdxsx/ajob1' in 3.8786 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamusdxsx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamusdxsx/G92/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.8316s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2042s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.6057s for 16384 events => throughput is 6.29E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0216s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamusdxsx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemggddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemggddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemggddx/ajob1' in 214.9989 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemggddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemggddx/G175/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 34.9884s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0832s + [COUNTERS] CudaCpp MEs ( 2 ) : 32.7763s for 16384 events => throughput is 5.00E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1289s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemggddx/G303/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 35.2885s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7163s + [COUNTERS] CudaCpp MEs ( 2 ) : 32.4348s for 16384 events => throughput is 5.05E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1373s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemggddx/G111/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 34.8587s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7752s + [COUNTERS] CudaCpp MEs ( 2 ) : 32.9511s for 16384 events => throughput is 4.97E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1324s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemggddx/G219/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 35.5155s + [COUNTERS] Fortran Overhead ( 0 ) : 3.0340s + [COUNTERS] CudaCpp MEs ( 2 ) : 32.3518s for 16384 events => throughput is 5.06E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1297s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemggddx/G318/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 38.5638s + [COUNTERS] Fortran Overhead ( 0 ) : 5.9645s + [COUNTERS] CudaCpp MEs ( 2 ) : 32.4695s for 16384 events => throughput is 5.05E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1298s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemggddx/G143/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 35.6591s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0135s + [COUNTERS] CudaCpp MEs ( 2 ) : 33.5086s for 16384 events => throughput is 4.89E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1370s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemggddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemguuxux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemguuxux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemguuxux/ajob1' in 39.6795 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemguuxux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemguuxux/G615/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 39.6258s + [COUNTERS] Fortran Overhead ( 0 ) : 7.0533s + [COUNTERS] CudaCpp MEs ( 2 ) : 32.4422s for 16384 events => throughput is 5.05E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1302s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemguuxux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamgggg/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamgggg/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamgggg/ajob1' in 112.6689 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamgggg/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamgggg/G66/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 112.6164s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6463s + [COUNTERS] CudaCpp MEs ( 2 ) : 109.7965s for 16384 events => throughput is 1.49E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1736s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamgggg/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgduxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgduxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgduxdx/ajob1' in 22.3181 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgduxdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgduxdx/G44/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 22.2874s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1192s + [COUNTERS] CudaCpp MEs ( 2 ) : 20.0923s for 16384 events => throughput is 8.15E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0759s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgduxdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgudux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgudux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgudux/ajob1' in 117.3252 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgudux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgudux/G39/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 16.3579s + [COUNTERS] Fortran Overhead ( 0 ) : 6.1266s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.1574s for 16384 events => throughput is 1.61E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0739s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgudux/G366/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.9763s + [COUNTERS] Fortran Overhead ( 0 ) : 4.7032s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.2004s for 16384 events => throughput is 1.61E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0726s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgudux/G259/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 12.5515s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3207s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.1584s for 16384 events => throughput is 1.61E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0725s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgudux/G28/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 12.3325s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0922s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.1680s for 16384 events => throughput is 1.61E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0724s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgudux/G32/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 12.7519s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4483s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.2326s for 16384 events => throughput is 1.60E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0710s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgudux/G61/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 18.6451s + [COUNTERS] Fortran Overhead ( 0 ) : 8.3436s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.2294s for 16384 events => throughput is 1.60E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0721s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgudux/G267/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 15.9207s + [COUNTERS] Fortran Overhead ( 0 ) : 5.6720s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.1757s for 16384 events => throughput is 1.61E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0730s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgudux/G144/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.6257s + [COUNTERS] Fortran Overhead ( 0 ) : 3.3893s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.1648s for 16384 events => throughput is 1.61E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0715s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgudux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemuudux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemuudux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemuudux/ajob1' in 38.1459 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemuudux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemuudux/G104/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 38.1156s + [COUNTERS] Fortran Overhead ( 0 ) : 19.8848s + [COUNTERS] CudaCpp MEs ( 2 ) : 18.1944s for 81920 events => throughput is 4.50E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0363s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemuudux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamggddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamggddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamggddx/ajob1' in 900.3714 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamggddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamggddx/G108/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 111.8662s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2124s + [COUNTERS] CudaCpp MEs ( 2 ) : 109.4813s for 16384 events => throughput is 1.50E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1725s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamggddx/G304/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 112.3292s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7364s + [COUNTERS] CudaCpp MEs ( 2 ) : 109.4199s for 16384 events => throughput is 1.50E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1728s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamggddx/G128/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 112.1908s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8438s + [COUNTERS] CudaCpp MEs ( 2 ) : 109.1753s for 16384 events => throughput is 1.50E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1717s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamggddx/G360/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 111.8015s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3739s + [COUNTERS] CudaCpp MEs ( 2 ) : 109.2564s for 16384 events => throughput is 1.50E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1713s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamggddx/G169/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 115.6646s + [COUNTERS] Fortran Overhead ( 0 ) : 6.1247s + [COUNTERS] CudaCpp MEs ( 2 ) : 109.3603s for 16384 events => throughput is 1.50E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1796s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamggddx/G176/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 111.6062s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0200s + [COUNTERS] CudaCpp MEs ( 2 ) : 109.4097s for 16384 events => throughput is 1.50E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1764s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamggddx/G303/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 112.3126s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8315s + [COUNTERS] CudaCpp MEs ( 2 ) : 109.3084s for 16384 events => throughput is 1.50E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1727s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamggddx/G513/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 112.4331s + [COUNTERS] Fortran Overhead ( 0 ) : 3.0815s + [COUNTERS] CudaCpp MEs ( 2 ) : 109.1786s for 16384 events => throughput is 1.50E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1730s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamggddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguccx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguccx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguccx/ajob1' in 65.1228 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguccx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguccx/G283/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 12.2733s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9971s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.2005s for 16384 events => throughput is 1.61E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0757s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguccx/G47/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 12.1719s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9627s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.1377s for 16384 events => throughput is 1.62E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0715s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguccx/G35/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 16.9302s + [COUNTERS] Fortran Overhead ( 0 ) : 6.6917s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.1662s for 16384 events => throughput is 1.61E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0722s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguccx/G373/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 11.8885s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7017s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.1156s for 16384 events => throughput is 1.62E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0712s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguccx/G249/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 11.7351s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5270s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.1369s for 16384 events => throughput is 1.62E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0711s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguccx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/ajob1' in 1850.9304 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G14/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.1741s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0057s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.9944s for 16384 events => throughput is 2.98E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1740s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G499/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.2910s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9531s + [COUNTERS] CudaCpp MEs ( 2 ) : 55.1631s for 16384 events => throughput is 2.97E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1748s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G56/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.9893s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6895s + [COUNTERS] CudaCpp MEs ( 2 ) : 55.1249s for 16384 events => throughput is 2.97E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1750s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G39/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.3769s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3003s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.9027s for 16384 events => throughput is 2.98E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1739s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G52/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 58.2951s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8154s + [COUNTERS] CudaCpp MEs ( 2 ) : 55.3071s for 16384 events => throughput is 2.96E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1726s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G304/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.7681s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6088s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.9862s for 16384 events => throughput is 2.98E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1731s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G64/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.3997s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3314s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.8918s for 16384 events => throughput is 2.98E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1766s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G3/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 60.0968s + [COUNTERS] Fortran Overhead ( 0 ) : 4.9463s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.9788s for 16384 events => throughput is 2.98E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1716s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G13/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 56.8412s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6517s + [COUNTERS] CudaCpp MEs ( 2 ) : 55.0171s for 16384 events => throughput is 2.98E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1723s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G33/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.1728s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1614s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.8402s for 16384 events => throughput is 2.99E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1713s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G484/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.7302s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5649s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.9930s for 16384 events => throughput is 2.98E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1723s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G68/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.2543s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1713s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.9111s for 16384 events => throughput is 2.98E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1719s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G49/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.0518s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7697s + [COUNTERS] CudaCpp MEs ( 2 ) : 55.1101s for 16384 events => throughput is 2.97E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1720s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G40/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 58.1363s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8163s + [COUNTERS] CudaCpp MEs ( 2 ) : 55.1437s for 16384 events => throughput is 2.97E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1763s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G523/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.2304s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1217s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.9356s for 16384 events => throughput is 2.98E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1731s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 56.7728s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5460s + [COUNTERS] CudaCpp MEs ( 2 ) : 55.0537s for 16384 events => throughput is 2.98E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1731s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G524/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.3237s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1071s + [COUNTERS] CudaCpp MEs ( 2 ) : 55.0446s for 16384 events => throughput is 2.98E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1720s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G65/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.2130s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1406s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.8999s for 16384 events => throughput is 2.98E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1725s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G44/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 61.0679s + [COUNTERS] Fortran Overhead ( 0 ) : 5.7543s + [COUNTERS] CudaCpp MEs ( 2 ) : 55.1425s for 16384 events => throughput is 2.97E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1712s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 56.7703s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5387s + [COUNTERS] CudaCpp MEs ( 2 ) : 55.0587s for 16384 events => throughput is 2.98E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1728s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G303/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 58.0664s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9725s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.9224s for 16384 events => throughput is 2.98E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1716s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G528/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 58.3342s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1192s + [COUNTERS] CudaCpp MEs ( 2 ) : 55.0423s for 16384 events => throughput is 2.98E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1727s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G55/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.9846s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9430s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.8668s for 16384 events => throughput is 2.99E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1748s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G298/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 58.2064s + [COUNTERS] Fortran Overhead ( 0 ) : 3.0106s + [COUNTERS] CudaCpp MEs ( 2 ) : 55.0239s for 16384 events => throughput is 2.98E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1720s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G34/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 56.9307s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7118s + [COUNTERS] CudaCpp MEs ( 2 ) : 55.0447s for 16384 events => throughput is 2.98E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1742s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G54/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 62.2785s + [COUNTERS] Fortran Overhead ( 0 ) : 7.0493s + [COUNTERS] CudaCpp MEs ( 2 ) : 55.0557s for 16384 events => throughput is 2.98E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1734s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G67/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.4746s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2839s + [COUNTERS] CudaCpp MEs ( 2 ) : 55.0176s for 16384 events => throughput is 2.98E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1731s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G512/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.0931s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0175s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.9007s for 16384 events => throughput is 2.98E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1749s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G511/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.6470s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6446s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.8297s for 16384 events => throughput is 2.99E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1728s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.5816s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4926s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.9157s for 16384 events => throughput is 2.98E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1733s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G21/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.1263s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7820s + [COUNTERS] CudaCpp MEs ( 2 ) : 55.1705s for 16384 events => throughput is 2.97E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1738s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G48/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.6433s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5138s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.9555s for 16384 events => throughput is 2.98E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1740s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_taptamggdux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_taptamggdux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_taptamggdux/ajob1' in 90.3961 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_taptamggdux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_taptamggdux/G68/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 24.1210s + [COUNTERS] Fortran Overhead ( 0 ) : 3.8673s + [COUNTERS] CudaCpp MEs ( 2 ) : 20.1812s for 16384 events => throughput is 8.12E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0724s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_taptamggdux/G167/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 22.0948s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7807s + [COUNTERS] CudaCpp MEs ( 2 ) : 20.2415s for 16384 events => throughput is 8.09E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0726s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_taptamggdux/G92/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 22.1846s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7437s + [COUNTERS] CudaCpp MEs ( 2 ) : 20.3678s for 16384 events => throughput is 8.04E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0730s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_taptamggdux/G60/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 21.9114s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4808s + [COUNTERS] CudaCpp MEs ( 2 ) : 20.3579s for 16384 events => throughput is 8.05E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0726s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_taptamggdux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/ajob1' in 1189.8108 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G56/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.7726s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8588s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.7390s for 16384 events => throughput is 2.99E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1747s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G39/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.8020s + [COUNTERS] Fortran Overhead ( 0 ) : 3.0457s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.5835s for 16384 events => throughput is 3.00E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1728s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G517/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 62.8685s + [COUNTERS] Fortran Overhead ( 0 ) : 8.0886s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.6084s for 16384 events => throughput is 3.00E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1716s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G497/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 56.9053s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9565s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.7776s for 16384 events => throughput is 2.99E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1712s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G13/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 56.4878s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7082s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.6077s for 16384 events => throughput is 3.00E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1720s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G484/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 61.3577s + [COUNTERS] Fortran Overhead ( 0 ) : 6.6572s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.5271s for 16384 events => throughput is 3.00E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1734s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G68/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.8586s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1250s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.5615s for 16384 events => throughput is 3.00E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1721s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G49/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.0295s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3500s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.5075s for 16384 events => throughput is 3.01E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1720s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G523/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.2543s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5206s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.5616s for 16384 events => throughput is 3.00E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1720s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G518/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 60.6709s + [COUNTERS] Fortran Overhead ( 0 ) : 5.8633s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.6347s for 16384 events => throughput is 3.00E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1729s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G42/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 62.7153s + [COUNTERS] Fortran Overhead ( 0 ) : 8.1026s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.4411s for 16384 events => throughput is 3.01E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1716s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G524/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 56.7185s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9599s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.5811s for 16384 events => throughput is 3.00E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1774s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G610/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.7801s + [COUNTERS] Fortran Overhead ( 0 ) : 3.0994s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.5065s for 16384 events => throughput is 3.01E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1743s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G26/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 61.1833s + [COUNTERS] Fortran Overhead ( 0 ) : 5.9907s + [COUNTERS] CudaCpp MEs ( 2 ) : 55.0194s for 16384 events => throughput is 2.98E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1732s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G58/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 62.8125s + [COUNTERS] Fortran Overhead ( 0 ) : 7.2406s + [COUNTERS] CudaCpp MEs ( 2 ) : 55.4006s for 16384 events => throughput is 2.96E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1713s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G511/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.0688s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1239s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.7702s for 16384 events => throughput is 2.99E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1747s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.6212s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8454s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.6026s for 16384 events => throughput is 3.00E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1733s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G32/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 62.8505s + [COUNTERS] Fortran Overhead ( 0 ) : 7.9365s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.7361s for 16384 events => throughput is 2.99E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1779s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G61/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 60.7656s + [COUNTERS] Fortran Overhead ( 0 ) : 6.0887s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.5048s for 16384 events => throughput is 3.01E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1721s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G390/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 63.8699s + [COUNTERS] Fortran Overhead ( 0 ) : 9.1919s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.5067s for 16384 events => throughput is 3.01E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1713s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_epemgguc/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_epemgguc/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_epemgguc/ajob1' in 41.5786 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_epemgguc/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_epemgguc/G164/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.3668s + [COUNTERS] Fortran Overhead ( 0 ) : 4.4202s + [COUNTERS] CudaCpp MEs ( 2 ) : 9.8749s for 16384 events => throughput is 1.66E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0718s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_epemgguc/G24/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.0058s + [COUNTERS] Fortran Overhead ( 0 ) : 3.9852s + [COUNTERS] CudaCpp MEs ( 2 ) : 9.9487s for 16384 events => throughput is 1.65E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0719s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_epemgguc/G162/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.1218s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1734s + [COUNTERS] CudaCpp MEs ( 2 ) : 9.8762s for 16384 events => throughput is 1.66E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0722s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_epemgguc/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/ajob1' in 1638.2619 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G336/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 58.5615s + [COUNTERS] Fortran Overhead ( 0 ) : 3.5216s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.8670s for 16384 events => throughput is 2.99E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1730s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G108/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.6247s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4709s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.9795s for 16384 events => throughput is 2.98E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1744s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G105/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.3282s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3402s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.8070s for 16384 events => throughput is 2.99E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1810s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G304/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 58.3397s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1202s + [COUNTERS] CudaCpp MEs ( 2 ) : 55.0456s for 16384 events => throughput is 2.98E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1738s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G175/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.2290s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2151s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.8413s for 16384 events => throughput is 2.99E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1726s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G128/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.5469s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5365s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.8375s for 16384 events => throughput is 2.99E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1729s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G360/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.6650s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5103s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.9821s for 16384 events => throughput is 2.98E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1726s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G120/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 58.0621s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2657s + [COUNTERS] CudaCpp MEs ( 2 ) : 55.6218s for 16384 events => throughput is 2.95E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1745s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G296/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.1066s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0523s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.8801s for 16384 events => throughput is 2.99E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1742s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G299/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.8054s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7900s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.8421s for 16384 events => throughput is 2.99E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1733s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G715/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.4723s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5912s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.7097s for 16384 events => throughput is 2.99E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1714s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G300/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.5910s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3609s + [COUNTERS] CudaCpp MEs ( 2 ) : 55.0552s for 16384 events => throughput is 2.98E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1749s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G303/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.8262s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8267s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.8273s for 16384 events => throughput is 2.99E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1722s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G127/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 58.0444s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8376s + [COUNTERS] CudaCpp MEs ( 2 ) : 55.0346s for 16384 events => throughput is 2.98E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1723s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G111/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.7249s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5142s + [COUNTERS] CudaCpp MEs ( 2 ) : 55.0377s for 16384 events => throughput is 2.98E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1729s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G126/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 65.4942s + [COUNTERS] Fortran Overhead ( 0 ) : 10.4898s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.8282s for 16384 events => throughput is 2.99E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1761s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G294/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 61.6497s + [COUNTERS] Fortran Overhead ( 0 ) : 6.5856s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.8917s for 16384 events => throughput is 2.98E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1724s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G598/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.6150s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3857s + [COUNTERS] CudaCpp MEs ( 2 ) : 55.0547s for 16384 events => throughput is 2.98E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1745s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G610/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 58.4681s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2731s + [COUNTERS] CudaCpp MEs ( 2 ) : 55.0229s for 16384 events => throughput is 2.98E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1721s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G178/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.6573s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6304s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.8467s for 16384 events => throughput is 2.99E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1803s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G121/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.5403s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4212s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.9460s for 16384 events => throughput is 2.98E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1731s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G685/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.1618s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1459s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.8454s for 16384 events => throughput is 2.99E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1705s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G690/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.0656s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0620s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.8301s for 16384 events => throughput is 2.99E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1735s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G103/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 63.2546s + [COUNTERS] Fortran Overhead ( 0 ) : 8.0887s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.9891s for 16384 events => throughput is 2.98E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1768s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G107/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.2932s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2992s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.8188s for 16384 events => throughput is 2.99E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1752s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G180/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.5728s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4519s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.9456s for 16384 events => throughput is 2.98E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1753s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G97/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 63.2143s + [COUNTERS] Fortran Overhead ( 0 ) : 8.1426s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.8978s for 16384 events => throughput is 2.98E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1739s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G179/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.7726s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7393s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.8489s for 16384 events => throughput is 2.99E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1843s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamggddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamggddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamggddx/ajob1' in 23.0753 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamggddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamggddx/G189/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 23.0439s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5288s + [COUNTERS] CudaCpp MEs ( 2 ) : 20.4437s for 16384 events => throughput is 8.01E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0714s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamggddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dd_taptamggdd/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dd_taptamggdd/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dd_taptamggdd/ajob1' in 66.0116 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dd_taptamggdd/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dd_taptamggdd/G463/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 65.9787s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1390s + [COUNTERS] CudaCpp MEs ( 2 ) : 63.7103s for 16384 events => throughput is 2.57E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1294s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dd_taptamggdd/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/ajob1' in 4528.4844 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G506/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 113.5876s + [COUNTERS] Fortran Overhead ( 0 ) : 4.1626s + [COUNTERS] CudaCpp MEs ( 2 ) : 109.2534s for 16384 events => throughput is 1.50E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1716s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G499/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 111.3914s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8022s + [COUNTERS] CudaCpp MEs ( 2 ) : 109.4163s for 16384 events => throughput is 1.50E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1729s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G46/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 117.3088s + [COUNTERS] Fortran Overhead ( 0 ) : 7.3950s + [COUNTERS] CudaCpp MEs ( 2 ) : 109.7381s for 16384 events => throughput is 1.49E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1757s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G56/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 111.8768s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4945s + [COUNTERS] CudaCpp MEs ( 2 ) : 109.2104s for 16384 events => throughput is 1.50E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1720s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G39/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 111.9935s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3580s + [COUNTERS] CudaCpp MEs ( 2 ) : 109.4638s for 16384 events => throughput is 1.50E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1716s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G517/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 114.5593s + [COUNTERS] Fortran Overhead ( 0 ) : 5.0036s + [COUNTERS] CudaCpp MEs ( 2 ) : 109.3853s for 16384 events => throughput is 1.50E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1704s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G52/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 112.5591s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8725s + [COUNTERS] CudaCpp MEs ( 2 ) : 109.5103s for 16384 events => throughput is 1.50E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1763s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G31/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 116.2018s + [COUNTERS] Fortran Overhead ( 0 ) : 6.5943s + [COUNTERS] CudaCpp MEs ( 2 ) : 109.4349s for 16384 events => throughput is 1.50E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1725s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G324/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 115.9349s + [COUNTERS] Fortran Overhead ( 0 ) : 6.2408s + [COUNTERS] CudaCpp MEs ( 2 ) : 109.5222s for 16384 events => throughput is 1.50E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1719s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G64/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 111.5217s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8050s + [COUNTERS] CudaCpp MEs ( 2 ) : 109.5424s for 16384 events => throughput is 1.50E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1743s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G3/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 113.9281s + [COUNTERS] Fortran Overhead ( 0 ) : 3.8686s + [COUNTERS] CudaCpp MEs ( 2 ) : 109.8856s for 16384 events => throughput is 1.49E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1738s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G33/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 112.3977s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1554s + [COUNTERS] CudaCpp MEs ( 2 ) : 110.0656s for 16384 events => throughput is 1.49E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1767s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G53/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 115.9561s + [COUNTERS] Fortran Overhead ( 0 ) : 6.3887s + [COUNTERS] CudaCpp MEs ( 2 ) : 109.3944s for 16384 events => throughput is 1.50E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1730s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G43/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 115.3053s + [COUNTERS] Fortran Overhead ( 0 ) : 5.2989s + [COUNTERS] CudaCpp MEs ( 2 ) : 109.8346s for 16384 events => throughput is 1.49E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1717s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G49/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 111.5643s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7763s + [COUNTERS] CudaCpp MEs ( 2 ) : 109.6133s for 16384 events => throughput is 1.49E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1748s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G40/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 113.3764s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7946s + [COUNTERS] CudaCpp MEs ( 2 ) : 110.4100s for 16384 events => throughput is 1.48E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1719s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G523/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 112.0809s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1084s + [COUNTERS] CudaCpp MEs ( 2 ) : 109.8018s for 16384 events => throughput is 1.49E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1708s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G518/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 114.8788s + [COUNTERS] Fortran Overhead ( 0 ) : 5.2268s + [COUNTERS] CudaCpp MEs ( 2 ) : 109.4807s for 16384 events => throughput is 1.50E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1712s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G35/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 111.7428s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8917s + [COUNTERS] CudaCpp MEs ( 2 ) : 109.6796s for 16384 events => throughput is 1.49E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1716s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 111.3659s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6059s + [COUNTERS] CudaCpp MEs ( 2 ) : 109.5868s for 16384 events => throughput is 1.50E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1732s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G42/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 114.9551s + [COUNTERS] Fortran Overhead ( 0 ) : 5.3443s + [COUNTERS] CudaCpp MEs ( 2 ) : 109.4395s for 16384 events => throughput is 1.50E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1713s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G524/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 111.8970s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0744s + [COUNTERS] CudaCpp MEs ( 2 ) : 109.6497s for 16384 events => throughput is 1.49E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1730s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G65/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 111.4656s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8518s + [COUNTERS] CudaCpp MEs ( 2 ) : 109.4418s for 16384 events => throughput is 1.50E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1720s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G27/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 115.5692s + [COUNTERS] Fortran Overhead ( 0 ) : 5.6480s + [COUNTERS] CudaCpp MEs ( 2 ) : 109.7500s for 16384 events => throughput is 1.49E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1712s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G44/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 115.7861s + [COUNTERS] Fortran Overhead ( 0 ) : 6.1905s + [COUNTERS] CudaCpp MEs ( 2 ) : 109.4244s for 16384 events => throughput is 1.50E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1712s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 111.4125s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5926s + [COUNTERS] CudaCpp MEs ( 2 ) : 109.6435s for 16384 events => throughput is 1.49E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1765s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G303/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 112.5961s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1027s + [COUNTERS] CudaCpp MEs ( 2 ) : 109.3211s for 16384 events => throughput is 1.50E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1723s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G55/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 112.3038s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6591s + [COUNTERS] CudaCpp MEs ( 2 ) : 109.4664s for 16384 events => throughput is 1.50E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1784s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G34/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 111.4196s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7721s + [COUNTERS] CudaCpp MEs ( 2 ) : 109.4769s for 16384 events => throughput is 1.50E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1705s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G26/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 114.9844s + [COUNTERS] Fortran Overhead ( 0 ) : 5.7795s + [COUNTERS] CudaCpp MEs ( 2 ) : 109.0341s for 16384 events => throughput is 1.50E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1709s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G62/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 115.4677s + [COUNTERS] Fortran Overhead ( 0 ) : 5.8381s + [COUNTERS] CudaCpp MEs ( 2 ) : 109.4571s for 16384 events => throughput is 1.50E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1725s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G37/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 111.6631s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1241s + [COUNTERS] CudaCpp MEs ( 2 ) : 109.3667s for 16384 events => throughput is 1.50E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1723s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G485/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 114.6683s + [COUNTERS] Fortran Overhead ( 0 ) : 5.1807s + [COUNTERS] CudaCpp MEs ( 2 ) : 109.3170s for 16384 events => throughput is 1.50E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1707s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G67/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 111.9891s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1112s + [COUNTERS] CudaCpp MEs ( 2 ) : 109.7068s for 16384 events => throughput is 1.49E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1711s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G512/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 111.4030s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7849s + [COUNTERS] CudaCpp MEs ( 2 ) : 109.4442s for 16384 events => throughput is 1.50E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1739s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G511/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 112.1426s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4736s + [COUNTERS] CudaCpp MEs ( 2 ) : 109.4976s for 16384 events => throughput is 1.50E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1714s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G51/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 112.3034s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4444s + [COUNTERS] CudaCpp MEs ( 2 ) : 109.6901s for 16384 events => throughput is 1.49E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1689s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 111.9526s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3693s + [COUNTERS] CudaCpp MEs ( 2 ) : 109.4125s for 16384 events => throughput is 1.50E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1708s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G63/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 111.8684s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2577s + [COUNTERS] CudaCpp MEs ( 2 ) : 109.4402s for 16384 events => throughput is 1.50E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1705s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G48/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 112.2893s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3358s + [COUNTERS] CudaCpp MEs ( 2 ) : 109.7797s for 16384 events => throughput is 1.49E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1739s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgggux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgggux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgggux/ajob1' in 799.6324 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgggux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgggux/G33/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 111.5222s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7434s + [COUNTERS] CudaCpp MEs ( 2 ) : 109.6066s for 16384 events => throughput is 1.49E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1721s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgggux/G68/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 113.3174s + [COUNTERS] Fortran Overhead ( 0 ) : 3.0662s + [COUNTERS] CudaCpp MEs ( 2 ) : 110.0629s for 16384 events => throughput is 1.49E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1883s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgggux/G25/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 116.2197s + [COUNTERS] Fortran Overhead ( 0 ) : 6.3004s + [COUNTERS] CudaCpp MEs ( 2 ) : 109.7413s for 16384 events => throughput is 1.49E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1779s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgggux/G494/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 111.2028s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6033s + [COUNTERS] CudaCpp MEs ( 2 ) : 109.4266s for 16384 events => throughput is 1.50E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1729s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgggux/G55/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 112.6659s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8617s + [COUNTERS] CudaCpp MEs ( 2 ) : 109.6310s for 16384 events => throughput is 1.49E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1732s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgggux/G509/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 118.9413s + [COUNTERS] Fortran Overhead ( 0 ) : 6.9097s + [COUNTERS] CudaCpp MEs ( 2 ) : 111.8561s for 16384 events => throughput is 1.46E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1755s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgggux/G29/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 115.5909s + [COUNTERS] Fortran Overhead ( 0 ) : 5.8937s + [COUNTERS] CudaCpp MEs ( 2 ) : 109.5240s for 16384 events => throughput is 1.50E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1732s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgggux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamgguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamgguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamgguux/ajob1' in 337.2395 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamgguux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamgguux/G260/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 66.2820s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5339s + [COUNTERS] CudaCpp MEs ( 2 ) : 64.6169s for 16384 events => throughput is 2.54E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1313s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamgguux/G447/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 66.2925s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5766s + [COUNTERS] CudaCpp MEs ( 2 ) : 64.5843s for 16384 events => throughput is 2.54E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1315s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamgguux/G229/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 68.7158s + [COUNTERS] Fortran Overhead ( 0 ) : 3.9671s + [COUNTERS] CudaCpp MEs ( 2 ) : 64.6196s for 16384 events => throughput is 2.54E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1292s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamgguux/G135/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 66.7853s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9477s + [COUNTERS] CudaCpp MEs ( 2 ) : 64.7075s for 16384 events => throughput is 2.53E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1301s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamgguux/G230/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 69.0319s + [COUNTERS] Fortran Overhead ( 0 ) : 4.2840s + [COUNTERS] CudaCpp MEs ( 2 ) : 64.6188s for 16384 events => throughput is 2.54E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1290s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamgguux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdddx/ajob1' in 68.3176 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdddx/G715/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 34.1903s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0393s + [COUNTERS] CudaCpp MEs ( 2 ) : 32.0158s for 16384 events => throughput is 5.12E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1352s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdddx/G111/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 34.0555s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0053s + [COUNTERS] CudaCpp MEs ( 2 ) : 31.9207s for 16384 events => throughput is 5.13E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1295s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemusdxsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemusdxsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemusdxsx/ajob1' in 7.4172 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemusdxsx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemusdxsx/G67/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.3624s + [COUNTERS] Fortran Overhead ( 0 ) : 6.0407s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.3007s for 16384 events => throughput is 1.26E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0209s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemusdxsx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdssx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdssx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdssx/ajob1' in 73.1353 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdssx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdssx/G31/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 12.1014s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9230s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.1061s for 16384 events => throughput is 1.62E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0724s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdssx/G27/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 12.0764s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9054s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.0986s for 16384 events => throughput is 1.62E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0724s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdssx/G119/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 12.3289s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1421s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.1153s for 16384 events => throughput is 1.62E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0715s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdssx/G284/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 12.5762s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3495s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.1543s for 16384 events => throughput is 1.61E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0725s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdssx/G375/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 11.7270s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5711s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.0846s for 16384 events => throughput is 1.62E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0713s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdssx/G102/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 12.1708s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9884s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.1107s for 16384 events => throughput is 1.62E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0717s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdssx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_epemgguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_epemgguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_epemgguux/ajob1' in 142.6171 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_epemgguux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_epemgguux/G105/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 33.8779s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5278s + [COUNTERS] CudaCpp MEs ( 2 ) : 32.2217s for 16384 events => throughput is 5.08E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1284s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_epemgguux/G223/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 34.6132s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1875s + [COUNTERS] CudaCpp MEs ( 2 ) : 32.2943s for 16384 events => throughput is 5.07E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1313s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_epemgguux/G103/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 36.7721s + [COUNTERS] Fortran Overhead ( 0 ) : 4.3879s + [COUNTERS] CudaCpp MEs ( 2 ) : 32.2559s for 16384 events => throughput is 5.08E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1283s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_epemgguux/G5/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 37.2454s + [COUNTERS] Fortran Overhead ( 0 ) : 4.8080s + [COUNTERS] CudaCpp MEs ( 2 ) : 32.3045s for 16384 events => throughput is 5.07E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1330s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_epemgguux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/ajob1' in 3588.8000 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G506/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 60.4338s + [COUNTERS] Fortran Overhead ( 0 ) : 5.3775s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.8792s for 16384 events => throughput is 2.99E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1771s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G499/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 56.5149s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8284s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.5141s for 16384 events => throughput is 3.01E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1724s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G38/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.3775s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6793s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.5258s for 16384 events => throughput is 3.00E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1724s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G46/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 62.3345s + [COUNTERS] Fortran Overhead ( 0 ) : 7.6321s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.5294s for 16384 events => throughput is 3.00E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1729s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G59/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 60.8557s + [COUNTERS] Fortran Overhead ( 0 ) : 6.1321s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.5516s for 16384 events => throughput is 3.00E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1720s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G56/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.3898s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7158s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.5012s for 16384 events => throughput is 3.01E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1728s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G525/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.0922s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4167s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.5009s for 16384 events => throughput is 3.01E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1746s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G39/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 56.8022s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3164s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.3129s for 16384 events => throughput is 3.02E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1729s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G517/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 60.5129s + [COUNTERS] Fortran Overhead ( 0 ) : 5.5610s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.7738s for 16384 events => throughput is 2.99E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1780s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G487/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 59.3311s + [COUNTERS] Fortran Overhead ( 0 ) : 4.7248s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.4306s for 16384 events => throughput is 3.01E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1758s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G52/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.6952s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8823s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.6398s for 16384 events => throughput is 3.00E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1731s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G31/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 62.4126s + [COUNTERS] Fortran Overhead ( 0 ) : 7.6343s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.6059s for 16384 events => throughput is 3.00E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1723s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G64/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 56.6471s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9428s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.5324s for 16384 events => throughput is 3.00E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1719s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G3/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 59.5486s + [COUNTERS] Fortran Overhead ( 0 ) : 4.8473s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.5300s for 16384 events => throughput is 3.00E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1713s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G497/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 56.9223s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2292s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.5197s for 16384 events => throughput is 3.01E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1734s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G13/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 56.3420s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6234s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.5474s for 16384 events => throughput is 3.00E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1713s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G66/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.1877s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3823s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.6300s for 16384 events => throughput is 3.00E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1755s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G33/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 56.6670s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9198s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.5759s for 16384 events => throughput is 3.00E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1713s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G53/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 62.6511s + [COUNTERS] Fortran Overhead ( 0 ) : 7.9293s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.5495s for 16384 events => throughput is 3.00E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1722s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G43/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 59.5976s + [COUNTERS] Fortran Overhead ( 0 ) : 4.7565s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.6701s for 16384 events => throughput is 3.00E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1710s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G68/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 56.7319s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1277s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.4317s for 16384 events => throughput is 3.01E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1725s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G49/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 56.6025s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8786s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.5521s for 16384 events => throughput is 3.00E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1718s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G302/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 63.0092s + [COUNTERS] Fortran Overhead ( 0 ) : 8.1419s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.6942s for 16384 events => throughput is 3.00E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1731s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G25/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 60.8870s + [COUNTERS] Fortran Overhead ( 0 ) : 6.1101s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.6024s for 16384 events => throughput is 3.00E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1745s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G40/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 58.0891s + [COUNTERS] Fortran Overhead ( 0 ) : 3.0222s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.8930s for 16384 events => throughput is 2.98E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1738s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G523/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.4080s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0123s + [COUNTERS] CudaCpp MEs ( 2 ) : 55.2238s for 16384 events => throughput is 2.97E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1720s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G518/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 59.7765s + [COUNTERS] Fortran Overhead ( 0 ) : 5.1201s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.4828s for 16384 events => throughput is 3.01E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1737s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G35/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 56.8225s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0816s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.5675s for 16384 events => throughput is 3.00E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1734s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 56.3748s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6084s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.5942s for 16384 events => throughput is 3.00E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1723s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G524/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 56.7201s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0483s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.5007s for 16384 events => throughput is 3.01E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1711s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G65/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.0676s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0901s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.7946s for 16384 events => throughput is 2.99E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1829s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G27/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 61.4551s + [COUNTERS] Fortran Overhead ( 0 ) : 6.3484s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.9345s for 16384 events => throughput is 2.98E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1722s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G44/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 60.8396s + [COUNTERS] Fortran Overhead ( 0 ) : 5.9283s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.7369s for 16384 events => throughput is 2.99E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1744s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G483/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 58.9818s + [COUNTERS] Fortran Overhead ( 0 ) : 4.1675s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.6371s for 16384 events => throughput is 3.00E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1771s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G482/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 60.2564s + [COUNTERS] Fortran Overhead ( 0 ) : 5.5883s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.4950s for 16384 events => throughput is 3.01E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1731s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G1/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 59.2287s + [COUNTERS] Fortran Overhead ( 0 ) : 4.4458s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.6099s for 16384 events => throughput is 3.00E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1729s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 56.3105s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5113s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.6258s for 16384 events => throughput is 3.00E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1734s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G55/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.1950s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4389s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.5851s for 16384 events => throughput is 3.00E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1710s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G298/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.6745s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9443s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.5591s for 16384 events => throughput is 3.00E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1711s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G513/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.8473s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8133s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.8605s for 16384 events => throughput is 2.99E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1735s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G60/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 62.0123s + [COUNTERS] Fortran Overhead ( 0 ) : 7.2847s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.5550s for 16384 events => throughput is 3.00E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1727s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G34/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 56.8706s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9752s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.7228s for 16384 events => throughput is 2.99E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1726s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G26/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 61.5826s + [COUNTERS] Fortran Overhead ( 0 ) : 6.7802s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.6304s for 16384 events => throughput is 3.00E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1720s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G58/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 60.6545s + [COUNTERS] Fortran Overhead ( 0 ) : 5.3826s + [COUNTERS] CudaCpp MEs ( 2 ) : 55.0947s for 16384 events => throughput is 2.97E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1771s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G54/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 60.6327s + [COUNTERS] Fortran Overhead ( 0 ) : 5.8585s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.5996s for 16384 events => throughput is 3.00E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1745s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G62/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 61.5462s + [COUNTERS] Fortran Overhead ( 0 ) : 6.8313s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.5357s for 16384 events => throughput is 3.00E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1793s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G37/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 56.6991s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8626s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.6616s for 16384 events => throughput is 3.00E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1748s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G493/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 56.7373s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1013s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.4631s for 16384 events => throughput is 3.01E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1729s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G4/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 60.9611s + [COUNTERS] Fortran Overhead ( 0 ) : 6.1008s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.6872s for 16384 events => throughput is 3.00E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1731s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G50/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 56.8519s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0674s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.6049s for 16384 events => throughput is 3.00E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1795s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G485/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 60.3770s + [COUNTERS] Fortran Overhead ( 0 ) : 5.6420s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.5635s for 16384 events => throughput is 3.00E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1714s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G41/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 61.6560s + [COUNTERS] Fortran Overhead ( 0 ) : 6.9331s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.5513s for 16384 events => throughput is 3.00E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1717s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G67/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 56.9419s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0941s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.6743s for 16384 events => throughput is 3.00E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1735s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G512/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 56.5026s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9860s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.3438s for 16384 events => throughput is 3.01E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1728s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G511/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.4510s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5110s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.7662s for 16384 events => throughput is 2.99E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1738s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G57/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 61.7746s + [COUNTERS] Fortran Overhead ( 0 ) : 6.7913s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.8082s for 16384 events => throughput is 2.99E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1751s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.4766s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8301s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.4736s for 16384 events => throughput is 3.01E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1728s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G63/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.0742s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3718s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.5298s for 16384 events => throughput is 3.00E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1727s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G32/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 62.1886s + [COUNTERS] Fortran Overhead ( 0 ) : 7.4921s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.5223s for 16384 events => throughput is 3.01E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1743s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G29/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 61.0161s + [COUNTERS] Fortran Overhead ( 0 ) : 6.1304s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.7128s for 16384 events => throughput is 2.99E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1728s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G48/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 56.9851s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3355s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.4768s for 16384 events => throughput is 3.01E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1729s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_taptamggdsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_taptamggdsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_taptamggdsx/ajob1' in 66.0501 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_taptamggdsx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_taptamggdsx/G92/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 22.1582s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3649s + [COUNTERS] CudaCpp MEs ( 2 ) : 19.7216s for 16384 events => throughput is 8.31E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0717s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_taptamggdsx/G144/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 21.9643s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0691s + [COUNTERS] CudaCpp MEs ( 2 ) : 19.8232s for 16384 events => throughput is 8.27E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0721s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_taptamggdsx/G207/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 21.8580s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0345s + [COUNTERS] CudaCpp MEs ( 2 ) : 19.7518s for 16384 events => throughput is 8.29E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0718s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_taptamggdsx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_epemggdsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_epemggdsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_epemggdsx/ajob1' in 16.8275 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_epemggdsx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_epemggdsx/G205/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 16.7951s + [COUNTERS] Fortran Overhead ( 0 ) : 6.5734s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.1488s for 16384 events => throughput is 1.61E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0729s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_epemggdsx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_taptamgggg/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_taptamgggg/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_taptamgggg/ajob1' in 111.5782 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_taptamgggg/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_taptamgggg/G67/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 111.5479s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6984s + [COUNTERS] CudaCpp MEs ( 2 ) : 108.6746s for 16384 events => throughput is 1.51E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1748s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_taptamgggg/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dc_epemdcssx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dc_epemdcssx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dc_epemdcssx/ajob1' in 3.5919 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dc_epemdcssx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dc_epemdcssx/G77/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.5455s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1709s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.3525s for 16384 events => throughput is 1.21E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0221s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dc_epemdcssx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguccx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguccx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguccx/ajob1' in 22.4731 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguccx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguccx/G48/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 22.4208s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4135s + [COUNTERS] CudaCpp MEs ( 2 ) : 19.9352s for 16384 events => throughput is 8.22E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0720s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguccx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ds_epemggds/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ds_epemggds/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ds_epemggds/ajob1' in 11.7701 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ds_epemggds/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ds_epemggds/G26/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 11.7388s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5794s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.0881s for 16384 events => throughput is 1.62E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0712s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ds_epemggds/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ucx_epemggucx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ucx_epemggucx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ucx_epemggucx/ajob1' in 25.8117 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ucx_epemggucx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ucx_epemggucx/G163/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.0843s + [COUNTERS] Fortran Overhead ( 0 ) : 3.9639s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.0470s for 16384 events => throughput is 1.63E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0735s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ucx_epemggucx/G60/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 11.6586s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5467s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.0383s for 16384 events => throughput is 1.63E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0735s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ucx_epemggucx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemggudx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemggudx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemggudx/ajob1' in 162.9500 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemggudx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemggudx/G109/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 11.6972s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4662s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.1574s for 16384 events => throughput is 1.61E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0735s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemggudx/G151/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 12.0076s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7972s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.1380s for 16384 events => throughput is 1.62E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0725s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemggudx/G106/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.3795s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1488s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.1581s for 16384 events => throughput is 1.61E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0726s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemggudx/G35/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 16.9837s + [COUNTERS] Fortran Overhead ( 0 ) : 6.7154s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.1961s for 16384 events => throughput is 1.61E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0721s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemggudx/G119/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 12.2207s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9775s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.1705s for 16384 events => throughput is 1.61E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0727s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemggudx/G187/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 12.0033s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8044s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.1263s for 16384 events => throughput is 1.62E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0726s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemggudx/G91/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 12.2489s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9730s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.2025s for 16384 events => throughput is 1.61E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0734s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemggudx/G363/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.6708s + [COUNTERS] Fortran Overhead ( 0 ) : 16.9260s + [COUNTERS] CudaCpp MEs ( 2 ) : 40.6728s for 65536 events => throughput is 1.61E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0720s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemggudx/G149/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.5550s + [COUNTERS] Fortran Overhead ( 0 ) : 4.3342s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.1481s for 16384 events => throughput is 1.61E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0727s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemggudx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgcuxcx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgcuxcx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgcuxcx/ajob1' in 12.4052 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgcuxcx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgcuxcx/G120/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 12.3554s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4086s + [COUNTERS] CudaCpp MEs ( 2 ) : 9.8754s for 16384 events => throughput is 1.66E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0714s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgcuxcx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemgggg/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemgggg/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemgggg/ajob1' in 116.9380 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemgggg/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemgggg/G65/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 58.3462s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1474s + [COUNTERS] CudaCpp MEs ( 2 ) : 55.0263s for 16384 events => throughput is 2.98E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1725s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemgggg/G48/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 58.5169s + [COUNTERS] Fortran Overhead ( 0 ) : 3.4139s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.9300s for 16384 events => throughput is 2.98E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1730s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemgggg/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_taptamggud/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_taptamggud/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_taptamggud/ajob1' in 198.5193 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_taptamggud/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_taptamggud/G46/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 21.9005s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1490s + [COUNTERS] CudaCpp MEs ( 2 ) : 19.6812s for 16384 events => throughput is 8.32E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0703s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_taptamggud/G39/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 23.8389s + [COUNTERS] Fortran Overhead ( 0 ) : 4.0772s + [COUNTERS] CudaCpp MEs ( 2 ) : 19.6915s for 16384 events => throughput is 8.32E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0702s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_taptamggud/G112/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 21.0194s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2292s + [COUNTERS] CudaCpp MEs ( 2 ) : 19.7181s for 16384 events => throughput is 8.31E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0721s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_taptamggud/G115/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 21.2075s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4557s + [COUNTERS] CudaCpp MEs ( 2 ) : 19.6793s for 16384 events => throughput is 8.33E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0725s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_taptamggud/G270/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 22.7258s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8885s + [COUNTERS] CudaCpp MEs ( 2 ) : 19.7650s for 16384 events => throughput is 8.29E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0723s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_taptamggud/G96/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 21.3432s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5593s + [COUNTERS] CudaCpp MEs ( 2 ) : 19.7133s for 16384 events => throughput is 8.31E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0706s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_taptamggud/G107/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 23.5841s + [COUNTERS] Fortran Overhead ( 0 ) : 3.7763s + [COUNTERS] CudaCpp MEs ( 2 ) : 19.7373s for 16384 events => throughput is 8.30E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0705s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_taptamggud/G63/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 21.6164s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8690s + [COUNTERS] CudaCpp MEs ( 2 ) : 19.6758s for 16384 events => throughput is 8.33E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0716s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_taptamggud/G375/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 21.0811s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2682s + [COUNTERS] CudaCpp MEs ( 2 ) : 19.7419s for 16384 events => throughput is 8.30E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0710s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_taptamggud/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguuux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguuux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguuux/ajob1' in 143.4455 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguuux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguuux/G122/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 33.8406s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8368s + [COUNTERS] CudaCpp MEs ( 2 ) : 31.8738s for 16384 events => throughput is 5.14E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1300s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguuux/G33/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 37.6201s + [COUNTERS] Fortran Overhead ( 0 ) : 5.6664s + [COUNTERS] CudaCpp MEs ( 2 ) : 31.8271s for 16384 events => throughput is 5.15E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1266s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguuux/G127/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 34.9826s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9548s + [COUNTERS] CudaCpp MEs ( 2 ) : 31.8973s for 16384 events => throughput is 5.14E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1305s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguuux/G116/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 36.9118s + [COUNTERS] Fortran Overhead ( 0 ) : 4.9633s + [COUNTERS] CudaCpp MEs ( 2 ) : 31.8170s for 16384 events => throughput is 5.15E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1314s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguuux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamgggdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamgggdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamgggdx/ajob1' in 1004.5281 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamgggdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamgggdx/G56/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 112.1195s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5498s + [COUNTERS] CudaCpp MEs ( 2 ) : 109.3969s for 16384 events => throughput is 1.50E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1728s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamgggdx/G64/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 111.9671s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3841s + [COUNTERS] CudaCpp MEs ( 2 ) : 109.4111s for 16384 events => throughput is 1.50E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1719s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamgggdx/G13/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 110.9781s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6195s + [COUNTERS] CudaCpp MEs ( 2 ) : 109.1852s for 16384 events => throughput is 1.50E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1735s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamgggdx/G33/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 110.8617s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8145s + [COUNTERS] CudaCpp MEs ( 2 ) : 108.8757s for 16384 events => throughput is 1.50E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1714s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamgggdx/G49/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 111.7569s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1059s + [COUNTERS] CudaCpp MEs ( 2 ) : 109.4791s for 16384 events => throughput is 1.50E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1719s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamgggdx/G523/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 112.1221s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5842s + [COUNTERS] CudaCpp MEs ( 2 ) : 109.3655s for 16384 events => throughput is 1.50E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1724s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamgggdx/G65/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 111.2227s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9576s + [COUNTERS] CudaCpp MEs ( 2 ) : 109.0924s for 16384 events => throughput is 1.50E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1727s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamgggdx/G34/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 111.5635s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0596s + [COUNTERS] CudaCpp MEs ( 2 ) : 109.3308s for 16384 events => throughput is 1.50E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1731s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamgggdx/G48/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 111.7250s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9816s + [COUNTERS] CudaCpp MEs ( 2 ) : 109.5710s for 16384 events => throughput is 1.50E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1724s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamgggdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgudux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgudux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgudux/ajob1' in 48.9529 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgudux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgudux/G59/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 22.4863s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8392s + [COUNTERS] CudaCpp MEs ( 2 ) : 20.5759s for 16384 events => throughput is 7.96E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0712s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgudux/G61/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 26.3974s + [COUNTERS] Fortran Overhead ( 0 ) : 5.5572s + [COUNTERS] CudaCpp MEs ( 2 ) : 20.7670s for 16384 events => throughput is 7.89E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0732s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgudux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgduxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgduxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgduxdx/ajob1' in 12.0241 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgduxdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgduxdx/G104/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 11.9733s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8755s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.0263s for 16384 events => throughput is 1.63E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0715s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgduxdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgsdxsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgsdxsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgsdxsx/ajob1' in 11.7674 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgsdxsx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgsdxsx/G25/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 11.7355s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7219s + [COUNTERS] CudaCpp MEs ( 2 ) : 9.9413s for 16384 events => throughput is 1.65E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0723s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgsdxsx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/ajob1' in 364.7243 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G208/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 11.9649s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8882s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.0054s for 16384 events => throughput is 1.64E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0713s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G235/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 11.8002s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6664s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.0621s for 16384 events => throughput is 1.63E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0717s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G140/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 12.0488s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9630s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.0132s for 16384 events => throughput is 1.64E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0726s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G109/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 11.7665s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6567s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.0381s for 16384 events => throughput is 1.63E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0717s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G33/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.8043s + [COUNTERS] Fortran Overhead ( 0 ) : 4.6122s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.1217s for 16384 events => throughput is 1.62E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0704s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G43/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 11.8090s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7240s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.0141s for 16384 events => throughput is 1.64E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0710s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G240/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 11.9646s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7477s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.1458s for 16384 events => throughput is 1.61E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0711s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G268/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.4659s + [COUNTERS] Fortran Overhead ( 0 ) : 4.3403s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.0543s for 16384 events => throughput is 1.63E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0713s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G139/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 12.0846s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9549s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.0584s for 16384 events => throughput is 1.63E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0713s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G120/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 11.9097s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7231s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.1149s for 16384 events => throughput is 1.62E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0718s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G236/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 11.8765s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7586s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.0468s for 16384 events => throughput is 1.63E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0711s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G181/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 15.3526s + [COUNTERS] Fortran Overhead ( 0 ) : 5.2213s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.0602s for 16384 events => throughput is 1.63E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0711s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G351/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 11.3873s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3401s + [COUNTERS] CudaCpp MEs ( 2 ) : 9.9748s for 16384 events => throughput is 1.64E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0724s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G65/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.0040s + [COUNTERS] Fortran Overhead ( 0 ) : 3.7951s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.1370s for 16384 events => throughput is 1.62E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0720s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G71/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 11.6657s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5330s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.0576s for 16384 events => throughput is 1.63E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0751s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G119/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 11.9749s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8841s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.0193s for 16384 events => throughput is 1.64E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0715s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G127/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 11.6941s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5575s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.0645s for 16384 events => throughput is 1.63E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0721s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G96/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 11.7863s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6699s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.0442s for 16384 events => throughput is 1.63E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0722s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G60/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 11.3288s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2128s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.0431s for 16384 events => throughput is 1.63E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0729s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G26/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 11.5871s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4418s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.0731s for 16384 events => throughput is 1.63E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0721s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G37/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 15.7502s + [COUNTERS] Fortran Overhead ( 0 ) : 5.6429s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.0354s for 16384 events => throughput is 1.63E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0719s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G125/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.3583s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2674s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.0190s for 16384 events => throughput is 1.64E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0719s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G91/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 11.6211s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4598s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.0904s for 16384 events => throughput is 1.62E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0709s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G143/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 11.4501s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3499s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.0284s for 16384 events => throughput is 1.63E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0718s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G116/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 11.3687s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2740s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.0234s for 16384 events => throughput is 1.63E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0713s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G61/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 15.6194s + [COUNTERS] Fortran Overhead ( 0 ) : 5.5228s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.0251s for 16384 events => throughput is 1.63E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0715s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G149/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.3090s + [COUNTERS] Fortran Overhead ( 0 ) : 4.2246s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.0125s for 16384 events => throughput is 1.64E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0720s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G144/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 11.5360s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4355s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.0296s for 16384 events => throughput is 1.63E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0708s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G207/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 11.8390s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7123s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.0544s for 16384 events => throughput is 1.63E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0722s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_epemgguu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_epemgguu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_epemgguu/ajob1' in 382.8558 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_epemgguu/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_epemgguu/G463/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 34.0246s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9825s + [COUNTERS] CudaCpp MEs ( 2 ) : 31.9153s for 16384 events => throughput is 5.13E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1268s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_epemgguu/G100/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 35.2033s + [COUNTERS] Fortran Overhead ( 0 ) : 3.0654s + [COUNTERS] CudaCpp MEs ( 2 ) : 32.0081s for 16384 events => throughput is 5.12E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1299s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_epemgguu/G110/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 33.3856s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4327s + [COUNTERS] CudaCpp MEs ( 2 ) : 31.8232s for 16384 events => throughput is 5.15E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1297s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_epemgguu/G124/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 33.9970s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5190s + [COUNTERS] CudaCpp MEs ( 2 ) : 32.3497s for 16384 events => throughput is 5.06E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1283s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_epemgguu/G367/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 33.9222s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8967s + [COUNTERS] CudaCpp MEs ( 2 ) : 31.8965s for 16384 events => throughput is 5.14E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1290s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_epemgguu/G126/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 35.9458s + [COUNTERS] Fortran Overhead ( 0 ) : 3.9492s + [COUNTERS] CudaCpp MEs ( 2 ) : 31.8678s for 16384 events => throughput is 5.14E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1288s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_epemgguu/G361/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 36.4473s + [COUNTERS] Fortran Overhead ( 0 ) : 4.3235s + [COUNTERS] CudaCpp MEs ( 2 ) : 31.9944s for 16384 events => throughput is 5.12E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1295s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_epemgguu/G116/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 34.7154s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6132s + [COUNTERS] CudaCpp MEs ( 2 ) : 31.9742s for 16384 events => throughput is 5.12E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1280s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_epemgguu/G363/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 36.0231s + [COUNTERS] Fortran Overhead ( 0 ) : 4.0045s + [COUNTERS] CudaCpp MEs ( 2 ) : 31.8905s for 16384 events => throughput is 5.14E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1281s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_epemgguu/G138/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 33.7139s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5588s + [COUNTERS] CudaCpp MEs ( 2 ) : 32.0254s for 16384 events => throughput is 5.12E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1298s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_epemgguu/G102/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 35.2602s + [COUNTERS] Fortran Overhead ( 0 ) : 3.0922s + [COUNTERS] CudaCpp MEs ( 2 ) : 32.0389s for 16384 events => throughput is 5.11E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1290s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_epemgguu/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgdssx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgdssx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgdssx/ajob1' in 22.0753 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgdssx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgdssx/G28/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 22.0189s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0030s + [COUNTERS] CudaCpp MEs ( 2 ) : 19.9437s for 16384 events => throughput is 8.22E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0721s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgdssx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_taptamgguu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_taptamgguu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_taptamgguu/ajob1' in 328.2254 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_taptamgguu/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_taptamgguu/G464/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 65.4459s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8704s + [COUNTERS] CudaCpp MEs ( 2 ) : 63.4479s for 16384 events => throughput is 2.58E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1276s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_taptamgguu/G59/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 65.6208s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0441s + [COUNTERS] CudaCpp MEs ( 2 ) : 63.4480s for 16384 events => throughput is 2.58E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1287s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_taptamgguu/G100/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 66.0257s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4739s + [COUNTERS] CudaCpp MEs ( 2 ) : 63.4235s for 16384 events => throughput is 2.58E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1283s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_taptamgguu/G106/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 65.4321s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7862s + [COUNTERS] CudaCpp MEs ( 2 ) : 63.5162s for 16384 events => throughput is 2.58E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1297s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_taptamgguu/G143/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 65.5728s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7716s + [COUNTERS] CudaCpp MEs ( 2 ) : 63.6732s for 16384 events => throughput is 2.57E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1280s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_taptamgguu/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dc_epemggdc/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dc_epemggdc/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dc_epemggdc/ajob1' in 15.3370 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dc_epemggdc/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dc_epemggdc/G18/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 15.2836s + [COUNTERS] Fortran Overhead ( 0 ) : 5.0774s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.1336s for 16384 events => throughput is 1.62E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0727s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dc_epemggdc/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ucx_taptamggucx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ucx_taptamggucx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ucx_taptamggucx/ajob1' in 24.9342 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ucx_taptamggucx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ucx_taptamggucx/G62/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 24.8871s + [COUNTERS] Fortran Overhead ( 0 ) : 4.7603s + [COUNTERS] CudaCpp MEs ( 2 ) : 20.0517s for 16384 events => throughput is 8.17E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0751s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ucx_taptamggucx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamucdxcx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamucdxcx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamucdxcx/ajob1' in 4.4679 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamucdxcx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamucdxcx/G47/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.4149s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7775s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.6158s for 16384 events => throughput is 6.26E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0216s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamucdxcx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_epemgggg/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_epemgggg/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_epemgggg/ajob1' in 58.4333 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_epemgggg/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_epemgggg/G49/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 58.3811s + [COUNTERS] Fortran Overhead ( 0 ) : 3.3405s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.8689s for 16384 events => throughput is 2.99E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1717s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_epemgggg/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemudddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemudddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemudddx/ajob1' in 5.1988 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemudddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemudddx/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.1466s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5687s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.5424s for 16384 events => throughput is 4.63E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0355s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemudddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamggudx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamggudx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamggudx/ajob1' in 21.5460 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamggudx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamggudx/G60/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 21.5144s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5983s + [COUNTERS] CudaCpp MEs ( 2 ) : 19.8436s for 16384 events => throughput is 8.26E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0725s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamggudx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_taptamuuddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_taptamuuddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_taptamuuddx/ajob1' in 8.9052 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_taptamuuddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_taptamuuddx/G11/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.8488s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6272s + [COUNTERS] CudaCpp MEs ( 2 ) : 7.1854s for 16384 events => throughput is 2.28E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0362s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_taptamuuddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished in 26596.4601 seconds +__CUDACPP_DEBUG: gen_ximprove.launch finished in 26600.4572 seconds +__CUDACPP_DEBUG: GridPackCmd.refine4grid finished in 26600.4580 seconds +__CUDACPP_DEBUG: GridPackCmd.launch (3a) combine_events +combine_events +INFO: fail to reach target 500 +__CUDACPP_DEBUG: GridPackCmd.launch (3b) store_events +store_events + === Results Summary for run: GridRun_21 tag: tag_1 === + + Cross-section : 0 +- 0 pb + Nb of events : 195 + +INFO: No version of lhapdf. Can not run systematics computation +decay_events -from_cards +__CUDACPP_DEBUG: GridPackCmd.launch finished in 26745.1639 seconds +__CUDACPP_DEBUG: GridPackCmd.__init__ back from self.launch +DONE +write ./events.lhe.gz +END: Sun Sep 15 07:36:07 AM CEST 2024 +ELAPSED: 26748 seconds diff --git a/epochX/cudacpp/tlau/fromgridpacks/pp_dy4j.mad/cppsse4/output.txt b/epochX/cudacpp/tlau/fromgridpacks/pp_dy4j.mad/cppsse4/output.txt new file mode 100644 index 0000000000..16710ed90e --- /dev/null +++ b/epochX/cudacpp/tlau/fromgridpacks/pp_dy4j.mad/cppsse4/output.txt @@ -0,0 +1,3327 @@ +START: Sun Sep 15 07:39:23 AM CEST 2024 +On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU (nproc=4)] [GPU: 1x Tesla V100S-PCIE-32GB]: +CUDACPP_RUNTIME_DISABLEFPE is set +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_bbx_epemdsdxsx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_bbx_taptamdsdxsx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_ccx_epemdcdxcx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_ccx_epemdsdxsx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_ccx_epemucuxcx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_ccx_epemuduxdx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_ccx_taptamdcdxcx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_ccx_taptamdsdxsx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_ccx_taptamucuxcx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_ccx_taptamuduxdx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_dc_epemdcccx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_dc_epemdcssx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_dc_epemddcdx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_dc_epemggdc/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_dc_epemudcux/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_dc_taptamdcccx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_dc_taptamdcssx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_dc_taptamddcdx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_dc_taptamggdc/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_dc_taptamudcux/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_dd_epemddddx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_dd_epemddssx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_dd_epemggdd/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_dd_epemuddux/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_dd_taptamddddx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_dd_taptamddssx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_dd_taptamggdd/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_dd_taptamuddux/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_ddx_epemdcdxcx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_ddx_epemdddxdx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_ddx_epemdsdxsx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_ddx_epemggddx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_ddx_epemgggg/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_ddx_epemggssx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_ddx_epemgguux/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_ddx_epemsbsxbx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_ddx_epemsssxsx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_ddx_epemucuxcx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_ddx_epemuduxdx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_ddx_epemusuxsx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_ddx_epemuuuxux/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_ddx_taptamdcdxcx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_ddx_taptamdddxdx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_ddx_taptamdsdxsx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_ddx_taptamggddx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_ddx_taptamgggg/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_ddx_taptamggssx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_ddx_taptamgguux/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_ddx_taptamsbsxbx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_ddx_taptamsssxsx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_ddx_taptamucuxcx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_ddx_taptamuduxdx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_ddx_taptamusuxsx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_ddx_taptamuuuxux/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_ds_epemddsdx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_ds_epemdsbbx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_ds_epemdsssx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_ds_epemggds/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_ds_epemudsux/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_ds_taptamddsdx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_ds_taptamdsbbx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_ds_taptamdsssx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_ds_taptamggds/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_ds_taptamudsux/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_dsx_epemdbsxbx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_dsx_epemdddxsx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_dsx_epemdssxsx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_dsx_epemggdsx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_dsx_epemuduxsx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_dsx_taptamdbsxbx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_dsx_taptamdddxsx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_dsx_taptamdssxsx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_dsx_taptamggdsx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_dsx_taptamuduxsx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_dux_epemdcuxcx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_dux_epemdduxdx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_dux_epemdsuxsx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_dux_epemggdux/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_dux_epemuduxux/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_dux_taptamdcuxcx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_dux_taptamdduxdx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_dux_taptamdsuxsx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_dux_taptamggdux/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_dux_taptamuduxux/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_dxcx_epemcdxcxcx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_dxcx_epemddxdxcx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_dxcx_epemggdxcx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_dxcx_epemsdxcxsx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_dxcx_epemuuxdxcx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_dxcx_taptamcdxcxcx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_dxcx_taptamddxdxcx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_dxcx_taptamggdxcx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_dxcx_taptamsdxcxsx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_dxcx_taptamuuxdxcx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_dxdx_epemddxdxdx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_dxdx_epemggdxdx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_dxdx_epemsdxdxsx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_dxdx_epemuuxdxdx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_dxdx_taptamddxdxdx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_dxdx_taptamggdxdx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_dxdx_taptamsdxdxsx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_dxdx_taptamuuxdxdx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_dxsx_epembdxsxbx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_dxsx_epemddxdxsx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_dxsx_epemggdxsx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_dxsx_epemsdxsxsx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_dxsx_epemuuxdxsx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_dxsx_taptambdxsxbx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_dxsx_taptamddxdxsx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_dxsx_taptamggdxsx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_dxsx_taptamsdxsxsx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_dxsx_taptamuuxdxsx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_gd_epemgdddx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_gd_epemgdssx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_gd_epemgggd/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_gd_epemgudux/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_gd_taptamgdddx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_gd_taptamgdssx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_gd_taptamgggd/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_gd_taptamgudux/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_gdx_epemgddxdx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_gdx_epemgggdx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_gdx_epemgsdxsx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_gdx_epemguuxdx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_gdx_taptamgddxdx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_gdx_taptamgggdx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_gdx_taptamgsdxsx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_gdx_taptamguuxdx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_gg_epemdddxdx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_gg_epemdsdxsx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_gg_epemggddx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_gg_epemgguux/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_gg_epemucuxcx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_gg_epemuduxdx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_gg_epemuuuxux/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_gg_taptamdddxdx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_gg_taptamdsdxsx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_gg_taptamggddx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_gg_taptamgguux/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_gg_taptamucuxcx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_gg_taptamuduxdx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_gg_taptamuuuxux/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_gu_epemgggu/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_gu_epemguccx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_gu_epemguddx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_gu_epemguuux/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_gu_taptamgggu/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_gu_taptamguccx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_gu_taptamguddx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_gu_taptamguuux/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_gux_epemgcuxcx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_gux_epemgduxdx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_gux_epemgggux/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_gux_epemguuxux/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_gux_taptamgcuxcx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_gux_taptamgduxdx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_gux_taptamgggux/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_gux_taptamguuxux/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_ssx_epemdbdxbx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_ssx_epemdcdxcx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_ssx_epemdsdxsx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_ssx_epemucuxcx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_ssx_epemuduxdx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_ssx_taptamdbdxbx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_ssx_taptamdcdxcx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_ssx_taptamdsdxsx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_ssx_taptamucuxcx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_ssx_taptamuduxdx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_uc_epemgguc/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_uc_epemucccx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_uc_epemudcdx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_uc_epemuucux/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_uc_taptamgguc/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_uc_taptamucccx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_uc_taptamudcdx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_uc_taptamuucux/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_ucx_epemggucx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_ucx_epemuccxcx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_ucx_epemuddxcx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_ucx_epemuuuxcx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_ucx_taptamggucx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_ucx_taptamuccxcx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_ucx_taptamuddxcx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_ucx_taptamuuuxcx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_ud_epemggud/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_ud_epemudccx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_ud_epemudddx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_ud_epemudssx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_ud_epemuudux/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_ud_taptamggud/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_ud_taptamudccx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_ud_taptamudddx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_ud_taptamudssx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_ud_taptamuudux/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_udx_epemggudx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_udx_epemucdxcx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_udx_epemuddxdx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_udx_epemusdxsx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_udx_epemuuuxdx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_udx_taptamggudx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_udx_taptamucdxcx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_udx_taptamuddxdx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_udx_taptamusdxsx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_udx_taptamuuuxdx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_uu_epemgguu/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_uu_epemuuccx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_uu_epemuuddx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_uu_epemuuuux/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_uu_taptamgguu/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_uu_taptamuuccx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_uu_taptamuuddx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_uu_taptamuuuux/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_uux_epemcccxcx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_uux_epemcscxsx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_uux_epemdcdxcx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_uux_epemdddxdx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_uux_epemdsdxsx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_uux_epemggccx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_uux_epemggddx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_uux_epemgggg/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_uux_epemgguux/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_uux_epemucuxcx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_uux_epemuduxdx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_uux_epemuuuxux/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_uux_taptamcccxcx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_uux_taptamcscxsx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_uux_taptamdcdxcx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_uux_taptamdddxdx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_uux_taptamdsdxsx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_uux_taptamggccx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_uux_taptamggddx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_uux_taptamgggg/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_uux_taptamgguux/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_uux_taptamucuxcx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_uux_taptamuduxdx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_uux_taptamuuuxux/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_uxcx_epemcuxcxcx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_uxcx_epemduxdxcx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_uxcx_epemgguxcx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_uxcx_epemuuxuxcx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_uxcx_taptamcuxcxcx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_uxcx_taptamduxdxcx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_uxcx_taptamgguxcx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_uxcx_taptamuuxuxcx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_uxdx_epemcuxdxcx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_uxdx_epemduxdxdx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_uxdx_epemgguxdx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_uxdx_epemsuxdxsx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_uxdx_epemuuxuxdx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_uxdx_taptamcuxdxcx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_uxdx_taptamduxdxdx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_uxdx_taptamgguxdx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_uxdx_taptamsuxdxsx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_uxdx_taptamuuxuxdx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_uxux_epemcuxuxcx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_uxux_epemduxuxdx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_uxux_epemgguxux/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_uxux_epemuuxuxux/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_uxux_taptamcuxuxcx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_uxux_taptamduxuxdx/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_uxux_taptamgguxux/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +lrwxrwxrwx. 1 avalassi zg 35 Sep 15 07:39 madevent/SubProcesses/P0_uxux_taptamuuxuxux/madevent -> build.sse4_d_inl0_hrd0/madevent_cpp +Now generating 500 events with random seed 21 and granularity 1 +************************************************************ +* * +* W E L C O M E to * +* M A D G R A P H 5 _ a M C @ N L O * +* M A D E V E N T * +* * +* * * * +* * * * * * +* * * * * 5 * * * * * +* * * * * * +* * * * +* * +* VERSION 3.5.3_lo_vect * +* * +* The MadGraph5_aMC@NLO Development Team - Find us at * +* https://server06.fynu.ucl.ac.be/projects/madgraph * +* * +* Type 'help' for in-line help. * +* * +************************************************************ +INFO: load configuration from /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/Cards/me5_configuration.txt +INFO: load configuration from /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/Cards/me5_configuration.txt +Using default text editor "vi". Set another one in ./input/mg5_configuration.txt +Using default eps viewer "evince". Set another one in ./input/mg5_configuration.txt +Using default web browser "firefox". Set another one in ./input/mg5_configuration.txt +WRITE GRIDCARD /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent +__CUDACPP_DEBUG: GridPackCmd.__init__ will call self.launch +__CUDACPP_DEBUG: GridPackCmd.launch starting +__CUDACPP_DEBUG: GridPackCmd.launch (1) restore_data +generate 500 events +P0_gg_epemgguux +P0_gg_epemggddx +P0_gg_taptamgguux +P0_gg_taptamggddx +P0_gu_epemgggu +P0_gd_epemgggd +P0_gux_epemgggux +P0_gdx_epemgggdx +P0_gu_taptamgggu +P0_gd_taptamgggd +P0_gux_taptamgggux +P0_gdx_taptamgggdx +P0_uux_epemgggg +P0_ddx_epemgggg +P0_uux_taptamgggg +P0_ddx_taptamgggg +P0_gg_epemuuuxux +P0_gg_epemdddxdx +P0_gg_epemuduxdx +P0_gg_epemucuxcx +P0_gg_epemdsdxsx +P0_gg_taptamuuuxux +P0_gg_taptamdddxdx +P0_gg_taptamuduxdx +P0_gg_taptamucuxcx +P0_gg_taptamdsdxsx +P0_gu_epemguuux +P0_gd_epemgdddx +P0_gux_epemguuxux +P0_gdx_epemgddxdx +P0_gu_epemguddx +P0_gu_epemguccx +P0_gd_epemgudux +P0_gd_epemgdssx +P0_gux_epemgduxdx +P0_gux_epemgcuxcx +P0_gdx_epemguuxdx +P0_gdx_epemgsdxsx +P0_gu_taptamguuux +P0_gd_taptamgdddx +P0_gux_taptamguuxux +P0_gdx_taptamgddxdx +P0_gu_taptamguddx +P0_gu_taptamguccx +P0_gd_taptamgudux +P0_gd_taptamgdssx +P0_gux_taptamgduxdx +P0_gux_taptamgcuxcx +P0_gdx_taptamguuxdx +P0_gdx_taptamgsdxsx +P0_uu_epemgguu +P0_uux_epemgguux +P0_dd_epemggdd +P0_ddx_epemggddx +P0_uxux_epemgguxux +P0_dxdx_epemggdxdx +P0_ud_epemggud +P0_uc_epemgguc +P0_uux_epemggddx +P0_uux_epemggccx +P0_udx_epemggudx +P0_ucx_epemggucx +P0_dc_epemggdc +P0_ds_epemggds +P0_dux_epemggdux +P0_ddx_epemgguux +P0_ddx_epemggssx +P0_dsx_epemggdsx +P0_uxdx_epemgguxdx +P0_uxcx_epemgguxcx +P0_dxcx_epemggdxcx +P0_dxsx_epemggdxsx +P0_uu_taptamgguu +P0_uux_taptamgguux +P0_dd_taptamggdd +P0_ddx_taptamggddx +P0_uxux_taptamgguxux +P0_dxdx_taptamggdxdx +P0_ud_taptamggud +P0_uc_taptamgguc +P0_uux_taptamggddx +P0_uux_taptamggccx +P0_udx_taptamggudx +P0_ucx_taptamggucx +P0_dc_taptamggdc +P0_ds_taptamggds +P0_dux_taptamggdux +P0_ddx_taptamgguux +P0_ddx_taptamggssx +P0_dsx_taptamggdsx +P0_uxdx_taptamgguxdx +P0_uxcx_taptamgguxcx +P0_dxcx_taptamggdxcx +P0_dxsx_taptamggdxsx +P0_uu_epemuuuux +P0_uux_epemuuuxux +P0_dd_epemddddx +P0_ddx_epemdddxdx +P0_uxux_epemuuxuxux +P0_dxdx_epemddxdxdx +P0_uu_epemuuddx +P0_uu_epemuuccx +P0_ud_epemuudux +P0_ud_epemudddx +P0_uc_epemuucux +P0_uc_epemucccx +P0_uux_epemuduxdx +P0_uux_epemucuxcx +P0_uux_epemdddxdx +P0_uux_epemcccxcx +P0_udx_epemuuuxdx +P0_udx_epemuddxdx +P0_ucx_epemuuuxcx +P0_ucx_epemuccxcx +P0_dd_epemuddux +P0_dd_epemddssx +P0_dc_epemddcdx +P0_dc_epemdcccx +P0_ds_epemddsdx +P0_ds_epemdsssx +P0_dux_epemuduxux +P0_dux_epemdduxdx +P0_ddx_epemuuuxux +P0_ddx_epemuduxdx +P0_ddx_epemdcdxcx +P0_ddx_epemdsdxsx +P0_ddx_epemsssxsx +P0_dsx_epemdddxsx +P0_dsx_epemdssxsx +P0_ccx_epemucuxcx +P0_ccx_epemdcdxcx +P0_ssx_epemdsdxsx +P0_uxux_epemduxuxdx +P0_uxux_epemcuxuxcx +P0_uxdx_epemuuxuxdx +P0_uxdx_epemduxdxdx +P0_uxcx_epemuuxuxcx +P0_uxcx_epemcuxcxcx +P0_dxdx_epemuuxdxdx +P0_dxdx_epemsdxdxsx +P0_dxcx_epemddxdxcx +P0_dxcx_epemcdxcxcx +P0_dxsx_epemddxdxsx +P0_dxsx_epemsdxsxsx +P0_ud_epemudccx +P0_ud_epemudssx +P0_uc_epemudcdx +P0_uux_epemdcdxcx +P0_uux_epemdsdxsx +P0_uux_epemcscxsx +P0_udx_epemucdxcx +P0_udx_epemusdxsx +P0_ucx_epemuddxcx +P0_dc_epemudcux +P0_dc_epemdcssx +P0_ds_epemudsux +P0_ds_epemdsbbx +P0_dux_epemdcuxcx +P0_dux_epemdsuxsx +P0_ddx_epemucuxcx +P0_ddx_epemusuxsx +P0_ddx_epemsbsxbx +P0_dsx_epemuduxsx +P0_dsx_epemdbsxbx +P0_ccx_epemuduxdx +P0_ccx_epemdsdxsx +P0_ssx_epemuduxdx +P0_ssx_epemucuxcx +P0_ssx_epemdcdxcx +P0_ssx_epemdbdxbx +P0_bbx_epemdsdxsx +P0_uxdx_epemcuxdxcx +P0_uxdx_epemsuxdxsx +P0_uxcx_epemduxdxcx +P0_dxcx_epemuuxdxcx +P0_dxcx_epemsdxcxsx +P0_dxsx_epemuuxdxsx +P0_dxsx_epembdxsxbx +P0_uu_taptamuuuux +P0_uux_taptamuuuxux +P0_dd_taptamddddx +P0_ddx_taptamdddxdx +P0_uxux_taptamuuxuxux +P0_dxdx_taptamddxdxdx +P0_uu_taptamuuddx +P0_uu_taptamuuccx +P0_ud_taptamuudux +P0_ud_taptamudddx +P0_uc_taptamuucux +P0_uc_taptamucccx +P0_uux_taptamuduxdx +P0_uux_taptamucuxcx +P0_uux_taptamdddxdx +P0_uux_taptamcccxcx +P0_udx_taptamuuuxdx +P0_udx_taptamuddxdx +P0_ucx_taptamuuuxcx +P0_ucx_taptamuccxcx +P0_dd_taptamuddux +P0_dd_taptamddssx +P0_dc_taptamddcdx +P0_dc_taptamdcccx +P0_ds_taptamddsdx +P0_ds_taptamdsssx +P0_dux_taptamuduxux +P0_dux_taptamdduxdx +P0_ddx_taptamuuuxux +P0_ddx_taptamuduxdx +P0_ddx_taptamdcdxcx +P0_ddx_taptamdsdxsx +P0_ddx_taptamsssxsx +P0_dsx_taptamdddxsx +P0_dsx_taptamdssxsx +P0_ccx_taptamucuxcx +P0_ccx_taptamdcdxcx +P0_ssx_taptamdsdxsx +P0_uxux_taptamduxuxdx +P0_uxux_taptamcuxuxcx +P0_uxdx_taptamuuxuxdx +P0_uxdx_taptamduxdxdx +P0_uxcx_taptamuuxuxcx +P0_uxcx_taptamcuxcxcx +P0_dxdx_taptamuuxdxdx +P0_dxdx_taptamsdxdxsx +P0_dxcx_taptamddxdxcx +P0_dxcx_taptamcdxcxcx +P0_dxsx_taptamddxdxsx +P0_dxsx_taptamsdxsxsx +P0_ud_taptamudccx +P0_ud_taptamudssx +P0_uc_taptamudcdx +P0_uux_taptamdcdxcx +P0_uux_taptamdsdxsx +P0_uux_taptamcscxsx +P0_udx_taptamucdxcx +P0_udx_taptamusdxsx +P0_ucx_taptamuddxcx +P0_dc_taptamudcux +P0_dc_taptamdcssx +P0_ds_taptamudsux +P0_ds_taptamdsbbx +P0_dux_taptamdcuxcx +P0_dux_taptamdsuxsx +P0_ddx_taptamucuxcx +P0_ddx_taptamusuxsx +P0_ddx_taptamsbsxbx +P0_dsx_taptamuduxsx +P0_dsx_taptamdbsxbx +P0_ccx_taptamuduxdx +P0_ccx_taptamdsdxsx +P0_ssx_taptamuduxdx +P0_ssx_taptamucuxcx +P0_ssx_taptamdcdxcx +P0_ssx_taptamdbdxbx +P0_bbx_taptamdsdxsx +P0_uxdx_taptamcuxdxcx +P0_uxdx_taptamsuxdxsx +P0_uxcx_taptamduxdxcx +P0_dxcx_taptamuuxdxcx +P0_dxcx_taptamsdxcxsx +P0_dxsx_taptamuuxdxsx +P0_dxsx_taptambdxsxbx +__CUDACPP_DEBUG: GridPackCmd.launch (2) refine4grid +__CUDACPP_DEBUG: GridPackCmd.refine4grid starting +Using random number seed offset = 21 +INFO: Generating 500.0 unweighted events. +__CUDACPP_DEBUG: gen_ximprove.launch starting +sum of cpu time of last step: 3 days,23h23m17s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event starting +INFO: need to improve 477 channels +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dxdx_taptamuuxdxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dxdx_taptamuuxdxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dxdx_taptamuuxdxdx/ajob1' in 7.7789 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dxdx_taptamuuxdxdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dxdx_taptamuuxdxdx/G21/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.7420s + [COUNTERS] Fortran Overhead ( 0 ) : 3.8195s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.9032s for 16384 events => throughput is 4.20E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0192s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dxdx_taptamuuxdxdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamguuxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamguuxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamguuxdx/ajob1' in 25.0440 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamguuxdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamguuxdx/G43/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 11.9900s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9983s + [COUNTERS] CudaCpp MEs ( 2 ) : 9.9562s for 16384 events => throughput is 1.65E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0355s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamguuxdx/G335/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.0042s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7424s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.2228s for 16384 events => throughput is 1.60E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0390s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamguuxdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemuuuxux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemuuuxux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemuuuxux/ajob1' in 18.8629 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemuuuxux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemuuuxux/G689/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 18.8045s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5177s + [COUNTERS] CudaCpp MEs ( 2 ) : 16.2236s for 16384 events => throughput is 1.01E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0632s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemuuuxux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamgguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamgguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamgguux/ajob1' in 522.0750 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamgguux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamgguux/G596/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 56.6518s + [COUNTERS] Fortran Overhead ( 0 ) : 3.3921s + [COUNTERS] CudaCpp MEs ( 2 ) : 53.1745s for 16384 events => throughput is 3.08E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0853s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamgguux/G304/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 56.3941s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9192s + [COUNTERS] CudaCpp MEs ( 2 ) : 53.3897s for 16384 events => throughput is 3.07E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0852s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamgguux/G128/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 56.2510s + [COUNTERS] Fortran Overhead ( 0 ) : 3.0273s + [COUNTERS] CudaCpp MEs ( 2 ) : 53.1376s for 16384 events => throughput is 3.08E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0861s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamgguux/G115/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 59.9559s + [COUNTERS] Fortran Overhead ( 0 ) : 6.6658s + [COUNTERS] CudaCpp MEs ( 2 ) : 53.2035s for 16384 events => throughput is 3.08E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0865s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamgguux/G599/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.0830s + [COUNTERS] Fortran Overhead ( 0 ) : 3.8244s + [COUNTERS] CudaCpp MEs ( 2 ) : 53.1725s for 16384 events => throughput is 3.08E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0861s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamgguux/G351/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 59.3910s + [COUNTERS] Fortran Overhead ( 0 ) : 6.0941s + [COUNTERS] CudaCpp MEs ( 2 ) : 53.2106s for 16384 events => throughput is 3.08E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0863s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamgguux/G111/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 55.5075s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3455s + [COUNTERS] CudaCpp MEs ( 2 ) : 53.0781s for 16384 events => throughput is 3.09E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0840s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamgguux/G354/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 60.0644s + [COUNTERS] Fortran Overhead ( 0 ) : 6.6200s + [COUNTERS] CudaCpp MEs ( 2 ) : 53.3592s for 16384 events => throughput is 3.07E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0852s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamgguux/G97/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 60.5804s + [COUNTERS] Fortran Overhead ( 0 ) : 7.2577s + [COUNTERS] CudaCpp MEs ( 2 ) : 53.2378s for 16384 events => throughput is 3.08E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0849s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamgguux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uxdx_epemsuxdxsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uxdx_epemsuxdxsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uxdx_epemsuxdxsx/ajob1' in 2.9467 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uxdx_epemsuxdxsx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uxdx_epemsuxdxsx/G74/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.9009s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1593s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.7301s for 16384 events => throughput is 2.24E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0115s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uxdx_epemsuxdxsx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemdddxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemdddxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemdddxdx/ajob1' in 37.8809 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemdddxdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemdddxdx/G513/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 18.7797s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4675s + [COUNTERS] CudaCpp MEs ( 2 ) : 16.2495s for 16384 events => throughput is 1.01E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0627s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemdddxdx/G180/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 19.0523s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7776s + [COUNTERS] CudaCpp MEs ( 2 ) : 16.2118s for 16384 events => throughput is 1.01E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0630s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemdddxdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgddxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgddxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgddxdx/ajob1' in 19.5559 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgddxdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgddxdx/G515/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 19.5110s + [COUNTERS] Fortran Overhead ( 0 ) : 3.0814s + [COUNTERS] CudaCpp MEs ( 2 ) : 16.3664s for 16384 events => throughput is 1.00E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0632s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgddxdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguddx/ajob1' in 113.8055 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguddx/G104/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 11.6120s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6225s + [COUNTERS] CudaCpp MEs ( 2 ) : 9.9530s for 16384 events => throughput is 1.65E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0364s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguddx/G120/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 12.1696s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1568s + [COUNTERS] CudaCpp MEs ( 2 ) : 9.9767s for 16384 events => throughput is 1.64E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0361s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguddx/G305/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 15.7944s + [COUNTERS] Fortran Overhead ( 0 ) : 5.8556s + [COUNTERS] CudaCpp MEs ( 2 ) : 9.9019s for 16384 events => throughput is 1.65E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0369s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguddx/G101/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 11.4899s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4538s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.0000s for 16384 events => throughput is 1.64E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0361s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguddx/G27/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 11.8829s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9009s + [COUNTERS] CudaCpp MEs ( 2 ) : 9.9462s for 16384 events => throughput is 1.65E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0358s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguddx/G273/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 12.2884s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2162s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.0356s for 16384 events => throughput is 1.63E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0367s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguddx/G349/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 11.4782s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3992s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.0429s for 16384 events => throughput is 1.63E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0362s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguddx/G379/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 11.7390s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7759s + [COUNTERS] CudaCpp MEs ( 2 ) : 9.9268s for 16384 events => throughput is 1.65E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0363s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguddx/G256/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 15.1595s + [COUNTERS] Fortran Overhead ( 0 ) : 5.1739s + [COUNTERS] CudaCpp MEs ( 2 ) : 9.9494s for 16384 events => throughput is 1.65E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0362s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemguuxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemguuxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemguuxdx/ajob1' in 29.1608 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemguuxdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemguuxdx/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 11.3763s + [COUNTERS] Fortran Overhead ( 0 ) : 6.1262s + [COUNTERS] CudaCpp MEs ( 2 ) : 5.2132s for 16384 events => throughput is 3.14E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0370s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemguuxdx/G68/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 10.4307s + [COUNTERS] Fortran Overhead ( 0 ) : 5.2417s + [COUNTERS] CudaCpp MEs ( 2 ) : 5.1531s for 16384 events => throughput is 3.18E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0360s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemguuxdx/G355/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.2725s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0646s + [COUNTERS] CudaCpp MEs ( 2 ) : 5.1711s for 16384 events => throughput is 3.17E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0368s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemguuxdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/ajob1' in 1309.9594 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G14/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 55.9621s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0644s + [COUNTERS] CudaCpp MEs ( 2 ) : 53.8073s for 16384 events => throughput is 3.04E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0903s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G56/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 56.5693s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5155s + [COUNTERS] CudaCpp MEs ( 2 ) : 53.9685s for 16384 events => throughput is 3.04E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0854s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G517/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 60.0422s + [COUNTERS] Fortran Overhead ( 0 ) : 6.1633s + [COUNTERS] CudaCpp MEs ( 2 ) : 53.7931s for 16384 events => throughput is 3.05E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0858s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G52/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 56.6900s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7812s + [COUNTERS] CudaCpp MEs ( 2 ) : 53.8237s for 16384 events => throughput is 3.04E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0850s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G304/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 56.5558s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6981s + [COUNTERS] CudaCpp MEs ( 2 ) : 53.7694s for 16384 events => throughput is 3.05E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0883s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G3/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 58.8894s + [COUNTERS] Fortran Overhead ( 0 ) : 4.8699s + [COUNTERS] CudaCpp MEs ( 2 ) : 53.9317s for 16384 events => throughput is 3.04E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0879s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G23/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 55.6295s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7234s + [COUNTERS] CudaCpp MEs ( 2 ) : 53.8201s for 16384 events => throughput is 3.04E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0859s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G33/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 56.3819s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1718s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.1226s for 16384 events => throughput is 3.03E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0875s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G53/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 60.5171s + [COUNTERS] Fortran Overhead ( 0 ) : 6.6874s + [COUNTERS] CudaCpp MEs ( 2 ) : 53.7437s for 16384 events => throughput is 3.05E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0860s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G495/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 55.4742s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6097s + [COUNTERS] CudaCpp MEs ( 2 ) : 53.7744s for 16384 events => throughput is 3.05E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0901s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G68/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 56.1924s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1863s + [COUNTERS] CudaCpp MEs ( 2 ) : 53.9214s for 16384 events => throughput is 3.04E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0847s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G49/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 56.1117s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0669s + [COUNTERS] CudaCpp MEs ( 2 ) : 53.9594s for 16384 events => throughput is 3.04E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0854s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G524/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 56.0151s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1716s + [COUNTERS] CudaCpp MEs ( 2 ) : 53.7573s for 16384 events => throughput is 3.05E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0863s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G65/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 56.1337s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0813s + [COUNTERS] CudaCpp MEs ( 2 ) : 53.9675s for 16384 events => throughput is 3.04E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0849s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G303/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.0976s + [COUNTERS] Fortran Overhead ( 0 ) : 3.0029s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.0093s for 16384 events => throughput is 3.03E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0855s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G55/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 56.4973s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6889s + [COUNTERS] CudaCpp MEs ( 2 ) : 53.7225s for 16384 events => throughput is 3.05E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0858s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G34/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 55.6931s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8100s + [COUNTERS] CudaCpp MEs ( 2 ) : 53.7973s for 16384 events => throughput is 3.05E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0858s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G509/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 61.6207s + [COUNTERS] Fortran Overhead ( 0 ) : 7.6850s + [COUNTERS] CudaCpp MEs ( 2 ) : 53.8493s for 16384 events => throughput is 3.04E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0864s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G512/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 55.7981s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8057s + [COUNTERS] CudaCpp MEs ( 2 ) : 53.9055s for 16384 events => throughput is 3.04E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0869s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G511/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 56.2713s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4246s + [COUNTERS] CudaCpp MEs ( 2 ) : 53.7616s for 16384 events => throughput is 3.05E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0851s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 56.6528s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8315s + [COUNTERS] CudaCpp MEs ( 2 ) : 53.7362s for 16384 events => throughput is 3.05E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0850s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G63/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 56.3644s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1070s + [COUNTERS] CudaCpp MEs ( 2 ) : 54.1705s for 16384 events => throughput is 3.02E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0870s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G48/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 56.3386s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3612s + [COUNTERS] CudaCpp MEs ( 2 ) : 53.8890s for 16384 events => throughput is 3.04E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0885s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/ajob1' in 741.0699 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G122/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 29.5499s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6956s + [COUNTERS] CudaCpp MEs ( 2 ) : 26.7682s for 16384 events => throughput is 6.12E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0860s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G108/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 29.3214s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3734s + [COUNTERS] CudaCpp MEs ( 2 ) : 26.8582s for 16384 events => throughput is 6.10E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0898s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G105/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 29.5353s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4865s + [COUNTERS] CudaCpp MEs ( 2 ) : 26.9622s for 16384 events => throughput is 6.08E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0866s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G304/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 30.0043s + [COUNTERS] Fortran Overhead ( 0 ) : 3.0555s + [COUNTERS] CudaCpp MEs ( 2 ) : 26.8619s for 16384 events => throughput is 6.10E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0869s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G301/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 34.4501s + [COUNTERS] Fortran Overhead ( 0 ) : 7.5840s + [COUNTERS] CudaCpp MEs ( 2 ) : 26.7804s for 16384 events => throughput is 6.12E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0857s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G128/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 30.1509s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1443s + [COUNTERS] CudaCpp MEs ( 2 ) : 26.9207s for 16384 events => throughput is 6.09E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0858s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G100/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 34.4998s + [COUNTERS] Fortran Overhead ( 0 ) : 7.5479s + [COUNTERS] CudaCpp MEs ( 2 ) : 26.8658s for 16384 events => throughput is 6.10E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0861s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G120/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 29.3447s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4313s + [COUNTERS] CudaCpp MEs ( 2 ) : 26.8265s for 16384 events => throughput is 6.11E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0869s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G302/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 38.0071s + [COUNTERS] Fortran Overhead ( 0 ) : 11.0798s + [COUNTERS] CudaCpp MEs ( 2 ) : 26.8357s for 16384 events => throughput is 6.11E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0916s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G299/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 29.5748s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6725s + [COUNTERS] CudaCpp MEs ( 2 ) : 26.8164s for 16384 events => throughput is 6.11E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0859s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G169/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 33.0498s + [COUNTERS] Fortran Overhead ( 0 ) : 6.0124s + [COUNTERS] CudaCpp MEs ( 2 ) : 26.9515s for 16384 events => throughput is 6.08E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0858s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G333/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 38.5782s + [COUNTERS] Fortran Overhead ( 0 ) : 11.5738s + [COUNTERS] CudaCpp MEs ( 2 ) : 26.9141s for 16384 events => throughput is 6.09E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0903s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G594/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 36.0981s + [COUNTERS] Fortran Overhead ( 0 ) : 9.2142s + [COUNTERS] CudaCpp MEs ( 2 ) : 26.7976s for 16384 events => throughput is 6.11E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0862s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G678/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 33.4939s + [COUNTERS] Fortran Overhead ( 0 ) : 6.3304s + [COUNTERS] CudaCpp MEs ( 2 ) : 27.0787s for 16384 events => throughput is 6.05E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0848s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G127/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 29.6365s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6857s + [COUNTERS] CudaCpp MEs ( 2 ) : 26.8651s for 16384 events => throughput is 6.10E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0857s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G126/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 36.3380s + [COUNTERS] Fortran Overhead ( 0 ) : 9.4607s + [COUNTERS] CudaCpp MEs ( 2 ) : 26.7916s for 16384 events => throughput is 6.12E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0857s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G598/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 29.1693s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1609s + [COUNTERS] CudaCpp MEs ( 2 ) : 26.9233s for 16384 events => throughput is 6.09E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0852s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G512/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 29.8865s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1842s + [COUNTERS] CudaCpp MEs ( 2 ) : 27.6135s for 16384 events => throughput is 5.93E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0888s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G103/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 35.8486s + [COUNTERS] Fortran Overhead ( 0 ) : 8.9752s + [COUNTERS] CudaCpp MEs ( 2 ) : 26.7869s for 16384 events => throughput is 6.12E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0866s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G107/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 29.2904s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3319s + [COUNTERS] CudaCpp MEs ( 2 ) : 26.8715s for 16384 events => throughput is 6.10E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0870s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G358/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 30.6463s + [COUNTERS] Fortran Overhead ( 0 ) : 3.7519s + [COUNTERS] CudaCpp MEs ( 2 ) : 26.8079s for 16384 events => throughput is 6.11E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0866s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G97/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 34.2559s + [COUNTERS] Fortran Overhead ( 0 ) : 7.4283s + [COUNTERS] CudaCpp MEs ( 2 ) : 26.7423s for 16384 events => throughput is 6.13E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0852s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G179/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 29.8872s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7427s + [COUNTERS] CudaCpp MEs ( 2 ) : 27.0590s for 16384 events => throughput is 6.05E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0855s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_epemudcdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_epemudcdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_epemudcdx/ajob1' in 2.4918 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_epemudcdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_epemudcdx/G42/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.4605s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7220s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.7271s for 16384 events => throughput is 2.25E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0113s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_epemudcdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ds_taptamggds/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ds_taptamggds/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ds_taptamggds/ajob1' in 26.3105 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ds_taptamggds/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ds_taptamggds/G17/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.5589s + [COUNTERS] Fortran Overhead ( 0 ) : 4.6473s + [COUNTERS] CudaCpp MEs ( 2 ) : 9.8751s for 16384 events => throughput is 1.66E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0365s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ds_taptamggds/G110/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 11.6815s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6625s + [COUNTERS] CudaCpp MEs ( 2 ) : 9.9835s for 16384 events => throughput is 1.64E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0355s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ds_taptamggds/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_epemggdux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_epemggdux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_epemggdux/ajob1' in 33.8044 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_epemggdux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_epemggdux/G208/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.2215s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9946s + [COUNTERS] CudaCpp MEs ( 2 ) : 5.1903s for 16384 events => throughput is 3.16E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0366s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_epemggdux/G192/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.6935s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3988s + [COUNTERS] CudaCpp MEs ( 2 ) : 5.2565s for 16384 events => throughput is 3.12E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0383s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_epemggdux/G380/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.7573s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6111s + [COUNTERS] CudaCpp MEs ( 2 ) : 5.1106s for 16384 events => throughput is 3.21E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0357s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_epemggdux/G127/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.4018s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2900s + [COUNTERS] CudaCpp MEs ( 2 ) : 5.0748s for 16384 events => throughput is 3.23E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0370s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_epemggdux/G223/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.3950s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2066s + [COUNTERS] CudaCpp MEs ( 2 ) : 5.1526s for 16384 events => throughput is 3.18E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0358s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_epemggdux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_taptamgguc/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_taptamgguc/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_taptamgguc/ajob1' in 40.8191 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_taptamgguc/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_taptamgguc/G20/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.2858s + [COUNTERS] Fortran Overhead ( 0 ) : 4.4259s + [COUNTERS] CudaCpp MEs ( 2 ) : 9.8238s for 16384 events => throughput is 1.67E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0362s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_taptamgguc/G110/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 11.1432s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3105s + [COUNTERS] CudaCpp MEs ( 2 ) : 9.7973s for 16384 events => throughput is 1.67E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0355s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_taptamgguc/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 15.3206s + [COUNTERS] Fortran Overhead ( 0 ) : 5.4461s + [COUNTERS] CudaCpp MEs ( 2 ) : 9.8385s for 16384 events => throughput is 1.67E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0360s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_taptamgguc/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uxcx_taptamgguxcx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uxcx_taptamgguxcx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uxcx_taptamgguxcx/ajob1' in 12.6256 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uxcx_taptamgguxcx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uxcx_taptamgguxcx/G187/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 12.5735s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5411s + [COUNTERS] CudaCpp MEs ( 2 ) : 9.9959s for 16384 events => throughput is 1.64E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0365s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uxcx_taptamgguxcx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/ajob1' in 782.3606 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G499/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 28.9296s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2052s + [COUNTERS] CudaCpp MEs ( 2 ) : 26.6334s for 16384 events => throughput is 6.15E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0911s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G56/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 29.6711s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9035s + [COUNTERS] CudaCpp MEs ( 2 ) : 26.6811s for 16384 events => throughput is 6.14E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0865s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G39/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 29.4492s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6934s + [COUNTERS] CudaCpp MEs ( 2 ) : 26.6696s for 16384 events => throughput is 6.14E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0862s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G64/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 29.4042s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6336s + [COUNTERS] CudaCpp MEs ( 2 ) : 26.6842s for 16384 events => throughput is 6.14E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0864s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G13/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 28.3917s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6417s + [COUNTERS] CudaCpp MEs ( 2 ) : 26.6615s for 16384 events => throughput is 6.15E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0884s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G33/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 28.4626s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7701s + [COUNTERS] CudaCpp MEs ( 2 ) : 26.6051s for 16384 events => throughput is 6.16E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0874s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G43/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 33.7317s + [COUNTERS] Fortran Overhead ( 0 ) : 7.0237s + [COUNTERS] CudaCpp MEs ( 2 ) : 26.6216s for 16384 events => throughput is 6.15E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0865s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G68/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 29.6644s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9113s + [COUNTERS] CudaCpp MEs ( 2 ) : 26.6666s for 16384 events => throughput is 6.14E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0865s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G49/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 28.8161s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0883s + [COUNTERS] CudaCpp MEs ( 2 ) : 26.6430s for 16384 events => throughput is 6.15E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0848s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G25/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 32.3933s + [COUNTERS] Fortran Overhead ( 0 ) : 5.7360s + [COUNTERS] CudaCpp MEs ( 2 ) : 26.5710s for 16384 events => throughput is 6.17E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0863s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G523/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 29.4657s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7715s + [COUNTERS] CudaCpp MEs ( 2 ) : 26.6090s for 16384 events => throughput is 6.16E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0852s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G395/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 29.2199s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4234s + [COUNTERS] CudaCpp MEs ( 2 ) : 26.7094s for 16384 events => throughput is 6.13E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0870s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G524/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 28.5502s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9075s + [COUNTERS] CudaCpp MEs ( 2 ) : 26.5571s for 16384 events => throughput is 6.17E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0857s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G65/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 28.8566s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0576s + [COUNTERS] CudaCpp MEs ( 2 ) : 26.7093s for 16384 events => throughput is 6.13E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0897s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G609/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 29.7133s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9224s + [COUNTERS] CudaCpp MEs ( 2 ) : 26.7055s for 16384 events => throughput is 6.14E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0854s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G611/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 29.9722s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2964s + [COUNTERS] CudaCpp MEs ( 2 ) : 26.5901s for 16384 events => throughput is 6.16E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0857s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G1/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 31.2892s + [COUNTERS] Fortran Overhead ( 0 ) : 4.5835s + [COUNTERS] CudaCpp MEs ( 2 ) : 26.6199s for 16384 events => throughput is 6.15E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0857s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G55/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 29.2308s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5864s + [COUNTERS] CudaCpp MEs ( 2 ) : 26.5591s for 16384 events => throughput is 6.17E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0853s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G34/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 28.8646s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1679s + [COUNTERS] CudaCpp MEs ( 2 ) : 26.6094s for 16384 events => throughput is 6.16E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0874s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G26/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 32.9707s + [COUNTERS] Fortran Overhead ( 0 ) : 6.2500s + [COUNTERS] CudaCpp MEs ( 2 ) : 26.6356s for 16384 events => throughput is 6.15E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0850s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G58/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 32.4494s + [COUNTERS] Fortran Overhead ( 0 ) : 5.6762s + [COUNTERS] CudaCpp MEs ( 2 ) : 26.6881s for 16384 events => throughput is 6.14E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0850s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G2/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 31.8490s + [COUNTERS] Fortran Overhead ( 0 ) : 5.0919s + [COUNTERS] CudaCpp MEs ( 2 ) : 26.6722s for 16384 events => throughput is 6.14E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0849s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G512/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 28.7854s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0855s + [COUNTERS] CudaCpp MEs ( 2 ) : 26.6149s for 16384 events => throughput is 6.16E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0849s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G51/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 29.4662s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6853s + [COUNTERS] CudaCpp MEs ( 2 ) : 26.6944s for 16384 events => throughput is 6.14E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0864s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G29/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 33.5099s + [COUNTERS] Fortran Overhead ( 0 ) : 6.6230s + [COUNTERS] CudaCpp MEs ( 2 ) : 26.7994s for 16384 events => throughput is 6.11E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0875s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G48/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 28.7223s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0550s + [COUNTERS] CudaCpp MEs ( 2 ) : 26.5812s for 16384 events => throughput is 6.16E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0861s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/ajob1' in 104.1814 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/G140/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.1201s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8359s + [COUNTERS] CudaCpp MEs ( 2 ) : 5.2487s for 16384 events => throughput is 3.12E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0356s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/G114/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 11.6804s + [COUNTERS] Fortran Overhead ( 0 ) : 6.4206s + [COUNTERS] CudaCpp MEs ( 2 ) : 5.2234s for 16384 events => throughput is 3.14E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0365s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/G49/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 12.5067s + [COUNTERS] Fortran Overhead ( 0 ) : 7.2351s + [COUNTERS] CudaCpp MEs ( 2 ) : 5.2357s for 16384 events => throughput is 3.13E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0359s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/G120/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.2844s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0112s + [COUNTERS] CudaCpp MEs ( 2 ) : 5.2371s for 16384 events => throughput is 3.13E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0360s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/G262/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.0331s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7479s + [COUNTERS] CudaCpp MEs ( 2 ) : 5.2485s for 16384 events => throughput is 3.12E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0367s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/G27/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.1682s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8902s + [COUNTERS] CudaCpp MEs ( 2 ) : 5.2420s for 16384 events => throughput is 3.13E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0360s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/G44/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.8606s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5702s + [COUNTERS] CudaCpp MEs ( 2 ) : 5.2543s for 16384 events => throughput is 3.12E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0361s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/G96/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.6090s + [COUNTERS] Fortran Overhead ( 0 ) : 3.4294s + [COUNTERS] CudaCpp MEs ( 2 ) : 5.1438s for 16384 events => throughput is 3.19E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0358s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/G261/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.3849s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1015s + [COUNTERS] CudaCpp MEs ( 2 ) : 5.2457s for 16384 events => throughput is 3.12E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0377s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/G37/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 10.1154s + [COUNTERS] Fortran Overhead ( 0 ) : 4.8691s + [COUNTERS] CudaCpp MEs ( 2 ) : 5.2102s for 16384 events => throughput is 3.14E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0361s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/G28/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.0149s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7834s + [COUNTERS] CudaCpp MEs ( 2 ) : 5.1955s for 16384 events => throughput is 3.15E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0360s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 11.1413s + [COUNTERS] Fortran Overhead ( 0 ) : 5.8470s + [COUNTERS] CudaCpp MEs ( 2 ) : 5.2583s for 16384 events => throughput is 3.12E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0360s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_epemuduxsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_epemuduxsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_epemuduxsx/ajob1' in 3.3529 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_epemuduxsx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_epemuduxsx/G42/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.3213s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5782s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.7314s for 16384 events => throughput is 2.24E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0116s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_epemuduxsx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dd_epemggdd/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dd_epemggdd/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dd_epemggdd/ajob1' in 53.8963 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dd_epemggdd/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dd_epemggdd/G467/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 17.9563s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8222s + [COUNTERS] CudaCpp MEs ( 2 ) : 16.0704s for 16384 events => throughput is 1.02E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0637s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dd_epemggdd/G351/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 17.4466s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2500s + [COUNTERS] CudaCpp MEs ( 2 ) : 16.1334s for 16384 events => throughput is 1.02E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0632s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dd_epemggdd/G44/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 18.4020s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2120s + [COUNTERS] CudaCpp MEs ( 2 ) : 16.1268s for 16384 events => throughput is 1.02E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0632s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dd_epemggdd/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamusdxsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamusdxsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamusdxsx/ajob1' in 2.7162 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamusdxsx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamusdxsx/G92/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.6713s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2003s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.4593s for 16384 events => throughput is 1.12E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0116s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamusdxsx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemggddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemggddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemggddx/ajob1' in 114.7440 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemggddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemggddx/G175/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 18.3016s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0895s + [COUNTERS] CudaCpp MEs ( 2 ) : 16.1484s for 16384 events => throughput is 1.01E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0637s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemggddx/G303/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 18.8186s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7095s + [COUNTERS] CudaCpp MEs ( 2 ) : 16.0459s for 16384 events => throughput is 1.02E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0632s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemggddx/G111/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 17.9827s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7702s + [COUNTERS] CudaCpp MEs ( 2 ) : 16.1477s for 16384 events => throughput is 1.01E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0648s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemggddx/G219/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 19.2678s + [COUNTERS] Fortran Overhead ( 0 ) : 3.0145s + [COUNTERS] CudaCpp MEs ( 2 ) : 16.1898s for 16384 events => throughput is 1.01E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0635s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemggddx/G318/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 22.1529s + [COUNTERS] Fortran Overhead ( 0 ) : 5.9565s + [COUNTERS] CudaCpp MEs ( 2 ) : 16.1331s for 16384 events => throughput is 1.02E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0633s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemggddx/G143/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 18.0955s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9662s + [COUNTERS] CudaCpp MEs ( 2 ) : 16.0658s for 16384 events => throughput is 1.02E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0635s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemggddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemguuxux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemguuxux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemguuxux/ajob1' in 23.3129 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemguuxux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemguuxux/G615/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 23.2570s + [COUNTERS] Fortran Overhead ( 0 ) : 7.0220s + [COUNTERS] CudaCpp MEs ( 2 ) : 16.1693s for 16384 events => throughput is 1.01E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0657s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemguuxux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamgggg/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamgggg/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamgggg/ajob1' in 56.2103 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamgggg/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamgggg/G66/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 56.1550s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6385s + [COUNTERS] CudaCpp MEs ( 2 ) : 53.4316s for 16384 events => throughput is 3.07E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0849s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamgggg/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgduxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgduxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgduxdx/ajob1' in 12.5453 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgduxdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgduxdx/G44/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 12.5145s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2273s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.2482s for 16384 events => throughput is 1.60E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0390s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgduxdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgudux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgudux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgudux/ajob1' in 76.5293 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgudux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgudux/G39/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 11.1825s + [COUNTERS] Fortran Overhead ( 0 ) : 6.1128s + [COUNTERS] CudaCpp MEs ( 2 ) : 5.0339s for 16384 events => throughput is 3.25E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0358s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgudux/G366/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 9.8246s + [COUNTERS] Fortran Overhead ( 0 ) : 4.6917s + [COUNTERS] CudaCpp MEs ( 2 ) : 5.0967s for 16384 events => throughput is 3.21E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0362s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgudux/G259/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.7480s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4209s + [COUNTERS] CudaCpp MEs ( 2 ) : 5.2884s for 16384 events => throughput is 3.10E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0387s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgudux/G28/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.1993s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0895s + [COUNTERS] CudaCpp MEs ( 2 ) : 5.0742s for 16384 events => throughput is 3.23E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0356s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgudux/G32/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.4078s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4471s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.9248s for 16384 events => throughput is 3.33E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0359s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgudux/G61/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.5901s + [COUNTERS] Fortran Overhead ( 0 ) : 8.4184s + [COUNTERS] CudaCpp MEs ( 2 ) : 5.1332s for 16384 events => throughput is 3.19E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0384s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgudux/G267/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 11.0298s + [COUNTERS] Fortran Overhead ( 0 ) : 5.7992s + [COUNTERS] CudaCpp MEs ( 2 ) : 5.1915s for 16384 events => throughput is 3.16E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0391s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgudux/G144/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.3839s + [COUNTERS] Fortran Overhead ( 0 ) : 3.3943s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.9536s for 16384 events => throughput is 3.31E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0359s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgudux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemuudux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemuudux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemuudux/ajob1' in 29.3141 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemuudux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemuudux/G104/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 29.2851s + [COUNTERS] Fortran Overhead ( 0 ) : 19.8339s + [COUNTERS] CudaCpp MEs ( 2 ) : 9.4326s for 81920 events => throughput is 8.68E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0185s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemuudux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamggddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamggddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamggddx/ajob1' in 466.9264 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamggddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamggddx/G108/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 58.7152s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3097s + [COUNTERS] CudaCpp MEs ( 2 ) : 56.3145s for 16384 events => throughput is 2.91E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0911s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamggddx/G304/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 59.3363s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8338s + [COUNTERS] CudaCpp MEs ( 2 ) : 56.4134s for 16384 events => throughput is 2.90E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0891s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamggddx/G128/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 58.9884s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9590s + [COUNTERS] CudaCpp MEs ( 2 ) : 55.9403s for 16384 events => throughput is 2.93E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0892s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamggddx/G360/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 55.9640s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3524s + [COUNTERS] CudaCpp MEs ( 2 ) : 53.5262s for 16384 events => throughput is 3.06E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0853s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamggddx/G169/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 59.7094s + [COUNTERS] Fortran Overhead ( 0 ) : 6.1302s + [COUNTERS] CudaCpp MEs ( 2 ) : 53.4924s for 16384 events => throughput is 3.06E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0868s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamggddx/G176/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 58.1936s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0855s + [COUNTERS] CudaCpp MEs ( 2 ) : 56.0185s for 16384 events => throughput is 2.92E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0896s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamggddx/G303/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 59.0801s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8484s + [COUNTERS] CudaCpp MEs ( 2 ) : 56.1456s for 16384 events => throughput is 2.92E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0861s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamggddx/G513/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 56.7757s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1049s + [COUNTERS] CudaCpp MEs ( 2 ) : 53.5843s for 16384 events => throughput is 3.06E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0865s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamggddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguccx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguccx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguccx/ajob1' in 42.7528 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguccx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguccx/G283/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.7700s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1372s + [COUNTERS] CudaCpp MEs ( 2 ) : 5.5943s for 16384 events => throughput is 2.93E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0385s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguccx/G47/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.6504s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0636s + [COUNTERS] CudaCpp MEs ( 2 ) : 5.5483s for 16384 events => throughput is 2.95E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0385s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguccx/G35/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 12.3709s + [COUNTERS] Fortran Overhead ( 0 ) : 6.9249s + [COUNTERS] CudaCpp MEs ( 2 ) : 5.4082s for 16384 events => throughput is 3.03E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0377s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguccx/G373/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.3103s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7762s + [COUNTERS] CudaCpp MEs ( 2 ) : 5.4953s for 16384 events => throughput is 2.98E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0388s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguccx/G249/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.2274s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6011s + [COUNTERS] CudaCpp MEs ( 2 ) : 5.5888s for 16384 events => throughput is 2.93E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0374s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguccx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/ajob1' in 938.3150 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G14/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 28.6683s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0057s + [COUNTERS] CudaCpp MEs ( 2 ) : 26.5783s for 16384 events => throughput is 6.16E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0843s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G499/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 28.5762s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9360s + [COUNTERS] CudaCpp MEs ( 2 ) : 26.5554s for 16384 events => throughput is 6.17E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0848s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G56/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 29.3319s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7057s + [COUNTERS] CudaCpp MEs ( 2 ) : 26.5392s for 16384 events => throughput is 6.17E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0870s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G39/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 29.1248s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2631s + [COUNTERS] CudaCpp MEs ( 2 ) : 26.7764s for 16384 events => throughput is 6.12E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0853s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G52/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 29.6324s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7993s + [COUNTERS] CudaCpp MEs ( 2 ) : 26.7479s for 16384 events => throughput is 6.13E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0852s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G304/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 29.2304s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6141s + [COUNTERS] CudaCpp MEs ( 2 ) : 26.5267s for 16384 events => throughput is 6.18E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0896s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G64/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 29.1005s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3414s + [COUNTERS] CudaCpp MEs ( 2 ) : 26.6731s for 16384 events => throughput is 6.14E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0860s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G3/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 31.5017s + [COUNTERS] Fortran Overhead ( 0 ) : 4.9448s + [COUNTERS] CudaCpp MEs ( 2 ) : 26.4722s for 16384 events => throughput is 6.19E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0847s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G13/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 28.2768s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6374s + [COUNTERS] CudaCpp MEs ( 2 ) : 26.5541s for 16384 events => throughput is 6.17E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0854s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G33/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 28.7103s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1556s + [COUNTERS] CudaCpp MEs ( 2 ) : 26.4689s for 16384 events => throughput is 6.19E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0859s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G484/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 29.3668s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5898s + [COUNTERS] CudaCpp MEs ( 2 ) : 26.6917s for 16384 events => throughput is 6.14E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0853s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G68/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 28.9768s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2243s + [COUNTERS] CudaCpp MEs ( 2 ) : 26.6675s for 16384 events => throughput is 6.14E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0849s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G49/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 28.3942s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7697s + [COUNTERS] CudaCpp MEs ( 2 ) : 26.5390s for 16384 events => throughput is 6.17E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0854s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G40/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 29.4846s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8972s + [COUNTERS] CudaCpp MEs ( 2 ) : 26.5027s for 16384 events => throughput is 6.18E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0848s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G523/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 28.6877s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1077s + [COUNTERS] CudaCpp MEs ( 2 ) : 26.4922s for 16384 events => throughput is 6.18E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0878s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 28.2450s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5541s + [COUNTERS] CudaCpp MEs ( 2 ) : 26.6054s for 16384 events => throughput is 6.16E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0856s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G524/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 28.6400s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0932s + [COUNTERS] CudaCpp MEs ( 2 ) : 26.4615s for 16384 events => throughput is 6.19E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0853s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G65/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 28.8570s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1491s + [COUNTERS] CudaCpp MEs ( 2 ) : 26.6232s for 16384 events => throughput is 6.15E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0848s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G44/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 32.3891s + [COUNTERS] Fortran Overhead ( 0 ) : 5.7674s + [COUNTERS] CudaCpp MEs ( 2 ) : 26.5364s for 16384 events => throughput is 6.17E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0853s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 28.1287s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5085s + [COUNTERS] CudaCpp MEs ( 2 ) : 26.5334s for 16384 events => throughput is 6.17E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0868s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G303/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 29.5540s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9604s + [COUNTERS] CudaCpp MEs ( 2 ) : 26.5085s for 16384 events => throughput is 6.18E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0851s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G528/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 29.7115s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1182s + [COUNTERS] CudaCpp MEs ( 2 ) : 26.5070s for 16384 events => throughput is 6.18E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0862s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G55/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 30.0015s + [COUNTERS] Fortran Overhead ( 0 ) : 3.0895s + [COUNTERS] CudaCpp MEs ( 2 ) : 26.8233s for 16384 events => throughput is 6.11E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0887s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G298/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 29.6404s + [COUNTERS] Fortran Overhead ( 0 ) : 3.0293s + [COUNTERS] CudaCpp MEs ( 2 ) : 26.5234s for 16384 events => throughput is 6.18E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0877s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G34/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 28.3543s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6837s + [COUNTERS] CudaCpp MEs ( 2 ) : 26.5834s for 16384 events => throughput is 6.16E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0871s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G54/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 33.6062s + [COUNTERS] Fortran Overhead ( 0 ) : 7.0163s + [COUNTERS] CudaCpp MEs ( 2 ) : 26.5045s for 16384 events => throughput is 6.18E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0854s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G67/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 28.9047s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2682s + [COUNTERS] CudaCpp MEs ( 2 ) : 26.5509s for 16384 events => throughput is 6.17E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0856s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G512/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 28.6652s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9706s + [COUNTERS] CudaCpp MEs ( 2 ) : 26.6096s for 16384 events => throughput is 6.16E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0850s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G511/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 29.1279s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5918s + [COUNTERS] CudaCpp MEs ( 2 ) : 26.4509s for 16384 events => throughput is 6.19E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0852s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 29.0329s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4962s + [COUNTERS] CudaCpp MEs ( 2 ) : 26.4509s for 16384 events => throughput is 6.19E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0859s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G21/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 28.4446s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7864s + [COUNTERS] CudaCpp MEs ( 2 ) : 26.5723s for 16384 events => throughput is 6.17E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0859s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G48/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 29.3044s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5023s + [COUNTERS] CudaCpp MEs ( 2 ) : 26.7163s for 16384 events => throughput is 6.13E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0859s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_taptamggdux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_taptamggdux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_taptamggdux/ajob1' in 48.9005 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_taptamggdux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_taptamggdux/G68/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.8676s + [COUNTERS] Fortran Overhead ( 0 ) : 3.8957s + [COUNTERS] CudaCpp MEs ( 2 ) : 9.9352s for 16384 events => throughput is 1.65E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0367s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_taptamggdux/G167/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 11.7531s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7862s + [COUNTERS] CudaCpp MEs ( 2 ) : 9.9302s for 16384 events => throughput is 1.65E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0368s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_taptamggdux/G92/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 11.6991s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7568s + [COUNTERS] CudaCpp MEs ( 2 ) : 9.9062s for 16384 events => throughput is 1.65E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0361s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_taptamggdux/G60/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 11.4918s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4896s + [COUNTERS] CudaCpp MEs ( 2 ) : 9.9656s for 16384 events => throughput is 1.64E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0366s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_taptamggdux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/ajob1' in 632.3164 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G56/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 29.7372s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8553s + [COUNTERS] CudaCpp MEs ( 2 ) : 26.7962s for 16384 events => throughput is 6.11E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0857s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G39/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 29.9315s + [COUNTERS] Fortran Overhead ( 0 ) : 3.0757s + [COUNTERS] CudaCpp MEs ( 2 ) : 26.7688s for 16384 events => throughput is 6.12E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0870s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G517/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 34.9710s + [COUNTERS] Fortran Overhead ( 0 ) : 8.1070s + [COUNTERS] CudaCpp MEs ( 2 ) : 26.7779s for 16384 events => throughput is 6.12E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0862s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G497/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 28.9785s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9909s + [COUNTERS] CudaCpp MEs ( 2 ) : 26.9021s for 16384 events => throughput is 6.09E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0854s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G13/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 28.5291s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7072s + [COUNTERS] CudaCpp MEs ( 2 ) : 26.7357s for 16384 events => throughput is 6.13E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0862s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G484/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 34.0056s + [COUNTERS] Fortran Overhead ( 0 ) : 6.6569s + [COUNTERS] CudaCpp MEs ( 2 ) : 27.2611s for 16384 events => throughput is 6.01E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0876s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G68/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 29.9893s + [COUNTERS] Fortran Overhead ( 0 ) : 3.0633s + [COUNTERS] CudaCpp MEs ( 2 ) : 26.8384s for 16384 events => throughput is 6.10E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0875s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G49/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 29.2897s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3476s + [COUNTERS] CudaCpp MEs ( 2 ) : 26.8543s for 16384 events => throughput is 6.10E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0878s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G523/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 29.5474s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5934s + [COUNTERS] CudaCpp MEs ( 2 ) : 26.8684s for 16384 events => throughput is 6.10E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0856s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G518/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 32.9665s + [COUNTERS] Fortran Overhead ( 0 ) : 5.9144s + [COUNTERS] CudaCpp MEs ( 2 ) : 26.9640s for 16384 events => throughput is 6.08E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0881s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G42/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 34.8458s + [COUNTERS] Fortran Overhead ( 0 ) : 8.0827s + [COUNTERS] CudaCpp MEs ( 2 ) : 26.6774s for 16384 events => throughput is 6.14E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0858s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G524/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 28.7812s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9048s + [COUNTERS] CudaCpp MEs ( 2 ) : 26.7904s for 16384 events => throughput is 6.12E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0859s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G610/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 29.9521s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1035s + [COUNTERS] CudaCpp MEs ( 2 ) : 26.7590s for 16384 events => throughput is 6.12E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0896s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G26/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 32.8489s + [COUNTERS] Fortran Overhead ( 0 ) : 5.9614s + [COUNTERS] CudaCpp MEs ( 2 ) : 26.8000s for 16384 events => throughput is 6.11E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0875s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G58/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 34.2141s + [COUNTERS] Fortran Overhead ( 0 ) : 7.2792s + [COUNTERS] CudaCpp MEs ( 2 ) : 26.8482s for 16384 events => throughput is 6.10E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0867s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G511/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 29.1783s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1394s + [COUNTERS] CudaCpp MEs ( 2 ) : 26.9526s for 16384 events => throughput is 6.08E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0863s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 29.7195s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8283s + [COUNTERS] CudaCpp MEs ( 2 ) : 26.8065s for 16384 events => throughput is 6.11E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0847s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G32/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 35.0916s + [COUNTERS] Fortran Overhead ( 0 ) : 8.0292s + [COUNTERS] CudaCpp MEs ( 2 ) : 26.9723s for 16384 events => throughput is 6.07E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0901s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G61/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 33.0677s + [COUNTERS] Fortran Overhead ( 0 ) : 6.1338s + [COUNTERS] CudaCpp MEs ( 2 ) : 26.8469s for 16384 events => throughput is 6.10E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0870s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G390/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 36.2580s + [COUNTERS] Fortran Overhead ( 0 ) : 9.2710s + [COUNTERS] CudaCpp MEs ( 2 ) : 26.9011s for 16384 events => throughput is 6.09E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0859s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_epemgguc/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_epemgguc/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_epemgguc/ajob1' in 28.1969 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_epemgguc/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_epemgguc/G164/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 9.6708s + [COUNTERS] Fortran Overhead ( 0 ) : 4.4734s + [COUNTERS] CudaCpp MEs ( 2 ) : 5.1606s for 16384 events => throughput is 3.17E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0369s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_epemgguc/G24/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 9.5326s + [COUNTERS] Fortran Overhead ( 0 ) : 4.0544s + [COUNTERS] CudaCpp MEs ( 2 ) : 5.4395s for 16384 events => throughput is 3.01E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0386s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_epemgguc/G162/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.9070s + [COUNTERS] Fortran Overhead ( 0 ) : 3.3827s + [COUNTERS] CudaCpp MEs ( 2 ) : 5.4855s for 16384 events => throughput is 2.99E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0387s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_epemgguc/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/ajob1' in 911.1455 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G336/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 32.7987s + [COUNTERS] Fortran Overhead ( 0 ) : 3.8037s + [COUNTERS] CudaCpp MEs ( 2 ) : 28.9043s for 16384 events => throughput is 5.67E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0907s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G108/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 31.3407s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5803s + [COUNTERS] CudaCpp MEs ( 2 ) : 28.6709s for 16384 events => throughput is 5.71E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0895s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G105/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 31.5204s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5343s + [COUNTERS] CudaCpp MEs ( 2 ) : 28.8931s for 16384 events => throughput is 5.67E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0930s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G304/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 32.1933s + [COUNTERS] Fortran Overhead ( 0 ) : 3.3840s + [COUNTERS] CudaCpp MEs ( 2 ) : 28.7184s for 16384 events => throughput is 5.71E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0908s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G175/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 31.2305s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3804s + [COUNTERS] CudaCpp MEs ( 2 ) : 28.7589s for 16384 events => throughput is 5.70E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0912s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G128/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 31.6207s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7211s + [COUNTERS] CudaCpp MEs ( 2 ) : 28.8083s for 16384 events => throughput is 5.69E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0913s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G360/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 31.2110s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6681s + [COUNTERS] CudaCpp MEs ( 2 ) : 28.4519s for 16384 events => throughput is 5.76E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0910s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G120/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 31.1599s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4340s + [COUNTERS] CudaCpp MEs ( 2 ) : 28.6331s for 16384 events => throughput is 5.72E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0927s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G296/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 31.0653s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3065s + [COUNTERS] CudaCpp MEs ( 2 ) : 28.6628s for 16384 events => throughput is 5.72E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0960s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G299/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 31.7413s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9276s + [COUNTERS] CudaCpp MEs ( 2 ) : 28.7235s for 16384 events => throughput is 5.70E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0902s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G715/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 31.7926s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7934s + [COUNTERS] CudaCpp MEs ( 2 ) : 28.9077s for 16384 events => throughput is 5.67E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0915s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G300/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 31.5695s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5714s + [COUNTERS] CudaCpp MEs ( 2 ) : 28.9065s for 16384 events => throughput is 5.67E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0916s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G303/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 32.2473s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1231s + [COUNTERS] CudaCpp MEs ( 2 ) : 29.0268s for 16384 events => throughput is 5.64E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0973s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G127/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 32.1934s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1223s + [COUNTERS] CudaCpp MEs ( 2 ) : 28.9795s for 16384 events => throughput is 5.65E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0916s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G111/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 31.5106s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7137s + [COUNTERS] CudaCpp MEs ( 2 ) : 28.7047s for 16384 events => throughput is 5.71E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0922s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G126/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 40.1261s + [COUNTERS] Fortran Overhead ( 0 ) : 11.2744s + [COUNTERS] CudaCpp MEs ( 2 ) : 28.7596s for 16384 events => throughput is 5.70E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0921s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G294/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 36.0063s + [COUNTERS] Fortran Overhead ( 0 ) : 7.1242s + [COUNTERS] CudaCpp MEs ( 2 ) : 28.7896s for 16384 events => throughput is 5.69E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0925s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G598/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 31.5383s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6121s + [COUNTERS] CudaCpp MEs ( 2 ) : 28.8344s for 16384 events => throughput is 5.68E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0918s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G610/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 32.1904s + [COUNTERS] Fortran Overhead ( 0 ) : 3.5130s + [COUNTERS] CudaCpp MEs ( 2 ) : 28.5855s for 16384 events => throughput is 5.73E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0919s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G178/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 31.5244s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7723s + [COUNTERS] CudaCpp MEs ( 2 ) : 28.6595s for 16384 events => throughput is 5.72E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0926s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G121/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 31.4509s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6512s + [COUNTERS] CudaCpp MEs ( 2 ) : 28.7084s for 16384 events => throughput is 5.71E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0913s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G685/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 31.3672s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4219s + [COUNTERS] CudaCpp MEs ( 2 ) : 28.8541s for 16384 events => throughput is 5.68E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0912s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G690/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 30.7808s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2315s + [COUNTERS] CudaCpp MEs ( 2 ) : 28.4579s for 16384 events => throughput is 5.76E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0914s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G103/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 37.8148s + [COUNTERS] Fortran Overhead ( 0 ) : 8.8175s + [COUNTERS] CudaCpp MEs ( 2 ) : 28.8985s for 16384 events => throughput is 5.67E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0987s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G107/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 31.2489s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4797s + [COUNTERS] CudaCpp MEs ( 2 ) : 28.6782s for 16384 events => throughput is 5.71E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0910s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G180/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 31.6257s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6461s + [COUNTERS] CudaCpp MEs ( 2 ) : 28.8879s for 16384 events => throughput is 5.67E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0917s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G97/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 37.6414s + [COUNTERS] Fortran Overhead ( 0 ) : 8.7609s + [COUNTERS] CudaCpp MEs ( 2 ) : 28.7891s for 16384 events => throughput is 5.69E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0914s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G179/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 31.8753s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9202s + [COUNTERS] CudaCpp MEs ( 2 ) : 28.8633s for 16384 events => throughput is 5.68E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0919s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamggddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamggddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamggddx/ajob1' in 13.3361 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamggddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamggddx/G189/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.2817s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6900s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.5526s for 16384 events => throughput is 1.55E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0390s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamggddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dd_taptamggdd/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dd_taptamggdd/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dd_taptamggdd/ajob1' in 37.6165 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dd_taptamggdd/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dd_taptamggdd/G463/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 37.5600s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3489s + [COUNTERS] CudaCpp MEs ( 2 ) : 35.1430s for 16384 events => throughput is 4.66E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0680s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dd_taptamggdd/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/ajob1' in 2366.5465 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G506/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 59.6064s + [COUNTERS] Fortran Overhead ( 0 ) : 4.3463s + [COUNTERS] CudaCpp MEs ( 2 ) : 55.1714s for 16384 events => throughput is 2.97E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0887s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G499/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.0192s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8801s + [COUNTERS] CudaCpp MEs ( 2 ) : 55.0499s for 16384 events => throughput is 2.98E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0891s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G46/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 62.9346s + [COUNTERS] Fortran Overhead ( 0 ) : 7.6567s + [COUNTERS] CudaCpp MEs ( 2 ) : 55.1888s for 16384 events => throughput is 2.97E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0891s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G56/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.8132s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5999s + [COUNTERS] CudaCpp MEs ( 2 ) : 55.1233s for 16384 events => throughput is 2.97E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0900s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G39/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.7217s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4267s + [COUNTERS] CudaCpp MEs ( 2 ) : 55.2054s for 16384 events => throughput is 2.97E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0897s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G517/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 60.5690s + [COUNTERS] Fortran Overhead ( 0 ) : 5.2344s + [COUNTERS] CudaCpp MEs ( 2 ) : 55.2462s for 16384 events => throughput is 2.97E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0884s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G52/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 58.2240s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9763s + [COUNTERS] CudaCpp MEs ( 2 ) : 55.1594s for 16384 events => throughput is 2.97E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0884s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G31/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 62.3494s + [COUNTERS] Fortran Overhead ( 0 ) : 6.9413s + [COUNTERS] CudaCpp MEs ( 2 ) : 55.3180s for 16384 events => throughput is 2.96E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0901s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G324/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 63.7630s + [COUNTERS] Fortran Overhead ( 0 ) : 6.7219s + [COUNTERS] CudaCpp MEs ( 2 ) : 56.9496s for 16384 events => throughput is 2.88E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0915s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G64/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.2504s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8659s + [COUNTERS] CudaCpp MEs ( 2 ) : 55.2957s for 16384 events => throughput is 2.96E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0888s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G3/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 59.4999s + [COUNTERS] Fortran Overhead ( 0 ) : 4.0268s + [COUNTERS] CudaCpp MEs ( 2 ) : 55.3826s for 16384 events => throughput is 2.96E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0905s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G33/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.5508s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2309s + [COUNTERS] CudaCpp MEs ( 2 ) : 55.2313s for 16384 events => throughput is 2.97E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0887s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G53/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 63.2851s + [COUNTERS] Fortran Overhead ( 0 ) : 6.8196s + [COUNTERS] CudaCpp MEs ( 2 ) : 56.3736s for 16384 events => throughput is 2.91E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0919s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G43/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 60.8499s + [COUNTERS] Fortran Overhead ( 0 ) : 5.4876s + [COUNTERS] CudaCpp MEs ( 2 ) : 55.2729s for 16384 events => throughput is 2.96E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0894s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G49/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.1300s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8264s + [COUNTERS] CudaCpp MEs ( 2 ) : 55.2141s for 16384 events => throughput is 2.97E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0895s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G40/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 58.3237s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9110s + [COUNTERS] CudaCpp MEs ( 2 ) : 55.3231s for 16384 events => throughput is 2.96E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0897s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G523/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.6480s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2208s + [COUNTERS] CudaCpp MEs ( 2 ) : 55.3372s for 16384 events => throughput is 2.96E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0899s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G518/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 60.8798s + [COUNTERS] Fortran Overhead ( 0 ) : 5.4597s + [COUNTERS] CudaCpp MEs ( 2 ) : 55.3305s for 16384 events => throughput is 2.96E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0896s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G35/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.2722s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9603s + [COUNTERS] CudaCpp MEs ( 2 ) : 55.2216s for 16384 events => throughput is 2.97E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0903s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.1942s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7079s + [COUNTERS] CudaCpp MEs ( 2 ) : 55.3944s for 16384 events => throughput is 2.96E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0920s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G42/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 60.8621s + [COUNTERS] Fortran Overhead ( 0 ) : 5.5518s + [COUNTERS] CudaCpp MEs ( 2 ) : 55.2205s for 16384 events => throughput is 2.97E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0898s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G524/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.5316s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1565s + [COUNTERS] CudaCpp MEs ( 2 ) : 55.2831s for 16384 events => throughput is 2.96E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0920s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G65/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.1976s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9132s + [COUNTERS] CudaCpp MEs ( 2 ) : 55.1939s for 16384 events => throughput is 2.97E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0905s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G27/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 61.3261s + [COUNTERS] Fortran Overhead ( 0 ) : 5.8830s + [COUNTERS] CudaCpp MEs ( 2 ) : 55.3535s for 16384 events => throughput is 2.96E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0895s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G44/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 61.8074s + [COUNTERS] Fortran Overhead ( 0 ) : 6.4154s + [COUNTERS] CudaCpp MEs ( 2 ) : 55.3024s for 16384 events => throughput is 2.96E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0896s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.4822s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6680s + [COUNTERS] CudaCpp MEs ( 2 ) : 55.7228s for 16384 events => throughput is 2.94E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0913s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G303/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 60.2576s + [COUNTERS] Fortran Overhead ( 0 ) : 3.3310s + [COUNTERS] CudaCpp MEs ( 2 ) : 56.8345s for 16384 events => throughput is 2.88E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0921s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G55/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 58.0631s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7752s + [COUNTERS] CudaCpp MEs ( 2 ) : 55.1988s for 16384 events => throughput is 2.97E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0891s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G34/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.8701s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8961s + [COUNTERS] CudaCpp MEs ( 2 ) : 55.8825s for 16384 events => throughput is 2.93E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0914s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G26/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 61.2903s + [COUNTERS] Fortran Overhead ( 0 ) : 6.0337s + [COUNTERS] CudaCpp MEs ( 2 ) : 55.1673s for 16384 events => throughput is 2.97E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0893s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G62/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 61.5358s + [COUNTERS] Fortran Overhead ( 0 ) : 6.0692s + [COUNTERS] CudaCpp MEs ( 2 ) : 55.3773s for 16384 events => throughput is 2.96E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0893s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G37/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.6907s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2158s + [COUNTERS] CudaCpp MEs ( 2 ) : 55.3859s for 16384 events => throughput is 2.96E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0890s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G485/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 60.6858s + [COUNTERS] Fortran Overhead ( 0 ) : 5.3948s + [COUNTERS] CudaCpp MEs ( 2 ) : 55.2012s for 16384 events => throughput is 2.97E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0897s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G67/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.5516s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2170s + [COUNTERS] CudaCpp MEs ( 2 ) : 55.2455s for 16384 events => throughput is 2.97E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0891s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G512/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.3866s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8661s + [COUNTERS] CudaCpp MEs ( 2 ) : 55.4307s for 16384 events => throughput is 2.96E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0897s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G511/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.8329s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5252s + [COUNTERS] CudaCpp MEs ( 2 ) : 55.2180s for 16384 events => throughput is 2.97E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0896s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G51/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 59.0296s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6283s + [COUNTERS] CudaCpp MEs ( 2 ) : 56.3098s for 16384 events => throughput is 2.91E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0915s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.9466s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4943s + [COUNTERS] CudaCpp MEs ( 2 ) : 55.3613s for 16384 events => throughput is 2.96E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0910s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G63/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.6646s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3694s + [COUNTERS] CudaCpp MEs ( 2 ) : 55.2058s for 16384 events => throughput is 2.97E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0894s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G48/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.8511s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3975s + [COUNTERS] CudaCpp MEs ( 2 ) : 55.3639s for 16384 events => throughput is 2.96E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0897s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgggux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgggux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgggux/ajob1' in 419.1141 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgggux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgggux/G33/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.4513s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9539s + [COUNTERS] CudaCpp MEs ( 2 ) : 55.4064s for 16384 events => throughput is 2.96E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0911s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgggux/G68/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 59.4316s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1873s + [COUNTERS] CudaCpp MEs ( 2 ) : 56.1550s for 16384 events => throughput is 2.92E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0893s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgggux/G25/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 62.0224s + [COUNTERS] Fortran Overhead ( 0 ) : 6.5265s + [COUNTERS] CudaCpp MEs ( 2 ) : 55.4042s for 16384 events => throughput is 2.96E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0917s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgggux/G494/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.2184s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7043s + [COUNTERS] CudaCpp MEs ( 2 ) : 55.4225s for 16384 events => throughput is 2.96E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0915s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgggux/G55/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 58.5185s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9666s + [COUNTERS] CudaCpp MEs ( 2 ) : 55.4630s for 16384 events => throughput is 2.95E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0889s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgggux/G509/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 62.6989s + [COUNTERS] Fortran Overhead ( 0 ) : 7.1740s + [COUNTERS] CudaCpp MEs ( 2 ) : 55.4355s for 16384 events => throughput is 2.96E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0895s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgggux/G29/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 61.6019s + [COUNTERS] Fortran Overhead ( 0 ) : 6.1275s + [COUNTERS] CudaCpp MEs ( 2 ) : 55.3847s for 16384 events => throughput is 2.96E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0897s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgggux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamgguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamgguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamgguux/ajob1' in 182.4375 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamgguux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamgguux/G260/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 35.3357s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5760s + [COUNTERS] CudaCpp MEs ( 2 ) : 33.6930s for 16384 events => throughput is 4.86E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0668s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamgguux/G447/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 35.4147s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6113s + [COUNTERS] CudaCpp MEs ( 2 ) : 33.7372s for 16384 events => throughput is 4.86E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0662s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamgguux/G229/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 37.8123s + [COUNTERS] Fortran Overhead ( 0 ) : 4.0969s + [COUNTERS] CudaCpp MEs ( 2 ) : 33.6492s for 16384 events => throughput is 4.87E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0662s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamgguux/G135/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 35.6571s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9984s + [COUNTERS] CudaCpp MEs ( 2 ) : 33.5930s for 16384 events => throughput is 4.88E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0657s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamgguux/G230/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 38.0915s + [COUNTERS] Fortran Overhead ( 0 ) : 4.4235s + [COUNTERS] CudaCpp MEs ( 2 ) : 33.6020s for 16384 events => throughput is 4.88E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0660s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamgguux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdddx/ajob1' in 38.0777 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdddx/G715/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 19.0223s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1082s + [COUNTERS] CudaCpp MEs ( 2 ) : 16.8490s for 16384 events => throughput is 9.72E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0651s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdddx/G111/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 18.9848s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0500s + [COUNTERS] CudaCpp MEs ( 2 ) : 16.8690s for 16384 events => throughput is 9.71E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0659s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemusdxsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemusdxsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemusdxsx/ajob1' in 7.1179 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemusdxsx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemusdxsx/G67/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.0730s + [COUNTERS] Fortran Overhead ( 0 ) : 6.2887s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.7723s for 16384 events => throughput is 2.12E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0119s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemusdxsx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdssx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdssx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdssx/ajob1' in 44.8342 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdssx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdssx/G31/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.3806s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9637s + [COUNTERS] CudaCpp MEs ( 2 ) : 5.3790s for 16384 events => throughput is 3.05E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0378s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdssx/G27/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.4675s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9994s + [COUNTERS] CudaCpp MEs ( 2 ) : 5.4292s for 16384 events => throughput is 3.02E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0389s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdssx/G119/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.5627s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1963s + [COUNTERS] CudaCpp MEs ( 2 ) : 5.3293s for 16384 events => throughput is 3.07E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0371s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdssx/G284/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.8551s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4230s + [COUNTERS] CudaCpp MEs ( 2 ) : 5.3947s for 16384 events => throughput is 3.04E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0375s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdssx/G375/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.9701s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5946s + [COUNTERS] CudaCpp MEs ( 2 ) : 5.3387s for 16384 events => throughput is 3.07E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0368s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdssx/G102/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.4480s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0716s + [COUNTERS] CudaCpp MEs ( 2 ) : 5.3380s for 16384 events => throughput is 3.07E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0384s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdssx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_epemgguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_epemgguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_epemgguux/ajob1' in 81.2139 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_epemgguux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_epemgguux/G105/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 18.4388s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5747s + [COUNTERS] CudaCpp MEs ( 2 ) : 16.7978s for 16384 events => throughput is 9.75E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0663s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_epemgguux/G223/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 19.0975s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2586s + [COUNTERS] CudaCpp MEs ( 2 ) : 16.7734s for 16384 events => throughput is 9.77E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0655s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_epemgguux/G103/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 21.7425s + [COUNTERS] Fortran Overhead ( 0 ) : 4.6503s + [COUNTERS] CudaCpp MEs ( 2 ) : 17.0253s for 16384 events => throughput is 9.62E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0670s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_epemgguux/G5/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 21.8200s + [COUNTERS] Fortran Overhead ( 0 ) : 4.9838s + [COUNTERS] CudaCpp MEs ( 2 ) : 16.7706s for 16384 events => throughput is 9.77E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0656s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_epemgguux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/ajob1' in 1962.5801 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G506/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 33.4520s + [COUNTERS] Fortran Overhead ( 0 ) : 5.5434s + [COUNTERS] CudaCpp MEs ( 2 ) : 27.8196s for 16384 events => throughput is 5.89E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0890s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G499/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 29.8725s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8859s + [COUNTERS] CudaCpp MEs ( 2 ) : 27.8968s for 16384 events => throughput is 5.87E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0898s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G38/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 30.7952s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7788s + [COUNTERS] CudaCpp MEs ( 2 ) : 27.9259s for 16384 events => throughput is 5.87E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0905s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G46/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 36.0485s + [COUNTERS] Fortran Overhead ( 0 ) : 7.9627s + [COUNTERS] CudaCpp MEs ( 2 ) : 27.9962s for 16384 events => throughput is 5.85E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0895s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G59/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 34.3236s + [COUNTERS] Fortran Overhead ( 0 ) : 6.3595s + [COUNTERS] CudaCpp MEs ( 2 ) : 27.8745s for 16384 events => throughput is 5.88E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0896s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G56/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 30.7632s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8153s + [COUNTERS] CudaCpp MEs ( 2 ) : 27.8589s for 16384 events => throughput is 5.88E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0890s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G525/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 30.4495s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4230s + [COUNTERS] CudaCpp MEs ( 2 ) : 27.9375s for 16384 events => throughput is 5.86E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0889s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G39/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 30.3584s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3505s + [COUNTERS] CudaCpp MEs ( 2 ) : 27.9180s for 16384 events => throughput is 5.87E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0900s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G517/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 33.8030s + [COUNTERS] Fortran Overhead ( 0 ) : 5.7983s + [COUNTERS] CudaCpp MEs ( 2 ) : 27.9157s for 16384 events => throughput is 5.87E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0889s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G487/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 32.8543s + [COUNTERS] Fortran Overhead ( 0 ) : 4.8920s + [COUNTERS] CudaCpp MEs ( 2 ) : 27.8733s for 16384 events => throughput is 5.88E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0890s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G52/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 31.0222s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9829s + [COUNTERS] CudaCpp MEs ( 2 ) : 27.9493s for 16384 events => throughput is 5.86E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0900s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G31/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 35.9025s + [COUNTERS] Fortran Overhead ( 0 ) : 7.9139s + [COUNTERS] CudaCpp MEs ( 2 ) : 27.8999s for 16384 events => throughput is 5.87E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0888s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G64/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 29.8731s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9961s + [COUNTERS] CudaCpp MEs ( 2 ) : 27.7891s for 16384 events => throughput is 5.90E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0879s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G3/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 32.9438s + [COUNTERS] Fortran Overhead ( 0 ) : 5.0290s + [COUNTERS] CudaCpp MEs ( 2 ) : 27.8255s for 16384 events => throughput is 5.89E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0893s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G497/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 30.2347s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3210s + [COUNTERS] CudaCpp MEs ( 2 ) : 27.8239s for 16384 events => throughput is 5.89E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0898s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G13/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 29.7382s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6731s + [COUNTERS] CudaCpp MEs ( 2 ) : 27.9757s for 16384 events => throughput is 5.86E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0894s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G66/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 30.3985s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4590s + [COUNTERS] CudaCpp MEs ( 2 ) : 27.8502s for 16384 events => throughput is 5.88E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0893s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G33/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 29.9779s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9692s + [COUNTERS] CudaCpp MEs ( 2 ) : 27.9198s for 16384 events => throughput is 5.87E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0889s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G53/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 36.1126s + [COUNTERS] Fortran Overhead ( 0 ) : 8.1829s + [COUNTERS] CudaCpp MEs ( 2 ) : 27.8409s for 16384 events => throughput is 5.88E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0889s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G43/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 33.0054s + [COUNTERS] Fortran Overhead ( 0 ) : 4.9883s + [COUNTERS] CudaCpp MEs ( 2 ) : 27.9280s for 16384 events => throughput is 5.87E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0891s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G68/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 30.1021s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1805s + [COUNTERS] CudaCpp MEs ( 2 ) : 27.8315s for 16384 events => throughput is 5.89E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0901s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G49/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 29.9022s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9455s + [COUNTERS] CudaCpp MEs ( 2 ) : 27.8657s for 16384 events => throughput is 5.88E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0911s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G302/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 36.3975s + [COUNTERS] Fortran Overhead ( 0 ) : 8.4573s + [COUNTERS] CudaCpp MEs ( 2 ) : 27.8518s for 16384 events => throughput is 5.88E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0884s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G25/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 34.2808s + [COUNTERS] Fortran Overhead ( 0 ) : 6.3213s + [COUNTERS] CudaCpp MEs ( 2 ) : 27.8700s for 16384 events => throughput is 5.88E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0895s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G40/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 31.0546s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1200s + [COUNTERS] CudaCpp MEs ( 2 ) : 27.8450s for 16384 events => throughput is 5.88E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0896s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G523/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 29.9409s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0439s + [COUNTERS] CudaCpp MEs ( 2 ) : 27.8087s for 16384 events => throughput is 5.89E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0882s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G518/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 33.1572s + [COUNTERS] Fortran Overhead ( 0 ) : 5.2070s + [COUNTERS] CudaCpp MEs ( 2 ) : 27.8612s for 16384 events => throughput is 5.88E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0890s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G35/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 30.1387s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1373s + [COUNTERS] CudaCpp MEs ( 2 ) : 27.9119s for 16384 events => throughput is 5.87E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0895s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 29.5894s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6476s + [COUNTERS] CudaCpp MEs ( 2 ) : 27.8530s for 16384 events => throughput is 5.88E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0888s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G524/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 30.0153s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1039s + [COUNTERS] CudaCpp MEs ( 2 ) : 27.8228s for 16384 events => throughput is 5.89E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0887s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G65/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 30.8312s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1981s + [COUNTERS] CudaCpp MEs ( 2 ) : 28.5418s for 16384 events => throughput is 5.74E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0912s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G27/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 34.7243s + [COUNTERS] Fortran Overhead ( 0 ) : 6.5280s + [COUNTERS] CudaCpp MEs ( 2 ) : 28.1076s for 16384 events => throughput is 5.83E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0888s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G44/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 34.5073s + [COUNTERS] Fortran Overhead ( 0 ) : 6.1914s + [COUNTERS] CudaCpp MEs ( 2 ) : 28.2277s for 16384 events => throughput is 5.80E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0882s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G483/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 32.2148s + [COUNTERS] Fortran Overhead ( 0 ) : 4.3086s + [COUNTERS] CudaCpp MEs ( 2 ) : 27.8165s for 16384 events => throughput is 5.89E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0897s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G482/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 33.6586s + [COUNTERS] Fortran Overhead ( 0 ) : 5.7053s + [COUNTERS] CudaCpp MEs ( 2 ) : 27.8635s for 16384 events => throughput is 5.88E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0898s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G1/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 32.5527s + [COUNTERS] Fortran Overhead ( 0 ) : 4.5901s + [COUNTERS] CudaCpp MEs ( 2 ) : 27.8746s for 16384 events => throughput is 5.88E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0879s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 29.5417s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5592s + [COUNTERS] CudaCpp MEs ( 2 ) : 27.8933s for 16384 events => throughput is 5.87E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0892s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G55/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 30.7127s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6567s + [COUNTERS] CudaCpp MEs ( 2 ) : 27.9652s for 16384 events => throughput is 5.86E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0908s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G298/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 31.0217s + [COUNTERS] Fortran Overhead ( 0 ) : 3.0575s + [COUNTERS] CudaCpp MEs ( 2 ) : 27.8753s for 16384 events => throughput is 5.88E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0889s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G513/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 30.9046s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9343s + [COUNTERS] CudaCpp MEs ( 2 ) : 27.8813s for 16384 events => throughput is 5.88E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0890s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G60/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 35.5346s + [COUNTERS] Fortran Overhead ( 0 ) : 7.5823s + [COUNTERS] CudaCpp MEs ( 2 ) : 27.8630s for 16384 events => throughput is 5.88E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0893s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G34/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 29.9645s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0280s + [COUNTERS] CudaCpp MEs ( 2 ) : 27.8479s for 16384 events => throughput is 5.88E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0885s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G26/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 35.0323s + [COUNTERS] Fortran Overhead ( 0 ) : 7.0192s + [COUNTERS] CudaCpp MEs ( 2 ) : 27.9233s for 16384 events => throughput is 5.87E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0898s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G58/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 33.4988s + [COUNTERS] Fortran Overhead ( 0 ) : 5.5354s + [COUNTERS] CudaCpp MEs ( 2 ) : 27.8741s for 16384 events => throughput is 5.88E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0892s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G54/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 33.9483s + [COUNTERS] Fortran Overhead ( 0 ) : 6.0654s + [COUNTERS] CudaCpp MEs ( 2 ) : 27.7937s for 16384 events => throughput is 5.89E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0892s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G62/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 34.9137s + [COUNTERS] Fortran Overhead ( 0 ) : 7.0116s + [COUNTERS] CudaCpp MEs ( 2 ) : 27.8127s for 16384 events => throughput is 5.89E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0894s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G37/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 30.4109s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9621s + [COUNTERS] CudaCpp MEs ( 2 ) : 28.3576s for 16384 events => throughput is 5.78E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0912s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G493/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 30.0871s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1522s + [COUNTERS] CudaCpp MEs ( 2 ) : 27.8462s for 16384 events => throughput is 5.88E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0887s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G4/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 34.2805s + [COUNTERS] Fortran Overhead ( 0 ) : 6.3223s + [COUNTERS] CudaCpp MEs ( 2 ) : 27.8695s for 16384 events => throughput is 5.88E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0887s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G50/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 30.1761s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1333s + [COUNTERS] CudaCpp MEs ( 2 ) : 27.9535s for 16384 events => throughput is 5.86E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0893s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G485/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 33.7968s + [COUNTERS] Fortran Overhead ( 0 ) : 5.8454s + [COUNTERS] CudaCpp MEs ( 2 ) : 27.8623s for 16384 events => throughput is 5.88E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0891s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G41/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 35.1195s + [COUNTERS] Fortran Overhead ( 0 ) : 7.2077s + [COUNTERS] CudaCpp MEs ( 2 ) : 27.8232s for 16384 events => throughput is 5.89E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0887s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G67/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 30.1464s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1440s + [COUNTERS] CudaCpp MEs ( 2 ) : 27.9140s for 16384 events => throughput is 5.87E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0884s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G512/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 30.1107s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0599s + [COUNTERS] CudaCpp MEs ( 2 ) : 27.9616s for 16384 events => throughput is 5.86E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0891s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G511/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 30.4474s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5450s + [COUNTERS] CudaCpp MEs ( 2 ) : 27.8133s for 16384 events => throughput is 5.89E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0891s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G57/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 35.0368s + [COUNTERS] Fortran Overhead ( 0 ) : 7.0497s + [COUNTERS] CudaCpp MEs ( 2 ) : 27.8980s for 16384 events => throughput is 5.87E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0890s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 30.9378s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9246s + [COUNTERS] CudaCpp MEs ( 2 ) : 27.9238s for 16384 events => throughput is 5.87E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0894s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G63/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 30.4565s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4427s + [COUNTERS] CudaCpp MEs ( 2 ) : 27.9248s for 16384 events => throughput is 5.87E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0890s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G32/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 35.7234s + [COUNTERS] Fortran Overhead ( 0 ) : 7.7710s + [COUNTERS] CudaCpp MEs ( 2 ) : 27.8634s for 16384 events => throughput is 5.88E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0889s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G29/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 34.3145s + [COUNTERS] Fortran Overhead ( 0 ) : 6.3252s + [COUNTERS] CudaCpp MEs ( 2 ) : 27.9000s for 16384 events => throughput is 5.87E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0892s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G48/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 30.3517s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3655s + [COUNTERS] CudaCpp MEs ( 2 ) : 27.8972s for 16384 events => throughput is 5.87E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0890s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_taptamggdsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_taptamggdsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_taptamggdsx/ajob1' in 37.8544 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_taptamggdsx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_taptamggdsx/G92/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 12.8184s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4446s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.3365s for 16384 events => throughput is 1.59E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0373s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_taptamggdsx/G144/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 12.4955s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1144s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.3434s for 16384 events => throughput is 1.58E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0377s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_taptamggdsx/G207/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 12.4610s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0702s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.3531s for 16384 events => throughput is 1.58E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0377s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_taptamggdsx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_epemggdsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_epemggdsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_epemggdsx/ajob1' in 12.3407 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_epemggdsx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_epemggdsx/G205/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 12.2960s + [COUNTERS] Fortran Overhead ( 0 ) : 6.8830s + [COUNTERS] CudaCpp MEs ( 2 ) : 5.3738s for 16384 events => throughput is 3.05E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0392s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_epemggdsx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_taptamgggg/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_taptamgggg/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_taptamgggg/ajob1' in 58.4817 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_taptamgggg/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_taptamgggg/G67/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 58.4566s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7566s + [COUNTERS] CudaCpp MEs ( 2 ) : 55.6092s for 16384 events => throughput is 2.95E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0907s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_taptamgggg/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dc_epemdcssx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dc_epemdcssx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dc_epemdcssx/ajob1' in 3.0501 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dc_epemdcssx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dc_epemdcssx/G77/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.0056s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2308s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.7630s for 16384 events => throughput is 2.15E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0119s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dc_epemdcssx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguccx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguccx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguccx/ajob1' in 13.1973 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguccx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguccx/G48/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.1514s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5624s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.5516s for 16384 events => throughput is 1.55E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0374s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguccx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ds_epemggds/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ds_epemggds/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ds_epemggds/ajob1' in 7.0532 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ds_epemggds/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ds_epemggds/G26/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.0052s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6314s + [COUNTERS] CudaCpp MEs ( 2 ) : 5.3360s for 16384 events => throughput is 3.07E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0378s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ds_epemggds/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ucx_epemggucx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ucx_epemggucx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ucx_epemggucx/ajob1' in 16.2281 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ucx_epemggucx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ucx_epemggucx/G163/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 9.2927s + [COUNTERS] Fortran Overhead ( 0 ) : 4.1072s + [COUNTERS] CudaCpp MEs ( 2 ) : 5.1479s for 16384 events => throughput is 3.18E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0376s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ucx_epemggucx/G60/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.8751s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5730s + [COUNTERS] CudaCpp MEs ( 2 ) : 5.2640s for 16384 events => throughput is 3.11E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0381s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ucx_epemggucx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemggudx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemggudx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemggudx/ajob1' in 105.9327 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemggudx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemggudx/G109/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.8498s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5023s + [COUNTERS] CudaCpp MEs ( 2 ) : 5.3100s for 16384 events => throughput is 3.09E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0374s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemggudx/G151/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.1829s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8446s + [COUNTERS] CudaCpp MEs ( 2 ) : 5.3009s for 16384 events => throughput is 3.09E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0374s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemggudx/G106/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.6253s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2779s + [COUNTERS] CudaCpp MEs ( 2 ) : 5.3090s for 16384 events => throughput is 3.09E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0384s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemggudx/G35/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 12.3881s + [COUNTERS] Fortran Overhead ( 0 ) : 7.0274s + [COUNTERS] CudaCpp MEs ( 2 ) : 5.3230s for 16384 events => throughput is 3.08E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0377s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemggudx/G119/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.3209s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0111s + [COUNTERS] CudaCpp MEs ( 2 ) : 5.2713s for 16384 events => throughput is 3.11E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0386s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemggudx/G187/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.1911s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8725s + [COUNTERS] CudaCpp MEs ( 2 ) : 5.2812s for 16384 events => throughput is 3.10E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0374s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemggudx/G91/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.3634s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0235s + [COUNTERS] CudaCpp MEs ( 2 ) : 5.3024s for 16384 events => throughput is 3.09E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0375s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemggudx/G363/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 38.9906s + [COUNTERS] Fortran Overhead ( 0 ) : 17.6954s + [COUNTERS] CudaCpp MEs ( 2 ) : 21.2578s for 65536 events => throughput is 3.08E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0374s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemggudx/G149/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 9.8675s + [COUNTERS] Fortran Overhead ( 0 ) : 4.5389s + [COUNTERS] CudaCpp MEs ( 2 ) : 5.2901s for 16384 events => throughput is 3.10E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0385s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemggudx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgcuxcx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgcuxcx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgcuxcx/ajob1' in 7.8018 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgcuxcx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgcuxcx/G120/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.7503s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5195s + [COUNTERS] CudaCpp MEs ( 2 ) : 5.1923s for 16384 events => throughput is 3.16E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0385s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgcuxcx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemgggg/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemgggg/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemgggg/ajob1' in 63.1267 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemgggg/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemgggg/G65/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 31.4703s + [COUNTERS] Fortran Overhead ( 0 ) : 3.3147s + [COUNTERS] CudaCpp MEs ( 2 ) : 28.0659s for 16384 events => throughput is 5.84E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0897s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemgggg/G48/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 31.5880s + [COUNTERS] Fortran Overhead ( 0 ) : 3.5285s + [COUNTERS] CudaCpp MEs ( 2 ) : 27.9704s for 16384 events => throughput is 5.86E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0891s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemgggg/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_taptamggud/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_taptamggud/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_taptamggud/ajob1' in 113.7942 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_taptamggud/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_taptamggud/G46/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 12.5497s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2095s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.3030s for 16384 events => throughput is 1.59E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0373s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_taptamggud/G39/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.4783s + [COUNTERS] Fortran Overhead ( 0 ) : 4.2208s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.2199s for 16384 events => throughput is 1.60E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0377s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_taptamggud/G112/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 11.5761s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2598s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.2767s for 16384 events => throughput is 1.59E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0396s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_taptamggud/G115/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 11.7640s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4864s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.2402s for 16384 events => throughput is 1.60E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0375s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_taptamggud/G270/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.3296s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9515s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.3405s for 16384 events => throughput is 1.58E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0376s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_taptamggud/G96/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 11.8979s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5909s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.2685s for 16384 events => throughput is 1.60E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0385s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_taptamggud/G107/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.2722s + [COUNTERS] Fortran Overhead ( 0 ) : 3.9242s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.3102s for 16384 events => throughput is 1.59E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0378s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_taptamggud/G63/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 12.1806s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8932s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.2494s for 16384 events => throughput is 1.60E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0380s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_taptamggud/G375/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 11.5910s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2893s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.2644s for 16384 events => throughput is 1.60E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0373s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_taptamggud/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguuux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguuux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguuux/ajob1' in 83.9647 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguuux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguuux/G122/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 18.8765s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8617s + [COUNTERS] CudaCpp MEs ( 2 ) : 16.9489s for 16384 events => throughput is 9.67E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0659s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguuux/G33/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 22.8445s + [COUNTERS] Fortran Overhead ( 0 ) : 5.8572s + [COUNTERS] CudaCpp MEs ( 2 ) : 16.9208s for 16384 events => throughput is 9.68E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0665s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguuux/G127/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 20.0489s + [COUNTERS] Fortran Overhead ( 0 ) : 3.0442s + [COUNTERS] CudaCpp MEs ( 2 ) : 16.9386s for 16384 events => throughput is 9.67E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0661s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguuux/G116/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 22.1116s + [COUNTERS] Fortran Overhead ( 0 ) : 5.1402s + [COUNTERS] CudaCpp MEs ( 2 ) : 16.9065s for 16384 events => throughput is 9.69E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0650s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguuux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamgggdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamgggdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamgggdx/ajob1' in 524.1329 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamgggdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamgggdx/G56/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 58.5587s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6759s + [COUNTERS] CudaCpp MEs ( 2 ) : 55.7930s for 16384 events => throughput is 2.94E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0899s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamgggdx/G64/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 58.3210s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4035s + [COUNTERS] CudaCpp MEs ( 2 ) : 55.8266s for 16384 events => throughput is 2.93E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0909s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamgggdx/G13/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.8701s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6414s + [COUNTERS] CudaCpp MEs ( 2 ) : 56.1389s for 16384 events => throughput is 2.92E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0898s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamgggdx/G33/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.8515s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8276s + [COUNTERS] CudaCpp MEs ( 2 ) : 55.9339s for 16384 events => throughput is 2.93E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0901s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamgggdx/G49/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 58.2647s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2269s + [COUNTERS] CudaCpp MEs ( 2 ) : 55.9495s for 16384 events => throughput is 2.93E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0883s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamgggdx/G523/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 59.0653s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6304s + [COUNTERS] CudaCpp MEs ( 2 ) : 56.3435s for 16384 events => throughput is 2.91E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0914s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamgggdx/G65/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.8338s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9767s + [COUNTERS] CudaCpp MEs ( 2 ) : 55.7684s for 16384 events => throughput is 2.94E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0886s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamgggdx/G34/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 58.3273s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1509s + [COUNTERS] CudaCpp MEs ( 2 ) : 56.0848s for 16384 events => throughput is 2.92E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0917s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamgggdx/G48/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 57.8900s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0171s + [COUNTERS] CudaCpp MEs ( 2 ) : 55.7840s for 16384 events => throughput is 2.94E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0890s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamgggdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgudux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgudux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgudux/ajob1' in 28.2631 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgudux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgudux/G59/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 12.1539s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8843s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.2317s for 16384 events => throughput is 1.60E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0378s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgudux/G61/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 16.0538s + [COUNTERS] Fortran Overhead ( 0 ) : 5.7525s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.2636s for 16384 events => throughput is 1.60E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0378s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgudux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgduxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgduxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgduxdx/ajob1' in 7.1799 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgduxdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgduxdx/G104/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.1367s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9137s + [COUNTERS] CudaCpp MEs ( 2 ) : 5.1857s for 16384 events => throughput is 3.16E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0373s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgduxdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgsdxsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgsdxsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgsdxsx/ajob1' in 7.1002 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgsdxsx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgsdxsx/G25/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.0749s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7484s + [COUNTERS] CudaCpp MEs ( 2 ) : 5.2892s for 16384 events => throughput is 3.10E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0372s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgsdxsx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/ajob1' in 234.0707 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G208/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.2190s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9810s + [COUNTERS] CudaCpp MEs ( 2 ) : 5.1996s for 16384 events => throughput is 3.15E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0385s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G235/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.3318s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7853s + [COUNTERS] CudaCpp MEs ( 2 ) : 5.5077s for 16384 events => throughput is 2.97E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0389s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G140/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.6018s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0631s + [COUNTERS] CudaCpp MEs ( 2 ) : 5.5002s for 16384 events => throughput is 2.98E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0385s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G109/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.8708s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6869s + [COUNTERS] CudaCpp MEs ( 2 ) : 5.1466s for 16384 events => throughput is 3.18E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0372s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G33/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 10.6360s + [COUNTERS] Fortran Overhead ( 0 ) : 5.0338s + [COUNTERS] CudaCpp MEs ( 2 ) : 5.5631s for 16384 events => throughput is 2.95E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0391s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G43/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.3768s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8727s + [COUNTERS] CudaCpp MEs ( 2 ) : 5.4625s for 16384 events => throughput is 3.00E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0416s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G240/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.2286s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7918s + [COUNTERS] CudaCpp MEs ( 2 ) : 5.3989s for 16384 events => throughput is 3.03E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0379s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G268/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 9.8860s + [COUNTERS] Fortran Overhead ( 0 ) : 4.4743s + [COUNTERS] CudaCpp MEs ( 2 ) : 5.3743s for 16384 events => throughput is 3.05E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0374s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G139/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.3315s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0106s + [COUNTERS] CudaCpp MEs ( 2 ) : 5.2829s for 16384 events => throughput is 3.10E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0381s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G120/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.4770s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8330s + [COUNTERS] CudaCpp MEs ( 2 ) : 5.6047s for 16384 events => throughput is 2.92E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0392s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G236/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.2989s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8571s + [COUNTERS] CudaCpp MEs ( 2 ) : 5.4033s for 16384 events => throughput is 3.03E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0386s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G181/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 11.2174s + [COUNTERS] Fortran Overhead ( 0 ) : 5.6124s + [COUNTERS] CudaCpp MEs ( 2 ) : 5.5665s for 16384 events => throughput is 2.94E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0386s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G351/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.6205s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3640s + [COUNTERS] CudaCpp MEs ( 2 ) : 5.2187s for 16384 events => throughput is 3.14E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0378s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G65/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 9.1497s + [COUNTERS] Fortran Overhead ( 0 ) : 3.9273s + [COUNTERS] CudaCpp MEs ( 2 ) : 5.1849s for 16384 events => throughput is 3.16E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0374s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G71/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.1954s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5777s + [COUNTERS] CudaCpp MEs ( 2 ) : 5.5796s for 16384 events => throughput is 2.94E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0381s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G119/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.5643s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9362s + [COUNTERS] CudaCpp MEs ( 2 ) : 5.5903s for 16384 events => throughput is 2.93E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0378s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G127/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.2639s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6371s + [COUNTERS] CudaCpp MEs ( 2 ) : 5.5880s for 16384 events => throughput is 2.93E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0388s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G96/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.9995s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7042s + [COUNTERS] CudaCpp MEs ( 2 ) : 5.2577s for 16384 events => throughput is 3.12E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0376s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G60/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.8531s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2828s + [COUNTERS] CudaCpp MEs ( 2 ) : 5.5314s for 16384 events => throughput is 2.96E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0390s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G26/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.1074s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5063s + [COUNTERS] CudaCpp MEs ( 2 ) : 5.5625s for 16384 events => throughput is 2.95E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0386s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G37/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 11.2181s + [COUNTERS] Fortran Overhead ( 0 ) : 5.8351s + [COUNTERS] CudaCpp MEs ( 2 ) : 5.3450s for 16384 events => throughput is 3.07E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0380s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G125/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 9.0990s + [COUNTERS] Fortran Overhead ( 0 ) : 3.5311s + [COUNTERS] CudaCpp MEs ( 2 ) : 5.5266s for 16384 events => throughput is 2.96E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0413s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G91/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.9692s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4950s + [COUNTERS] CudaCpp MEs ( 2 ) : 5.4359s for 16384 events => throughput is 3.01E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0383s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G143/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.1817s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4434s + [COUNTERS] CudaCpp MEs ( 2 ) : 5.6974s for 16384 events => throughput is 2.88E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0409s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G116/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.9560s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3694s + [COUNTERS] CudaCpp MEs ( 2 ) : 5.5474s for 16384 events => throughput is 2.95E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0392s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G61/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 11.4080s + [COUNTERS] Fortran Overhead ( 0 ) : 5.7991s + [COUNTERS] CudaCpp MEs ( 2 ) : 5.5705s for 16384 events => throughput is 2.94E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0383s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G149/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 10.1282s + [COUNTERS] Fortran Overhead ( 0 ) : 4.5438s + [COUNTERS] CudaCpp MEs ( 2 ) : 5.5459s for 16384 events => throughput is 2.95E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0384s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G144/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.7132s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4618s + [COUNTERS] CudaCpp MEs ( 2 ) : 5.2140s for 16384 events => throughput is 3.14E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0375s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G207/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.4221s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8284s + [COUNTERS] CudaCpp MEs ( 2 ) : 5.5550s for 16384 events => throughput is 2.95E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0387s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_epemgguu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_epemgguu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_epemgguu/ajob1' in 219.9957 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_epemgguu/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_epemgguu/G463/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 19.5477s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0821s + [COUNTERS] CudaCpp MEs ( 2 ) : 17.3967s for 16384 events => throughput is 9.42E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0690s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_epemgguu/G100/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 20.3167s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2783s + [COUNTERS] CudaCpp MEs ( 2 ) : 16.9723s for 16384 events => throughput is 9.65E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0660s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_epemgguu/G110/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 18.4683s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4637s + [COUNTERS] CudaCpp MEs ( 2 ) : 16.9390s for 16384 events => throughput is 9.67E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0657s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_epemgguu/G124/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 19.0315s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5264s + [COUNTERS] CudaCpp MEs ( 2 ) : 17.4393s for 16384 events => throughput is 9.39E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0659s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_epemgguu/G367/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 19.5232s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9594s + [COUNTERS] CudaCpp MEs ( 2 ) : 17.4956s for 16384 events => throughput is 9.36E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0683s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_epemgguu/G126/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 21.1021s + [COUNTERS] Fortran Overhead ( 0 ) : 4.0744s + [COUNTERS] CudaCpp MEs ( 2 ) : 16.9621s for 16384 events => throughput is 9.66E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0656s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_epemgguu/G361/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 21.4852s + [COUNTERS] Fortran Overhead ( 0 ) : 4.4633s + [COUNTERS] CudaCpp MEs ( 2 ) : 16.9562s for 16384 events => throughput is 9.66E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0657s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_epemgguu/G116/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 19.7173s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6970s + [COUNTERS] CudaCpp MEs ( 2 ) : 16.9528s for 16384 events => throughput is 9.66E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0675s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_epemgguu/G363/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 21.1903s + [COUNTERS] Fortran Overhead ( 0 ) : 4.1777s + [COUNTERS] CudaCpp MEs ( 2 ) : 16.9471s for 16384 events => throughput is 9.67E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0655s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_epemgguu/G138/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 19.1144s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5931s + [COUNTERS] CudaCpp MEs ( 2 ) : 17.4555s for 16384 events => throughput is 9.39E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0658s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_epemgguu/G102/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 20.2019s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1275s + [COUNTERS] CudaCpp MEs ( 2 ) : 17.0086s for 16384 events => throughput is 9.63E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0659s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_epemgguu/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgdssx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgdssx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgdssx/ajob1' in 12.4888 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgdssx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgdssx/G28/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 12.4375s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0543s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.3452s for 16384 events => throughput is 1.58E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0381s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgdssx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_taptamgguu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_taptamgguu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_taptamgguu/ajob1' in 180.4466 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_taptamgguu/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_taptamgguu/G464/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 36.0069s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9058s + [COUNTERS] CudaCpp MEs ( 2 ) : 34.0355s for 16384 events => throughput is 4.81E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0656s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_taptamgguu/G59/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 36.1612s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0892s + [COUNTERS] CudaCpp MEs ( 2 ) : 34.0044s for 16384 events => throughput is 4.82E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0676s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_taptamgguu/G100/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 36.5060s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5244s + [COUNTERS] CudaCpp MEs ( 2 ) : 33.9153s for 16384 events => throughput is 4.83E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0663s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_taptamgguu/G106/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 35.8420s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8260s + [COUNTERS] CudaCpp MEs ( 2 ) : 33.9503s for 16384 events => throughput is 4.83E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0657s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_taptamgguu/G143/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 35.8263s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7863s + [COUNTERS] CudaCpp MEs ( 2 ) : 33.9743s for 16384 events => throughput is 4.82E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0657s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_taptamgguu/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dc_epemggdc/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dc_epemggdc/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dc_epemggdc/ajob1' in 10.7437 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dc_epemggdc/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dc_epemggdc/G18/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 10.6979s + [COUNTERS] Fortran Overhead ( 0 ) : 5.2594s + [COUNTERS] CudaCpp MEs ( 2 ) : 5.4007s for 16384 events => throughput is 3.03E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0378s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dc_epemggdc/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ucx_taptamggucx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ucx_taptamggucx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ucx_taptamggucx/ajob1' in 15.1225 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ucx_taptamggucx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ucx_taptamggucx/G62/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 15.0755s + [COUNTERS] Fortran Overhead ( 0 ) : 4.7376s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.3002s for 16384 events => throughput is 1.59E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0377s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ucx_taptamggucx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamucdxcx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamucdxcx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamucdxcx/ajob1' in 3.4000 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamucdxcx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamucdxcx/G47/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.3552s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8099s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.5333s for 16384 events => throughput is 1.07E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0120s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamucdxcx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_epemgggg/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_epemgggg/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_epemgggg/ajob1' in 31.3128 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_epemgggg/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_epemgggg/G49/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 31.2669s + [COUNTERS] Fortran Overhead ( 0 ) : 3.4223s + [COUNTERS] CudaCpp MEs ( 2 ) : 27.7558s for 16384 events => throughput is 5.90E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0887s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_epemgggg/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemudddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemudddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemudddx/ajob1' in 3.6547 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemudddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemudddx/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.6115s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6084s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.9839s for 16384 events => throughput is 8.26E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0192s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemudddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamggudx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamggudx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamggudx/ajob1' in 11.8761 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamggudx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamggudx/G60/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 11.8335s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6363s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.1593s for 16384 events => throughput is 1.61E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0379s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamggudx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_taptamuuddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_taptamuuddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_taptamuuddx/ajob1' in 5.4996 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_taptamuuddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_taptamuuddx/G11/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.4579s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7520s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.6865s for 16384 events => throughput is 4.44E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0194s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_taptamuuddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished in 14243.3670 seconds +__CUDACPP_DEBUG: gen_ximprove.launch finished in 14247.4407 seconds +__CUDACPP_DEBUG: GridPackCmd.refine4grid finished in 14247.4415 seconds +__CUDACPP_DEBUG: GridPackCmd.launch (3a) combine_events +combine_events +INFO: fail to reach target 500 +__CUDACPP_DEBUG: GridPackCmd.launch (3b) store_events +store_events + === Results Summary for run: GridRun_21 tag: tag_1 === + + Cross-section : 0 +- 0 pb + Nb of events : 195 + +INFO: No version of lhapdf. Can not run systematics computation +decay_events -from_cards +__CUDACPP_DEBUG: GridPackCmd.launch finished in 14398.4664 seconds +__CUDACPP_DEBUG: GridPackCmd.__init__ back from self.launch +DONE +write ./events.lhe.gz +END: Sun Sep 15 11:39:24 AM CEST 2024 +ELAPSED: 14401 seconds diff --git a/epochX/cudacpp/tlau/fromgridpacks/pp_dy4j.mad/cuda/output.txt b/epochX/cudacpp/tlau/fromgridpacks/pp_dy4j.mad/cuda/output.txt new file mode 100644 index 0000000000..54240604c7 --- /dev/null +++ b/epochX/cudacpp/tlau/fromgridpacks/pp_dy4j.mad/cuda/output.txt @@ -0,0 +1,3327 @@ +START: Sat Sep 14 11:24:46 PM CEST 2024 +On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU (nproc=4)] [GPU: 1x Tesla V100S-PCIE-32GB]: +CUDACPP_RUNTIME_DISABLEFPE is set +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_bbx_epemdsdxsx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_bbx_taptamdsdxsx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_ccx_epemdcdxcx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_ccx_epemdsdxsx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_ccx_epemucuxcx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_ccx_epemuduxdx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_ccx_taptamdcdxcx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_ccx_taptamdsdxsx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_ccx_taptamucuxcx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_ccx_taptamuduxdx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_dc_epemdcccx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_dc_epemdcssx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_dc_epemddcdx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_dc_epemggdc/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_dc_epemudcux/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_dc_taptamdcccx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_dc_taptamdcssx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_dc_taptamddcdx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_dc_taptamggdc/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_dc_taptamudcux/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_dd_epemddddx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_dd_epemddssx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_dd_epemggdd/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_dd_epemuddux/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_dd_taptamddddx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_dd_taptamddssx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_dd_taptamggdd/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_dd_taptamuddux/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_ddx_epemdcdxcx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_ddx_epemdddxdx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_ddx_epemdsdxsx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_ddx_epemggddx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_ddx_epemgggg/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_ddx_epemggssx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_ddx_epemgguux/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_ddx_epemsbsxbx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_ddx_epemsssxsx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_ddx_epemucuxcx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_ddx_epemuduxdx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_ddx_epemusuxsx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_ddx_epemuuuxux/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_ddx_taptamdcdxcx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_ddx_taptamdddxdx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_ddx_taptamdsdxsx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_ddx_taptamggddx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_ddx_taptamgggg/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_ddx_taptamggssx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_ddx_taptamgguux/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_ddx_taptamsbsxbx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_ddx_taptamsssxsx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_ddx_taptamucuxcx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_ddx_taptamuduxdx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_ddx_taptamusuxsx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_ddx_taptamuuuxux/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_ds_epemddsdx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_ds_epemdsbbx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_ds_epemdsssx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_ds_epemggds/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_ds_epemudsux/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_ds_taptamddsdx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_ds_taptamdsbbx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_ds_taptamdsssx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_ds_taptamggds/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_ds_taptamudsux/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_dsx_epemdbsxbx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_dsx_epemdddxsx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_dsx_epemdssxsx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_dsx_epemggdsx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_dsx_epemuduxsx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_dsx_taptamdbsxbx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_dsx_taptamdddxsx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_dsx_taptamdssxsx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_dsx_taptamggdsx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_dsx_taptamuduxsx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_dux_epemdcuxcx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_dux_epemdduxdx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_dux_epemdsuxsx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_dux_epemggdux/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_dux_epemuduxux/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_dux_taptamdcuxcx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_dux_taptamdduxdx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_dux_taptamdsuxsx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_dux_taptamggdux/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_dux_taptamuduxux/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_dxcx_epemcdxcxcx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_dxcx_epemddxdxcx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_dxcx_epemggdxcx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_dxcx_epemsdxcxsx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_dxcx_epemuuxdxcx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_dxcx_taptamcdxcxcx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_dxcx_taptamddxdxcx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_dxcx_taptamggdxcx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_dxcx_taptamsdxcxsx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_dxcx_taptamuuxdxcx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_dxdx_epemddxdxdx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_dxdx_epemggdxdx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_dxdx_epemsdxdxsx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_dxdx_epemuuxdxdx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_dxdx_taptamddxdxdx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_dxdx_taptamggdxdx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_dxdx_taptamsdxdxsx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_dxdx_taptamuuxdxdx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_dxsx_epembdxsxbx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_dxsx_epemddxdxsx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_dxsx_epemggdxsx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_dxsx_epemsdxsxsx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_dxsx_epemuuxdxsx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_dxsx_taptambdxsxbx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_dxsx_taptamddxdxsx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_dxsx_taptamggdxsx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_dxsx_taptamsdxsxsx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_dxsx_taptamuuxdxsx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_gd_epemgdddx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_gd_epemgdssx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_gd_epemgggd/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_gd_epemgudux/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_gd_taptamgdddx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_gd_taptamgdssx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_gd_taptamgggd/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_gd_taptamgudux/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_gdx_epemgddxdx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_gdx_epemgggdx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_gdx_epemgsdxsx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_gdx_epemguuxdx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_gdx_taptamgddxdx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_gdx_taptamgggdx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_gdx_taptamgsdxsx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_gdx_taptamguuxdx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_gg_epemdddxdx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_gg_epemdsdxsx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_gg_epemggddx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_gg_epemgguux/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_gg_epemucuxcx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_gg_epemuduxdx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_gg_epemuuuxux/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_gg_taptamdddxdx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_gg_taptamdsdxsx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_gg_taptamggddx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_gg_taptamgguux/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_gg_taptamucuxcx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_gg_taptamuduxdx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_gg_taptamuuuxux/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_gu_epemgggu/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_gu_epemguccx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_gu_epemguddx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_gu_epemguuux/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_gu_taptamgggu/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_gu_taptamguccx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_gu_taptamguddx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_gu_taptamguuux/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_gux_epemgcuxcx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_gux_epemgduxdx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_gux_epemgggux/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_gux_epemguuxux/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_gux_taptamgcuxcx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_gux_taptamgduxdx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_gux_taptamgggux/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_gux_taptamguuxux/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_ssx_epemdbdxbx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_ssx_epemdcdxcx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_ssx_epemdsdxsx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_ssx_epemucuxcx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_ssx_epemuduxdx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_ssx_taptamdbdxbx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_ssx_taptamdcdxcx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_ssx_taptamdsdxsx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_ssx_taptamucuxcx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_ssx_taptamuduxdx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_uc_epemgguc/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_uc_epemucccx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_uc_epemudcdx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_uc_epemuucux/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_uc_taptamgguc/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_uc_taptamucccx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_uc_taptamudcdx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_uc_taptamuucux/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_ucx_epemggucx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_ucx_epemuccxcx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_ucx_epemuddxcx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_ucx_epemuuuxcx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_ucx_taptamggucx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_ucx_taptamuccxcx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_ucx_taptamuddxcx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_ucx_taptamuuuxcx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_ud_epemggud/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_ud_epemudccx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_ud_epemudddx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_ud_epemudssx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_ud_epemuudux/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_ud_taptamggud/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_ud_taptamudccx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_ud_taptamudddx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_ud_taptamudssx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_ud_taptamuudux/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_udx_epemggudx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_udx_epemucdxcx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_udx_epemuddxdx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_udx_epemusdxsx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_udx_epemuuuxdx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_udx_taptamggudx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_udx_taptamucdxcx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_udx_taptamuddxdx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_udx_taptamusdxsx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_udx_taptamuuuxdx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_uu_epemgguu/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_uu_epemuuccx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_uu_epemuuddx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_uu_epemuuuux/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_uu_taptamgguu/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_uu_taptamuuccx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_uu_taptamuuddx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_uu_taptamuuuux/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_uux_epemcccxcx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_uux_epemcscxsx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_uux_epemdcdxcx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_uux_epemdddxdx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_uux_epemdsdxsx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_uux_epemggccx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_uux_epemggddx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_uux_epemgggg/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_uux_epemgguux/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_uux_epemucuxcx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_uux_epemuduxdx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_uux_epemuuuxux/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_uux_taptamcccxcx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_uux_taptamcscxsx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_uux_taptamdcdxcx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_uux_taptamdddxdx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_uux_taptamdsdxsx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_uux_taptamggccx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_uux_taptamggddx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_uux_taptamgggg/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_uux_taptamgguux/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_uux_taptamucuxcx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_uux_taptamuduxdx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_uux_taptamuuuxux/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_uxcx_epemcuxcxcx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_uxcx_epemduxdxcx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_uxcx_epemgguxcx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_uxcx_epemuuxuxcx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_uxcx_taptamcuxcxcx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_uxcx_taptamduxdxcx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_uxcx_taptamgguxcx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_uxcx_taptamuuxuxcx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_uxdx_epemcuxdxcx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_uxdx_epemduxdxdx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_uxdx_epemgguxdx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_uxdx_epemsuxdxsx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_uxdx_epemuuxuxdx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_uxdx_taptamcuxdxcx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_uxdx_taptamduxdxdx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_uxdx_taptamgguxdx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_uxdx_taptamsuxdxsx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_uxdx_taptamuuxuxdx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_uxux_epemcuxuxcx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_uxux_epemduxuxdx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_uxux_epemgguxux/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_uxux_epemuuxuxux/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_uxux_taptamcuxuxcx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_uxux_taptamduxuxdx/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_uxux_taptamgguxux/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +lrwxrwxrwx. 1 avalassi zg 36 Sep 14 23:24 madevent/SubProcesses/P0_uxux_taptamuuxuxux/madevent -> build.cuda_d_inl0_hrd0/madevent_cuda +Now generating 500 events with random seed 21 and granularity 1 +************************************************************ +* * +* W E L C O M E to * +* M A D G R A P H 5 _ a M C @ N L O * +* M A D E V E N T * +* * +* * * * +* * * * * * +* * * * * 5 * * * * * +* * * * * * +* * * * +* * +* VERSION 3.5.3_lo_vect * +* * +* The MadGraph5_aMC@NLO Development Team - Find us at * +* https://server06.fynu.ucl.ac.be/projects/madgraph * +* * +* Type 'help' for in-line help. * +* * +************************************************************ +INFO: load configuration from /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/Cards/me5_configuration.txt +INFO: load configuration from /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/Cards/me5_configuration.txt +Using default text editor "vi". Set another one in ./input/mg5_configuration.txt +Using default eps viewer "evince". Set another one in ./input/mg5_configuration.txt +Using default web browser "firefox". Set another one in ./input/mg5_configuration.txt +WRITE GRIDCARD /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent +__CUDACPP_DEBUG: GridPackCmd.__init__ will call self.launch +__CUDACPP_DEBUG: GridPackCmd.launch starting +__CUDACPP_DEBUG: GridPackCmd.launch (1) restore_data +generate 500 events +P0_gg_epemgguux +P0_gg_epemggddx +P0_gg_taptamgguux +P0_gg_taptamggddx +P0_gu_epemgggu +P0_gd_epemgggd +P0_gux_epemgggux +P0_gdx_epemgggdx +P0_gu_taptamgggu +P0_gd_taptamgggd +P0_gux_taptamgggux +P0_gdx_taptamgggdx +P0_uux_epemgggg +P0_ddx_epemgggg +P0_uux_taptamgggg +P0_ddx_taptamgggg +P0_gg_epemuuuxux +P0_gg_epemdddxdx +P0_gg_epemuduxdx +P0_gg_epemucuxcx +P0_gg_epemdsdxsx +P0_gg_taptamuuuxux +P0_gg_taptamdddxdx +P0_gg_taptamuduxdx +P0_gg_taptamucuxcx +P0_gg_taptamdsdxsx +P0_gu_epemguuux +P0_gd_epemgdddx +P0_gux_epemguuxux +P0_gdx_epemgddxdx +P0_gu_epemguddx +P0_gu_epemguccx +P0_gd_epemgudux +P0_gd_epemgdssx +P0_gux_epemgduxdx +P0_gux_epemgcuxcx +P0_gdx_epemguuxdx +P0_gdx_epemgsdxsx +P0_gu_taptamguuux +P0_gd_taptamgdddx +P0_gux_taptamguuxux +P0_gdx_taptamgddxdx +P0_gu_taptamguddx +P0_gu_taptamguccx +P0_gd_taptamgudux +P0_gd_taptamgdssx +P0_gux_taptamgduxdx +P0_gux_taptamgcuxcx +P0_gdx_taptamguuxdx +P0_gdx_taptamgsdxsx +P0_uu_epemgguu +P0_uux_epemgguux +P0_dd_epemggdd +P0_ddx_epemggddx +P0_uxux_epemgguxux +P0_dxdx_epemggdxdx +P0_ud_epemggud +P0_uc_epemgguc +P0_uux_epemggddx +P0_uux_epemggccx +P0_udx_epemggudx +P0_ucx_epemggucx +P0_dc_epemggdc +P0_ds_epemggds +P0_dux_epemggdux +P0_ddx_epemgguux +P0_ddx_epemggssx +P0_dsx_epemggdsx +P0_uxdx_epemgguxdx +P0_uxcx_epemgguxcx +P0_dxcx_epemggdxcx +P0_dxsx_epemggdxsx +P0_uu_taptamgguu +P0_uux_taptamgguux +P0_dd_taptamggdd +P0_ddx_taptamggddx +P0_uxux_taptamgguxux +P0_dxdx_taptamggdxdx +P0_ud_taptamggud +P0_uc_taptamgguc +P0_uux_taptamggddx +P0_uux_taptamggccx +P0_udx_taptamggudx +P0_ucx_taptamggucx +P0_dc_taptamggdc +P0_ds_taptamggds +P0_dux_taptamggdux +P0_ddx_taptamgguux +P0_ddx_taptamggssx +P0_dsx_taptamggdsx +P0_uxdx_taptamgguxdx +P0_uxcx_taptamgguxcx +P0_dxcx_taptamggdxcx +P0_dxsx_taptamggdxsx +P0_uu_epemuuuux +P0_uux_epemuuuxux +P0_dd_epemddddx +P0_ddx_epemdddxdx +P0_uxux_epemuuxuxux +P0_dxdx_epemddxdxdx +P0_uu_epemuuddx +P0_uu_epemuuccx +P0_ud_epemuudux +P0_ud_epemudddx +P0_uc_epemuucux +P0_uc_epemucccx +P0_uux_epemuduxdx +P0_uux_epemucuxcx +P0_uux_epemdddxdx +P0_uux_epemcccxcx +P0_udx_epemuuuxdx +P0_udx_epemuddxdx +P0_ucx_epemuuuxcx +P0_ucx_epemuccxcx +P0_dd_epemuddux +P0_dd_epemddssx +P0_dc_epemddcdx +P0_dc_epemdcccx +P0_ds_epemddsdx +P0_ds_epemdsssx +P0_dux_epemuduxux +P0_dux_epemdduxdx +P0_ddx_epemuuuxux +P0_ddx_epemuduxdx +P0_ddx_epemdcdxcx +P0_ddx_epemdsdxsx +P0_ddx_epemsssxsx +P0_dsx_epemdddxsx +P0_dsx_epemdssxsx +P0_ccx_epemucuxcx +P0_ccx_epemdcdxcx +P0_ssx_epemdsdxsx +P0_uxux_epemduxuxdx +P0_uxux_epemcuxuxcx +P0_uxdx_epemuuxuxdx +P0_uxdx_epemduxdxdx +P0_uxcx_epemuuxuxcx +P0_uxcx_epemcuxcxcx +P0_dxdx_epemuuxdxdx +P0_dxdx_epemsdxdxsx +P0_dxcx_epemddxdxcx +P0_dxcx_epemcdxcxcx +P0_dxsx_epemddxdxsx +P0_dxsx_epemsdxsxsx +P0_ud_epemudccx +P0_ud_epemudssx +P0_uc_epemudcdx +P0_uux_epemdcdxcx +P0_uux_epemdsdxsx +P0_uux_epemcscxsx +P0_udx_epemucdxcx +P0_udx_epemusdxsx +P0_ucx_epemuddxcx +P0_dc_epemudcux +P0_dc_epemdcssx +P0_ds_epemudsux +P0_ds_epemdsbbx +P0_dux_epemdcuxcx +P0_dux_epemdsuxsx +P0_ddx_epemucuxcx +P0_ddx_epemusuxsx +P0_ddx_epemsbsxbx +P0_dsx_epemuduxsx +P0_dsx_epemdbsxbx +P0_ccx_epemuduxdx +P0_ccx_epemdsdxsx +P0_ssx_epemuduxdx +P0_ssx_epemucuxcx +P0_ssx_epemdcdxcx +P0_ssx_epemdbdxbx +P0_bbx_epemdsdxsx +P0_uxdx_epemcuxdxcx +P0_uxdx_epemsuxdxsx +P0_uxcx_epemduxdxcx +P0_dxcx_epemuuxdxcx +P0_dxcx_epemsdxcxsx +P0_dxsx_epemuuxdxsx +P0_dxsx_epembdxsxbx +P0_uu_taptamuuuux +P0_uux_taptamuuuxux +P0_dd_taptamddddx +P0_ddx_taptamdddxdx +P0_uxux_taptamuuxuxux +P0_dxdx_taptamddxdxdx +P0_uu_taptamuuddx +P0_uu_taptamuuccx +P0_ud_taptamuudux +P0_ud_taptamudddx +P0_uc_taptamuucux +P0_uc_taptamucccx +P0_uux_taptamuduxdx +P0_uux_taptamucuxcx +P0_uux_taptamdddxdx +P0_uux_taptamcccxcx +P0_udx_taptamuuuxdx +P0_udx_taptamuddxdx +P0_ucx_taptamuuuxcx +P0_ucx_taptamuccxcx +P0_dd_taptamuddux +P0_dd_taptamddssx +P0_dc_taptamddcdx +P0_dc_taptamdcccx +P0_ds_taptamddsdx +P0_ds_taptamdsssx +P0_dux_taptamuduxux +P0_dux_taptamdduxdx +P0_ddx_taptamuuuxux +P0_ddx_taptamuduxdx +P0_ddx_taptamdcdxcx +P0_ddx_taptamdsdxsx +P0_ddx_taptamsssxsx +P0_dsx_taptamdddxsx +P0_dsx_taptamdssxsx +P0_ccx_taptamucuxcx +P0_ccx_taptamdcdxcx +P0_ssx_taptamdsdxsx +P0_uxux_taptamduxuxdx +P0_uxux_taptamcuxuxcx +P0_uxdx_taptamuuxuxdx +P0_uxdx_taptamduxdxdx +P0_uxcx_taptamuuxuxcx +P0_uxcx_taptamcuxcxcx +P0_dxdx_taptamuuxdxdx +P0_dxdx_taptamsdxdxsx +P0_dxcx_taptamddxdxcx +P0_dxcx_taptamcdxcxcx +P0_dxsx_taptamddxdxsx +P0_dxsx_taptamsdxsxsx +P0_ud_taptamudccx +P0_ud_taptamudssx +P0_uc_taptamudcdx +P0_uux_taptamdcdxcx +P0_uux_taptamdsdxsx +P0_uux_taptamcscxsx +P0_udx_taptamucdxcx +P0_udx_taptamusdxsx +P0_ucx_taptamuddxcx +P0_dc_taptamudcux +P0_dc_taptamdcssx +P0_ds_taptamudsux +P0_ds_taptamdsbbx +P0_dux_taptamdcuxcx +P0_dux_taptamdsuxsx +P0_ddx_taptamucuxcx +P0_ddx_taptamusuxsx +P0_ddx_taptamsbsxbx +P0_dsx_taptamuduxsx +P0_dsx_taptamdbsxbx +P0_ccx_taptamuduxdx +P0_ccx_taptamdsdxsx +P0_ssx_taptamuduxdx +P0_ssx_taptamucuxcx +P0_ssx_taptamdcdxcx +P0_ssx_taptamdbdxbx +P0_bbx_taptamdsdxsx +P0_uxdx_taptamcuxdxcx +P0_uxdx_taptamsuxdxsx +P0_uxcx_taptamduxdxcx +P0_dxcx_taptamuuxdxcx +P0_dxcx_taptamsdxcxsx +P0_dxsx_taptamuuxdxsx +P0_dxsx_taptambdxsxbx +__CUDACPP_DEBUG: GridPackCmd.launch (2) refine4grid +__CUDACPP_DEBUG: GridPackCmd.refine4grid starting +Using random number seed offset = 21 +INFO: Generating 500.0 unweighted events. +__CUDACPP_DEBUG: gen_ximprove.launch starting +sum of cpu time of last step: 3 days,23h23m17s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event starting +INFO: need to improve 477 channels +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dxdx_taptamuuxdxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dxdx_taptamuuxdxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dxdx_taptamuuxdxdx/ajob1' in 4.8479 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dxdx_taptamuuxdxdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dxdx_taptamuuxdxdx/G21/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.4354s + [COUNTERS] Fortran Overhead ( 0 ) : 4.2154s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0403s for 16384 events => throughput is 4.06E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1797s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dxdx_taptamuuxdxdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamguuxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamguuxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamguuxdx/ajob1' in 7.0712 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamguuxdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamguuxdx/G43/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.9084s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4524s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0944s for 16384 events => throughput is 1.74E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3616s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamguuxdx/G335/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.5809s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1350s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0936s for 16384 events => throughput is 1.75E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3523s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamguuxdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemuuuxux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemuuuxux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemuuuxux/ajob1' in 4.1095 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemuuuxux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemuuuxux/G689/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.7701s + [COUNTERS] Fortran Overhead ( 0 ) : 3.0198s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1235s for 16384 events => throughput is 1.33E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.6268s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemuuuxux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamgguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamgguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamgguux/ajob1' in 60.1926 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamgguux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamgguux/G596/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.0929s + [COUNTERS] Fortran Overhead ( 0 ) : 3.8460s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4173s for 16384 events => throughput is 3.93E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8296s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamgguux/G304/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.6715s + [COUNTERS] Fortran Overhead ( 0 ) : 3.4171s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4186s for 16384 events => throughput is 3.91E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8359s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamgguux/G128/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.7484s + [COUNTERS] Fortran Overhead ( 0 ) : 3.4950s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4175s for 16384 events => throughput is 3.92E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8359s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamgguux/G115/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.3716s + [COUNTERS] Fortran Overhead ( 0 ) : 7.1200s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4169s for 16384 events => throughput is 3.93E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8347s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamgguux/G599/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.5080s + [COUNTERS] Fortran Overhead ( 0 ) : 4.2541s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4169s for 16384 events => throughput is 3.93E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8369s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamgguux/G351/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.8626s + [COUNTERS] Fortran Overhead ( 0 ) : 6.6093s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4183s for 16384 events => throughput is 3.92E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8350s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamgguux/G111/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.0478s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7983s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4173s for 16384 events => throughput is 3.93E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8323s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamgguux/G354/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.3192s + [COUNTERS] Fortran Overhead ( 0 ) : 7.0673s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4173s for 16384 events => throughput is 3.93E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8346s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamgguux/G97/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.9492s + [COUNTERS] Fortran Overhead ( 0 ) : 7.6980s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4189s for 16384 events => throughput is 3.91E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8323s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamgguux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uxdx_epemsuxdxsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uxdx_epemsuxdxsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uxdx_epemsuxdxsx/ajob1' in 3.0778 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uxdx_epemsuxdxsx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uxdx_epemsuxdxsx/G74/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.7617s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6231s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0137s for 16384 events => throughput is 1.20E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1249s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uxdx_epemsuxdxsx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemdddxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemdddxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemdddxdx/ajob1' in 8.2288 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemdddxdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemdddxdx/G513/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.6798s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9293s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1239s for 16384 events => throughput is 1.32E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.6266s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemdddxdx/G180/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.9470s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1996s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1238s for 16384 events => throughput is 1.32E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.6236s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemdddxdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgddxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgddxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgddxdx/ajob1' in 4.6155 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgddxdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgddxdx/G515/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.2785s + [COUNTERS] Fortran Overhead ( 0 ) : 3.5290s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1236s for 16384 events => throughput is 1.33E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.6259s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgddxdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguddx/ajob1' in 33.9611 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguddx/G104/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.4772s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0287s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0938s for 16384 events => throughput is 1.75E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3547s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguddx/G120/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.0394s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5824s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0949s for 16384 events => throughput is 1.73E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3621s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguddx/G305/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.7799s + [COUNTERS] Fortran Overhead ( 0 ) : 6.3218s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0949s for 16384 events => throughput is 1.73E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3631s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguddx/G101/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.3364s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8801s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0947s for 16384 events => throughput is 1.73E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3615s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguddx/G27/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.7608s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3074s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0941s for 16384 events => throughput is 1.74E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3593s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguddx/G273/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.0300s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5790s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0941s for 16384 events => throughput is 1.74E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3569s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguddx/G349/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.2671s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8151s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0940s for 16384 events => throughput is 1.74E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3579s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguddx/G379/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.6495s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1931s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0948s for 16384 events => throughput is 1.73E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3616s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguddx/G256/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.0608s + [COUNTERS] Fortran Overhead ( 0 ) : 5.6046s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0960s for 16384 events => throughput is 1.71E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3602s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemguuxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemguuxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemguuxdx/ajob1' in 16.8604 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemguuxdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemguuxdx/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.9074s + [COUNTERS] Fortran Overhead ( 0 ) : 6.4946s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0510s for 16384 events => throughput is 3.22E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3619s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemguuxdx/G68/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.1403s + [COUNTERS] Fortran Overhead ( 0 ) : 5.7333s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0505s for 16384 events => throughput is 3.24E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3564s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemguuxdx/G355/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.9353s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5243s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0505s for 16384 events => throughput is 3.25E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3606s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemguuxdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/ajob1' in 114.1815 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G14/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.7427s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4909s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4223s for 16384 events => throughput is 3.88E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8296s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G56/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.1873s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9386s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4175s for 16384 events => throughput is 3.92E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8313s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G517/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.9011s + [COUNTERS] Fortran Overhead ( 0 ) : 6.6424s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4238s for 16384 events => throughput is 3.87E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8349s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G52/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.5085s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2542s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4190s for 16384 events => throughput is 3.91E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8353s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G304/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.4089s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1563s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4192s for 16384 events => throughput is 3.91E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8335s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G3/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.5593s + [COUNTERS] Fortran Overhead ( 0 ) : 5.3091s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4196s for 16384 events => throughput is 3.90E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8305s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G23/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.4023s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1590s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4198s for 16384 events => throughput is 3.90E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8235s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G33/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.9388s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6854s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4222s for 16384 events => throughput is 3.88E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8312s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G53/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.3597s + [COUNTERS] Fortran Overhead ( 0 ) : 7.1081s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4202s for 16384 events => throughput is 3.90E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8314s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G495/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.3054s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0589s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4197s for 16384 events => throughput is 3.90E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8268s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G68/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.8887s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6404s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4189s for 16384 events => throughput is 3.91E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8295s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G49/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.7845s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5280s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4224s for 16384 events => throughput is 3.88E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8340s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G524/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.8620s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6173s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4190s for 16384 events => throughput is 3.91E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8257s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G65/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.8072s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5583s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4200s for 16384 events => throughput is 3.90E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8290s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G303/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.6767s + [COUNTERS] Fortran Overhead ( 0 ) : 3.4307s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4197s for 16384 events => throughput is 3.90E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8264s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G55/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.3823s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1355s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4188s for 16384 events => throughput is 3.91E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8280s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G34/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.5016s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2481s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4206s for 16384 events => throughput is 3.90E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8329s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G509/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 9.3210s + [COUNTERS] Fortran Overhead ( 0 ) : 8.0706s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4219s for 16384 events => throughput is 3.88E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8284s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G512/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.5033s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2523s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4202s for 16384 events => throughput is 3.90E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8308s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G511/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.1240s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8722s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4232s for 16384 events => throughput is 3.87E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8286s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.5758s + [COUNTERS] Fortran Overhead ( 0 ) : 3.3273s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4184s for 16384 events => throughput is 3.92E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8302s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G63/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.8115s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5676s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4207s for 16384 events => throughput is 3.89E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8231s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G48/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.0071s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7551s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4210s for 16384 events => throughput is 3.89E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8310s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/ajob1' in 159.9469 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G122/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.1581s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1248s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2134s for 16384 events => throughput is 7.68E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8199s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G108/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.8072s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7733s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2127s for 16384 events => throughput is 7.70E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8212s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G105/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.9536s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9162s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2129s for 16384 events => throughput is 7.70E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8245s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G304/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.5391s + [COUNTERS] Fortran Overhead ( 0 ) : 3.5062s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2132s for 16384 events => throughput is 7.68E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8197s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G301/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 9.1343s + [COUNTERS] Fortran Overhead ( 0 ) : 8.0985s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2124s for 16384 events => throughput is 7.72E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8234s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G128/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.6388s + [COUNTERS] Fortran Overhead ( 0 ) : 3.6091s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2130s for 16384 events => throughput is 7.69E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8167s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G100/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 9.1035s + [COUNTERS] Fortran Overhead ( 0 ) : 8.0672s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2128s for 16384 events => throughput is 7.70E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8235s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G120/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.9160s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8838s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2133s for 16384 events => throughput is 7.68E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8189s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G302/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 12.2022s + [COUNTERS] Fortran Overhead ( 0 ) : 11.1665s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2129s for 16384 events => throughput is 7.70E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8229s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G299/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.1579s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1198s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2142s for 16384 events => throughput is 7.65E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8239s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G169/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.5841s + [COUNTERS] Fortran Overhead ( 0 ) : 6.5496s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2148s for 16384 events => throughput is 7.63E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8197s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G333/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 12.3823s + [COUNTERS] Fortran Overhead ( 0 ) : 11.3523s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2122s for 16384 events => throughput is 7.72E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8179s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G594/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 10.7675s + [COUNTERS] Fortran Overhead ( 0 ) : 9.7319s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2123s for 16384 events => throughput is 7.72E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8233s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G678/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.8298s + [COUNTERS] Fortran Overhead ( 0 ) : 6.7924s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2150s for 16384 events => throughput is 7.62E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8224s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G127/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.1827s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1475s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2147s for 16384 events => throughput is 7.63E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8205s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G126/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 11.1270s + [COUNTERS] Fortran Overhead ( 0 ) : 10.0901s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2135s for 16384 events => throughput is 7.67E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8234s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G598/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.6670s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6337s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2121s for 16384 events => throughput is 7.72E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8212s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G512/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.5818s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5502s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2133s for 16384 events => throughput is 7.68E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8184s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G103/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 10.4590s + [COUNTERS] Fortran Overhead ( 0 ) : 9.4220s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2121s for 16384 events => throughput is 7.72E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8249s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G107/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.8499s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8132s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2144s for 16384 events => throughput is 7.64E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8223s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G358/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.1844s + [COUNTERS] Fortran Overhead ( 0 ) : 4.1538s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2131s for 16384 events => throughput is 7.69E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8176s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G97/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.8508s + [COUNTERS] Fortran Overhead ( 0 ) : 7.8198s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2123s for 16384 events => throughput is 7.72E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8186s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G179/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.2181s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1838s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2128s for 16384 events => throughput is 7.70E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8215s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_epemudcdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_epemudcdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_epemudcdx/ajob1' in 2.5759 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_epemudcdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_epemudcdx/G42/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.2643s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1262s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0138s for 16384 events => throughput is 1.18E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1242s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_epemudcdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ds_taptamggds/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ds_taptamggds/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ds_taptamggds/ajob1' in 8.6856 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ds_taptamggds/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ds_taptamggds/G17/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.5109s + [COUNTERS] Fortran Overhead ( 0 ) : 5.0619s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0932s for 16384 events => throughput is 1.76E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3558s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ds_taptamggds/G110/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.5834s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1346s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0933s for 16384 events => throughput is 1.76E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3554s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ds_taptamggds/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_epemggdux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_epemggdux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_epemggdux/ajob1' in 13.1166 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_epemggdux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_epemggdux/G208/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.8466s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4377s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0504s for 16384 events => throughput is 3.25E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3585s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_epemggdux/G192/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.2842s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8703s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0510s for 16384 events => throughput is 3.21E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3629s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_epemggdux/G380/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.4298s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0294s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0499s for 16384 events => throughput is 3.28E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3504s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_epemggdux/G127/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.0988s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6853s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0509s for 16384 events => throughput is 3.22E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3626s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_epemggdux/G223/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.0298s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6187s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0506s for 16384 events => throughput is 3.24E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3604s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_epemggdux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_taptamgguc/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_taptamgguc/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_taptamgguc/ajob1' in 14.6073 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_taptamgguc/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_taptamgguc/G20/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.2598s + [COUNTERS] Fortran Overhead ( 0 ) : 4.8130s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0929s for 16384 events => throughput is 1.76E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3538s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_taptamgguc/G110/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.1871s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7400s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0928s for 16384 events => throughput is 1.77E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3543s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_taptamgguc/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.2810s + [COUNTERS] Fortran Overhead ( 0 ) : 5.8369s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0923s for 16384 events => throughput is 1.77E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3518s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_taptamgguc/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uxcx_taptamgguxcx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uxcx_taptamgguxcx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uxcx_taptamgguxcx/ajob1' in 3.7413 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uxcx_taptamgguxcx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uxcx_taptamgguxcx/G187/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.4176s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9701s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0932s for 16384 events => throughput is 1.76E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3543s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uxcx_taptamgguxcx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/ajob1' in 132.9646 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G499/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.6717s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6364s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2119s for 16384 events => throughput is 7.73E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8235s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G56/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.4011s + [COUNTERS] Fortran Overhead ( 0 ) : 3.3609s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2151s for 16384 events => throughput is 7.62E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8251s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G39/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.1641s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1347s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2118s for 16384 events => throughput is 7.74E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8176s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G64/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.0036s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9671s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2141s for 16384 events => throughput is 7.65E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8224s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G13/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.1790s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1438s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2132s for 16384 events => throughput is 7.69E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8221s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G33/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.2472s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2142s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2124s for 16384 events => throughput is 7.71E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8206s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G43/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.5097s + [COUNTERS] Fortran Overhead ( 0 ) : 7.4727s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2120s for 16384 events => throughput is 7.73E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8249s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G68/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.3835s + [COUNTERS] Fortran Overhead ( 0 ) : 3.3478s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2140s for 16384 events => throughput is 7.66E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8217s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G49/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.5747s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5387s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2128s for 16384 events => throughput is 7.70E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8232s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G25/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.3581s + [COUNTERS] Fortran Overhead ( 0 ) : 6.3220s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2123s for 16384 events => throughput is 7.72E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8237s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G523/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.2059s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1671s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2145s for 16384 events => throughput is 7.64E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8244s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G395/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.8675s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8298s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2126s for 16384 events => throughput is 7.71E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8251s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G524/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.4520s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4138s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2127s for 16384 events => throughput is 7.70E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8255s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G65/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.4933s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4608s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2134s for 16384 events => throughput is 7.68E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8191s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G609/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.3990s + [COUNTERS] Fortran Overhead ( 0 ) : 3.3601s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2129s for 16384 events => throughput is 7.69E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8260s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G611/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.7760s + [COUNTERS] Fortran Overhead ( 0 ) : 3.7399s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2132s for 16384 events => throughput is 7.68E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8229s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G1/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.1241s + [COUNTERS] Fortran Overhead ( 0 ) : 5.0876s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2124s for 16384 events => throughput is 7.71E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8241s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G55/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.0691s + [COUNTERS] Fortran Overhead ( 0 ) : 3.0347s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2138s for 16384 events => throughput is 7.66E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8205s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G34/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.6183s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5824s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2133s for 16384 events => throughput is 7.68E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8226s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G26/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.7510s + [COUNTERS] Fortran Overhead ( 0 ) : 6.7161s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2117s for 16384 events => throughput is 7.74E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8232s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G58/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.1919s + [COUNTERS] Fortran Overhead ( 0 ) : 6.1541s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2141s for 16384 events => throughput is 7.65E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8236s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G2/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.5575s + [COUNTERS] Fortran Overhead ( 0 ) : 5.5281s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2118s for 16384 events => throughput is 7.74E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8177s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G512/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.5794s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5432s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2151s for 16384 events => throughput is 7.62E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8211s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G51/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.1490s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1120s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2121s for 16384 events => throughput is 7.73E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8249s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G29/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.2076s + [COUNTERS] Fortran Overhead ( 0 ) : 7.1689s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2145s for 16384 events => throughput is 7.64E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8242s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G48/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.5314s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4999s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2124s for 16384 events => throughput is 7.71E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8191s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/ajob1' in 54.1580 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/G140/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.6626s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2486s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0511s for 16384 events => throughput is 3.20E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3628s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/G114/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.2673s + [COUNTERS] Fortran Overhead ( 0 ) : 6.8554s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0505s for 16384 events => throughput is 3.25E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3614s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/G49/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.0914s + [COUNTERS] Fortran Overhead ( 0 ) : 7.6818s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0505s for 16384 events => throughput is 3.24E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3591s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/G120/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.8300s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4148s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0538s for 16384 events => throughput is 3.05E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3615s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/G262/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.5881s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1837s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0503s for 16384 events => throughput is 3.26E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3542s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/G27/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.6889s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2779s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0504s for 16384 events => throughput is 3.25E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3607s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/G44/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.4084s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9978s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0505s for 16384 events => throughput is 3.24E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3600s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/G96/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.2217s + [COUNTERS] Fortran Overhead ( 0 ) : 3.8166s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0503s for 16384 events => throughput is 3.26E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3548s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/G261/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.9208s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5107s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0506s for 16384 events => throughput is 3.24E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3595s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/G37/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.6991s + [COUNTERS] Fortran Overhead ( 0 ) : 5.2932s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0508s for 16384 events => throughput is 3.22E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3552s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/G28/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.6132s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1994s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0514s for 16384 events => throughput is 3.19E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3624s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.7507s + [COUNTERS] Fortran Overhead ( 0 ) : 6.3392s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0507s for 16384 events => throughput is 3.23E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3609s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_epemuduxsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_epemuduxsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_epemuduxsx/ajob1' in 3.4674 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_epemuduxsx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_epemuduxsx/G42/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.1705s + [COUNTERS] Fortran Overhead ( 0 ) : 3.0307s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0139s for 16384 events => throughput is 1.17E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1259s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_epemuduxsx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dd_epemggdd/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dd_epemggdd/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dd_epemggdd/ajob1' in 9.7727 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dd_epemggdd/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dd_epemggdd/G467/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.0314s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2862s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1236s for 16384 events => throughput is 1.33E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.6216s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dd_epemggdd/G351/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.4481s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6979s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1236s for 16384 events => throughput is 1.33E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.6266s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dd_epemggdd/G44/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.3901s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6420s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1243s for 16384 events => throughput is 1.32E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.6239s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dd_epemggdd/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamusdxsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamusdxsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamusdxsx/ajob1' in 2.0885 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamusdxsx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamusdxsx/G92/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 1.7722s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6248s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0221s for 16384 events => throughput is 7.41E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1254s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamusdxsx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemggddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemggddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemggddx/ajob1' in 26.4032 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemggddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemggddx/G175/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.2843s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5288s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1231s for 16384 events => throughput is 1.33E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.6324s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemggddx/G303/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.9035s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1487s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1230s for 16384 events => throughput is 1.33E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.6319s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemggddx/G111/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.9681s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2188s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1230s for 16384 events => throughput is 1.33E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.6263s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemggddx/G219/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.2153s + [COUNTERS] Fortran Overhead ( 0 ) : 3.4618s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1228s for 16384 events => throughput is 1.33E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.6307s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemggddx/G318/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.1505s + [COUNTERS] Fortran Overhead ( 0 ) : 6.3926s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1233s for 16384 events => throughput is 1.33E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.6346s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemggddx/G143/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.1278s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3702s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1230s for 16384 events => throughput is 1.33E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.6346s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemggddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemguuxux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemguuxux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemguuxux/ajob1' in 8.8114 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemguuxux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemguuxux/G615/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.2638s + [COUNTERS] Fortran Overhead ( 0 ) : 7.5148s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1238s for 16384 events => throughput is 1.32E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.6252s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemguuxux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamgggg/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamgggg/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamgggg/ajob1' in 4.6590 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamgggg/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamgggg/G66/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.3296s + [COUNTERS] Fortran Overhead ( 0 ) : 3.0760s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4169s for 16384 events => throughput is 3.93E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8367s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamgggg/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgduxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgduxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgduxdx/ajob1' in 3.2860 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgduxdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgduxdx/G44/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.9758s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5190s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0945s for 16384 events => throughput is 1.73E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3623s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgduxdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgudux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgudux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgudux/ajob1' in 44.1290 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgudux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgudux/G39/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.9687s + [COUNTERS] Fortran Overhead ( 0 ) : 6.5561s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0510s for 16384 events => throughput is 3.22E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3616s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgudux/G366/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.5351s + [COUNTERS] Fortran Overhead ( 0 ) : 5.1237s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0507s for 16384 events => throughput is 3.23E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3606s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgudux/G259/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.1661s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7403s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0640s for 16384 events => throughput is 2.56E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3618s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgudux/G28/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.9223s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5132s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0503s for 16384 events => throughput is 3.26E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3589s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgudux/G32/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.3134s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9111s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0504s for 16384 events => throughput is 3.25E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3519s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgudux/G61/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 9.2490s + [COUNTERS] Fortran Overhead ( 0 ) : 8.8408s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0508s for 16384 events => throughput is 3.22E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3574s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgudux/G267/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.5175s + [COUNTERS] Fortran Overhead ( 0 ) : 6.0990s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0565s for 16384 events => throughput is 2.90E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3621s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgudux/G144/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.2072s + [COUNTERS] Fortran Overhead ( 0 ) : 3.8027s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0502s for 16384 events => throughput is 3.26E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3544s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgudux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemuudux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemuudux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemuudux/ajob1' in 20.7390 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemuudux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemuudux/G104/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 20.4168s + [COUNTERS] Fortran Overhead ( 0 ) : 20.1198s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1030s for 81920 events => throughput is 7.95E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1940s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemuudux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamggddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamggddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamggddx/ajob1' in 40.2006 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamggddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamggddx/G108/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.9121s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6683s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4179s for 16384 events => throughput is 3.92E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8259s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamggddx/G304/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.4317s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1762s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4210s for 16384 events => throughput is 3.89E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8345s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamggddx/G128/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.5892s + [COUNTERS] Fortran Overhead ( 0 ) : 3.3408s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4178s for 16384 events => throughput is 3.92E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8305s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamggddx/G360/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.0487s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7984s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4233s for 16384 events => throughput is 3.87E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8270s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamggddx/G169/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.8323s + [COUNTERS] Fortran Overhead ( 0 ) : 6.5827s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4215s for 16384 events => throughput is 3.89E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8281s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamggddx/G176/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.7025s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4543s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4181s for 16384 events => throughput is 3.92E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8301s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamggddx/G303/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.5867s + [COUNTERS] Fortran Overhead ( 0 ) : 3.3316s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4228s for 16384 events => throughput is 3.87E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8323s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamggddx/G513/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.7891s + [COUNTERS] Fortran Overhead ( 0 ) : 3.5338s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4200s for 16384 events => throughput is 3.90E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8354s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamggddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguccx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguccx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguccx/ajob1' in 19.3837 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguccx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguccx/G283/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.8222s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4148s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0506s for 16384 events => throughput is 3.24E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3567s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguccx/G47/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.7603s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3629s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0499s for 16384 events => throughput is 3.28E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3475s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguccx/G35/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.4526s + [COUNTERS] Fortran Overhead ( 0 ) : 7.0508s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0498s for 16384 events => throughput is 3.29E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3520s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguccx/G373/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.5303s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1232s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0502s for 16384 events => throughput is 3.26E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3569s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguccx/G249/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.3573s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9538s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0498s for 16384 events => throughput is 3.29E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3537s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguccx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/ajob1' in 141.2046 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G14/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.4943s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4593s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2115s for 16384 events => throughput is 7.75E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8235s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G499/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.4375s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3996s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2137s for 16384 events => throughput is 7.67E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8242s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G56/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.1722s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1349s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2132s for 16384 events => throughput is 7.69E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8241s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G39/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.7445s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7102s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2138s for 16384 events => throughput is 7.66E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8205s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G52/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.3261s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2893s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2147s for 16384 events => throughput is 7.63E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8222s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G304/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.1056s + [COUNTERS] Fortran Overhead ( 0 ) : 3.0703s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2140s for 16384 events => throughput is 7.66E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8213s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G64/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.8094s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7713s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2137s for 16384 events => throughput is 7.67E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8244s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G3/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.4630s + [COUNTERS] Fortran Overhead ( 0 ) : 5.4235s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2136s for 16384 events => throughput is 7.67E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8258s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G13/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.1413s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1061s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2120s for 16384 events => throughput is 7.73E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8232s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G33/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.6344s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6009s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2135s for 16384 events => throughput is 7.68E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8200s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G484/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.0625s + [COUNTERS] Fortran Overhead ( 0 ) : 3.0234s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2138s for 16384 events => throughput is 7.66E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8252s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G68/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.6602s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6239s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2137s for 16384 events => throughput is 7.67E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8227s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G49/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.2438s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2139s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2103s for 16384 events => throughput is 7.79E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8196s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G40/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.3133s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2739s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2136s for 16384 events => throughput is 7.67E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8258s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G523/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.6200s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5814s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2138s for 16384 events => throughput is 7.66E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8248s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.0245s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9928s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2117s for 16384 events => throughput is 7.74E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8200s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G524/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.5604s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5216s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2151s for 16384 events => throughput is 7.62E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8237s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G65/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.6108s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5724s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2131s for 16384 events => throughput is 7.69E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8253s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G44/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.3645s + [COUNTERS] Fortran Overhead ( 0 ) : 6.3265s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2147s for 16384 events => throughput is 7.63E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8232s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.9813s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9454s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2138s for 16384 events => throughput is 7.66E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8222s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G303/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.4929s + [COUNTERS] Fortran Overhead ( 0 ) : 3.4533s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2135s for 16384 events => throughput is 7.67E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8262s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G528/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.6028s + [COUNTERS] Fortran Overhead ( 0 ) : 3.5681s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2127s for 16384 events => throughput is 7.70E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8220s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G55/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.4250s + [COUNTERS] Fortran Overhead ( 0 ) : 3.3922s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2116s for 16384 events => throughput is 7.74E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8211s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G298/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.5292s + [COUNTERS] Fortran Overhead ( 0 ) : 3.4924s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2136s for 16384 events => throughput is 7.67E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8231s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G34/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.1757s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1379s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2136s for 16384 events => throughput is 7.67E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8242s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G54/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.5807s + [COUNTERS] Fortran Overhead ( 0 ) : 7.5483s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2118s for 16384 events => throughput is 7.73E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8206s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G67/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.7703s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7299s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2147s for 16384 events => throughput is 7.63E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8258s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G512/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.4531s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4158s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2132s for 16384 events => throughput is 7.69E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8242s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G511/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.0772s + [COUNTERS] Fortran Overhead ( 0 ) : 3.0405s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2134s for 16384 events => throughput is 7.68E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8233s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.9769s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9427s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2131s for 16384 events => throughput is 7.69E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8211s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G21/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.2235s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1916s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2118s for 16384 events => throughput is 7.74E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8201s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G48/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.9659s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9294s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2132s for 16384 events => throughput is 7.68E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8233s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_taptamggdux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_taptamggdux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_taptamggdux/ajob1' in 13.5507 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_taptamggdux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_taptamggdux/G68/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.7547s + [COUNTERS] Fortran Overhead ( 0 ) : 4.3025s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0941s for 16384 events => throughput is 1.74E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3581s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_taptamggdux/G167/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.6595s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2074s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0943s for 16384 events => throughput is 1.74E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3578s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_taptamggdux/G92/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.6194s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1685s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0941s for 16384 events => throughput is 1.74E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3568s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_taptamggdux/G60/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.3554s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8985s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0950s for 16384 events => throughput is 1.73E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3620s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_taptamggdux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/ajob1' in 127.8872 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G56/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.3097s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2779s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2145s for 16384 events => throughput is 7.64E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8173s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G39/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.5227s + [COUNTERS] Fortran Overhead ( 0 ) : 3.4883s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2142s for 16384 events => throughput is 7.65E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8203s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G517/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 9.6036s + [COUNTERS] Fortran Overhead ( 0 ) : 8.5693s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2151s for 16384 events => throughput is 7.62E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8192s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G497/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.4411s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4058s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2141s for 16384 events => throughput is 7.65E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8212s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G13/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.1664s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1408s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2123s for 16384 events => throughput is 7.72E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8133s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G484/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.0913s + [COUNTERS] Fortran Overhead ( 0 ) : 7.0607s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2143s for 16384 events => throughput is 7.65E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8163s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G68/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.5261s + [COUNTERS] Fortran Overhead ( 0 ) : 3.4939s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2125s for 16384 events => throughput is 7.71E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8198s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G49/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.8195s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7858s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2144s for 16384 events => throughput is 7.64E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8193s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G523/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.0000s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9655s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2139s for 16384 events => throughput is 7.66E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8207s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G518/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.3159s + [COUNTERS] Fortran Overhead ( 0 ) : 6.2799s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2144s for 16384 events => throughput is 7.64E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8216s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G42/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 9.5590s + [COUNTERS] Fortran Overhead ( 0 ) : 8.5275s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2142s for 16384 events => throughput is 7.65E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8173s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G524/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.3904s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3561s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2165s for 16384 events => throughput is 7.57E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8178s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G610/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.6008s + [COUNTERS] Fortran Overhead ( 0 ) : 3.5646s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2139s for 16384 events => throughput is 7.66E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8222s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G26/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.3995s + [COUNTERS] Fortran Overhead ( 0 ) : 6.3655s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2146s for 16384 events => throughput is 7.63E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8194s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G58/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.6867s + [COUNTERS] Fortran Overhead ( 0 ) : 7.6526s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2125s for 16384 events => throughput is 7.71E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8216s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G511/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.6518s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6184s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2122s for 16384 events => throughput is 7.72E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8212s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.3055s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2707s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2148s for 16384 events => throughput is 7.63E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8200s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G32/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 9.4359s + [COUNTERS] Fortran Overhead ( 0 ) : 8.3984s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2125s for 16384 events => throughput is 7.71E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8250s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G61/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.5465s + [COUNTERS] Fortran Overhead ( 0 ) : 6.5128s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2141s for 16384 events => throughput is 7.65E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8195s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G390/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 10.7326s + [COUNTERS] Fortran Overhead ( 0 ) : 9.6977s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2123s for 16384 events => throughput is 7.72E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8225s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_epemgguc/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_epemgguc/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_epemgguc/ajob1' in 15.2383 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_epemgguc/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_epemgguc/G164/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.2850s + [COUNTERS] Fortran Overhead ( 0 ) : 4.8839s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0500s for 16384 events => throughput is 3.28E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3510s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_epemgguc/G24/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.7941s + [COUNTERS] Fortran Overhead ( 0 ) : 4.3866s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0504s for 16384 events => throughput is 3.25E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3571s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_epemgguc/G162/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.9755s + [COUNTERS] Fortran Overhead ( 0 ) : 3.5718s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0497s for 16384 events => throughput is 3.30E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3540s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_epemgguc/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/ajob1' in 144.2270 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G336/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.0002s + [COUNTERS] Fortran Overhead ( 0 ) : 3.9612s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2139s for 16384 events => throughput is 7.66E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8251s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G108/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.9479s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9181s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2148s for 16384 events => throughput is 7.63E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8151s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G105/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.8400s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8015s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2135s for 16384 events => throughput is 7.67E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8251s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G304/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.6140s + [COUNTERS] Fortran Overhead ( 0 ) : 3.5761s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2120s for 16384 events => throughput is 7.73E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8259s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G175/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.7129s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6811s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2122s for 16384 events => throughput is 7.72E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8196s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G128/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.0131s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9773s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2119s for 16384 events => throughput is 7.73E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8238s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G360/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.9768s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9389s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2140s for 16384 events => throughput is 7.66E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8239s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G120/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.6953s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6596s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2130s for 16384 events => throughput is 7.69E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8227s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G296/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.5234s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4960s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2140s for 16384 events => throughput is 7.66E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8134s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G299/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.2904s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2526s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2133s for 16384 events => throughput is 7.68E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8245s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G715/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.0705s + [COUNTERS] Fortran Overhead ( 0 ) : 3.0381s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2120s for 16384 events => throughput is 7.73E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8205s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G300/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.8364s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8010s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2134s for 16384 events => throughput is 7.68E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8220s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G303/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.3199s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2856s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2118s for 16384 events => throughput is 7.74E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8226s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G127/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.2949s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2627s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2134s for 16384 events => throughput is 7.68E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8187s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G111/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.9938s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9545s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2138s for 16384 events => throughput is 7.66E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8254s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G126/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 11.9214s + [COUNTERS] Fortran Overhead ( 0 ) : 10.8815s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2149s for 16384 events => throughput is 7.62E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8250s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G294/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.0444s + [COUNTERS] Fortran Overhead ( 0 ) : 7.0072s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2138s for 16384 events => throughput is 7.66E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8235s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G598/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.8584s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8228s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2133s for 16384 events => throughput is 7.68E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8223s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G610/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.7637s + [COUNTERS] Fortran Overhead ( 0 ) : 3.7296s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2139s for 16384 events => throughput is 7.66E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8203s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G178/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.0289s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9889s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2148s for 16384 events => throughput is 7.63E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8251s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G121/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.9073s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8682s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2141s for 16384 events => throughput is 7.65E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8249s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G685/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.6240s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5906s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2142s for 16384 events => throughput is 7.65E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8193s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G690/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.5276s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4898s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2146s for 16384 events => throughput is 7.63E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8232s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G103/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 9.6007s + [COUNTERS] Fortran Overhead ( 0 ) : 8.5619s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2131s for 16384 events => throughput is 7.69E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8257s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G107/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.7730s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7341s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2142s for 16384 events => throughput is 7.65E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8247s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G180/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.9168s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8863s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2125s for 16384 events => throughput is 7.71E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8179s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G97/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 9.5902s + [COUNTERS] Fortran Overhead ( 0 ) : 8.5518s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2131s for 16384 events => throughput is 7.69E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8253s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G179/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.2663s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2349s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2120s for 16384 events => throughput is 7.73E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8194s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamggddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamggddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamggddx/ajob1' in 3.6678 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamggddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamggddx/G189/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.3552s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9027s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0935s for 16384 events => throughput is 1.75E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3590s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamggddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dd_taptamggdd/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dd_taptamggdd/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dd_taptamggdd/ajob1' in 3.7474 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dd_taptamggdd/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dd_taptamggdd/G463/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.4141s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5472s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2392s for 16384 events => throughput is 6.85E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.6276s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dd_taptamggdd/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/ajob1' in 218.8344 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G506/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.8779s + [COUNTERS] Fortran Overhead ( 0 ) : 4.6271s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4188s for 16384 events => throughput is 3.91E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8320s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G499/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.5013s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2510s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4173s for 16384 events => throughput is 3.93E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8331s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G46/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 9.0558s + [COUNTERS] Fortran Overhead ( 0 ) : 7.8070s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4181s for 16384 events => throughput is 3.92E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8307s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G56/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.1841s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9321s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4175s for 16384 events => throughput is 3.92E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8344s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G39/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.0682s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8181s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4188s for 16384 events => throughput is 3.91E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8313s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G517/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.7199s + [COUNTERS] Fortran Overhead ( 0 ) : 5.4710s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4171s for 16384 events => throughput is 3.93E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8318s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G52/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.5909s + [COUNTERS] Fortran Overhead ( 0 ) : 3.3398s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4173s for 16384 events => throughput is 3.93E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8338s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G31/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.3347s + [COUNTERS] Fortran Overhead ( 0 ) : 7.0832s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4171s for 16384 events => throughput is 3.93E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8345s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G324/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.9635s + [COUNTERS] Fortran Overhead ( 0 ) : 6.7161s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4176s for 16384 events => throughput is 3.92E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8298s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G64/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.5041s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2542s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4175s for 16384 events => throughput is 3.92E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8323s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G3/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.5728s + [COUNTERS] Fortran Overhead ( 0 ) : 4.3250s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4171s for 16384 events => throughput is 3.93E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8307s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G33/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.8506s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5987s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4178s for 16384 events => throughput is 3.92E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8341s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G53/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.1671s + [COUNTERS] Fortran Overhead ( 0 ) : 6.9134s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4187s for 16384 events => throughput is 3.91E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8350s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G43/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.9801s + [COUNTERS] Fortran Overhead ( 0 ) : 5.7383s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4173s for 16384 events => throughput is 3.93E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8244s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G49/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.4650s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2131s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4171s for 16384 events => throughput is 3.93E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8348s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G40/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.4948s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2530s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4175s for 16384 events => throughput is 3.92E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8244s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G523/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.8974s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6469s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4176s for 16384 events => throughput is 3.92E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8329s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G518/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.9485s + [COUNTERS] Fortran Overhead ( 0 ) : 5.6992s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4175s for 16384 events => throughput is 3.92E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8319s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G35/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.5869s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3370s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4191s for 16384 events => throughput is 3.91E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8308s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.2899s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0423s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4191s for 16384 events => throughput is 3.91E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8285s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G42/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.0103s + [COUNTERS] Fortran Overhead ( 0 ) : 5.7609s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4183s for 16384 events => throughput is 3.92E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8311s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G524/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.8256s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5739s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4192s for 16384 events => throughput is 3.91E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8325s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G65/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.5632s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3161s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4173s for 16384 events => throughput is 3.93E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8298s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G27/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.3901s + [COUNTERS] Fortran Overhead ( 0 ) : 6.1393s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4171s for 16384 events => throughput is 3.93E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8336s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G44/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.8274s + [COUNTERS] Fortran Overhead ( 0 ) : 6.5806s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4169s for 16384 events => throughput is 3.93E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8298s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.2835s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0419s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4179s for 16384 events => throughput is 3.92E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8237s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G303/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.7865s + [COUNTERS] Fortran Overhead ( 0 ) : 3.5376s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4183s for 16384 events => throughput is 3.92E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8306s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G55/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.3454s + [COUNTERS] Fortran Overhead ( 0 ) : 3.0957s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4207s for 16384 events => throughput is 3.89E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8291s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G34/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.4635s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2145s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4186s for 16384 events => throughput is 3.91E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8304s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G26/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.4759s + [COUNTERS] Fortran Overhead ( 0 ) : 6.2291s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4178s for 16384 events => throughput is 3.92E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8290s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G62/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.5324s + [COUNTERS] Fortran Overhead ( 0 ) : 6.2822s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4176s for 16384 events => throughput is 3.92E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8326s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G37/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.8166s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5694s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4180s for 16384 events => throughput is 3.92E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8292s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G485/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.9184s + [COUNTERS] Fortran Overhead ( 0 ) : 5.6687s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4171s for 16384 events => throughput is 3.93E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8327s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G67/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.8214s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5700s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4178s for 16384 events => throughput is 3.92E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8337s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G512/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.5078s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2554s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4175s for 16384 events => throughput is 3.92E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8348s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G511/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.1263s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8770s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4171s for 16384 events => throughput is 3.93E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8322s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G51/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.1672s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9119s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4198s for 16384 events => throughput is 3.90E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8355s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.0768s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8240s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4209s for 16384 events => throughput is 3.89E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8319s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G63/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.0413s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7928s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4191s for 16384 events => throughput is 3.91E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8294s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G48/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.0272s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7761s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4179s for 16384 events => throughput is 3.92E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8331s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgggux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgggux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgggux/ajob1' in 42.3458 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgggux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgggux/G33/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.4294s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1808s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4172s for 16384 events => throughput is 3.93E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8314s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgggux/G68/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.7473s + [COUNTERS] Fortran Overhead ( 0 ) : 3.4971s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4175s for 16384 events => throughput is 3.92E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8327s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgggux/G25/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.9795s + [COUNTERS] Fortran Overhead ( 0 ) : 6.7289s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4194s for 16384 events => throughput is 3.91E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8312s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgggux/G494/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.2777s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0325s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4178s for 16384 events => throughput is 3.92E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8274s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgggux/G55/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.5619s + [COUNTERS] Fortran Overhead ( 0 ) : 3.3096s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4182s for 16384 events => throughput is 3.92E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8341s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgggux/G509/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.5883s + [COUNTERS] Fortran Overhead ( 0 ) : 7.3357s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4188s for 16384 events => throughput is 3.91E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8338s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgggux/G29/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.7122s + [COUNTERS] Fortran Overhead ( 0 ) : 6.4636s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4175s for 16384 events => throughput is 3.92E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8311s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgggux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamgguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamgguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamgguux/ajob1' in 21.4757 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamgguux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamgguux/G260/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.8838s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0159s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2375s for 16384 events => throughput is 6.90E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.6304s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamgguux/G447/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.8666s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0044s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2379s for 16384 events => throughput is 6.89E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.6243s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamgguux/G229/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.2737s + [COUNTERS] Fortran Overhead ( 0 ) : 4.4082s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2378s for 16384 events => throughput is 6.89E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.6276s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamgguux/G135/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.2623s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3937s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2378s for 16384 events => throughput is 6.89E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.6307s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamgguux/G230/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.6948s + [COUNTERS] Fortran Overhead ( 0 ) : 4.8323s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2380s for 16384 events => throughput is 6.88E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.6245s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamgguux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdddx/ajob1' in 7.2452 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdddx/G715/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.2218s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4758s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1241s for 16384 events => throughput is 1.32E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.6219s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdddx/G111/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.2142s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4699s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1243s for 16384 events => throughput is 1.32E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.6199s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemusdxsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemusdxsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemusdxsx/ajob1' in 6.9336 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemusdxsx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemusdxsx/G67/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.6193s + [COUNTERS] Fortran Overhead ( 0 ) : 6.4809s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0138s for 16384 events => throughput is 1.18E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1245s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemusdxsx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdssx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdssx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdssx/ajob1' in 18.4869 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdssx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdssx/G31/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.7199s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3167s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0499s for 16384 events => throughput is 3.28E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3533s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdssx/G27/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.7240s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3221s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0500s for 16384 events => throughput is 3.28E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3520s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdssx/G119/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.9455s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5417s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0501s for 16384 events => throughput is 3.27E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3538s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdssx/G284/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.1615s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7591s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0499s for 16384 events => throughput is 3.28E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3524s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdssx/G375/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.3906s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9860s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0499s for 16384 events => throughput is 3.28E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3546s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdssx/G102/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.8086s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4072s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0496s for 16384 events => throughput is 3.30E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3518s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdssx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_epemgguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_epemgguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_epemgguux/ajob1' in 19.2162 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_epemgguux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_epemgguux/G105/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.7091s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9577s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1229s for 16384 events => throughput is 1.33E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.6286s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_epemgguux/G223/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.4218s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6750s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1227s for 16384 events => throughput is 1.34E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.6241s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_epemgguux/G103/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.5930s + [COUNTERS] Fortran Overhead ( 0 ) : 4.8413s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1231s for 16384 events => throughput is 1.33E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.6285s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_epemgguux/G5/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.0661s + [COUNTERS] Fortran Overhead ( 0 ) : 5.3139s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1232s for 16384 events => throughput is 1.33E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.6290s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_epemgguux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/ajob1' in 353.2107 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G506/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.8311s + [COUNTERS] Fortran Overhead ( 0 ) : 5.7998s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2123s for 16384 events => throughput is 7.72E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8190s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G499/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.3248s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2907s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2143s for 16384 events => throughput is 7.65E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8198s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G38/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.1632s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1312s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2137s for 16384 events => throughput is 7.67E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8183s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G46/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 9.1899s + [COUNTERS] Fortran Overhead ( 0 ) : 8.1532s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2145s for 16384 events => throughput is 7.64E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8223s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G59/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.6496s + [COUNTERS] Fortran Overhead ( 0 ) : 6.6094s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2161s for 16384 events => throughput is 7.58E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8241s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G56/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.2137s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1824s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2126s for 16384 events => throughput is 7.71E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8186s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G525/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.8348s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8027s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2123s for 16384 events => throughput is 7.72E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8199s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G39/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.7466s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7095s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2131s for 16384 events => throughput is 7.69E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8239s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G517/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.0443s + [COUNTERS] Fortran Overhead ( 0 ) : 6.0145s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2124s for 16384 events => throughput is 7.72E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8174s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G487/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.2125s + [COUNTERS] Fortran Overhead ( 0 ) : 5.1838s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2136s for 16384 events => throughput is 7.67E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8151s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G52/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.3603s + [COUNTERS] Fortran Overhead ( 0 ) : 3.3270s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2130s for 16384 events => throughput is 7.69E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8203s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G31/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 9.1246s + [COUNTERS] Fortran Overhead ( 0 ) : 8.0886s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2122s for 16384 events => throughput is 7.72E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8238s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G64/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.4316s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3960s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2133s for 16384 events => throughput is 7.68E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8223s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G3/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.3315s + [COUNTERS] Fortran Overhead ( 0 ) : 5.2970s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2125s for 16384 events => throughput is 7.71E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8220s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G497/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.7163s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6820s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2145s for 16384 events => throughput is 7.64E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8197s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G13/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.1118s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0741s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2140s for 16384 events => throughput is 7.65E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8237s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G66/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.8744s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8401s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2131s for 16384 events => throughput is 7.69E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8211s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G33/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.3927s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3609s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2144s for 16384 events => throughput is 7.64E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8174s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G53/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 9.4383s + [COUNTERS] Fortran Overhead ( 0 ) : 8.4041s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2134s for 16384 events => throughput is 7.68E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8208s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G43/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.2650s + [COUNTERS] Fortran Overhead ( 0 ) : 5.2273s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2141s for 16384 events => throughput is 7.65E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8236s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G68/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.5979s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5623s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2123s for 16384 events => throughput is 7.72E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8233s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G49/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.4352s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4051s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2123s for 16384 events => throughput is 7.72E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8177s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G302/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 9.6208s + [COUNTERS] Fortran Overhead ( 0 ) : 8.5860s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2140s for 16384 events => throughput is 7.66E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8208s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G25/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.5847s + [COUNTERS] Fortran Overhead ( 0 ) : 6.5543s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2123s for 16384 events => throughput is 7.72E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8181s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G40/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.4625s + [COUNTERS] Fortran Overhead ( 0 ) : 3.4305s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2128s for 16384 events => throughput is 7.70E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8191s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G523/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.4556s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4204s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2121s for 16384 events => throughput is 7.73E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8232s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G518/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.4984s + [COUNTERS] Fortran Overhead ( 0 ) : 5.4599s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2133s for 16384 events => throughput is 7.68E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8252s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G35/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.5549s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5223s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2122s for 16384 events => throughput is 7.72E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8203s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.0876s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0493s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2144s for 16384 events => throughput is 7.64E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8240s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G524/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.5066s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4711s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2140s for 16384 events => throughput is 7.66E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8215s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G65/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.5533s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5192s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2136s for 16384 events => throughput is 7.67E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8204s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G27/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.7993s + [COUNTERS] Fortran Overhead ( 0 ) : 6.7631s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2122s for 16384 events => throughput is 7.72E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8239s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G44/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.4826s + [COUNTERS] Fortran Overhead ( 0 ) : 6.4511s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2145s for 16384 events => throughput is 7.64E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8169s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G483/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.6572s + [COUNTERS] Fortran Overhead ( 0 ) : 4.6238s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2127s for 16384 events => throughput is 7.70E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8207s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G482/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.9853s + [COUNTERS] Fortran Overhead ( 0 ) : 5.9506s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2135s for 16384 events => throughput is 7.67E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8212s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G1/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.9270s + [COUNTERS] Fortran Overhead ( 0 ) : 4.8895s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2124s for 16384 events => throughput is 7.71E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8251s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.9863s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9525s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2130s for 16384 events => throughput is 7.69E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8208s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G55/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.9030s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8679s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2129s for 16384 events => throughput is 7.69E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8222s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G298/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.4303s + [COUNTERS] Fortran Overhead ( 0 ) : 3.3942s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2122s for 16384 events => throughput is 7.72E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8238s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G513/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.3100s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2779s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2146s for 16384 events => throughput is 7.63E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8175s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G60/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.7744s + [COUNTERS] Fortran Overhead ( 0 ) : 7.7375s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2125s for 16384 events => throughput is 7.71E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8243s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G34/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.4382s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4051s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2130s for 16384 events => throughput is 7.69E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8201s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G26/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.2293s + [COUNTERS] Fortran Overhead ( 0 ) : 7.1970s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2134s for 16384 events => throughput is 7.68E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8189s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G58/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.8162s + [COUNTERS] Fortran Overhead ( 0 ) : 5.7830s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2122s for 16384 events => throughput is 7.72E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8210s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G54/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.3514s + [COUNTERS] Fortran Overhead ( 0 ) : 6.3167s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2123s for 16384 events => throughput is 7.72E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8224s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G62/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.3801s + [COUNTERS] Fortran Overhead ( 0 ) : 7.3459s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2146s for 16384 events => throughput is 7.63E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8195s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G37/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.3814s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3485s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2125s for 16384 events => throughput is 7.71E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8204s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G493/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.5563s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5220s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2144s for 16384 events => throughput is 7.64E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8199s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G4/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.6164s + [COUNTERS] Fortran Overhead ( 0 ) : 6.5817s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2113s for 16384 events => throughput is 7.75E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8234s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G50/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.5401s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5062s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2123s for 16384 events => throughput is 7.72E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8216s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G485/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.1270s + [COUNTERS] Fortran Overhead ( 0 ) : 6.0985s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2136s for 16384 events => throughput is 7.67E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8149s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G41/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.4480s + [COUNTERS] Fortran Overhead ( 0 ) : 7.4094s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2145s for 16384 events => throughput is 7.64E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8241s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G67/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.5467s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5140s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2137s for 16384 events => throughput is 7.67E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8189s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G512/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.4720s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4373s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2123s for 16384 events => throughput is 7.72E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8224s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G511/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.9589s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9277s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2135s for 16384 events => throughput is 7.68E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8177s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G57/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.2316s + [COUNTERS] Fortran Overhead ( 0 ) : 7.1988s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2122s for 16384 events => throughput is 7.72E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8206s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.3104s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2761s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2128s for 16384 events => throughput is 7.70E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8215s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G63/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.8379s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8018s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2140s for 16384 events => throughput is 7.65E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8221s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G32/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.9834s + [COUNTERS] Fortran Overhead ( 0 ) : 7.9480s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2124s for 16384 events => throughput is 7.71E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8230s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G29/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.6021s + [COUNTERS] Fortran Overhead ( 0 ) : 6.5639s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2135s for 16384 events => throughput is 7.67E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8246s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G48/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.7722s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7375s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2147s for 16384 events => throughput is 7.63E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8199s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_taptamggdsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_taptamggdsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_taptamggdsx/ajob1' in 10.2337 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_taptamggdsx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_taptamggdsx/G92/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.3259s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8828s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0927s for 16384 events => throughput is 1.77E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3504s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_taptamggdsx/G144/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.0273s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5771s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0935s for 16384 events => throughput is 1.75E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3566s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_taptamggdsx/G207/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.9948s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5427s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0939s for 16384 events => throughput is 1.74E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3582s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_taptamggdsx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_epemggdsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_epemggdsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_epemggdsx/ajob1' in 8.0152 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_epemggdsx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_epemggdsx/G205/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.6993s + [COUNTERS] Fortran Overhead ( 0 ) : 7.2940s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0500s for 16384 events => throughput is 3.28E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3553s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_epemggdsx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_taptamgggg/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_taptamgggg/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_taptamgggg/ajob1' in 4.9357 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_taptamgggg/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_taptamgggg/G67/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.4068s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1627s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4202s for 16384 events => throughput is 3.90E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8239s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_taptamgggg/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dc_epemdcssx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dc_epemdcssx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dc_epemdcssx/ajob1' in 3.2083 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dc_epemdcssx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dc_epemdcssx/G77/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.7249s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5876s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0141s for 16384 events => throughput is 1.16E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1232s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dc_epemdcssx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguccx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguccx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguccx/ajob1' in 3.6012 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguccx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguccx/G48/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.2963s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8494s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0930s for 16384 events => throughput is 1.76E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3539s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguccx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ds_epemggds/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ds_epemggds/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ds_epemggds/ajob1' in 2.7314 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ds_epemggds/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ds_epemggds/G26/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.4247s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0200s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0500s for 16384 events => throughput is 3.28E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3547s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ds_epemggds/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ucx_epemggucx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ucx_epemggucx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ucx_epemggucx/ajob1' in 7.8203 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ucx_epemggucx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ucx_epemggucx/G163/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.8172s + [COUNTERS] Fortran Overhead ( 0 ) : 4.4104s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0504s for 16384 events => throughput is 3.25E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3564s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ucx_epemggucx/G60/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.4017s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9978s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0503s for 16384 events => throughput is 3.26E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3536s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ucx_epemggucx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemggudx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemggudx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemggudx/ajob1' in 50.6108 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemggudx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemggudx/G109/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.3057s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8908s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0522s for 16384 events => throughput is 3.14E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3627s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemggudx/G151/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.6380s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2249s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0515s for 16384 events => throughput is 3.18E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3616s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemggudx/G106/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.0801s + [COUNTERS] Fortran Overhead ( 0 ) : 3.6716s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0508s for 16384 events => throughput is 3.23E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3578s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemggudx/G35/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.5856s + [COUNTERS] Fortran Overhead ( 0 ) : 7.1755s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0505s for 16384 events => throughput is 3.24E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3596s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemggudx/G119/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.8098s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3964s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0510s for 16384 events => throughput is 3.21E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3625s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemggudx/G187/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.6668s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2472s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0588s for 16384 events => throughput is 2.78E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3607s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemggudx/G91/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.8084s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3949s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0509s for 16384 events => throughput is 3.22E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3625s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemggudx/G363/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 17.9296s + [COUNTERS] Fortran Overhead ( 0 ) : 17.3772s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1900s for 65536 events => throughput is 3.45E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3624s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemggudx/G149/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.1980s + [COUNTERS] Fortran Overhead ( 0 ) : 4.7889s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0506s for 16384 events => throughput is 3.24E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3586s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemggudx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgcuxcx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgcuxcx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgcuxcx/ajob1' in 3.5924 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgcuxcx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgcuxcx/G120/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.2650s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8613s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0500s for 16384 events => throughput is 3.28E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3537s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgcuxcx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemgggg/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemgggg/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemgggg/ajob1' in 10.1907 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemgggg/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemgggg/G65/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.6574s + [COUNTERS] Fortran Overhead ( 0 ) : 3.6193s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2155s for 16384 events => throughput is 7.60E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8226s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemgggg/G48/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.9120s + [COUNTERS] Fortran Overhead ( 0 ) : 3.8717s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2158s for 16384 events => throughput is 7.59E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8245s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemgggg/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_taptamggud/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_taptamggud/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_taptamggud/ajob1' in 30.8330 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_taptamggud/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_taptamggud/G46/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.0730s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6199s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0941s for 16384 events => throughput is 1.74E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3591s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_taptamggud/G39/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.9677s + [COUNTERS] Fortran Overhead ( 0 ) : 4.5199s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0938s for 16384 events => throughput is 1.75E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3540s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_taptamggud/G112/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.1302s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6809s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0932s for 16384 events => throughput is 1.76E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3560s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_taptamggud/G115/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.3572s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8998s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0967s for 16384 events => throughput is 1.69E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3606s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_taptamggud/G270/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.7467s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2951s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0938s for 16384 events => throughput is 1.75E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3578s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_taptamggud/G96/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.4395s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9855s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0939s for 16384 events => throughput is 1.74E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3601s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_taptamggud/G107/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.6686s + [COUNTERS] Fortran Overhead ( 0 ) : 4.2124s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0942s for 16384 events => throughput is 1.74E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3620s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_taptamggud/G63/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.7245s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2716s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0934s for 16384 events => throughput is 1.75E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3595s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_taptamggud/G375/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.1701s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7144s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0949s for 16384 events => throughput is 1.73E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3608s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_taptamggud/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguuux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguuux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguuux/ajob1' in 21.4408 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguuux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguuux/G122/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.0433s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2977s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1238s for 16384 events => throughput is 1.32E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.6218s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguuux/G33/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.8792s + [COUNTERS] Fortran Overhead ( 0 ) : 6.1322s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1235s for 16384 events => throughput is 1.33E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.6235s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguuux/G127/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.1635s + [COUNTERS] Fortran Overhead ( 0 ) : 3.4193s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1234s for 16384 events => throughput is 1.33E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.6209s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguuux/G116/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.1840s + [COUNTERS] Fortran Overhead ( 0 ) : 5.4371s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1236s for 16384 events => throughput is 1.33E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.6232s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguuux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamgggdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamgggdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamgggdx/ajob1' in 37.3375 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamgggdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamgggdx/G56/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.2652s + [COUNTERS] Fortran Overhead ( 0 ) : 3.0116s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4205s for 16384 events => throughput is 3.90E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8331s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamgggdx/G64/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.0841s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8326s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4187s for 16384 events => throughput is 3.91E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8328s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamgggdx/G13/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.3326s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0817s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4182s for 16384 events => throughput is 3.92E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8327s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamgggdx/G33/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.5685s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3178s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4178s for 16384 events => throughput is 3.92E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8328s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamgggdx/G49/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.8221s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5716s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4187s for 16384 events => throughput is 3.91E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8318s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamgggdx/G523/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.2671s + [COUNTERS] Fortran Overhead ( 0 ) : 3.0171s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4191s for 16384 events => throughput is 3.91E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8309s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamgggdx/G65/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.6646s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4141s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4194s for 16384 events => throughput is 3.91E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8312s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamgggdx/G34/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.7991s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5459s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4213s for 16384 events => throughput is 3.89E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8319s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamgggdx/G48/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.7159s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4677s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4185s for 16384 events => throughput is 3.91E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8296s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamgggdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgudux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgudux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgudux/ajob1' in 9.8117 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgudux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgudux/G59/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.7398s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2729s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1042s for 16384 events => throughput is 1.57E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3628s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgudux/G61/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.4552s + [COUNTERS] Fortran Overhead ( 0 ) : 6.0030s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0942s for 16384 events => throughput is 1.74E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3580s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgudux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgduxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgduxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgduxdx/ajob1' in 3.0549 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgduxdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgduxdx/G104/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.7295s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3169s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0510s for 16384 events => throughput is 3.21E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3616s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgduxdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgsdxsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgsdxsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgsdxsx/ajob1' in 3.0775 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgsdxsx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgsdxsx/G25/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.5639s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1621s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0500s for 16384 events => throughput is 3.28E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3517s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgsdxsx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/ajob1' in 103.3841 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G208/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.7432s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3376s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0505s for 16384 events => throughput is 3.24E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3552s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G235/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.5245s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1166s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0504s for 16384 events => throughput is 3.25E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3574s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G140/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.8862s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4775s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0507s for 16384 events => throughput is 3.23E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3581s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G109/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.5011s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0911s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0507s for 16384 events => throughput is 3.23E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3594s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G33/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.5282s + [COUNTERS] Fortran Overhead ( 0 ) : 5.1193s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0505s for 16384 events => throughput is 3.25E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3585s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G43/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.5768s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1693s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0504s for 16384 events => throughput is 3.25E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3572s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G240/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.5821s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1719s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0505s for 16384 events => throughput is 3.24E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3596s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G268/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.1844s + [COUNTERS] Fortran Overhead ( 0 ) : 4.7750s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0507s for 16384 events => throughput is 3.23E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3587s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G139/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.7885s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3768s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0505s for 16384 events => throughput is 3.24E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3612s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G120/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.5630s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1561s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0507s for 16384 events => throughput is 3.23E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3562s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G236/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.6102s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1975s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0515s for 16384 events => throughput is 3.18E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3611s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G181/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.0669s + [COUNTERS] Fortran Overhead ( 0 ) : 5.6607s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0502s for 16384 events => throughput is 3.26E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3560s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G351/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.1933s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7809s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0515s for 16384 events => throughput is 3.18E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3609s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G65/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.6499s + [COUNTERS] Fortran Overhead ( 0 ) : 4.2382s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0505s for 16384 events => throughput is 3.24E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3612s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G71/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.3770s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9693s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0503s for 16384 events => throughput is 3.26E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3574s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G119/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.7236s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3170s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0504s for 16384 events => throughput is 3.25E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3561s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G127/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.3907s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9839s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0505s for 16384 events => throughput is 3.24E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3562s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G96/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.5003s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0900s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0503s for 16384 events => throughput is 3.26E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3600s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G60/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.0525s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6448s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0507s for 16384 events => throughput is 3.23E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3570s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G26/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.2862s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8797s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0503s for 16384 events => throughput is 3.25E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3561s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G37/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.5040s + [COUNTERS] Fortran Overhead ( 0 ) : 6.0943s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0506s for 16384 events => throughput is 3.24E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3590s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G125/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.1120s + [COUNTERS] Fortran Overhead ( 0 ) : 3.6989s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0511s for 16384 events => throughput is 3.20E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3619s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G91/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.3003s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8898s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0506s for 16384 events => throughput is 3.24E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3599s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G143/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.2293s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8226s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0505s for 16384 events => throughput is 3.24E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3561s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G116/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.1154s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7039s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0506s for 16384 events => throughput is 3.24E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3609s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G61/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 6.3292s + [COUNTERS] Fortran Overhead ( 0 ) : 5.9197s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0502s for 16384 events => throughput is 3.26E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3594s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G149/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.1000s + [COUNTERS] Fortran Overhead ( 0 ) : 4.6936s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0505s for 16384 events => throughput is 3.24E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3559s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G144/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.2617s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8583s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0503s for 16384 events => throughput is 3.26E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3531s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G207/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.5599s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1522s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0505s for 16384 events => throughput is 3.24E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3572s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_epemgguu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_epemgguu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_epemgguu/ajob1' in 45.7406 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_epemgguu/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_epemgguu/G463/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.1833s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4351s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1236s for 16384 events => throughput is 1.33E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.6246s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_epemgguu/G100/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.2223s + [COUNTERS] Fortran Overhead ( 0 ) : 3.4781s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1235s for 16384 events => throughput is 1.33E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.6206s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_epemgguu/G110/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.6542s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9054s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1235s for 16384 events => throughput is 1.33E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.6253s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_epemgguu/G124/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.7111s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9663s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1235s for 16384 events => throughput is 1.33E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.6212s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_epemgguu/G367/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.0836s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3427s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1235s for 16384 events => throughput is 1.33E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.6174s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_epemgguu/G126/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.1438s + [COUNTERS] Fortran Overhead ( 0 ) : 4.3959s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1235s for 16384 events => throughput is 1.33E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.6245s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_epemgguu/G361/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.5179s + [COUNTERS] Fortran Overhead ( 0 ) : 4.7712s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1236s for 16384 events => throughput is 1.33E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.6231s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_epemgguu/G116/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.7959s + [COUNTERS] Fortran Overhead ( 0 ) : 3.0471s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1235s for 16384 events => throughput is 1.33E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.6253s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_epemgguu/G363/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.2031s + [COUNTERS] Fortran Overhead ( 0 ) : 4.4567s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1237s for 16384 events => throughput is 1.32E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.6227s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_epemgguu/G138/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.8058s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0565s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1235s for 16384 events => throughput is 1.33E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.6258s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_epemgguu/G102/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.2377s + [COUNTERS] Fortran Overhead ( 0 ) : 3.4933s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1236s for 16384 events => throughput is 1.33E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.6209s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_epemgguu/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgdssx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgdssx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgdssx/ajob1' in 3.4583 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgdssx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgdssx/G28/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.9257s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4769s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0931s for 16384 events => throughput is 1.76E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3557s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgdssx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_taptamgguu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_taptamgguu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_taptamgguu/ajob1' in 18.1422 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_taptamgguu/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_taptamgguu/G464/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.2514s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3773s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2396s for 16384 events => throughput is 6.84E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.6345s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_taptamgguu/G59/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.3787s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5170s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2390s for 16384 events => throughput is 6.86E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.6228s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_taptamgguu/G100/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.7940s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9236s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2396s for 16384 events => throughput is 6.84E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.6309s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_taptamgguu/G106/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.1278s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2571s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2389s for 16384 events => throughput is 6.86E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.6318s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_taptamgguu/G143/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.1065s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2355s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2388s for 16384 events => throughput is 6.86E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.6322s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_taptamgguu/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dc_epemggdc/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dc_epemggdc/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dc_epemggdc/ajob1' in 6.2298 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dc_epemggdc/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dc_epemggdc/G18/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.9068s + [COUNTERS] Fortran Overhead ( 0 ) : 5.4974s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0509s for 16384 events => throughput is 3.22E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3584s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dc_epemggdc/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ucx_taptamggucx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ucx_taptamggucx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ucx_taptamggucx/ajob1' in 5.8156 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ucx_taptamggucx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ucx_taptamggucx/G62/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 5.4862s + [COUNTERS] Fortran Overhead ( 0 ) : 5.0362s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0939s for 16384 events => throughput is 1.74E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3561s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ucx_taptamggucx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamucdxcx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamucdxcx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamucdxcx/ajob1' in 2.6337 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamucdxcx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamucdxcx/G47/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.3332s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1881s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0215s for 16384 events => throughput is 7.62E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1236s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamucdxcx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_epemgggg/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_epemgggg/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_epemgggg/ajob1' in 5.5868 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_epemgggg/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_epemgggg/G49/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.8317s + [COUNTERS] Fortran Overhead ( 0 ) : 3.8078s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2120s for 16384 events => throughput is 7.73E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8120s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_epemgggg/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemudddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemudddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemudddx/ajob1' in 2.5148 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemudddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemudddx/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.1909s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9890s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0231s for 16384 events => throughput is 7.10E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1789s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemudddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamggudx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamggudx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamggudx/ajob1' in 2.8296 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamggudx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamggudx/G60/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.5185s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0630s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0944s for 16384 events => throughput is 1.74E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.3611s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamggudx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_taptamuuddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_taptamuuddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_taptamuuddx/ajob1' in 2.8473 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_taptamuuddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_taptamuuddx/G11/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.3212s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0809s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0430s for 16384 events => throughput is 3.81E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1974s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_taptamuuddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished in 2377.2825 seconds +__CUDACPP_DEBUG: gen_ximprove.launch finished in 2381.3226 seconds +__CUDACPP_DEBUG: GridPackCmd.refine4grid finished in 2381.3234 seconds +__CUDACPP_DEBUG: GridPackCmd.launch (3a) combine_events +combine_events +INFO: fail to reach target 500 +__CUDACPP_DEBUG: GridPackCmd.launch (3b) store_events +store_events + === Results Summary for run: GridRun_21 tag: tag_1 === + + Cross-section : 0 +- 0 pb + Nb of events : 195 + +INFO: No version of lhapdf. Can not run systematics computation +decay_events -from_cards +__CUDACPP_DEBUG: GridPackCmd.launch finished in 2523.7488 seconds +__CUDACPP_DEBUG: GridPackCmd.__init__ back from self.launch +DONE +write ./events.lhe.gz +END: Sun Sep 15 12:06:52 AM CEST 2024 +ELAPSED: 2526 seconds diff --git a/epochX/cudacpp/tlau/fromgridpacks/pp_dy4j.mad/fortran/output.txt b/epochX/cudacpp/tlau/fromgridpacks/pp_dy4j.mad/fortran/output.txt new file mode 100644 index 0000000000..6d22ecf4ae --- /dev/null +++ b/epochX/cudacpp/tlau/fromgridpacks/pp_dy4j.mad/fortran/output.txt @@ -0,0 +1,2850 @@ +START: Sat Sep 14 05:19:42 PM CEST 2024 +On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU (nproc=4)] [GPU: 1x Tesla V100S-PCIE-32GB]: +CUDACPP_RUNTIME_DISABLEFPE is set +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_bbx_epemdsdxsx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_bbx_taptamdsdxsx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_ccx_epemdcdxcx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_ccx_epemdsdxsx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_ccx_epemucuxcx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_ccx_epemuduxdx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_ccx_taptamdcdxcx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_ccx_taptamdsdxsx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_ccx_taptamucuxcx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_ccx_taptamuduxdx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_dc_epemdcccx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_dc_epemdcssx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_dc_epemddcdx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_dc_epemggdc/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_dc_epemudcux/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_dc_taptamdcccx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_dc_taptamdcssx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_dc_taptamddcdx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_dc_taptamggdc/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_dc_taptamudcux/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_dd_epemddddx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_dd_epemddssx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_dd_epemggdd/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_dd_epemuddux/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_dd_taptamddddx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_dd_taptamddssx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_dd_taptamggdd/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_dd_taptamuddux/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_ddx_epemdcdxcx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_ddx_epemdddxdx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_ddx_epemdsdxsx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_ddx_epemggddx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_ddx_epemgggg/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_ddx_epemggssx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_ddx_epemgguux/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_ddx_epemsbsxbx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_ddx_epemsssxsx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_ddx_epemucuxcx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_ddx_epemuduxdx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_ddx_epemusuxsx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_ddx_epemuuuxux/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_ddx_taptamdcdxcx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_ddx_taptamdddxdx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_ddx_taptamdsdxsx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_ddx_taptamggddx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_ddx_taptamgggg/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_ddx_taptamggssx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_ddx_taptamgguux/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_ddx_taptamsbsxbx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_ddx_taptamsssxsx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_ddx_taptamucuxcx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_ddx_taptamuduxdx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_ddx_taptamusuxsx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_ddx_taptamuuuxux/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_ds_epemddsdx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_ds_epemdsbbx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_ds_epemdsssx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_ds_epemggds/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_ds_epemudsux/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_ds_taptamddsdx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_ds_taptamdsbbx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_ds_taptamdsssx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_ds_taptamggds/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_ds_taptamudsux/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_dsx_epemdbsxbx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_dsx_epemdddxsx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_dsx_epemdssxsx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_dsx_epemggdsx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_dsx_epemuduxsx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_dsx_taptamdbsxbx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_dsx_taptamdddxsx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_dsx_taptamdssxsx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_dsx_taptamggdsx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_dsx_taptamuduxsx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_dux_epemdcuxcx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_dux_epemdduxdx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_dux_epemdsuxsx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_dux_epemggdux/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_dux_epemuduxux/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_dux_taptamdcuxcx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_dux_taptamdduxdx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_dux_taptamdsuxsx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_dux_taptamggdux/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_dux_taptamuduxux/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_dxcx_epemcdxcxcx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_dxcx_epemddxdxcx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_dxcx_epemggdxcx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_dxcx_epemsdxcxsx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_dxcx_epemuuxdxcx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_dxcx_taptamcdxcxcx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_dxcx_taptamddxdxcx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_dxcx_taptamggdxcx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_dxcx_taptamsdxcxsx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_dxcx_taptamuuxdxcx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_dxdx_epemddxdxdx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_dxdx_epemggdxdx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_dxdx_epemsdxdxsx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_dxdx_epemuuxdxdx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_dxdx_taptamddxdxdx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_dxdx_taptamggdxdx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_dxdx_taptamsdxdxsx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_dxdx_taptamuuxdxdx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_dxsx_epembdxsxbx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_dxsx_epemddxdxsx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_dxsx_epemggdxsx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_dxsx_epemsdxsxsx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_dxsx_epemuuxdxsx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_dxsx_taptambdxsxbx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_dxsx_taptamddxdxsx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_dxsx_taptamggdxsx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_dxsx_taptamsdxsxsx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_dxsx_taptamuuxdxsx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_gd_epemgdddx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_gd_epemgdssx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_gd_epemgggd/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_gd_epemgudux/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_gd_taptamgdddx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_gd_taptamgdssx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_gd_taptamgggd/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_gd_taptamgudux/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_gdx_epemgddxdx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_gdx_epemgggdx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_gdx_epemgsdxsx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_gdx_epemguuxdx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_gdx_taptamgddxdx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_gdx_taptamgggdx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_gdx_taptamgsdxsx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_gdx_taptamguuxdx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_gg_epemdddxdx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_gg_epemdsdxsx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_gg_epemggddx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_gg_epemgguux/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_gg_epemucuxcx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_gg_epemuduxdx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_gg_epemuuuxux/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_gg_taptamdddxdx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_gg_taptamdsdxsx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_gg_taptamggddx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_gg_taptamgguux/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_gg_taptamucuxcx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_gg_taptamuduxdx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_gg_taptamuuuxux/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_gu_epemgggu/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_gu_epemguccx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_gu_epemguddx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_gu_epemguuux/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_gu_taptamgggu/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_gu_taptamguccx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_gu_taptamguddx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_gu_taptamguuux/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_gux_epemgcuxcx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_gux_epemgduxdx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_gux_epemgggux/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_gux_epemguuxux/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_gux_taptamgcuxcx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_gux_taptamgduxdx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_gux_taptamgggux/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_gux_taptamguuxux/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_ssx_epemdbdxbx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_ssx_epemdcdxcx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_ssx_epemdsdxsx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_ssx_epemucuxcx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_ssx_epemuduxdx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_ssx_taptamdbdxbx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_ssx_taptamdcdxcx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_ssx_taptamdsdxsx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_ssx_taptamucuxcx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_ssx_taptamuduxdx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_uc_epemgguc/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_uc_epemucccx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_uc_epemudcdx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_uc_epemuucux/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_uc_taptamgguc/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_uc_taptamucccx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_uc_taptamudcdx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_uc_taptamuucux/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_ucx_epemggucx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_ucx_epemuccxcx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_ucx_epemuddxcx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_ucx_epemuuuxcx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_ucx_taptamggucx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_ucx_taptamuccxcx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_ucx_taptamuddxcx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_ucx_taptamuuuxcx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_ud_epemggud/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_ud_epemudccx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_ud_epemudddx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_ud_epemudssx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_ud_epemuudux/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_ud_taptamggud/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_ud_taptamudccx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_ud_taptamudddx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_ud_taptamudssx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_ud_taptamuudux/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_udx_epemggudx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_udx_epemucdxcx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_udx_epemuddxdx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_udx_epemusdxsx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_udx_epemuuuxdx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_udx_taptamggudx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_udx_taptamucdxcx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_udx_taptamuddxdx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_udx_taptamusdxsx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_udx_taptamuuuxdx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_uu_epemgguu/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_uu_epemuuccx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_uu_epemuuddx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_uu_epemuuuux/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_uu_taptamgguu/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_uu_taptamuuccx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_uu_taptamuuddx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_uu_taptamuuuux/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_uux_epemcccxcx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_uux_epemcscxsx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_uux_epemdcdxcx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_uux_epemdddxdx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_uux_epemdsdxsx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_uux_epemggccx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_uux_epemggddx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_uux_epemgggg/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_uux_epemgguux/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_uux_epemucuxcx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_uux_epemuduxdx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_uux_epemuuuxux/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_uux_taptamcccxcx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_uux_taptamcscxsx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_uux_taptamdcdxcx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_uux_taptamdddxdx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_uux_taptamdsdxsx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_uux_taptamggccx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_uux_taptamggddx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_uux_taptamgggg/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_uux_taptamgguux/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_uux_taptamucuxcx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_uux_taptamuduxdx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_uux_taptamuuuxux/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_uxcx_epemcuxcxcx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_uxcx_epemduxdxcx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_uxcx_epemgguxcx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_uxcx_epemuuxuxcx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_uxcx_taptamcuxcxcx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_uxcx_taptamduxdxcx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_uxcx_taptamgguxcx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_uxcx_taptamuuxuxcx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_uxdx_epemcuxdxcx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_uxdx_epemduxdxdx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_uxdx_epemgguxdx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_uxdx_epemsuxdxsx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_uxdx_epemuuxuxdx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_uxdx_taptamcuxdxcx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_uxdx_taptamduxdxdx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_uxdx_taptamgguxdx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_uxdx_taptamsuxdxsx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_uxdx_taptamuuxuxdx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_uxux_epemcuxuxcx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_uxux_epemduxuxdx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_uxux_epemgguxux/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_uxux_epemuuxuxux/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_uxux_taptamcuxuxcx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_uxux_taptamduxuxdx/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_uxux_taptamgguxux/madevent -> madevent_fortran +lrwxrwxrwx. 1 avalassi zg 16 Sep 14 17:19 madevent/SubProcesses/P0_uxux_taptamuuxuxux/madevent -> madevent_fortran +Now generating 500 events with random seed 21 and granularity 1 +************************************************************ +* * +* W E L C O M E to * +* M A D G R A P H 5 _ a M C @ N L O * +* M A D E V E N T * +* * +* * * * +* * * * * * +* * * * * 5 * * * * * +* * * * * * +* * * * +* * +* VERSION 3.5.3_lo_vect * +* * +* The MadGraph5_aMC@NLO Development Team - Find us at * +* https://server06.fynu.ucl.ac.be/projects/madgraph * +* * +* Type 'help' for in-line help. * +* * +************************************************************ +INFO: load configuration from /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/Cards/me5_configuration.txt +INFO: load configuration from /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/Cards/me5_configuration.txt +Using default text editor "vi". Set another one in ./input/mg5_configuration.txt +Using default eps viewer "evince". Set another one in ./input/mg5_configuration.txt +Using default web browser "firefox". Set another one in ./input/mg5_configuration.txt +WRITE GRIDCARD /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent +__CUDACPP_DEBUG: GridPackCmd.__init__ will call self.launch +__CUDACPP_DEBUG: GridPackCmd.launch starting +__CUDACPP_DEBUG: GridPackCmd.launch (1) restore_data +generate 500 events +P0_gg_epemgguux +P0_gg_epemggddx +P0_gg_taptamgguux +P0_gg_taptamggddx +P0_gu_epemgggu +P0_gd_epemgggd +P0_gux_epemgggux +P0_gdx_epemgggdx +P0_gu_taptamgggu +P0_gd_taptamgggd +P0_gux_taptamgggux +P0_gdx_taptamgggdx +P0_uux_epemgggg +P0_ddx_epemgggg +P0_uux_taptamgggg +P0_ddx_taptamgggg +P0_gg_epemuuuxux +P0_gg_epemdddxdx +P0_gg_epemuduxdx +P0_gg_epemucuxcx +P0_gg_epemdsdxsx +P0_gg_taptamuuuxux +P0_gg_taptamdddxdx +P0_gg_taptamuduxdx +P0_gg_taptamucuxcx +P0_gg_taptamdsdxsx +P0_gu_epemguuux +P0_gd_epemgdddx +P0_gux_epemguuxux +P0_gdx_epemgddxdx +P0_gu_epemguddx +P0_gu_epemguccx +P0_gd_epemgudux +P0_gd_epemgdssx +P0_gux_epemgduxdx +P0_gux_epemgcuxcx +P0_gdx_epemguuxdx +P0_gdx_epemgsdxsx +P0_gu_taptamguuux +P0_gd_taptamgdddx +P0_gux_taptamguuxux +P0_gdx_taptamgddxdx +P0_gu_taptamguddx +P0_gu_taptamguccx +P0_gd_taptamgudux +P0_gd_taptamgdssx +P0_gux_taptamgduxdx +P0_gux_taptamgcuxcx +P0_gdx_taptamguuxdx +P0_gdx_taptamgsdxsx +P0_uu_epemgguu +P0_uux_epemgguux +P0_dd_epemggdd +P0_ddx_epemggddx +P0_uxux_epemgguxux +P0_dxdx_epemggdxdx +P0_ud_epemggud +P0_uc_epemgguc +P0_uux_epemggddx +P0_uux_epemggccx +P0_udx_epemggudx +P0_ucx_epemggucx +P0_dc_epemggdc +P0_ds_epemggds +P0_dux_epemggdux +P0_ddx_epemgguux +P0_ddx_epemggssx +P0_dsx_epemggdsx +P0_uxdx_epemgguxdx +P0_uxcx_epemgguxcx +P0_dxcx_epemggdxcx +P0_dxsx_epemggdxsx +P0_uu_taptamgguu +P0_uux_taptamgguux +P0_dd_taptamggdd +P0_ddx_taptamggddx +P0_uxux_taptamgguxux +P0_dxdx_taptamggdxdx +P0_ud_taptamggud +P0_uc_taptamgguc +P0_uux_taptamggddx +P0_uux_taptamggccx +P0_udx_taptamggudx +P0_ucx_taptamggucx +P0_dc_taptamggdc +P0_ds_taptamggds +P0_dux_taptamggdux +P0_ddx_taptamgguux +P0_ddx_taptamggssx +P0_dsx_taptamggdsx +P0_uxdx_taptamgguxdx +P0_uxcx_taptamgguxcx +P0_dxcx_taptamggdxcx +P0_dxsx_taptamggdxsx +P0_uu_epemuuuux +P0_uux_epemuuuxux +P0_dd_epemddddx +P0_ddx_epemdddxdx +P0_uxux_epemuuxuxux +P0_dxdx_epemddxdxdx +P0_uu_epemuuddx +P0_uu_epemuuccx +P0_ud_epemuudux +P0_ud_epemudddx +P0_uc_epemuucux +P0_uc_epemucccx +P0_uux_epemuduxdx +P0_uux_epemucuxcx +P0_uux_epemdddxdx +P0_uux_epemcccxcx +P0_udx_epemuuuxdx +P0_udx_epemuddxdx +P0_ucx_epemuuuxcx +P0_ucx_epemuccxcx +P0_dd_epemuddux +P0_dd_epemddssx +P0_dc_epemddcdx +P0_dc_epemdcccx +P0_ds_epemddsdx +P0_ds_epemdsssx +P0_dux_epemuduxux +P0_dux_epemdduxdx +P0_ddx_epemuuuxux +P0_ddx_epemuduxdx +P0_ddx_epemdcdxcx +P0_ddx_epemdsdxsx +P0_ddx_epemsssxsx +P0_dsx_epemdddxsx +P0_dsx_epemdssxsx +P0_ccx_epemucuxcx +P0_ccx_epemdcdxcx +P0_ssx_epemdsdxsx +P0_uxux_epemduxuxdx +P0_uxux_epemcuxuxcx +P0_uxdx_epemuuxuxdx +P0_uxdx_epemduxdxdx +P0_uxcx_epemuuxuxcx +P0_uxcx_epemcuxcxcx +P0_dxdx_epemuuxdxdx +P0_dxdx_epemsdxdxsx +P0_dxcx_epemddxdxcx +P0_dxcx_epemcdxcxcx +P0_dxsx_epemddxdxsx +P0_dxsx_epemsdxsxsx +P0_ud_epemudccx +P0_ud_epemudssx +P0_uc_epemudcdx +P0_uux_epemdcdxcx +P0_uux_epemdsdxsx +P0_uux_epemcscxsx +P0_udx_epemucdxcx +P0_udx_epemusdxsx +P0_ucx_epemuddxcx +P0_dc_epemudcux +P0_dc_epemdcssx +P0_ds_epemudsux +P0_ds_epemdsbbx +P0_dux_epemdcuxcx +P0_dux_epemdsuxsx +P0_ddx_epemucuxcx +P0_ddx_epemusuxsx +P0_ddx_epemsbsxbx +P0_dsx_epemuduxsx +P0_dsx_epemdbsxbx +P0_ccx_epemuduxdx +P0_ccx_epemdsdxsx +P0_ssx_epemuduxdx +P0_ssx_epemucuxcx +P0_ssx_epemdcdxcx +P0_ssx_epemdbdxbx +P0_bbx_epemdsdxsx +P0_uxdx_epemcuxdxcx +P0_uxdx_epemsuxdxsx +P0_uxcx_epemduxdxcx +P0_dxcx_epemuuxdxcx +P0_dxcx_epemsdxcxsx +P0_dxsx_epemuuxdxsx +P0_dxsx_epembdxsxbx +P0_uu_taptamuuuux +P0_uux_taptamuuuxux +P0_dd_taptamddddx +P0_ddx_taptamdddxdx +P0_uxux_taptamuuxuxux +P0_dxdx_taptamddxdxdx +P0_uu_taptamuuddx +P0_uu_taptamuuccx +P0_ud_taptamuudux +P0_ud_taptamudddx +P0_uc_taptamuucux +P0_uc_taptamucccx +P0_uux_taptamuduxdx +P0_uux_taptamucuxcx +P0_uux_taptamdddxdx +P0_uux_taptamcccxcx +P0_udx_taptamuuuxdx +P0_udx_taptamuddxdx +P0_ucx_taptamuuuxcx +P0_ucx_taptamuccxcx +P0_dd_taptamuddux +P0_dd_taptamddssx +P0_dc_taptamddcdx +P0_dc_taptamdcccx +P0_ds_taptamddsdx +P0_ds_taptamdsssx +P0_dux_taptamuduxux +P0_dux_taptamdduxdx +P0_ddx_taptamuuuxux +P0_ddx_taptamuduxdx +P0_ddx_taptamdcdxcx +P0_ddx_taptamdsdxsx +P0_ddx_taptamsssxsx +P0_dsx_taptamdddxsx +P0_dsx_taptamdssxsx +P0_ccx_taptamucuxcx +P0_ccx_taptamdcdxcx +P0_ssx_taptamdsdxsx +P0_uxux_taptamduxuxdx +P0_uxux_taptamcuxuxcx +P0_uxdx_taptamuuxuxdx +P0_uxdx_taptamduxdxdx +P0_uxcx_taptamuuxuxcx +P0_uxcx_taptamcuxcxcx +P0_dxdx_taptamuuxdxdx +P0_dxdx_taptamsdxdxsx +P0_dxcx_taptamddxdxcx +P0_dxcx_taptamcdxcxcx +P0_dxsx_taptamddxdxsx +P0_dxsx_taptamsdxsxsx +P0_ud_taptamudccx +P0_ud_taptamudssx +P0_uc_taptamudcdx +P0_uux_taptamdcdxcx +P0_uux_taptamdsdxsx +P0_uux_taptamcscxsx +P0_udx_taptamucdxcx +P0_udx_taptamusdxsx +P0_ucx_taptamuddxcx +P0_dc_taptamudcux +P0_dc_taptamdcssx +P0_ds_taptamudsux +P0_ds_taptamdsbbx +P0_dux_taptamdcuxcx +P0_dux_taptamdsuxsx +P0_ddx_taptamucuxcx +P0_ddx_taptamusuxsx +P0_ddx_taptamsbsxbx +P0_dsx_taptamuduxsx +P0_dsx_taptamdbsxbx +P0_ccx_taptamuduxdx +P0_ccx_taptamdsdxsx +P0_ssx_taptamuduxdx +P0_ssx_taptamucuxcx +P0_ssx_taptamdcdxcx +P0_ssx_taptamdbdxbx +P0_bbx_taptamdsdxsx +P0_uxdx_taptamcuxdxcx +P0_uxdx_taptamsuxdxsx +P0_uxcx_taptamduxdxcx +P0_dxcx_taptamuuxdxcx +P0_dxcx_taptamsdxcxsx +P0_dxsx_taptamuuxdxsx +P0_dxsx_taptambdxsxbx +__CUDACPP_DEBUG: GridPackCmd.launch (2) refine4grid +__CUDACPP_DEBUG: GridPackCmd.refine4grid starting +Using random number seed offset = 21 +INFO: Generating 500.0 unweighted events. +__CUDACPP_DEBUG: gen_ximprove.launch starting +sum of cpu time of last step: 3 days,23h23m17s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event starting +INFO: need to improve 477 channels +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dxdx_taptamuuxdxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dxdx_taptamuuxdxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dxdx_taptamuuxdxdx/ajob1' in 10.0534 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dxdx_taptamuuxdxdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dxdx_taptamuuxdxdx/G21/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 10.0216s + [COUNTERS] Fortran Overhead ( 0 ) : 3.6551s + [COUNTERS] Fortran MEs ( 1 ) : 6.3665s for 16384 events => throughput is 2.57E+03 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dxdx_taptamuuxdxdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamguuxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamguuxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamguuxdx/ajob1' in 40.6649 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamguuxdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamguuxdx/G43/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 19.9808s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9869s + [COUNTERS] Fortran MEs ( 1 ) : 17.9939s for 16384 events => throughput is 9.11E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamguuxdx/G335/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 20.6432s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6045s + [COUNTERS] Fortran MEs ( 1 ) : 18.0387s for 16384 events => throughput is 9.08E+02 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamguuxdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemuuuxux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemuuuxux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemuuuxux/ajob1' in 25.9736 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemuuuxux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemuuuxux/G689/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 25.9304s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4931s + [COUNTERS] Fortran MEs ( 1 ) : 23.4372s for 16384 events => throughput is 6.99E+02 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemuuuxux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamgguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamgguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamgguux/ajob1' in 836.7948 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamgguux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamgguux/G596/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 91.4605s + [COUNTERS] Fortran Overhead ( 0 ) : 3.3821s + [COUNTERS] Fortran MEs ( 1 ) : 88.0785s for 16384 events => throughput is 1.86E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamgguux/G304/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 91.5477s + [COUNTERS] Fortran Overhead ( 0 ) : 3.0633s + [COUNTERS] Fortran MEs ( 1 ) : 88.4844s for 16384 events => throughput is 1.85E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamgguux/G128/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 91.4352s + [COUNTERS] Fortran Overhead ( 0 ) : 3.0495s + [COUNTERS] Fortran MEs ( 1 ) : 88.3857s for 16384 events => throughput is 1.85E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamgguux/G115/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 94.7790s + [COUNTERS] Fortran Overhead ( 0 ) : 6.6736s + [COUNTERS] Fortran MEs ( 1 ) : 88.1054s for 16384 events => throughput is 1.86E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamgguux/G599/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 92.5037s + [COUNTERS] Fortran Overhead ( 0 ) : 3.8581s + [COUNTERS] Fortran MEs ( 1 ) : 88.6456s for 16384 events => throughput is 1.85E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamgguux/G351/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 94.2785s + [COUNTERS] Fortran Overhead ( 0 ) : 6.0938s + [COUNTERS] Fortran MEs ( 1 ) : 88.1847s for 16384 events => throughput is 1.86E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamgguux/G111/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 90.5002s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4095s + [COUNTERS] Fortran MEs ( 1 ) : 88.0907s for 16384 events => throughput is 1.86E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamgguux/G354/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 94.8235s + [COUNTERS] Fortran Overhead ( 0 ) : 6.5803s + [COUNTERS] Fortran MEs ( 1 ) : 88.2432s for 16384 events => throughput is 1.86E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamgguux/G97/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 95.2940s + [COUNTERS] Fortran Overhead ( 0 ) : 7.2272s + [COUNTERS] Fortran MEs ( 1 ) : 88.0668s for 16384 events => throughput is 1.86E+02 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamgguux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uxdx_epemsuxdxsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uxdx_epemsuxdxsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uxdx_epemsuxdxsx/ajob1' in 3.5239 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uxdx_epemsuxdxsx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uxdx_epemsuxdxsx/G74/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.4675s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1722s + [COUNTERS] Fortran MEs ( 1 ) : 1.2952s for 16384 events => throughput is 1.26E+04 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uxdx_epemsuxdxsx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemdddxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemdddxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemdddxdx/ajob1' in 52.2432 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemdddxdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemdddxdx/G513/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 25.9937s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4673s + [COUNTERS] Fortran MEs ( 1 ) : 23.5265s for 16384 events => throughput is 6.96E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemdddxdx/G180/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 26.2069s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7746s + [COUNTERS] Fortran MEs ( 1 ) : 23.4322s for 16384 events => throughput is 6.99E+02 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemdddxdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgddxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgddxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgddxdx/ajob1' in 26.5740 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgddxdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgddxdx/G515/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 26.5167s + [COUNTERS] Fortran Overhead ( 0 ) : 3.0807s + [COUNTERS] Fortran MEs ( 1 ) : 23.4360s for 16384 events => throughput is 6.99E+02 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgddxdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguddx/ajob1' in 185.6400 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguddx/G104/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 19.5394s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5937s + [COUNTERS] Fortran MEs ( 1 ) : 17.9457s for 16384 events => throughput is 9.13E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguddx/G120/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 20.2166s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1718s + [COUNTERS] Fortran MEs ( 1 ) : 18.0448s for 16384 events => throughput is 9.08E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguddx/G305/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 23.8594s + [COUNTERS] Fortran Overhead ( 0 ) : 5.8943s + [COUNTERS] Fortran MEs ( 1 ) : 17.9651s for 16384 events => throughput is 9.12E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguddx/G101/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 19.4399s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4614s + [COUNTERS] Fortran MEs ( 1 ) : 17.9786s for 16384 events => throughput is 9.11E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguddx/G27/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 19.9759s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8797s + [COUNTERS] Fortran MEs ( 1 ) : 18.0961s for 16384 events => throughput is 9.05E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguddx/G273/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 20.1059s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1511s + [COUNTERS] Fortran MEs ( 1 ) : 17.9548s for 16384 events => throughput is 9.13E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguddx/G349/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 19.4471s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3908s + [COUNTERS] Fortran MEs ( 1 ) : 18.0563s for 16384 events => throughput is 9.07E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguddx/G379/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 19.6589s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7691s + [COUNTERS] Fortran MEs ( 1 ) : 17.8898s for 16384 events => throughput is 9.16E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguddx/G256/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 23.2081s + [COUNTERS] Fortran Overhead ( 0 ) : 5.2166s + [COUNTERS] Fortran MEs ( 1 ) : 17.9915s for 16384 events => throughput is 9.11E+02 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemguuxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemguuxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemguuxdx/ajob1' in 40.6885 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemguuxdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemguuxdx/G19/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 15.1389s + [COUNTERS] Fortran Overhead ( 0 ) : 6.0820s + [COUNTERS] Fortran MEs ( 1 ) : 9.0568s for 16384 events => throughput is 1.81E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemguuxdx/G68/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.3916s + [COUNTERS] Fortran Overhead ( 0 ) : 5.3710s + [COUNTERS] Fortran MEs ( 1 ) : 9.0206s for 16384 events => throughput is 1.82E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemguuxdx/G355/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 11.0821s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0802s + [COUNTERS] Fortran MEs ( 1 ) : 9.0019s for 16384 events => throughput is 1.82E+03 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemguuxdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/ajob1' in 2086.2863 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G14/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 89.8679s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0813s + [COUNTERS] Fortran MEs ( 1 ) : 87.7866s for 16384 events => throughput is 1.87E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G56/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 90.2957s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4649s + [COUNTERS] Fortran MEs ( 1 ) : 87.8308s for 16384 events => throughput is 1.87E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G517/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 93.6333s + [COUNTERS] Fortran Overhead ( 0 ) : 6.0937s + [COUNTERS] Fortran MEs ( 1 ) : 87.5396s for 16384 events => throughput is 1.87E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G52/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 90.7669s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7622s + [COUNTERS] Fortran MEs ( 1 ) : 88.0046s for 16384 events => throughput is 1.86E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G304/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 90.4082s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7169s + [COUNTERS] Fortran MEs ( 1 ) : 87.6913s for 16384 events => throughput is 1.87E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G3/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 92.5562s + [COUNTERS] Fortran Overhead ( 0 ) : 4.8257s + [COUNTERS] Fortran MEs ( 1 ) : 87.7305s for 16384 events => throughput is 1.87E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G23/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 89.2057s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7149s + [COUNTERS] Fortran MEs ( 1 ) : 87.4908s for 16384 events => throughput is 1.87E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G33/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 89.9041s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1571s + [COUNTERS] Fortran MEs ( 1 ) : 87.7470s for 16384 events => throughput is 1.87E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G53/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 94.4256s + [COUNTERS] Fortran Overhead ( 0 ) : 6.6932s + [COUNTERS] Fortran MEs ( 1 ) : 87.7325s for 16384 events => throughput is 1.87E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G495/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 89.3186s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5954s + [COUNTERS] Fortran MEs ( 1 ) : 87.7232s for 16384 events => throughput is 1.87E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G68/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 89.6403s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1721s + [COUNTERS] Fortran MEs ( 1 ) : 87.4682s for 16384 events => throughput is 1.87E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G49/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 89.8428s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0788s + [COUNTERS] Fortran MEs ( 1 ) : 87.7641s for 16384 events => throughput is 1.87E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G524/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 89.9220s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1532s + [COUNTERS] Fortran MEs ( 1 ) : 87.7688s for 16384 events => throughput is 1.87E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G65/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 89.5789s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0702s + [COUNTERS] Fortran MEs ( 1 ) : 87.5086s for 16384 events => throughput is 1.87E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G303/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 90.7346s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9904s + [COUNTERS] Fortran MEs ( 1 ) : 87.7442s for 16384 events => throughput is 1.87E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G55/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 90.5491s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6735s + [COUNTERS] Fortran MEs ( 1 ) : 87.8756s for 16384 events => throughput is 1.86E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G34/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 89.7723s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7978s + [COUNTERS] Fortran MEs ( 1 ) : 87.9744s for 16384 events => throughput is 1.86E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G509/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 95.2680s + [COUNTERS] Fortran Overhead ( 0 ) : 7.6401s + [COUNTERS] Fortran MEs ( 1 ) : 87.6279s for 16384 events => throughput is 1.87E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G512/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 89.5019s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8254s + [COUNTERS] Fortran MEs ( 1 ) : 87.6764s for 16384 events => throughput is 1.87E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G511/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 90.2786s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4382s + [COUNTERS] Fortran MEs ( 1 ) : 87.8404s for 16384 events => throughput is 1.87E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 90.4389s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7788s + [COUNTERS] Fortran MEs ( 1 ) : 87.6602s for 16384 events => throughput is 1.87E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G63/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 89.8801s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0964s + [COUNTERS] Fortran MEs ( 1 ) : 87.7837s for 16384 events => throughput is 1.87E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/G48/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 90.0510s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3190s + [COUNTERS] Fortran MEs ( 1 ) : 87.7320s for 16384 events => throughput is 1.87E+02 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgggd/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/ajob1' in 1130.1713 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G122/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 46.5678s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7597s + [COUNTERS] Fortran MEs ( 1 ) : 43.8082s for 16384 events => throughput is 3.74E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G108/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 46.3199s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3269s + [COUNTERS] Fortran MEs ( 1 ) : 43.9930s for 16384 events => throughput is 3.72E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G105/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 46.8621s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5275s + [COUNTERS] Fortran MEs ( 1 ) : 44.3346s for 16384 events => throughput is 3.70E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G304/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 46.9645s + [COUNTERS] Fortran Overhead ( 0 ) : 3.0559s + [COUNTERS] Fortran MEs ( 1 ) : 43.9086s for 16384 events => throughput is 3.73E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G301/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 51.6109s + [COUNTERS] Fortran Overhead ( 0 ) : 7.5751s + [COUNTERS] Fortran MEs ( 1 ) : 44.0358s for 16384 events => throughput is 3.72E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G128/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 47.0865s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1624s + [COUNTERS] Fortran MEs ( 1 ) : 43.9240s for 16384 events => throughput is 3.73E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G100/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 51.5785s + [COUNTERS] Fortran Overhead ( 0 ) : 7.6852s + [COUNTERS] Fortran MEs ( 1 ) : 43.8933s for 16384 events => throughput is 3.73E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G120/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 46.3728s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4392s + [COUNTERS] Fortran MEs ( 1 ) : 43.9336s for 16384 events => throughput is 3.73E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G302/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 54.5038s + [COUNTERS] Fortran Overhead ( 0 ) : 10.6601s + [COUNTERS] Fortran MEs ( 1 ) : 43.8438s for 16384 events => throughput is 3.74E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G299/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 46.5031s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6744s + [COUNTERS] Fortran MEs ( 1 ) : 43.8288s for 16384 events => throughput is 3.74E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G169/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 49.8521s + [COUNTERS] Fortran Overhead ( 0 ) : 5.9976s + [COUNTERS] Fortran MEs ( 1 ) : 43.8545s for 16384 events => throughput is 3.74E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G333/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 54.8695s + [COUNTERS] Fortran Overhead ( 0 ) : 11.0187s + [COUNTERS] Fortran MEs ( 1 ) : 43.8508s for 16384 events => throughput is 3.74E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G594/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 53.1203s + [COUNTERS] Fortran Overhead ( 0 ) : 9.2621s + [COUNTERS] Fortran MEs ( 1 ) : 43.8581s for 16384 events => throughput is 3.74E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G678/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 50.3101s + [COUNTERS] Fortran Overhead ( 0 ) : 6.3169s + [COUNTERS] Fortran MEs ( 1 ) : 43.9932s for 16384 events => throughput is 3.72E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G127/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 46.8569s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7053s + [COUNTERS] Fortran MEs ( 1 ) : 44.1517s for 16384 events => throughput is 3.71E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G126/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 53.5912s + [COUNTERS] Fortran Overhead ( 0 ) : 9.6736s + [COUNTERS] Fortran MEs ( 1 ) : 43.9176s for 16384 events => throughput is 3.73E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G598/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 46.0797s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1524s + [COUNTERS] Fortran MEs ( 1 ) : 43.9273s for 16384 events => throughput is 3.73E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G512/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 45.9111s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1079s + [COUNTERS] Fortran MEs ( 1 ) : 43.8032s for 16384 events => throughput is 3.74E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G103/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 52.9247s + [COUNTERS] Fortran Overhead ( 0 ) : 8.9416s + [COUNTERS] Fortran MEs ( 1 ) : 43.9831s for 16384 events => throughput is 3.73E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G107/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 46.2835s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3119s + [COUNTERS] Fortran MEs ( 1 ) : 43.9716s for 16384 events => throughput is 3.73E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G358/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 47.6450s + [COUNTERS] Fortran Overhead ( 0 ) : 3.7648s + [COUNTERS] Fortran MEs ( 1 ) : 43.8802s for 16384 events => throughput is 3.73E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G97/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 51.2481s + [COUNTERS] Fortran Overhead ( 0 ) : 7.3957s + [COUNTERS] Fortran MEs ( 1 ) : 43.8524s for 16384 events => throughput is 3.74E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/G179/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 46.6845s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7368s + [COUNTERS] Fortran MEs ( 1 ) : 43.9477s for 16384 events => throughput is 3.73E+02 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemgguux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_epemudcdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_epemudcdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_epemudcdx/ajob1' in 3.0169 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_epemudcdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_epemudcdx/G42/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 2.9919s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6940s + [COUNTERS] Fortran MEs ( 1 ) : 1.2979s for 16384 events => throughput is 1.26E+04 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_epemudcdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ds_taptamggds/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ds_taptamggds/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ds_taptamggds/ajob1' in 42.0741 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ds_taptamggds/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ds_taptamggds/G17/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 22.5295s + [COUNTERS] Fortran Overhead ( 0 ) : 4.6363s + [COUNTERS] Fortran MEs ( 1 ) : 17.8932s for 16384 events => throughput is 9.16E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ds_taptamggds/G110/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 19.4826s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6508s + [COUNTERS] Fortran MEs ( 1 ) : 17.8319s for 16384 events => throughput is 9.19E+02 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ds_taptamggds/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_epemggdux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_epemggdux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_epemggdux/ajob1' in 52.4047 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_epemggdux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_epemggdux/G208/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 10.9807s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0133s + [COUNTERS] Fortran MEs ( 1 ) : 8.9674s for 16384 events => throughput is 1.83E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_epemggdux/G192/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 10.3832s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4039s + [COUNTERS] Fortran MEs ( 1 ) : 8.9793s for 16384 events => throughput is 1.82E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_epemggdux/G380/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 10.5899s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6061s + [COUNTERS] Fortran MEs ( 1 ) : 8.9837s for 16384 events => throughput is 1.82E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_epemggdux/G127/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 10.1764s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2499s + [COUNTERS] Fortran MEs ( 1 ) : 8.9266s for 16384 events => throughput is 1.84E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_epemggdux/G223/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 10.1596s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2020s + [COUNTERS] Fortran MEs ( 1 ) : 8.9576s for 16384 events => throughput is 1.83E+03 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_epemggdux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_taptamgguc/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_taptamgguc/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_taptamgguc/ajob1' in 64.5806 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_taptamgguc/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_taptamgguc/G20/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 22.2010s + [COUNTERS] Fortran Overhead ( 0 ) : 4.4006s + [COUNTERS] Fortran MEs ( 1 ) : 17.8004s for 16384 events => throughput is 9.20E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_taptamgguc/G110/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 19.0537s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3058s + [COUNTERS] Fortran MEs ( 1 ) : 17.7478s for 16384 events => throughput is 9.23E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_taptamgguc/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 23.2652s + [COUNTERS] Fortran Overhead ( 0 ) : 5.4453s + [COUNTERS] Fortran MEs ( 1 ) : 17.8199s for 16384 events => throughput is 9.19E+02 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_taptamgguc/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uxcx_taptamgguxcx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uxcx_taptamgguxcx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uxcx_taptamgguxcx/ajob1' in 20.4209 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uxcx_taptamgguxcx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uxcx_taptamgguxcx/G187/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 20.3803s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5233s + [COUNTERS] Fortran MEs ( 1 ) : 17.8570s for 16384 events => throughput is 9.18E+02 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uxcx_taptamgguxcx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/ajob1' in 1226.7286 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G499/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 45.9915s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2124s + [COUNTERS] Fortran MEs ( 1 ) : 43.7791s for 16384 events => throughput is 3.74E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G56/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 46.7095s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9767s + [COUNTERS] Fortran MEs ( 1 ) : 43.7328s for 16384 events => throughput is 3.75E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G39/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 46.5104s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6931s + [COUNTERS] Fortran MEs ( 1 ) : 43.8172s for 16384 events => throughput is 3.74E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G64/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 46.3442s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5197s + [COUNTERS] Fortran MEs ( 1 ) : 43.8245s for 16384 events => throughput is 3.74E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G13/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 45.3963s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5907s + [COUNTERS] Fortran MEs ( 1 ) : 43.8057s for 16384 events => throughput is 3.74E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G33/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 46.3717s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7750s + [COUNTERS] Fortran MEs ( 1 ) : 44.5967s for 16384 events => throughput is 3.67E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G43/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 50.8691s + [COUNTERS] Fortran Overhead ( 0 ) : 7.1130s + [COUNTERS] Fortran MEs ( 1 ) : 43.7561s for 16384 events => throughput is 3.74E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G68/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 46.7627s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9108s + [COUNTERS] Fortran MEs ( 1 ) : 43.8519s for 16384 events => throughput is 3.74E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G49/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 46.2237s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1041s + [COUNTERS] Fortran MEs ( 1 ) : 44.1196s for 16384 events => throughput is 3.71E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G25/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 49.5688s + [COUNTERS] Fortran Overhead ( 0 ) : 5.7374s + [COUNTERS] Fortran MEs ( 1 ) : 43.8315s for 16384 events => throughput is 3.74E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G523/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 46.5423s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7202s + [COUNTERS] Fortran MEs ( 1 ) : 43.8220s for 16384 events => throughput is 3.74E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G395/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 46.0910s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3919s + [COUNTERS] Fortran MEs ( 1 ) : 43.6991s for 16384 events => throughput is 3.75E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G524/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 45.6411s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9090s + [COUNTERS] Fortran MEs ( 1 ) : 43.7321s for 16384 events => throughput is 3.75E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G65/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 45.8140s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0352s + [COUNTERS] Fortran MEs ( 1 ) : 43.7788s for 16384 events => throughput is 3.74E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G609/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 46.7362s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9736s + [COUNTERS] Fortran MEs ( 1 ) : 43.7626s for 16384 events => throughput is 3.74E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G611/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 47.0227s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2528s + [COUNTERS] Fortran MEs ( 1 ) : 43.7698s for 16384 events => throughput is 3.74E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G1/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 48.3256s + [COUNTERS] Fortran Overhead ( 0 ) : 4.6418s + [COUNTERS] Fortran MEs ( 1 ) : 43.6838s for 16384 events => throughput is 3.75E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G55/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 46.4422s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5956s + [COUNTERS] Fortran MEs ( 1 ) : 43.8466s for 16384 events => throughput is 3.74E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G34/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 46.0062s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1564s + [COUNTERS] Fortran MEs ( 1 ) : 43.8499s for 16384 events => throughput is 3.74E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G26/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 50.1294s + [COUNTERS] Fortran Overhead ( 0 ) : 6.2378s + [COUNTERS] Fortran MEs ( 1 ) : 43.8916s for 16384 events => throughput is 3.73E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G58/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 49.3265s + [COUNTERS] Fortran Overhead ( 0 ) : 5.6570s + [COUNTERS] Fortran MEs ( 1 ) : 43.6695s for 16384 events => throughput is 3.75E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G2/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 48.8432s + [COUNTERS] Fortran Overhead ( 0 ) : 5.0593s + [COUNTERS] Fortran MEs ( 1 ) : 43.7839s for 16384 events => throughput is 3.74E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G512/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 45.9623s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1851s + [COUNTERS] Fortran MEs ( 1 ) : 43.7772s for 16384 events => throughput is 3.74E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G51/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 46.3574s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6610s + [COUNTERS] Fortran MEs ( 1 ) : 43.6964s for 16384 events => throughput is 3.75E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G29/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 50.3252s + [COUNTERS] Fortran Overhead ( 0 ) : 6.5907s + [COUNTERS] Fortran MEs ( 1 ) : 43.7345s for 16384 events => throughput is 3.75E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/G48/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 45.9114s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0540s + [COUNTERS] Fortran MEs ( 1 ) : 43.8574s for 16384 events => throughput is 3.74E+02 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgggdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/ajob1' in 149.0260 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/G140/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 10.7991s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8322s + [COUNTERS] Fortran MEs ( 1 ) : 8.9669s for 16384 events => throughput is 1.83E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/G114/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 15.4630s + [COUNTERS] Fortran Overhead ( 0 ) : 6.4959s + [COUNTERS] Fortran MEs ( 1 ) : 8.9671s for 16384 events => throughput is 1.83E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/G49/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 16.2622s + [COUNTERS] Fortran Overhead ( 0 ) : 7.2638s + [COUNTERS] Fortran MEs ( 1 ) : 8.9984s for 16384 events => throughput is 1.82E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/G120/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 10.9846s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0019s + [COUNTERS] Fortran MEs ( 1 ) : 8.9827s for 16384 events => throughput is 1.82E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/G262/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 10.7289s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7533s + [COUNTERS] Fortran MEs ( 1 ) : 8.9756s for 16384 events => throughput is 1.83E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/G27/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 10.8347s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8708s + [COUNTERS] Fortran MEs ( 1 ) : 8.9639s for 16384 events => throughput is 1.83E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/G44/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 10.5825s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5668s + [COUNTERS] Fortran MEs ( 1 ) : 9.0157s for 16384 events => throughput is 1.82E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/G96/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 12.4895s + [COUNTERS] Fortran Overhead ( 0 ) : 3.4377s + [COUNTERS] Fortran MEs ( 1 ) : 9.0517s for 16384 events => throughput is 1.81E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/G261/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 11.0920s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1039s + [COUNTERS] Fortran MEs ( 1 ) : 8.9881s for 16384 events => throughput is 1.82E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/G37/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.8525s + [COUNTERS] Fortran Overhead ( 0 ) : 4.8814s + [COUNTERS] Fortran MEs ( 1 ) : 8.9711s for 16384 events => throughput is 1.83E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/G28/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 10.8202s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7758s + [COUNTERS] Fortran MEs ( 1 ) : 9.0444s for 16384 events => throughput is 1.81E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.8675s + [COUNTERS] Fortran Overhead ( 0 ) : 5.8733s + [COUNTERS] Fortran MEs ( 1 ) : 8.9943s for 16384 events => throughput is 1.82E+03 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_epemuduxsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_epemuduxsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_epemuduxsx/ajob1' in 3.8926 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_epemuduxsx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_epemuduxsx/G42/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.8664s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5703s + [COUNTERS] Fortran MEs ( 1 ) : 1.2961s for 16384 events => throughput is 1.26E+04 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_epemuduxsx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dd_epemggdd/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dd_epemggdd/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dd_epemggdd/ajob1' in 75.9550 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dd_epemggdd/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dd_epemggdd/G467/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 25.3229s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8480s + [COUNTERS] Fortran MEs ( 1 ) : 23.4749s for 16384 events => throughput is 6.98E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dd_epemggdd/G351/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 24.8538s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2810s + [COUNTERS] Fortran MEs ( 1 ) : 23.5728s for 16384 events => throughput is 6.95E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dd_epemggdd/G44/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 25.6963s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2279s + [COUNTERS] Fortran MEs ( 1 ) : 23.4684s for 16384 events => throughput is 6.98E+02 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dd_epemggdd/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamusdxsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamusdxsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamusdxsx/ajob1' in 3.8296 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamusdxsx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamusdxsx/G92/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.7854s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2040s + [COUNTERS] Fortran MEs ( 1 ) : 2.5815s for 16384 events => throughput is 6.35E+03 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamusdxsx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemggddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemggddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemggddx/ajob1' in 158.5166 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemggddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemggddx/G175/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 25.6370s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1126s + [COUNTERS] Fortran MEs ( 1 ) : 23.5244s for 16384 events => throughput is 6.96E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemggddx/G303/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 26.1569s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7121s + [COUNTERS] Fortran MEs ( 1 ) : 23.4448s for 16384 events => throughput is 6.99E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemggddx/G111/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 25.2165s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7893s + [COUNTERS] Fortran MEs ( 1 ) : 23.4272s for 16384 events => throughput is 6.99E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemggddx/G219/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 26.4581s + [COUNTERS] Fortran Overhead ( 0 ) : 3.0419s + [COUNTERS] Fortran MEs ( 1 ) : 23.4162s for 16384 events => throughput is 7.00E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemggddx/G318/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 29.4930s + [COUNTERS] Fortran Overhead ( 0 ) : 6.0094s + [COUNTERS] Fortran MEs ( 1 ) : 23.4836s for 16384 events => throughput is 6.98E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemggddx/G143/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 25.4398s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9378s + [COUNTERS] Fortran MEs ( 1 ) : 23.5019s for 16384 events => throughput is 6.97E+02 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemggddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemguuxux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemguuxux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemguuxux/ajob1' in 30.7083 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemguuxux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemguuxux/G615/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 30.6568s + [COUNTERS] Fortran Overhead ( 0 ) : 7.0910s + [COUNTERS] Fortran MEs ( 1 ) : 23.5658s for 16384 events => throughput is 6.95E+02 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemguuxux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamgggg/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamgggg/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamgggg/ajob1' in 90.7381 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamgggg/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamgggg/G66/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 90.6904s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6586s + [COUNTERS] Fortran MEs ( 1 ) : 88.0318s for 16384 events => throughput is 1.86E+02 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamgggg/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgduxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgduxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgduxdx/ajob1' in 20.0008 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgduxdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgduxdx/G44/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 19.9752s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1069s + [COUNTERS] Fortran MEs ( 1 ) : 17.8683s for 16384 events => throughput is 9.17E+02 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgduxdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgudux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgudux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgudux/ajob1' in 107.5490 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgudux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgudux/G39/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 15.1717s + [COUNTERS] Fortran Overhead ( 0 ) : 6.0766s + [COUNTERS] Fortran MEs ( 1 ) : 9.0950s for 16384 events => throughput is 1.80E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgudux/G366/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.6864s + [COUNTERS] Fortran Overhead ( 0 ) : 4.6739s + [COUNTERS] Fortran MEs ( 1 ) : 9.0124s for 16384 events => throughput is 1.82E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgudux/G259/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 11.3201s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2970s + [COUNTERS] Fortran MEs ( 1 ) : 9.0231s for 16384 events => throughput is 1.82E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgudux/G28/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 11.1306s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0835s + [COUNTERS] Fortran MEs ( 1 ) : 9.0471s for 16384 events => throughput is 1.81E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgudux/G32/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 11.4948s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4522s + [COUNTERS] Fortran MEs ( 1 ) : 9.0425s for 16384 events => throughput is 1.81E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgudux/G61/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 17.3689s + [COUNTERS] Fortran Overhead ( 0 ) : 8.3472s + [COUNTERS] Fortran MEs ( 1 ) : 9.0217s for 16384 events => throughput is 1.82E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgudux/G267/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.8296s + [COUNTERS] Fortran Overhead ( 0 ) : 5.7805s + [COUNTERS] Fortran MEs ( 1 ) : 9.0491s for 16384 events => throughput is 1.81E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgudux/G144/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 12.3952s + [COUNTERS] Fortran Overhead ( 0 ) : 3.3839s + [COUNTERS] Fortran MEs ( 1 ) : 9.0113s for 16384 events => throughput is 1.82E+03 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgudux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemuudux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemuudux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemuudux/ajob1' in 7.2391 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemuudux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemuudux/G104/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.2137s + [COUNTERS] Fortran Overhead ( 0 ) : 3.9674s + [COUNTERS] Fortran MEs ( 1 ) : 3.2462s for 16384 events => throughput is 5.05E+03 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemuudux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamggddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamggddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamggddx/ajob1' in 725.6078 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamggddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamggddx/G108/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 89.8534s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1986s + [COUNTERS] Fortran MEs ( 1 ) : 87.6548s for 16384 events => throughput is 1.87E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamggddx/G304/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 90.3551s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7271s + [COUNTERS] Fortran MEs ( 1 ) : 87.6280s for 16384 events => throughput is 1.87E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamggddx/G128/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 90.4463s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8373s + [COUNTERS] Fortran MEs ( 1 ) : 87.6090s for 16384 events => throughput is 1.87E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamggddx/G360/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 90.0890s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3544s + [COUNTERS] Fortran MEs ( 1 ) : 87.7347s for 16384 events => throughput is 1.87E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamggddx/G169/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 93.8165s + [COUNTERS] Fortran Overhead ( 0 ) : 6.0837s + [COUNTERS] Fortran MEs ( 1 ) : 87.7327s for 16384 events => throughput is 1.87E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamggddx/G176/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 89.5404s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9865s + [COUNTERS] Fortran MEs ( 1 ) : 87.5539s for 16384 events => throughput is 1.87E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamggddx/G303/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 90.4764s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8285s + [COUNTERS] Fortran MEs ( 1 ) : 87.6478s for 16384 events => throughput is 1.87E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamggddx/G513/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 90.8779s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1470s + [COUNTERS] Fortran MEs ( 1 ) : 87.7309s for 16384 events => throughput is 1.87E+02 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_taptamggddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguccx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguccx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguccx/ajob1' in 58.6608 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguccx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguccx/G283/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 10.9581s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9879s + [COUNTERS] Fortran MEs ( 1 ) : 8.9703s for 16384 events => throughput is 1.83E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguccx/G47/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 10.8905s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9360s + [COUNTERS] Fortran MEs ( 1 ) : 8.9545s for 16384 events => throughput is 1.83E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguccx/G35/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 15.5775s + [COUNTERS] Fortran Overhead ( 0 ) : 6.6189s + [COUNTERS] Fortran MEs ( 1 ) : 8.9585s for 16384 events => throughput is 1.83E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguccx/G373/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 10.6485s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6965s + [COUNTERS] Fortran MEs ( 1 ) : 8.9520s for 16384 events => throughput is 1.83E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguccx/G249/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 10.4712s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5269s + [COUNTERS] Fortran MEs ( 1 ) : 8.9442s for 16384 events => throughput is 1.83E+03 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguccx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/ajob1' in 1482.5574 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G14/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 45.7222s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0261s + [COUNTERS] Fortran MEs ( 1 ) : 43.6961s for 16384 events => throughput is 3.75E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G499/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 45.6552s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9437s + [COUNTERS] Fortran MEs ( 1 ) : 43.7115s for 16384 events => throughput is 3.75E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G56/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 46.3617s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7009s + [COUNTERS] Fortran MEs ( 1 ) : 43.6608s for 16384 events => throughput is 3.75E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G39/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 45.8665s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2786s + [COUNTERS] Fortran MEs ( 1 ) : 43.5879s for 16384 events => throughput is 3.76E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G52/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 46.4705s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8526s + [COUNTERS] Fortran MEs ( 1 ) : 43.6179s for 16384 events => throughput is 3.76E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G304/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 46.2397s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6063s + [COUNTERS] Fortran MEs ( 1 ) : 43.6334s for 16384 events => throughput is 3.75E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G64/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 45.9282s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3019s + [COUNTERS] Fortran MEs ( 1 ) : 43.6263s for 16384 events => throughput is 3.76E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G3/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 48.6948s + [COUNTERS] Fortran Overhead ( 0 ) : 4.9517s + [COUNTERS] Fortran MEs ( 1 ) : 43.7431s for 16384 events => throughput is 3.75E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G13/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 45.2092s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6341s + [COUNTERS] Fortran MEs ( 1 ) : 43.5751s for 16384 events => throughput is 3.76E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G33/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 45.8897s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1612s + [COUNTERS] Fortran MEs ( 1 ) : 43.7286s for 16384 events => throughput is 3.75E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G484/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 46.2253s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5493s + [COUNTERS] Fortran MEs ( 1 ) : 43.6760s for 16384 events => throughput is 3.75E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G68/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 45.7726s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1718s + [COUNTERS] Fortran MEs ( 1 ) : 43.6008s for 16384 events => throughput is 3.76E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G49/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 45.4841s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7763s + [COUNTERS] Fortran MEs ( 1 ) : 43.7078s for 16384 events => throughput is 3.75E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G40/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 46.3710s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8042s + [COUNTERS] Fortran MEs ( 1 ) : 43.5668s for 16384 events => throughput is 3.76E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G523/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 45.8745s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1113s + [COUNTERS] Fortran MEs ( 1 ) : 43.7632s for 16384 events => throughput is 3.74E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 45.2497s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5512s + [COUNTERS] Fortran MEs ( 1 ) : 43.6986s for 16384 events => throughput is 3.75E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G524/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 45.8884s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0896s + [COUNTERS] Fortran MEs ( 1 ) : 43.7988s for 16384 events => throughput is 3.74E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G65/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 45.8365s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1445s + [COUNTERS] Fortran MEs ( 1 ) : 43.6920s for 16384 events => throughput is 3.75E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G44/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 49.3990s + [COUNTERS] Fortran Overhead ( 0 ) : 5.7328s + [COUNTERS] Fortran MEs ( 1 ) : 43.6662s for 16384 events => throughput is 3.75E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 45.1349s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5003s + [COUNTERS] Fortran MEs ( 1 ) : 43.6345s for 16384 events => throughput is 3.75E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G303/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 46.5211s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9632s + [COUNTERS] Fortran MEs ( 1 ) : 43.5579s for 16384 events => throughput is 3.76E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G528/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 47.2327s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1221s + [COUNTERS] Fortran MEs ( 1 ) : 44.1106s for 16384 events => throughput is 3.71E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G55/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 46.5967s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9336s + [COUNTERS] Fortran MEs ( 1 ) : 43.6630s for 16384 events => throughput is 3.75E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G298/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 46.5709s + [COUNTERS] Fortran Overhead ( 0 ) : 3.0149s + [COUNTERS] Fortran MEs ( 1 ) : 43.5560s for 16384 events => throughput is 3.76E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G34/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 45.3822s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7008s + [COUNTERS] Fortran MEs ( 1 ) : 43.6814s for 16384 events => throughput is 3.75E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G54/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 50.7016s + [COUNTERS] Fortran Overhead ( 0 ) : 7.0370s + [COUNTERS] Fortran MEs ( 1 ) : 43.6646s for 16384 events => throughput is 3.75E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G67/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 45.9053s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2700s + [COUNTERS] Fortran MEs ( 1 ) : 43.6353s for 16384 events => throughput is 3.75E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G512/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 45.5190s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9695s + [COUNTERS] Fortran MEs ( 1 ) : 43.5495s for 16384 events => throughput is 3.76E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G511/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 46.2046s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5727s + [COUNTERS] Fortran MEs ( 1 ) : 43.6319s for 16384 events => throughput is 3.76E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 46.4027s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4987s + [COUNTERS] Fortran MEs ( 1 ) : 43.9040s for 16384 events => throughput is 3.73E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G21/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 45.4917s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7585s + [COUNTERS] Fortran MEs ( 1 ) : 43.7333s for 16384 events => throughput is 3.75E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/G48/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 46.1685s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4980s + [COUNTERS] Fortran MEs ( 1 ) : 43.6705s for 16384 events => throughput is 3.75E+02 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgggd/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_taptamggdux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_taptamggdux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_taptamggdux/ajob1' in 80.4777 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_taptamggdux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_taptamggdux/G68/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 21.7574s + [COUNTERS] Fortran Overhead ( 0 ) : 3.8420s + [COUNTERS] Fortran MEs ( 1 ) : 17.9154s for 16384 events => throughput is 9.15E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_taptamggdux/G167/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 19.6391s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7669s + [COUNTERS] Fortran MEs ( 1 ) : 17.8722s for 16384 events => throughput is 9.17E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_taptamggdux/G92/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 19.6434s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7210s + [COUNTERS] Fortran MEs ( 1 ) : 17.9224s for 16384 events => throughput is 9.14E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_taptamggdux/G60/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 19.3571s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4716s + [COUNTERS] Fortran MEs ( 1 ) : 17.8855s for 16384 events => throughput is 9.16E+02 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dux_taptamggdux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/ajob1' in 975.4535 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G56/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 46.8592s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8317s + [COUNTERS] Fortran MEs ( 1 ) : 44.0275s for 16384 events => throughput is 3.72E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G39/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 47.1302s + [COUNTERS] Fortran Overhead ( 0 ) : 3.0811s + [COUNTERS] Fortran MEs ( 1 ) : 44.0492s for 16384 events => throughput is 3.72E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G517/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 52.1301s + [COUNTERS] Fortran Overhead ( 0 ) : 8.1025s + [COUNTERS] Fortran MEs ( 1 ) : 44.0276s for 16384 events => throughput is 3.72E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G497/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 45.9444s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9508s + [COUNTERS] Fortran MEs ( 1 ) : 43.9936s for 16384 events => throughput is 3.72E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G13/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 45.7224s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6959s + [COUNTERS] Fortran MEs ( 1 ) : 44.0265s for 16384 events => throughput is 3.72E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G484/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 50.7569s + [COUNTERS] Fortran Overhead ( 0 ) : 6.6088s + [COUNTERS] Fortran MEs ( 1 ) : 44.1481s for 16384 events => throughput is 3.71E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G68/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 47.0987s + [COUNTERS] Fortran Overhead ( 0 ) : 3.0689s + [COUNTERS] Fortran MEs ( 1 ) : 44.0298s for 16384 events => throughput is 3.72E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G49/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 46.4392s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3345s + [COUNTERS] Fortran MEs ( 1 ) : 44.1046s for 16384 events => throughput is 3.71E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G523/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 47.2060s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5797s + [COUNTERS] Fortran MEs ( 1 ) : 44.6263s for 16384 events => throughput is 3.67E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G518/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 49.9084s + [COUNTERS] Fortran Overhead ( 0 ) : 5.8536s + [COUNTERS] Fortran MEs ( 1 ) : 44.0548s for 16384 events => throughput is 3.72E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G42/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 52.1450s + [COUNTERS] Fortran Overhead ( 0 ) : 8.0636s + [COUNTERS] Fortran MEs ( 1 ) : 44.0814s for 16384 events => throughput is 3.72E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G524/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 45.9038s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9018s + [COUNTERS] Fortran MEs ( 1 ) : 44.0019s for 16384 events => throughput is 3.72E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G610/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 47.1374s + [COUNTERS] Fortran Overhead ( 0 ) : 3.0835s + [COUNTERS] Fortran MEs ( 1 ) : 44.0540s for 16384 events => throughput is 3.72E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G26/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 49.8927s + [COUNTERS] Fortran Overhead ( 0 ) : 5.9647s + [COUNTERS] Fortran MEs ( 1 ) : 43.9280s for 16384 events => throughput is 3.73E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G58/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 51.3816s + [COUNTERS] Fortran Overhead ( 0 ) : 7.2738s + [COUNTERS] Fortran MEs ( 1 ) : 44.1077s for 16384 events => throughput is 3.71E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G511/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 46.2548s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1132s + [COUNTERS] Fortran MEs ( 1 ) : 44.1416s for 16384 events => throughput is 3.71E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 46.9017s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8274s + [COUNTERS] Fortran MEs ( 1 ) : 44.0743s for 16384 events => throughput is 3.72E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G32/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 52.0395s + [COUNTERS] Fortran Overhead ( 0 ) : 7.9885s + [COUNTERS] Fortran MEs ( 1 ) : 44.0510s for 16384 events => throughput is 3.72E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G61/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 50.1707s + [COUNTERS] Fortran Overhead ( 0 ) : 6.0855s + [COUNTERS] Fortran MEs ( 1 ) : 44.0852s for 16384 events => throughput is 3.72E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/G390/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 54.0374s + [COUNTERS] Fortran Overhead ( 0 ) : 9.5784s + [COUNTERS] Fortran MEs ( 1 ) : 44.4589s for 16384 events => throughput is 3.69E+02 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgggux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_epemgguc/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_epemgguc/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_epemgguc/ajob1' in 38.5680 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_epemgguc/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_epemgguc/G164/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.4539s + [COUNTERS] Fortran Overhead ( 0 ) : 4.3997s + [COUNTERS] Fortran MEs ( 1 ) : 9.0541s for 16384 events => throughput is 1.81E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_epemgguc/G24/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 12.9177s + [COUNTERS] Fortran Overhead ( 0 ) : 3.9392s + [COUNTERS] Fortran MEs ( 1 ) : 8.9785s for 16384 events => throughput is 1.82E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_epemgguc/G162/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 12.1138s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1404s + [COUNTERS] Fortran MEs ( 1 ) : 8.9734s for 16384 events => throughput is 1.83E+03 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uc_epemgguc/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/ajob1' in 1322.6902 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G336/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 47.2847s + [COUNTERS] Fortran Overhead ( 0 ) : 3.4975s + [COUNTERS] Fortran MEs ( 1 ) : 43.7872s for 16384 events => throughput is 3.74E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G108/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 46.3074s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4724s + [COUNTERS] Fortran MEs ( 1 ) : 43.8350s for 16384 events => throughput is 3.74E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G105/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 46.1340s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3754s + [COUNTERS] Fortran MEs ( 1 ) : 43.7586s for 16384 events => throughput is 3.74E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G304/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 46.8184s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1039s + [COUNTERS] Fortran MEs ( 1 ) : 43.7146s for 16384 events => throughput is 3.75E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G175/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 46.5982s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2074s + [COUNTERS] Fortran MEs ( 1 ) : 44.3908s for 16384 events => throughput is 3.69E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G128/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 47.2639s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5317s + [COUNTERS] Fortran MEs ( 1 ) : 44.7322s for 16384 events => throughput is 3.66E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G360/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 46.0983s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4743s + [COUNTERS] Fortran MEs ( 1 ) : 43.6240s for 16384 events => throughput is 3.76E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G120/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 45.9608s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1851s + [COUNTERS] Fortran MEs ( 1 ) : 43.7757s for 16384 events => throughput is 3.74E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G296/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 45.7038s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0461s + [COUNTERS] Fortran MEs ( 1 ) : 43.6578s for 16384 events => throughput is 3.75E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G299/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 46.4736s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7634s + [COUNTERS] Fortran MEs ( 1 ) : 43.7102s for 16384 events => throughput is 3.75E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G715/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 46.4135s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6012s + [COUNTERS] Fortran MEs ( 1 ) : 43.8122s for 16384 events => throughput is 3.74E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G300/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 46.9946s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3315s + [COUNTERS] Fortran MEs ( 1 ) : 44.6631s for 16384 events => throughput is 3.67E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G303/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 46.5049s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7926s + [COUNTERS] Fortran MEs ( 1 ) : 43.7123s for 16384 events => throughput is 3.75E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G127/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 46.6877s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8020s + [COUNTERS] Fortran MEs ( 1 ) : 43.8857s for 16384 events => throughput is 3.73E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G111/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 46.3014s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4949s + [COUNTERS] Fortran MEs ( 1 ) : 43.8065s for 16384 events => throughput is 3.74E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G126/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 54.1562s + [COUNTERS] Fortran Overhead ( 0 ) : 10.4648s + [COUNTERS] Fortran MEs ( 1 ) : 43.6913s for 16384 events => throughput is 3.75E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G294/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 50.2530s + [COUNTERS] Fortran Overhead ( 0 ) : 6.5307s + [COUNTERS] Fortran MEs ( 1 ) : 43.7223s for 16384 events => throughput is 3.75E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G598/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 45.9616s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3739s + [COUNTERS] Fortran MEs ( 1 ) : 43.5877s for 16384 events => throughput is 3.76E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G610/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 46.9974s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2848s + [COUNTERS] Fortran MEs ( 1 ) : 43.7126s for 16384 events => throughput is 3.75E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G178/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 46.4159s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5381s + [COUNTERS] Fortran MEs ( 1 ) : 43.8778s for 16384 events => throughput is 3.73E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G121/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 46.6022s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5640s + [COUNTERS] Fortran MEs ( 1 ) : 44.0383s for 16384 events => throughput is 3.72E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G685/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 45.9394s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1340s + [COUNTERS] Fortran MEs ( 1 ) : 43.8054s for 16384 events => throughput is 3.74E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G690/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 45.7090s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0314s + [COUNTERS] Fortran MEs ( 1 ) : 43.6776s for 16384 events => throughput is 3.75E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G103/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 51.9680s + [COUNTERS] Fortran Overhead ( 0 ) : 8.0449s + [COUNTERS] Fortran MEs ( 1 ) : 43.9231s for 16384 events => throughput is 3.73E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G107/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 46.0008s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2507s + [COUNTERS] Fortran MEs ( 1 ) : 43.7501s for 16384 events => throughput is 3.74E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G180/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 46.1617s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4293s + [COUNTERS] Fortran MEs ( 1 ) : 43.7324s for 16384 events => throughput is 3.75E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G97/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 52.0794s + [COUNTERS] Fortran Overhead ( 0 ) : 8.0889s + [COUNTERS] Fortran MEs ( 1 ) : 43.9905s for 16384 events => throughput is 3.72E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/G179/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 46.3579s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6945s + [COUNTERS] Fortran MEs ( 1 ) : 43.6634s for 16384 events => throughput is 3.75E+02 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gg_epemggddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamggddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamggddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamggddx/ajob1' in 20.3546 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamggddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamggddx/G189/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 20.3283s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4854s + [COUNTERS] Fortran MEs ( 1 ) : 17.8429s for 16384 events => throughput is 9.18E+02 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamggddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dd_taptamggdd/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dd_taptamggdd/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dd_taptamggdd/ajob1' in 49.2851 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dd_taptamggdd/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dd_taptamggdd/G463/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 49.2599s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1359s + [COUNTERS] Fortran MEs ( 1 ) : 47.1239s for 16384 events => throughput is 3.48E+02 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dd_taptamggdd/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/ajob1' in 3637.8094 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G506/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 91.8123s + [COUNTERS] Fortran Overhead ( 0 ) : 4.1900s + [COUNTERS] Fortran MEs ( 1 ) : 87.6223s for 16384 events => throughput is 1.87E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G499/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 89.3329s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7835s + [COUNTERS] Fortran MEs ( 1 ) : 87.5494s for 16384 events => throughput is 1.87E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G46/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 94.8118s + [COUNTERS] Fortran Overhead ( 0 ) : 7.3716s + [COUNTERS] Fortran MEs ( 1 ) : 87.4402s for 16384 events => throughput is 1.87E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G56/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 89.9988s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4793s + [COUNTERS] Fortran MEs ( 1 ) : 87.5194s for 16384 events => throughput is 1.87E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G39/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 89.9348s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3238s + [COUNTERS] Fortran MEs ( 1 ) : 87.6110s for 16384 events => throughput is 1.87E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G517/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 92.3924s + [COUNTERS] Fortran Overhead ( 0 ) : 5.0385s + [COUNTERS] Fortran MEs ( 1 ) : 87.3538s for 16384 events => throughput is 1.88E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G52/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 90.2907s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8355s + [COUNTERS] Fortran MEs ( 1 ) : 87.4553s for 16384 events => throughput is 1.87E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G31/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 94.0027s + [COUNTERS] Fortran Overhead ( 0 ) : 6.6223s + [COUNTERS] Fortran MEs ( 1 ) : 87.3804s for 16384 events => throughput is 1.88E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G324/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 93.8293s + [COUNTERS] Fortran Overhead ( 0 ) : 6.2354s + [COUNTERS] Fortran MEs ( 1 ) : 87.5939s for 16384 events => throughput is 1.87E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G64/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 89.3040s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8215s + [COUNTERS] Fortran MEs ( 1 ) : 87.4825s for 16384 events => throughput is 1.87E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G3/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 91.3417s + [COUNTERS] Fortran Overhead ( 0 ) : 3.8648s + [COUNTERS] Fortran MEs ( 1 ) : 87.4769s for 16384 events => throughput is 1.87E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G33/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 89.3513s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1207s + [COUNTERS] Fortran MEs ( 1 ) : 87.2306s for 16384 events => throughput is 1.88E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G53/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 93.8266s + [COUNTERS] Fortran Overhead ( 0 ) : 6.3838s + [COUNTERS] Fortran MEs ( 1 ) : 87.4428s for 16384 events => throughput is 1.87E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G43/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 92.8092s + [COUNTERS] Fortran Overhead ( 0 ) : 5.3568s + [COUNTERS] Fortran MEs ( 1 ) : 87.4524s for 16384 events => throughput is 1.87E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G49/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 89.0558s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7293s + [COUNTERS] Fortran MEs ( 1 ) : 87.3265s for 16384 events => throughput is 1.88E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G40/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 90.3250s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7698s + [COUNTERS] Fortran MEs ( 1 ) : 87.5552s for 16384 events => throughput is 1.87E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G523/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 89.5763s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1012s + [COUNTERS] Fortran MEs ( 1 ) : 87.4752s for 16384 events => throughput is 1.87E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G518/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 92.7883s + [COUNTERS] Fortran Overhead ( 0 ) : 5.2374s + [COUNTERS] Fortran MEs ( 1 ) : 87.5509s for 16384 events => throughput is 1.87E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G35/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 89.2327s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8623s + [COUNTERS] Fortran MEs ( 1 ) : 87.3704s for 16384 events => throughput is 1.88E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 88.9043s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5728s + [COUNTERS] Fortran MEs ( 1 ) : 87.3315s for 16384 events => throughput is 1.88E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G42/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 92.6538s + [COUNTERS] Fortran Overhead ( 0 ) : 5.3100s + [COUNTERS] Fortran MEs ( 1 ) : 87.3438s for 16384 events => throughput is 1.88E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G524/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 89.3964s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0523s + [COUNTERS] Fortran MEs ( 1 ) : 87.3441s for 16384 events => throughput is 1.88E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G65/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 89.9335s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8205s + [COUNTERS] Fortran MEs ( 1 ) : 88.1130s for 16384 events => throughput is 1.86E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G27/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 93.1121s + [COUNTERS] Fortran Overhead ( 0 ) : 5.6894s + [COUNTERS] Fortran MEs ( 1 ) : 87.4227s for 16384 events => throughput is 1.87E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G44/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 93.5299s + [COUNTERS] Fortran Overhead ( 0 ) : 6.1108s + [COUNTERS] Fortran MEs ( 1 ) : 87.4191s for 16384 events => throughput is 1.87E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 89.0306s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5746s + [COUNTERS] Fortran MEs ( 1 ) : 87.4560s for 16384 events => throughput is 1.87E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G303/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 90.7654s + [COUNTERS] Fortran Overhead ( 0 ) : 3.0981s + [COUNTERS] Fortran MEs ( 1 ) : 87.6673s for 16384 events => throughput is 1.87E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G55/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 90.1044s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6240s + [COUNTERS] Fortran MEs ( 1 ) : 87.4804s for 16384 events => throughput is 1.87E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G34/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 89.2395s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7494s + [COUNTERS] Fortran MEs ( 1 ) : 87.4901s for 16384 events => throughput is 1.87E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G26/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 93.3101s + [COUNTERS] Fortran Overhead ( 0 ) : 5.8496s + [COUNTERS] Fortran MEs ( 1 ) : 87.4605s for 16384 events => throughput is 1.87E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G62/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 93.3231s + [COUNTERS] Fortran Overhead ( 0 ) : 5.8358s + [COUNTERS] Fortran MEs ( 1 ) : 87.4872s for 16384 events => throughput is 1.87E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G37/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 89.4518s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0956s + [COUNTERS] Fortran MEs ( 1 ) : 87.3561s for 16384 events => throughput is 1.88E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G485/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 92.6873s + [COUNTERS] Fortran Overhead ( 0 ) : 5.1617s + [COUNTERS] Fortran MEs ( 1 ) : 87.5256s for 16384 events => throughput is 1.87E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G67/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 89.5666s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0895s + [COUNTERS] Fortran MEs ( 1 ) : 87.4771s for 16384 events => throughput is 1.87E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G512/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 89.2511s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7766s + [COUNTERS] Fortran MEs ( 1 ) : 87.4745s for 16384 events => throughput is 1.87E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G511/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 89.8383s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3837s + [COUNTERS] Fortran MEs ( 1 ) : 87.4546s for 16384 events => throughput is 1.87E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G51/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 89.8636s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4138s + [COUNTERS] Fortran MEs ( 1 ) : 87.4498s for 16384 events => throughput is 1.87E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 89.7468s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3404s + [COUNTERS] Fortran MEs ( 1 ) : 87.4063s for 16384 events => throughput is 1.87E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G63/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 89.6699s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2391s + [COUNTERS] Fortran MEs ( 1 ) : 87.4308s for 16384 events => throughput is 1.87E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/G48/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 89.6359s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2839s + [COUNTERS] Fortran MEs ( 1 ) : 87.3521s for 16384 events => throughput is 1.88E+02 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamgggu/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgggux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgggux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgggux/ajob1' in 645.4991 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgggux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgggux/G33/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 89.6620s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7258s + [COUNTERS] Fortran MEs ( 1 ) : 87.9363s for 16384 events => throughput is 1.86E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgggux/G68/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 91.2514s + [COUNTERS] Fortran Overhead ( 0 ) : 3.0646s + [COUNTERS] Fortran MEs ( 1 ) : 88.1868s for 16384 events => throughput is 1.86E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgggux/G25/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 94.7217s + [COUNTERS] Fortran Overhead ( 0 ) : 6.3213s + [COUNTERS] Fortran MEs ( 1 ) : 88.4004s for 16384 events => throughput is 1.85E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgggux/G494/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 89.5833s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6035s + [COUNTERS] Fortran MEs ( 1 ) : 87.9799s for 16384 events => throughput is 1.86E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgggux/G55/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 91.0043s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8273s + [COUNTERS] Fortran MEs ( 1 ) : 88.1770s for 16384 events => throughput is 1.86E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgggux/G509/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 95.0584s + [COUNTERS] Fortran Overhead ( 0 ) : 6.9541s + [COUNTERS] Fortran MEs ( 1 ) : 88.1043s for 16384 events => throughput is 1.86E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgggux/G29/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 94.0561s + [COUNTERS] Fortran Overhead ( 0 ) : 5.9079s + [COUNTERS] Fortran MEs ( 1 ) : 88.1482s for 16384 events => throughput is 1.86E+02 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_taptamgggux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamgguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamgguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamgguux/ajob1' in 247.8848 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamgguux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamgguux/G260/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 48.4410s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5612s + [COUNTERS] Fortran MEs ( 1 ) : 46.8799s for 16384 events => throughput is 3.49E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamgguux/G447/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 48.5040s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6018s + [COUNTERS] Fortran MEs ( 1 ) : 46.9023s for 16384 events => throughput is 3.49E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamgguux/G229/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 50.8175s + [COUNTERS] Fortran Overhead ( 0 ) : 3.9757s + [COUNTERS] Fortran MEs ( 1 ) : 46.8419s for 16384 events => throughput is 3.50E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamgguux/G135/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 48.8388s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9694s + [COUNTERS] Fortran MEs ( 1 ) : 46.8695s for 16384 events => throughput is 3.50E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamgguux/G230/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 51.1593s + [COUNTERS] Fortran Overhead ( 0 ) : 4.2795s + [COUNTERS] Fortran MEs ( 1 ) : 46.8798s for 16384 events => throughput is 3.49E+02 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_taptamgguux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdddx/ajob1' in 51.0860 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdddx/G715/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 25.5329s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0321s + [COUNTERS] Fortran MEs ( 1 ) : 23.5008s for 16384 events => throughput is 6.97E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdddx/G111/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 25.4841s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9761s + [COUNTERS] Fortran MEs ( 1 ) : 23.5080s for 16384 events => throughput is 6.97E+02 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemusdxsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemusdxsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemusdxsx/ajob1' in 7.4379 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemusdxsx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemusdxsx/G67/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 7.3942s + [COUNTERS] Fortran Overhead ( 0 ) : 6.0946s + [COUNTERS] Fortran MEs ( 1 ) : 1.2996s for 16384 events => throughput is 1.26E+04 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemusdxsx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdssx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdssx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdssx/ajob1' in 65.6237 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdssx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdssx/G31/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 10.8582s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8958s + [COUNTERS] Fortran MEs ( 1 ) : 8.9623s for 16384 events => throughput is 1.83E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdssx/G27/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 10.8329s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9009s + [COUNTERS] Fortran MEs ( 1 ) : 8.9319s for 16384 events => throughput is 1.83E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdssx/G119/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 11.0442s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1274s + [COUNTERS] Fortran MEs ( 1 ) : 8.9169s for 16384 events => throughput is 1.84E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdssx/G284/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 11.2652s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3282s + [COUNTERS] Fortran MEs ( 1 ) : 8.9370s for 16384 events => throughput is 1.83E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdssx/G375/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 10.5802s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5890s + [COUNTERS] Fortran MEs ( 1 ) : 8.9912s for 16384 events => throughput is 1.82E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdssx/G102/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 10.9018s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9921s + [COUNTERS] Fortran MEs ( 1 ) : 8.9096s for 16384 events => throughput is 1.84E+03 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_epemgdssx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_epemgguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_epemgguux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_epemgguux/ajob1' in 107.4136 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_epemgguux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_epemgguux/G105/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 25.1063s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5246s + [COUNTERS] Fortran MEs ( 1 ) : 23.5817s for 16384 events => throughput is 6.95E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_epemgguux/G223/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 25.7827s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1939s + [COUNTERS] Fortran MEs ( 1 ) : 23.5888s for 16384 events => throughput is 6.95E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_epemgguux/G103/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 28.0240s + [COUNTERS] Fortran Overhead ( 0 ) : 4.4298s + [COUNTERS] Fortran MEs ( 1 ) : 23.5942s for 16384 events => throughput is 6.94E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_epemgguux/G5/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 28.3915s + [COUNTERS] Fortran Overhead ( 0 ) : 4.8109s + [COUNTERS] Fortran MEs ( 1 ) : 23.5806s for 16384 events => throughput is 6.95E+02 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_epemgguux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/ajob1' in 2931.0743 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G506/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 49.6445s + [COUNTERS] Fortran Overhead ( 0 ) : 5.3728s + [COUNTERS] Fortran MEs ( 1 ) : 44.2717s for 16384 events => throughput is 3.70E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G499/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 45.7669s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8403s + [COUNTERS] Fortran MEs ( 1 ) : 43.9266s for 16384 events => throughput is 3.73E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G38/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 46.8372s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6916s + [COUNTERS] Fortran MEs ( 1 ) : 44.1456s for 16384 events => throughput is 3.71E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G46/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 51.6661s + [COUNTERS] Fortran Overhead ( 0 ) : 7.6876s + [COUNTERS] Fortran MEs ( 1 ) : 43.9786s for 16384 events => throughput is 3.73E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G59/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 50.0205s + [COUNTERS] Fortran Overhead ( 0 ) : 6.1348s + [COUNTERS] Fortran MEs ( 1 ) : 43.8857s for 16384 events => throughput is 3.73E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G56/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 46.7994s + [COUNTERS] Fortran Overhead ( 0 ) : 2.7425s + [COUNTERS] Fortran MEs ( 1 ) : 44.0569s for 16384 events => throughput is 3.72E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G525/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 46.3588s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3849s + [COUNTERS] Fortran MEs ( 1 ) : 43.9739s for 16384 events => throughput is 3.73E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G39/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 46.2534s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2962s + [COUNTERS] Fortran MEs ( 1 ) : 43.9573s for 16384 events => throughput is 3.73E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G517/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 49.6729s + [COUNTERS] Fortran Overhead ( 0 ) : 5.6008s + [COUNTERS] Fortran MEs ( 1 ) : 44.0720s for 16384 events => throughput is 3.72E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G487/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 48.6697s + [COUNTERS] Fortran Overhead ( 0 ) : 4.7388s + [COUNTERS] Fortran MEs ( 1 ) : 43.9310s for 16384 events => throughput is 3.73E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G52/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 47.0117s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9116s + [COUNTERS] Fortran MEs ( 1 ) : 44.1000s for 16384 events => throughput is 3.72E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G31/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 51.5291s + [COUNTERS] Fortran Overhead ( 0 ) : 7.6236s + [COUNTERS] Fortran MEs ( 1 ) : 43.9055s for 16384 events => throughput is 3.73E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G64/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 46.0791s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9522s + [COUNTERS] Fortran MEs ( 1 ) : 44.1268s for 16384 events => throughput is 3.71E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G3/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 48.7488s + [COUNTERS] Fortran Overhead ( 0 ) : 4.8802s + [COUNTERS] Fortran MEs ( 1 ) : 43.8685s for 16384 events => throughput is 3.73E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G497/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 46.1968s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2579s + [COUNTERS] Fortran MEs ( 1 ) : 43.9389s for 16384 events => throughput is 3.73E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G13/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 45.6608s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6404s + [COUNTERS] Fortran MEs ( 1 ) : 44.0204s for 16384 events => throughput is 3.72E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G66/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 47.1545s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4882s + [COUNTERS] Fortran MEs ( 1 ) : 44.6663s for 16384 events => throughput is 3.67E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G33/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 45.8458s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9642s + [COUNTERS] Fortran MEs ( 1 ) : 43.8816s for 16384 events => throughput is 3.73E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G53/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 51.7747s + [COUNTERS] Fortran Overhead ( 0 ) : 7.9050s + [COUNTERS] Fortran MEs ( 1 ) : 43.8698s for 16384 events => throughput is 3.73E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G43/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 48.9976s + [COUNTERS] Fortran Overhead ( 0 ) : 4.7908s + [COUNTERS] Fortran MEs ( 1 ) : 44.2069s for 16384 events => throughput is 3.71E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G68/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 46.1144s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1203s + [COUNTERS] Fortran MEs ( 1 ) : 43.9941s for 16384 events => throughput is 3.72E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G49/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 45.8974s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8955s + [COUNTERS] Fortran MEs ( 1 ) : 44.0019s for 16384 events => throughput is 3.72E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G302/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 52.2783s + [COUNTERS] Fortran Overhead ( 0 ) : 8.2206s + [COUNTERS] Fortran MEs ( 1 ) : 44.0577s for 16384 events => throughput is 3.72E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G25/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 50.1157s + [COUNTERS] Fortran Overhead ( 0 ) : 6.2115s + [COUNTERS] Fortran MEs ( 1 ) : 43.9042s for 16384 events => throughput is 3.73E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G40/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 46.9489s + [COUNTERS] Fortran Overhead ( 0 ) : 3.0145s + [COUNTERS] Fortran MEs ( 1 ) : 43.9344s for 16384 events => throughput is 3.73E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G523/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 46.0122s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9974s + [COUNTERS] Fortran MEs ( 1 ) : 44.0148s for 16384 events => throughput is 3.72E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G518/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 48.9756s + [COUNTERS] Fortran Overhead ( 0 ) : 5.0628s + [COUNTERS] Fortran MEs ( 1 ) : 43.9128s for 16384 events => throughput is 3.73E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G35/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 46.2436s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1011s + [COUNTERS] Fortran MEs ( 1 ) : 44.1425s for 16384 events => throughput is 3.71E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G16/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 45.5602s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6078s + [COUNTERS] Fortran MEs ( 1 ) : 43.9524s for 16384 events => throughput is 3.73E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G524/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 46.1087s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0412s + [COUNTERS] Fortran MEs ( 1 ) : 44.0675s for 16384 events => throughput is 3.72E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G65/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 46.4590s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0901s + [COUNTERS] Fortran MEs ( 1 ) : 44.3689s for 16384 events => throughput is 3.69E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G27/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 50.2311s + [COUNTERS] Fortran Overhead ( 0 ) : 6.3129s + [COUNTERS] Fortran MEs ( 1 ) : 43.9182s for 16384 events => throughput is 3.73E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G44/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 49.9098s + [COUNTERS] Fortran Overhead ( 0 ) : 5.9419s + [COUNTERS] Fortran MEs ( 1 ) : 43.9678s for 16384 events => throughput is 3.73E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G483/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 48.0826s + [COUNTERS] Fortran Overhead ( 0 ) : 4.2204s + [COUNTERS] Fortran MEs ( 1 ) : 43.8621s for 16384 events => throughput is 3.74E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G482/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 49.4478s + [COUNTERS] Fortran Overhead ( 0 ) : 5.5410s + [COUNTERS] Fortran MEs ( 1 ) : 43.9068s for 16384 events => throughput is 3.73E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G1/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 48.5843s + [COUNTERS] Fortran Overhead ( 0 ) : 4.4957s + [COUNTERS] Fortran MEs ( 1 ) : 44.0886s for 16384 events => throughput is 3.72E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 45.3824s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5184s + [COUNTERS] Fortran MEs ( 1 ) : 43.8639s for 16384 events => throughput is 3.74E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G55/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 46.5275s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4391s + [COUNTERS] Fortran MEs ( 1 ) : 44.0884s for 16384 events => throughput is 3.72E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G298/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 46.8554s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9685s + [COUNTERS] Fortran MEs ( 1 ) : 43.8869s for 16384 events => throughput is 3.73E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G513/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 46.9078s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8385s + [COUNTERS] Fortran MEs ( 1 ) : 44.0693s for 16384 events => throughput is 3.72E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G60/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 51.2306s + [COUNTERS] Fortran Overhead ( 0 ) : 7.2786s + [COUNTERS] Fortran MEs ( 1 ) : 43.9520s for 16384 events => throughput is 3.73E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G34/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 46.0627s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9748s + [COUNTERS] Fortran MEs ( 1 ) : 44.0879s for 16384 events => throughput is 3.72E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G26/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 50.6481s + [COUNTERS] Fortran Overhead ( 0 ) : 6.7849s + [COUNTERS] Fortran MEs ( 1 ) : 43.8632s for 16384 events => throughput is 3.74E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G58/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 49.2848s + [COUNTERS] Fortran Overhead ( 0 ) : 5.3662s + [COUNTERS] Fortran MEs ( 1 ) : 43.9187s for 16384 events => throughput is 3.73E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G54/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 49.9163s + [COUNTERS] Fortran Overhead ( 0 ) : 5.8887s + [COUNTERS] Fortran MEs ( 1 ) : 44.0276s for 16384 events => throughput is 3.72E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G62/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 50.9351s + [COUNTERS] Fortran Overhead ( 0 ) : 6.8550s + [COUNTERS] Fortran MEs ( 1 ) : 44.0801s for 16384 events => throughput is 3.72E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G37/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 45.8140s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8578s + [COUNTERS] Fortran MEs ( 1 ) : 43.9562s for 16384 events => throughput is 3.73E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G493/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 45.9818s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1109s + [COUNTERS] Fortran MEs ( 1 ) : 43.8709s for 16384 events => throughput is 3.73E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G4/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 49.9758s + [COUNTERS] Fortran Overhead ( 0 ) : 6.1154s + [COUNTERS] Fortran MEs ( 1 ) : 43.8603s for 16384 events => throughput is 3.74E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G50/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 46.0931s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1183s + [COUNTERS] Fortran MEs ( 1 ) : 43.9749s for 16384 events => throughput is 3.73E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G485/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 49.5820s + [COUNTERS] Fortran Overhead ( 0 ) : 5.6646s + [COUNTERS] Fortran MEs ( 1 ) : 43.9174s for 16384 events => throughput is 3.73E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G41/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 51.0578s + [COUNTERS] Fortran Overhead ( 0 ) : 6.9688s + [COUNTERS] Fortran MEs ( 1 ) : 44.0889s for 16384 events => throughput is 3.72E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G67/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 45.9832s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0887s + [COUNTERS] Fortran MEs ( 1 ) : 43.8945s for 16384 events => throughput is 3.73E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G512/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 46.0395s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0075s + [COUNTERS] Fortran MEs ( 1 ) : 44.0320s for 16384 events => throughput is 3.72E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G511/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 46.5852s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4811s + [COUNTERS] Fortran MEs ( 1 ) : 44.1042s for 16384 events => throughput is 3.71E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G57/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 50.6715s + [COUNTERS] Fortran Overhead ( 0 ) : 6.7819s + [COUNTERS] Fortran MEs ( 1 ) : 43.8896s for 16384 events => throughput is 3.73E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G36/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 46.7149s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8315s + [COUNTERS] Fortran MEs ( 1 ) : 43.8834s for 16384 events => throughput is 3.73E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G63/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 46.1637s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3698s + [COUNTERS] Fortran MEs ( 1 ) : 43.7939s for 16384 events => throughput is 3.74E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G32/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 51.6371s + [COUNTERS] Fortran Overhead ( 0 ) : 7.5770s + [COUNTERS] Fortran MEs ( 1 ) : 44.0602s for 16384 events => throughput is 3.72E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G29/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 50.0635s + [COUNTERS] Fortran Overhead ( 0 ) : 6.1757s + [COUNTERS] Fortran MEs ( 1 ) : 43.8878s for 16384 events => throughput is 3.73E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/G48/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 46.1284s + [COUNTERS] Fortran Overhead ( 0 ) : 2.2988s + [COUNTERS] Fortran MEs ( 1 ) : 43.8296s for 16384 events => throughput is 3.74E+02 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemgggu/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_taptamggdsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_taptamggdsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_taptamggdsx/ajob1' in 59.8752 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_taptamggdsx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_taptamggdsx/G92/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 20.1589s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3583s + [COUNTERS] Fortran MEs ( 1 ) : 17.8005s for 16384 events => throughput is 9.20E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_taptamggdsx/G144/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 19.8315s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0477s + [COUNTERS] Fortran MEs ( 1 ) : 17.7838s for 16384 events => throughput is 9.21E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_taptamggdsx/G207/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 19.8216s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0119s + [COUNTERS] Fortran MEs ( 1 ) : 17.8097s for 16384 events => throughput is 9.20E+02 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_taptamggdsx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_epemggdsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_epemggdsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_epemggdsx/ajob1' in 15.5588 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_epemggdsx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_epemggdsx/G205/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 15.5333s + [COUNTERS] Fortran Overhead ( 0 ) : 6.6063s + [COUNTERS] Fortran MEs ( 1 ) : 8.9270s for 16384 events => throughput is 1.84E+03 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dsx_epemggdsx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_taptamgggg/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_taptamgggg/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_taptamgggg/ajob1' in 90.3448 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_taptamgggg/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_taptamgggg/G67/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 90.3181s + [COUNTERS] Fortran Overhead ( 0 ) : 2.6772s + [COUNTERS] Fortran MEs ( 1 ) : 87.6409s for 16384 events => throughput is 1.87E+02 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_taptamgggg/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dc_epemdcssx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dc_epemdcssx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dc_epemdcssx/ajob1' in 3.4950 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dc_epemdcssx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dc_epemdcssx/G77/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 3.4483s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1582s + [COUNTERS] Fortran MEs ( 1 ) : 1.2901s for 16384 events => throughput is 1.27E+04 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dc_epemdcssx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguccx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguccx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguccx/ajob1' in 20.2425 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguccx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguccx/G48/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 20.2168s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4187s + [COUNTERS] Fortran MEs ( 1 ) : 17.7981s for 16384 events => throughput is 9.21E+02 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_taptamguccx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ds_epemggds/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ds_epemggds/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ds_epemggds/ajob1' in 10.5861 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ds_epemggds/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ds_epemggds/G26/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 10.5612s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5920s + [COUNTERS] Fortran MEs ( 1 ) : 8.9692s for 16384 events => throughput is 1.83E+03 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ds_epemggds/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ucx_epemggucx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ucx_epemggucx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ucx_epemggucx/ajob1' in 23.5367 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ucx_epemggucx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ucx_epemggucx/G163/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 12.9455s + [COUNTERS] Fortran Overhead ( 0 ) : 3.9831s + [COUNTERS] Fortran MEs ( 1 ) : 8.9623s for 16384 events => throughput is 1.83E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ucx_epemggucx/G60/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 10.5285s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5483s + [COUNTERS] Fortran MEs ( 1 ) : 8.9802s for 16384 events => throughput is 1.82E+03 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ucx_epemggucx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemggudx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemggudx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemggudx/ajob1' in 108.7355 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemggudx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemggudx/G109/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 10.5113s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4739s + [COUNTERS] Fortran MEs ( 1 ) : 9.0375s for 16384 events => throughput is 1.81E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemggudx/G151/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 10.7884s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8130s + [COUNTERS] Fortran MEs ( 1 ) : 8.9754s for 16384 events => throughput is 1.83E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemggudx/G106/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 12.1764s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1642s + [COUNTERS] Fortran MEs ( 1 ) : 9.0122s for 16384 events => throughput is 1.82E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemggudx/G35/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 15.7447s + [COUNTERS] Fortran Overhead ( 0 ) : 6.7619s + [COUNTERS] Fortran MEs ( 1 ) : 8.9828s for 16384 events => throughput is 1.82E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemggudx/G119/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 10.9799s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9688s + [COUNTERS] Fortran MEs ( 1 ) : 9.0110s for 16384 events => throughput is 1.82E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemggudx/G187/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 10.7886s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8154s + [COUNTERS] Fortran MEs ( 1 ) : 8.9732s for 16384 events => throughput is 1.83E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemggudx/G91/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 10.9730s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9802s + [COUNTERS] Fortran MEs ( 1 ) : 8.9927s for 16384 events => throughput is 1.82E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemggudx/G363/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.2699s + [COUNTERS] Fortran Overhead ( 0 ) : 4.2848s + [COUNTERS] Fortran MEs ( 1 ) : 8.9851s for 16384 events => throughput is 1.82E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemggudx/G149/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.3275s + [COUNTERS] Fortran Overhead ( 0 ) : 4.3639s + [COUNTERS] Fortran MEs ( 1 ) : 8.9636s for 16384 events => throughput is 1.83E+03 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_epemggudx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgcuxcx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgcuxcx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgcuxcx/ajob1' in 11.4948 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgcuxcx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgcuxcx/G120/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 11.4515s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4249s + [COUNTERS] Fortran MEs ( 1 ) : 9.0267s for 16384 events => throughput is 1.82E+03 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgcuxcx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemgggg/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemgggg/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemgggg/ajob1' in 94.3665 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemgggg/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemgggg/G65/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 47.0622s + [COUNTERS] Fortran Overhead ( 0 ) : 3.1745s + [COUNTERS] Fortran MEs ( 1 ) : 43.8877s for 16384 events => throughput is 3.73E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemgggg/G48/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 47.2374s + [COUNTERS] Fortran Overhead ( 0 ) : 3.3939s + [COUNTERS] Fortran MEs ( 1 ) : 43.8435s for 16384 events => throughput is 3.74E+02 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ddx_epemgggg/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_taptamggud/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_taptamggud/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_taptamggud/ajob1' in 182.1103 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_taptamggud/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_taptamggud/G46/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 20.1423s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1638s + [COUNTERS] Fortran MEs ( 1 ) : 17.9785s for 16384 events => throughput is 9.11E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_taptamggud/G39/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 22.0599s + [COUNTERS] Fortran Overhead ( 0 ) : 4.0919s + [COUNTERS] Fortran MEs ( 1 ) : 17.9680s for 16384 events => throughput is 9.12E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_taptamggud/G112/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 19.1302s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2397s + [COUNTERS] Fortran MEs ( 1 ) : 17.8905s for 16384 events => throughput is 9.16E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_taptamggud/G115/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 19.3844s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4477s + [COUNTERS] Fortran MEs ( 1 ) : 17.9367s for 16384 events => throughput is 9.13E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_taptamggud/G270/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 20.7945s + [COUNTERS] Fortran Overhead ( 0 ) : 2.8545s + [COUNTERS] Fortran MEs ( 1 ) : 17.9400s for 16384 events => throughput is 9.13E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_taptamggud/G96/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 19.5452s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5900s + [COUNTERS] Fortran MEs ( 1 ) : 17.9552s for 16384 events => throughput is 9.12E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_taptamggud/G107/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 21.8128s + [COUNTERS] Fortran Overhead ( 0 ) : 3.8035s + [COUNTERS] Fortran MEs ( 1 ) : 18.0093s for 16384 events => throughput is 9.10E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_taptamggud/G63/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 19.7998s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8378s + [COUNTERS] Fortran MEs ( 1 ) : 17.9620s for 16384 events => throughput is 9.12E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_taptamggud/G375/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 19.2435s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2736s + [COUNTERS] Fortran MEs ( 1 ) : 17.9699s for 16384 events => throughput is 9.12E+02 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_taptamggud/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguuux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguuux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguuux/ajob1' in 109.3048 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguuux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguuux/G122/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 25.2710s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8556s + [COUNTERS] Fortran MEs ( 1 ) : 23.4154s for 16384 events => throughput is 7.00E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguuux/G33/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 29.0973s + [COUNTERS] Fortran Overhead ( 0 ) : 5.6491s + [COUNTERS] Fortran MEs ( 1 ) : 23.4481s for 16384 events => throughput is 6.99E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguuux/G127/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 26.4202s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9921s + [COUNTERS] Fortran MEs ( 1 ) : 23.4281s for 16384 events => throughput is 6.99E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguuux/G116/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 28.4343s + [COUNTERS] Fortran Overhead ( 0 ) : 4.9955s + [COUNTERS] Fortran MEs ( 1 ) : 23.4388s for 16384 events => throughput is 6.99E+02 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gu_epemguuux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamgggdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamgggdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamgggdx/ajob1' in 810.5104 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamgggdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamgggdx/G56/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 90.4514s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5621s + [COUNTERS] Fortran MEs ( 1 ) : 87.8893s for 16384 events => throughput is 1.86E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamgggdx/G64/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 90.2399s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3811s + [COUNTERS] Fortran MEs ( 1 ) : 87.8588s for 16384 events => throughput is 1.86E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamgggdx/G13/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 89.5192s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6217s + [COUNTERS] Fortran MEs ( 1 ) : 87.8975s for 16384 events => throughput is 1.86E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamgggdx/G33/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 89.6914s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7906s + [COUNTERS] Fortran MEs ( 1 ) : 87.9008s for 16384 events => throughput is 1.86E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamgggdx/G49/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 90.2634s + [COUNTERS] Fortran Overhead ( 0 ) : 2.1015s + [COUNTERS] Fortran MEs ( 1 ) : 88.1618s for 16384 events => throughput is 1.86E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamgggdx/G523/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 90.4073s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5464s + [COUNTERS] Fortran MEs ( 1 ) : 87.8609s for 16384 events => throughput is 1.86E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamgggdx/G65/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 90.0205s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9411s + [COUNTERS] Fortran MEs ( 1 ) : 88.0794s for 16384 events => throughput is 1.86E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamgggdx/G34/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 89.8112s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0540s + [COUNTERS] Fortran MEs ( 1 ) : 87.7573s for 16384 events => throughput is 1.87E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamgggdx/G48/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 89.9083s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9683s + [COUNTERS] Fortran MEs ( 1 ) : 87.9400s for 16384 events => throughput is 1.86E+02 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_taptamgggdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgudux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgudux/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgudux/ajob1' in 43.9475 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgudux/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgudux/G59/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 20.2383s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8555s + [COUNTERS] Fortran MEs ( 1 ) : 18.3827s for 16384 events => throughput is 8.91E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgudux/G61/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 23.6489s + [COUNTERS] Fortran Overhead ( 0 ) : 5.5811s + [COUNTERS] Fortran MEs ( 1 ) : 18.0678s for 16384 events => throughput is 9.07E+02 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgudux/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgduxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgduxdx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgduxdx/ajob1' in 10.8690 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgduxdx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgduxdx/G104/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 10.8230s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8820s + [COUNTERS] Fortran MEs ( 1 ) : 8.9411s for 16384 events => throughput is 1.83E+03 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gux_epemgduxdx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgsdxsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgsdxsx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgsdxsx/ajob1' in 10.6731 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgsdxsx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgsdxsx/G25/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 10.6458s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7163s + [COUNTERS] Fortran MEs ( 1 ) : 8.9295s for 16384 events => throughput is 1.83E+03 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gdx_epemgsdxsx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/ajob1' in 331.4404 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G208/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 10.8347s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8920s + [COUNTERS] Fortran MEs ( 1 ) : 8.9427s for 16384 events => throughput is 1.83E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G235/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 10.6802s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6790s + [COUNTERS] Fortran MEs ( 1 ) : 9.0012s for 16384 events => throughput is 1.82E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G140/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 10.9400s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9795s + [COUNTERS] Fortran MEs ( 1 ) : 8.9605s for 16384 events => throughput is 1.83E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G109/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 10.6183s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6655s + [COUNTERS] Fortran MEs ( 1 ) : 8.9529s for 16384 events => throughput is 1.83E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G33/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.5154s + [COUNTERS] Fortran Overhead ( 0 ) : 4.5968s + [COUNTERS] Fortran MEs ( 1 ) : 8.9186s for 16384 events => throughput is 1.84E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G43/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 10.7068s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7286s + [COUNTERS] Fortran MEs ( 1 ) : 8.9782s for 16384 events => throughput is 1.82E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G240/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 10.7115s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7488s + [COUNTERS] Fortran MEs ( 1 ) : 8.9627s for 16384 events => throughput is 1.83E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G268/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.2986s + [COUNTERS] Fortran Overhead ( 0 ) : 4.3247s + [COUNTERS] Fortran MEs ( 1 ) : 8.9739s for 16384 events => throughput is 1.83E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G139/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 10.9816s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9614s + [COUNTERS] Fortran MEs ( 1 ) : 9.0201s for 16384 events => throughput is 1.82E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G120/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 10.7020s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7335s + [COUNTERS] Fortran MEs ( 1 ) : 8.9684s for 16384 events => throughput is 1.83E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G236/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 10.7282s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7742s + [COUNTERS] Fortran MEs ( 1 ) : 8.9540s for 16384 events => throughput is 1.83E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G181/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.1559s + [COUNTERS] Fortran Overhead ( 0 ) : 5.2175s + [COUNTERS] Fortran MEs ( 1 ) : 8.9384s for 16384 events => throughput is 1.83E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G351/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 10.4157s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3554s + [COUNTERS] Fortran MEs ( 1 ) : 9.0603s for 16384 events => throughput is 1.81E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G65/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 12.7918s + [COUNTERS] Fortran Overhead ( 0 ) : 3.8246s + [COUNTERS] Fortran MEs ( 1 ) : 8.9671s for 16384 events => throughput is 1.83E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G71/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 10.4958s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5328s + [COUNTERS] Fortran MEs ( 1 ) : 8.9630s for 16384 events => throughput is 1.83E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G119/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 10.8493s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8827s + [COUNTERS] Fortran MEs ( 1 ) : 8.9666s for 16384 events => throughput is 1.83E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G127/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 10.5217s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5583s + [COUNTERS] Fortran MEs ( 1 ) : 8.9634s for 16384 events => throughput is 1.83E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G96/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 10.6108s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6612s + [COUNTERS] Fortran MEs ( 1 ) : 8.9496s for 16384 events => throughput is 1.83E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G60/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 10.1764s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2193s + [COUNTERS] Fortran MEs ( 1 ) : 8.9571s for 16384 events => throughput is 1.83E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G26/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 10.4340s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4539s + [COUNTERS] Fortran MEs ( 1 ) : 8.9800s for 16384 events => throughput is 1.82E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G37/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.6328s + [COUNTERS] Fortran Overhead ( 0 ) : 5.6603s + [COUNTERS] Fortran MEs ( 1 ) : 8.9725s for 16384 events => throughput is 1.83E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G125/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 12.2639s + [COUNTERS] Fortran Overhead ( 0 ) : 3.2810s + [COUNTERS] Fortran MEs ( 1 ) : 8.9829s for 16384 events => throughput is 1.82E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G91/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 10.4490s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4872s + [COUNTERS] Fortran MEs ( 1 ) : 8.9619s for 16384 events => throughput is 1.83E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G143/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 10.3002s + [COUNTERS] Fortran Overhead ( 0 ) : 1.3500s + [COUNTERS] Fortran MEs ( 1 ) : 8.9502s for 16384 events => throughput is 1.83E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G116/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 10.2472s + [COUNTERS] Fortran Overhead ( 0 ) : 1.2783s + [COUNTERS] Fortran MEs ( 1 ) : 8.9688s for 16384 events => throughput is 1.83E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G61/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.4536s + [COUNTERS] Fortran Overhead ( 0 ) : 5.4934s + [COUNTERS] Fortran MEs ( 1 ) : 8.9602s for 16384 events => throughput is 1.83E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G149/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 13.1685s + [COUNTERS] Fortran Overhead ( 0 ) : 4.2284s + [COUNTERS] Fortran MEs ( 1 ) : 8.9401s for 16384 events => throughput is 1.83E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G144/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 10.4670s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4445s + [COUNTERS] Fortran MEs ( 1 ) : 9.0226s for 16384 events => throughput is 1.82E+03 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/G207/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 10.7202s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7031s + [COUNTERS] Fortran MEs ( 1 ) : 9.0171s for 16384 events => throughput is 1.82E+03 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemggud/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_epemgguu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_epemgguu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_epemgguu/ajob1' in 288.8541 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_epemgguu/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_epemgguu/G463/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 25.5418s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9598s + [COUNTERS] Fortran MEs ( 1 ) : 23.5820s for 16384 events => throughput is 6.95E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_epemgguu/G100/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 26.5778s + [COUNTERS] Fortran Overhead ( 0 ) : 2.9947s + [COUNTERS] Fortran MEs ( 1 ) : 23.5831s for 16384 events => throughput is 6.95E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_epemgguu/G110/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 24.9645s + [COUNTERS] Fortran Overhead ( 0 ) : 1.4239s + [COUNTERS] Fortran MEs ( 1 ) : 23.5406s for 16384 events => throughput is 6.96E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_epemgguu/G124/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 25.1798s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5053s + [COUNTERS] Fortran MEs ( 1 ) : 23.6744s for 16384 events => throughput is 6.92E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_epemgguu/G367/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 25.4391s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8794s + [COUNTERS] Fortran MEs ( 1 ) : 23.5598s for 16384 events => throughput is 6.95E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_epemgguu/G126/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 27.5174s + [COUNTERS] Fortran Overhead ( 0 ) : 3.9060s + [COUNTERS] Fortran MEs ( 1 ) : 23.6114s for 16384 events => throughput is 6.94E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_epemgguu/G361/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 27.8684s + [COUNTERS] Fortran Overhead ( 0 ) : 4.2613s + [COUNTERS] Fortran MEs ( 1 ) : 23.6071s for 16384 events => throughput is 6.94E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_epemgguu/G116/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 26.1262s + [COUNTERS] Fortran Overhead ( 0 ) : 2.5575s + [COUNTERS] Fortran MEs ( 1 ) : 23.5688s for 16384 events => throughput is 6.95E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_epemgguu/G363/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 27.5657s + [COUNTERS] Fortran Overhead ( 0 ) : 3.9406s + [COUNTERS] Fortran MEs ( 1 ) : 23.6251s for 16384 events => throughput is 6.93E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_epemgguu/G138/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 25.2244s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5423s + [COUNTERS] Fortran MEs ( 1 ) : 23.6821s for 16384 events => throughput is 6.92E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_epemgguu/G102/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 26.6376s + [COUNTERS] Fortran Overhead ( 0 ) : 3.0074s + [COUNTERS] Fortran MEs ( 1 ) : 23.6302s for 16384 events => throughput is 6.93E+02 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_epemgguu/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgdssx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgdssx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgdssx/ajob1' in 19.8626 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgdssx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgdssx/G28/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 19.8153s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0166s + [COUNTERS] Fortran MEs ( 1 ) : 17.7987s for 16384 events => throughput is 9.21E+02 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_gd_taptamgdssx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_taptamgguu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_taptamgguu/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_taptamgguu/ajob1' in 245.5883 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_taptamgguu/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_taptamgguu/G464/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 48.5882s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8821s + [COUNTERS] Fortran MEs ( 1 ) : 46.7061s for 16384 events => throughput is 3.51E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_taptamgguu/G59/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 49.1324s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0560s + [COUNTERS] Fortran MEs ( 1 ) : 47.0764s for 16384 events => throughput is 3.48E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_taptamgguu/G100/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 49.8882s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4939s + [COUNTERS] Fortran MEs ( 1 ) : 47.3943s for 16384 events => throughput is 3.46E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_taptamgguu/G106/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 48.5111s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7546s + [COUNTERS] Fortran MEs ( 1 ) : 46.7565s for 16384 events => throughput is 3.50E+02 events/s +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_taptamgguu/G143/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 49.3490s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7555s + [COUNTERS] Fortran MEs ( 1 ) : 47.5935s for 16384 events => throughput is 3.44E+02 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_taptamgguu/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dc_epemggdc/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dc_epemggdc/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dc_epemggdc/ajob1' in 14.6389 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dc_epemggdc/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dc_epemggdc/G18/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 14.5966s + [COUNTERS] Fortran Overhead ( 0 ) : 5.2461s + [COUNTERS] Fortran MEs ( 1 ) : 9.3505s for 16384 events => throughput is 1.75E+03 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_dc_epemggdc/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ucx_taptamggucx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ucx_taptamggucx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ucx_taptamggucx/ajob1' in 23.3710 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ucx_taptamggucx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ucx_taptamggucx/G62/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 23.3268s + [COUNTERS] Fortran Overhead ( 0 ) : 4.8127s + [COUNTERS] Fortran MEs ( 1 ) : 18.5141s for 16384 events => throughput is 8.85E+02 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ucx_taptamggucx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamucdxcx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamucdxcx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamucdxcx/ajob1' in 4.3808 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamucdxcx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamucdxcx/G47/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.3550s + [COUNTERS] Fortran Overhead ( 0 ) : 1.7830s + [COUNTERS] Fortran MEs ( 1 ) : 2.5720s for 16384 events => throughput is 6.37E+03 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamucdxcx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_epemgggg/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_epemgggg/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_epemgggg/ajob1' in 47.3211 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_epemgggg/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_epemgggg/G49/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 47.2717s + [COUNTERS] Fortran Overhead ( 0 ) : 3.3542s + [COUNTERS] Fortran MEs ( 1 ) : 43.9175s for 16384 events => throughput is 3.73E+02 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uux_epemgggg/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemudddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemudddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemudddx/ajob1' in 4.7721 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemudddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemudddx/G15/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 4.7234s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5612s + [COUNTERS] Fortran MEs ( 1 ) : 3.1622s for 16384 events => throughput is 5.18E+03 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_ud_epemudddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamggudx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamggudx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamggudx/ajob1' in 19.5487 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamggudx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamggudx/G60/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 19.5240s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6141s + [COUNTERS] Fortran MEs ( 1 ) : 17.9099s for 16384 events => throughput is 9.15E+02 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_udx_taptamggudx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_taptamuuddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_taptamuuddx/ajob1' +__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_taptamuuddx/ajob1' in 8.0583 seconds +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_taptamuuddx/ajob1' +/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_taptamuuddx/G11/GridRun_21_app.log + [COUNTERS] PROGRAM TOTAL : 8.0172s + [COUNTERS] Fortran Overhead ( 0 ) : 1.6427s + [COUNTERS] Fortran MEs ( 1 ) : 6.3746s for 16384 events => throughput is 2.57E+03 events/s +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/madevent/SubProcesses/P0_uu_taptamuuddx/ajob1' +__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished in 21556.7202 seconds +__CUDACPP_DEBUG: gen_ximprove.launch finished in 21560.7551 seconds +__CUDACPP_DEBUG: GridPackCmd.refine4grid finished in 21560.7559 seconds +__CUDACPP_DEBUG: GridPackCmd.launch (3a) combine_events +combine_events +INFO: fail to reach target 500 +__CUDACPP_DEBUG: GridPackCmd.launch (3b) store_events +store_events + === Results Summary for run: GridRun_21 tag: tag_1 === + + Cross-section : 0 +- 0 pb + Nb of events : 81 + +INFO: No version of lhapdf. Can not run systematics computation +decay_events -from_cards +__CUDACPP_DEBUG: GridPackCmd.launch finished in 21707.6095 seconds +__CUDACPP_DEBUG: GridPackCmd.__init__ back from self.launch +DONE +write ./events.lhe.gz +END: Sat Sep 14 11:21:32 PM CEST 2024 +ELAPSED: 21710 seconds diff --git a/epochX/cudacpp/tlau/fromgridpacks/pp_dy4j.mad/summary.txt b/epochX/cudacpp/tlau/fromgridpacks/pp_dy4j.mad/summary.txt new file mode 100644 index 0000000000..819250ce38 --- /dev/null +++ b/epochX/cudacpp/tlau/fromgridpacks/pp_dy4j.mad/summary.txt @@ -0,0 +1,51 @@ +pp_dy4j.mad/fortran/output.txt (#events: 81) +[GridPackCmd.launch] OVERALL TOTAL 21707.6095 seconds +[madevent COUNTERS] PROGRAM TOTAL 21546.1 +[madevent COUNTERS] Fortran Overhead 1579.09 +[madevent COUNTERS] Fortran MEs 19967 +-------------------------------------------------------------------------------- +pp_dy4j.mad/cppnone/output.txt (#events: 195) +[GridPackCmd.launch] OVERALL TOTAL 26745.1639 seconds +[madevent COUNTERS] PROGRAM TOTAL 26584.9 +[madevent COUNTERS] Fortran Overhead 1608.51 +[madevent COUNTERS] CudaCpp MEs 24910.4 +[madevent COUNTERS] CudaCpp HEL 66.0341 +-------------------------------------------------------------------------------- +pp_dy4j.mad/cppsse4/output.txt (#events: 195) +[GridPackCmd.launch] OVERALL TOTAL 14398.4664 seconds +[madevent COUNTERS] PROGRAM TOTAL 14231.3 +[madevent COUNTERS] Fortran Overhead 1647.03 +[madevent COUNTERS] CudaCpp MEs 12550.6 +[madevent COUNTERS] CudaCpp HEL 33.7035 +-------------------------------------------------------------------------------- +pp_dy4j.mad/cppavx2/output.txt (#events: 195) +[GridPackCmd.launch] OVERALL TOTAL 7335.2356 seconds +[madevent COUNTERS] PROGRAM TOTAL 7114.43 +[madevent COUNTERS] Fortran Overhead 1683.7 +[madevent COUNTERS] CudaCpp MEs 5415.48 +[madevent COUNTERS] CudaCpp HEL 15.2596 +-------------------------------------------------------------------------------- +pp_dy4j.mad/cpp512y/output.txt (#events: 195) +[GridPackCmd.launch] OVERALL TOTAL 6831.8971 seconds +[madevent COUNTERS] PROGRAM TOTAL 6649.98 +[madevent COUNTERS] Fortran Overhead 1669.94 +[madevent COUNTERS] CudaCpp MEs 4966.24 +[madevent COUNTERS] CudaCpp HEL 13.8066 +-------------------------------------------------------------------------------- +pp_dy4j.mad/cpp512z/output.txt (#events: 195) +[GridPackCmd.launch] OVERALL TOTAL 7136.2962 seconds +[madevent COUNTERS] PROGRAM TOTAL 6958.96 +[madevent COUNTERS] Fortran Overhead 1636.28 +[madevent COUNTERS] CudaCpp MEs 5305.14 +[madevent COUNTERS] CudaCpp HEL 17.5447 +-------------------------------------------------------------------------------- +pp_dy4j.mad/cuda/output.txt (#events: 195) +[GridPackCmd.launch] OVERALL TOTAL 2523.7488 seconds +[madevent COUNTERS] PROGRAM TOTAL 2234.93 +[madevent COUNTERS] Fortran Overhead 1820.36 +[madevent COUNTERS] CudaCpp MEs 97.9622 +[madevent COUNTERS] CudaCpp HEL 316.613 +-------------------------------------------------------------------------------- +pp_dy4j.mad/hip/output.txt +File not found: SKIP backend hip +-------------------------------------------------------------------------------- diff --git a/epochX/cudacpp/tlau/gridpacks/.gitignore b/epochX/cudacpp/tlau/gridpacks/.gitignore new file mode 100644 index 0000000000..30a22396e5 --- /dev/null +++ b/epochX/cudacpp/tlau/gridpacks/.gitignore @@ -0,0 +1,2 @@ +madevent +run.sh diff --git a/epochX/cudacpp/tlau/gridpacks/MG5aMC_patches/cluster.py b/epochX/cudacpp/tlau/gridpacks/MG5aMC_patches/cluster.py new file mode 100755 index 0000000000..e62fe24559 --- /dev/null +++ b/epochX/cudacpp/tlau/gridpacks/MG5aMC_patches/cluster.py @@ -0,0 +1,2221 @@ +################################################################################ +# Copyright (c) 2009 The MadGraph5_aMC@NLO Development team and Contributors +# +# This file is a part of the MadGraph5_aMC@NLO project, an application which +# automatically generates Feynman diagrams and matrix elements for arbitrary +# high-energy processes in the Standard Model and beyond. +# +# It is subject to the MadGraph5_aMC@NLO license which should accompany this +# distribution. +# +# For more information, visit madgraph.phys.ucl.ac.be and amcatnlo.web.cern.ch +# +################################################################################ +from __future__ import absolute_import +import subprocess +import logging +import os +import time +import re +import glob +import inspect +import sys +import six +from six.moves import range +from six.moves import input + +logger = logging.getLogger('madgraph.cluster') + +try: + from madgraph import MadGraph5Error, MG5DIR + import madgraph.various.misc as misc + MADEVENT=False +except Exception as error: + if __debug__: + print(str(error)) + from internal import MadGraph5Error + import internal.misc as misc + MADEVENT=True + LOCALDIR = os.path.realpath(os.path.join(os.path.dirname(__file__), os.path.pardir, + os.path.pardir)) + + + +pjoin = os.path.join + +class ClusterManagmentError(MadGraph5Error): + pass + +class NotImplemented(MadGraph5Error): + pass + + +multiple_try = misc.multiple_try +pjoin = os.path.join + + +def check_interupt(error=KeyboardInterrupt): + + def deco_interupt(f): + def deco_f_interupt(self, *args, **opt): + try: + return f(self, *args, **opt) + except error: + try: + self.remove(*args, **opt) + except Exception: + pass + raise error + return deco_f_interupt + return deco_interupt + +def store_input(arg=''): + + def deco_store(f): + def deco_f_store(self, prog, argument=[], cwd=None, stdout=None, stderr=None, log=None, + input_files=[], output_files=[], required_output=[], nb_submit=0): + frame = inspect.currentframe() + args, _, _, values = inspect.getargvalues(frame) + args = dict([(i, values[i]) for i in args if i != 'self']) + id = f(self, **args) + if self.nb_retry > 0: + self.retry_args[id] = args + return id + return deco_f_store + return deco_store + +def need_transfer(options): + """ This function checks whether compression of input files are necessary + given the running options given. """ + + if options['run_mode'] != 1 and options['cluster_temp_path'] is None: + return False + else: + return True + +class Cluster(object): + """Basic Class for all cluster type submission""" + name = 'mother class' + identifier_length = 14 + + def __init__(self,*args, **opts): + """Init the cluster""" + + self.submitted = 0 + self.submitted_ids = [] + self.finish = 0 + self.submitted_dirs = [] #HTCaaS + self.submitted_exes = [] #HTCaaS + self.submitted_args = [] #HTCaaS + + if 'cluster_queue' in opts: + self.cluster_queue = opts['cluster_queue'] + else: + self.cluster_queue = 'madgraph' + if 'cluster_temp_path' in opts: + self.temp_dir = opts['cluster_temp_path'] + else: + self.temp_dir = None + self.options = {'cluster_status_update': (600, 30)} + for key,value in opts.items(): + self.options[key] = value + self.nb_retry = opts['cluster_nb_retry'] if 'cluster_nb_retry' in opts else 0 + self.cluster_retry_wait = float(opts['cluster_retry_wait']) if 'cluster_retry_wait' in opts else 300 + self.options = dict(opts) + self.retry_args = {} + # controlling jobs in controlled type submision + self.packet = {} + self.id_to_packet = {} + + def submit(self, prog, argument=[], cwd=None, stdout=None, stderr=None, + log=None, required_output=[], nb_submit=0): + """How to make one submission. Return status id on the cluster.""" + raise NotImplemented('No implementation of how to submit a job to cluster \'%s\'' % self.name) + + + @store_input() + def submit2(self, prog, argument=[], cwd=None, stdout=None, stderr=None, + log=None, input_files=[], output_files=[], required_output=[], + nb_submit=0): + """How to make one submission. Return status id on the cluster. + NO SHARE DISK""" + + if cwd is None: + cwd = os.getcwd() + if not os.path.exists(prog): + prog = os.path.join(cwd, prog) + + if not required_output and output_files: + required_output = output_files + + if not hasattr(self, 'temp_dir') or not self.temp_dir or \ + (input_files == [] == output_files): + + return self.submit(prog, argument, cwd, stdout, stderr, log, + required_output=required_output, nb_submit=nb_submit) + + if not input_files and not output_files: + # not input/output so not using submit2 + return self.submit(prog, argument, cwd, stdout, stderr, log, + required_output=required_output, nb_submit=nb_submit) + + if cwd is None: + cwd = os.getcwd() + if not os.path.exists(prog): + prog = os.path.join(cwd, prog) + temp_file_name = "sub." + os.path.basename(prog) + '.'.join(argument) + + text = """#!/bin/bash + MYTMP=%(tmpdir)s/run$%(job_id)s + MYPWD=%(cwd)s + mkdir -p $MYTMP + cd $MYPWD + input_files=( %(input_files)s ) + for i in ${input_files[@]} + do + cp -R -L $i $MYTMP + done + cd $MYTMP + echo '%(arguments)s' > arguments + chmod +x ./%(script)s + %(program)s ./%(script)s %(arguments)s + exit=$? + output_files=( %(output_files)s ) + for i in ${output_files[@]} + do + cp -r $MYTMP/$i $MYPWD + done +# if [ "$exit" -eq "0" ] +# then + rm -rf $MYTMP +# fi + """ + + dico = {'tmpdir' : self.temp_dir, 'script': os.path.basename(prog), + 'cwd': cwd, 'job_id': self.job_id, + 'input_files': ' '.join(input_files + [prog]), + 'output_files': ' '.join(output_files), + 'arguments': ' '.join([str(a) for a in argument]), + 'program': sys.executable if '.py' in prog else 'bash'} + + temp_file_name = temp_file_name.replace("/","_") + # writing a new script for the submission + new_prog = pjoin(cwd, temp_file_name) + open(new_prog, 'w').write(text % dico) + misc.Popen(['chmod','+x',new_prog],cwd=cwd) + + return self.submit(new_prog, argument, cwd, stdout, stderr, log, + required_output=required_output, nb_submit=nb_submit) + + + def cluster_submit(self, prog, argument=[], cwd=None, stdout=None, stderr=None, + log=None, input_files=[], output_files=[], required_output=[], + nb_submit=0, packet_member=None): + """This function wrap the cluster submition with cluster independant + method should not be overwritten (but for DAG type submission)""" + + id = self.submit2(prog, argument, cwd, stdout, stderr, log, input_files, + output_files, required_output, nb_submit) + + + if not packet_member: + return id + else: + if isinstance(packet_member, Packet): + self.id_to_packet[id] = packet_member + packet_member.put(id) + if packet_member.tag not in self.packet: + self.packet[packet_member.tag] = packet_member + else: + if packet_member in self.packet: + packet = self.packet[packet_member] + packet.put(id) + self.id_to_packet[id] = packet + return id + + def control(self, me_dir=None): + """Check the status of job associated to directory me_dir. return (idle, run, finish, fail)""" + if not self.submitted_ids: + raise NotImplemented('No implementation of how to control the job status to cluster \'%s\'' % self.name) + idle, run, fail = 0, 0, 0 + for pid in self.submitted_ids[:]: + status = self.control_one_job(id) + if status == 'I': + idle += 1 + elif status == 'R': + run += 1 + elif status == 'F': + self.finish +=1 + self.submitted_ids.remove(pid) + else: + fail += 1 + + return idle, run, self.finish, fail + + def control_one_job(self, pid): + """ control the status of a single job with it's cluster id """ + raise NotImplemented('No implementation of how to control the job status to cluster \'%s\'' % self.name) + + def get_jobs_identifier(self, path, second_path=None): + """get a unique run_name for all the jobs helps to identify the runs + in the controller for some cluster.""" + + if second_path: + path = os.path.realpath(pjoin(path, second_path)) + elif not os.path.exists(path): + return path # job already done + + if 'SubProcesses' in path: + target = path.rsplit('/SubProcesses',1)[0] + elif 'MCatNLO' in path: + target = path.rsplit('/MCatNLO',1)[0] + elif 'PY8_parallelization' in path: + target = path.rsplit('/PY8_parallelization',1)[0] + elif second_path: + target=path + logger.warning("cluster.get_job_identifier runs unexpectedly. This should be fine but report this message if you have problem.") + else: + target = path + + if target.endswith('/'): + target = target[:-1] + + target = misc.digest(target.encode())[-self.identifier_length:] + if not target[0].isalpha(): + target = 'a' + target[1:] + + return target + + + @check_interupt() + def wait(self, me_dir, fct, minimal_job=0, update_first=None): + """Wait that all job are finish. + if minimal_job set, then return if idle + run is lower than that number""" + + + mode = 1 # 0 is long waiting/ 1 is short waiting + nb_iter = 0 + nb_short = 0 + change_at = 5 # number of iteration from which we wait longer between update. + + if update_first: + idle, run, finish, fail = self.control(me_dir) + update_first(idle, run, finish) + + #usefull shortcut for readibility + longtime, shorttime = self.options['cluster_status_update'] + + nb_job = 0 + + if self.options['cluster_type'] == 'htcaas2': + me_dir = self.metasubmit(self) + + while 1: + old_mode = mode + nb_iter += 1 + idle, run, finish, fail = self.control(me_dir) + if nb_job: + if idle + run + finish + fail != nb_job: + nb_job = idle + run + finish + fail + nb_iter = 1 # since some packet finish prevent to pass in long waiting mode + else: + nb_job = idle + run + finish + fail + if fail: + raise ClusterManagmentError('Some Jobs are in a Hold/... state. Please try to investigate or contact the IT team') + if idle + run == 0: + #time.sleep(20) #security to ensure that the file are really written on the disk + logger.info('All jobs finished') + fct(idle, run, finish) + break + if idle + run < minimal_job: + return + fct(idle, run, finish) + #Determine how much we have to wait (mode=0->long time, mode=1->short time) + if nb_iter < change_at: + mode = 1 + elif idle < run: + if old_mode == 0: + if nb_short: + mode = 0 #we already be back from short to long so stay in long + #check if we need to go back to short mode + elif idle: + if nb_iter > change_at + int(longtime)//shorttime: + mode = 0 #stay in long waiting mode + else: + mode = 1 # pass in short waiting mode + nb_short =0 + else: + mode = 1 # pass in short waiting mode + nb_short = 0 + elif old_mode == 1: + nb_short +=1 + if nb_short > 3* max(change_at, int(longtime)//shorttime): + mode = 0 #go back in slow waiting + else: + mode = 0 + + #if pass from fast(mode=1) to slow(mode=0) make a print statement: + if old_mode > mode: + logger.info('''Start to wait %ss between checking status. +Note that you can change this time in the configuration file. +Press ctrl-C to force the update.''' % self.options['cluster_status_update'][0]) + + #now Waiting! + if mode == 0: + try: + time.sleep(self.options['cluster_status_update'][0]) + except KeyboardInterrupt: + logger.info('start to update the status') + nb_iter = min(0, change_at -2) + nb_short = 0 + else: + time.sleep(self.options['cluster_status_update'][1]) + + + self.submitted = 0 + self.submitted_ids = [] + self.id_to_packet = {} + + def check_termination(self, job_id): + """Check the termination of the jobs with job_id and relaunch it if needed.""" + + + if job_id not in self.retry_args: + if job_id in self.id_to_packet: + nb_in_packet = self.id_to_packet[job_id].remove_one() + if nb_in_packet == 0: + # packet done run the associate function + packet = self.id_to_packet[job_id] + # fully ensure that the packet is finished (thread safe) + packet.queue.join() + #running the function + packet.fct(*packet.args) + del self.id_to_packet[job_id] + return 'resubmit' + else: + return True + + args = self.retry_args[job_id] + if 'time_check' in args: + time_check = args['time_check'] + else: + time_check = 0 + + for path in args['required_output']: + if args['cwd']: + path = pjoin(args['cwd'], path) +# check that file exists and is not empty. + if not (os.path.exists(path) and os.stat(path).st_size != 0) : + break + else: + # all requested output are present + if time_check > 0: + logger.info('Job %s Finally found the missing output.' % (job_id)) + del self.retry_args[job_id] + self.submitted_ids.remove(job_id) + # check if the job_id is in a packet + if job_id in self.id_to_packet: + nb_in_packet = self.id_to_packet[job_id].remove_one() + if nb_in_packet == 0: + # packet done run the associate function + packet = self.id_to_packet[job_id] + # fully ensure that the packet is finished (thread safe) + packet.queue.join() + #running the function + packet.fct(*packet.args) + del self.id_to_packet[job_id] + return 'resubmit' + + return 'done' + + if time_check == 0: + logger.debug('''Job %s: missing output:%s''' % (job_id,path)) + args['time_check'] = time.time() + return 'wait' + elif self.cluster_retry_wait > time.time() - time_check: + return 'wait' + + #jobs failed to be completed even after waiting time!! + if self.nb_retry < 0: + logger.critical('''Fail to run correctly job %s. + with option: %s + file missing: %s''' % (job_id, args, path)) + input('press enter to continue.') + elif self.nb_retry == 0: + logger.critical('''Fail to run correctly job %s. + with option: %s + file missing: %s. + Stopping all runs.''' % (job_id, args, path)) + self.remove() + elif args['nb_submit'] >= self.nb_retry: + logger.critical('''Fail to run correctly job %s. + with option: %s + file missing: %s + Fails %s times + No resubmition. ''' % (job_id, args, path, args['nb_submit'])) + self.remove() + else: + args['nb_submit'] += 1 + logger.warning('resubmit job (for the %s times)' % args['nb_submit']) + del self.retry_args[job_id] + self.submitted_ids.remove(job_id) + if 'time_check' in args: + del args['time_check'] + if job_id in self.id_to_packet: + self.id_to_packet[job_id].remove_one() + args['packet_member'] = self.id_to_packet[job_id] + del self.id_to_packet[job_id] + self.cluster_submit(**args) + else: + self.submit2(**args) + return 'resubmit' + return 'done' + + @check_interupt() + def launch_and_wait(self, prog, argument=[], cwd=None, stdout=None, + stderr=None, log=None, required_output=[], nb_submit=0, + input_files=[], output_files=[]): + """launch one job on the cluster and wait for it""" + + special_output = False # tag for concatenate the error with the output. + if stderr == -2 and stdout: + #We are suppose to send the output to stdout + special_output = True + stderr = stdout + '.err' + + id = self.submit2(prog, argument, cwd, stdout, stderr, log, + required_output=required_output, input_files=input_files, + output_files=output_files) + + if self.options['cluster_type']=='htcaas2': + if self.submitted == self.submitted_ids[-1]: + id = self.metasubmit(self) + + frame = inspect.currentframe() + args, _, _, values = inspect.getargvalues(frame) + args = dict([(i, values[i]) for i in args if i != 'self']) + self.retry_args[id] = args + + nb_wait=0 + while 1: + nb_wait+=1 + status = self.control_one_job(id) + if not status in ['R','I']: + status = self.check_termination(id) + if status in ['wait']: + time.sleep(30) + continue + elif status in ['resubmit']: + id = self.submitted_ids[0] + time.sleep(30) + continue + #really stop! + time.sleep(30) #security to ensure that the file are really written on the disk + break + time.sleep(self.options['cluster_status_update'][1]) + + if required_output: + status = self.check_termination(id) + if status == 'wait': + run += 1 + elif status == 'resubmit': + idle += 1 + + + if special_output: + # combine the stdout and the stderr + #wait up to 50 s to see if those files exists + for i in range(5): + if os.path.exists(stdout): + if not os.path.exists(stderr): + time.sleep(5) + if os.path.exists(stderr): + err_text = open(stderr).read() + if not err_text: + return + logger.warning(err_text) + text = open(stdout).read() + open(stdout,'w').write(text + err_text) + else: + return + time.sleep(10) + + def remove(self, *args, **opts): + """ """ + logger.warning("""This cluster didn't support job removal, + the jobs are still running on the cluster.""") + + @store_input() + def metasubmit(self, me_dir): + logger.warning("""This cluster didn't support metajob submit.""") + return 0 + + def modify_interface(self, run_interface): + """routine which allow to modify the run_card/mg5cmd object to change the + default behavior of the runs. + This is called at the time of the compilation of the run_card. + Note that this function can be called multiple times by run. + """ + #run_card = run_interface.run_card + return + +class Packet(object): + """ an object for handling packet of job, it is designed to be thread safe + """ + + def __init__(self, name, fct, args, opts={}): + import six.moves.queue + import threading + self.queue = six.moves.queue.Queue() + self.tag = name + self.fct = fct + self.args = args + self.opts = opts + self.done = threading.Event() + + def put(self, *args, **opts): + self.queue.put(*args, **opts) + + append = put + + def remove_one(self): + self.queue.get(True) + self.queue.task_done() + return self.queue.qsize() + +class MultiCore(Cluster): + """class for dealing with the submission in multiple node""" + + job_id = "$" + + def __init__(self, *args, **opt): + """Init the cluster """ + + + super(MultiCore, self).__init__(self, *args, **opt) + + import six.moves.queue + import threading + import six.moves._thread + self.queue = six.moves.queue.Queue() # list of job to do + self.done = six.moves.queue.Queue() # list of job finisned + self.submitted = six.moves.queue.Queue() # one entry by job submitted + self.stoprequest = threading.Event() #flag to ensure everything to close + self.demons = [] + self.nb_done =0 + if 'nb_core' in opt: + self.nb_core = opt['nb_core'] + elif isinstance(args[0],int): + self.nb_core = args[0] + else: + self.nb_core = 1 + # flag controlling if one keep the thread open or not after a wait() + if 'keep_thread' in opt: + self.keep_thread = opt['keep_thread'] + else: + self.keep_thread = False + + self.update_fct = None + + self.lock = threading.Event() # allow nice lock of the main thread + self.pids = six.moves.queue.Queue() # allow to clean jobs submit via subprocess + self.done_pid = [] # list of job finisned + self.done_pid_queue = six.moves.queue.Queue() + self.fail_msg = None + + + + def start_demon(self): + import threading + t = threading.Thread(target=self.worker) + t.daemon = True + t.start() + self.demons.append(t) + + + def worker(self): + import six.moves.queue + import six.moves._thread + while not self.stoprequest.isSet(): + try: + args = self.queue.get(timeout=10) + tag, exe, arg, opt = args + try: + # check for executable case + if isinstance(exe,str): + if os.path.exists(exe) and not exe.startswith('/'): + exe = './' + exe + if isinstance(opt['stdout'],str): + opt['stdout'] = open(opt['stdout'],'w') + if opt['stderr'] == None: + opt['stderr'] = subprocess.STDOUT + if arg: + proc = misc.Popen([exe] + arg, **opt) + else: + proc = misc.Popen(exe, **opt) + pid = proc.pid + self.pids.put(pid) + proc.wait() + if proc.returncode not in [0, 143, -15] and not self.stoprequest.isSet(): + fail_msg = 'program %s launch ends with non zero status: %s. Stop all computation' % \ + (' '.join([exe]+arg), proc.returncode) + logger.warning(fail_msg) + self.stoprequest.set() + self.remove(fail_msg) + # handle the case when this is a python function. Note that + # this use Thread so they are NO built-in parralelization this is + # going to work on a single core! (but this is fine for IO intensive + # function. for CPU intensive fct this will slow down the computation + else: + pid = tag + self.pids.put(pid) + # the function should return 0 if everything is fine + # the error message otherwise + returncode = exe(*arg, **opt) + if returncode != 0: + logger.warning("fct %s does not return 0. Stopping the code in a clean way. The error was:\n%s", exe, returncode) + self.stoprequest.set() + self.remove("fct %s does not return 0:\n %s" % (exe, returncode)) + except Exception as error: + self.fail_msg = sys.exc_info() + logger.warning(str(error)) + self.stoprequest.set() + self.remove(error) + + if __debug__: + six.reraise(self.fail_msg[0], self.fail_msg[1], self.fail_msg[2]) + + self.queue.task_done() + self.done.put(tag) + self.done_pid_queue.put(pid) + #release the mother to print the status on the screen + try: + self.lock.set() + except six.moves._thread.error: + continue + except six.moves.queue.Empty: + continue + import threading + self.demons.remove(threading.current_thread()) + + + def submit(self, prog, argument=[], cwd=None, stdout=None, stderr=None, + log=None, required_output=[], nb_submit=0): + """submit a job on multicore machine""" + + # open threads if needed + self.stoprequest.clear() + if len(self.demons) < self.nb_core: + self.start_demon() + + tag = (prog, tuple(argument), cwd, nb_submit) + if isinstance(prog, str): + + opt = {'cwd': cwd, + 'stdout':stdout, + 'stderr': stderr} + + self.queue.put((tag, prog, argument, opt)) + self.submitted.put(1) + return tag + else: + # python function + self.queue.put((tag, prog, argument, {})) + self.submitted.put(1) + return tag + + def launch_and_wait(self, prog, argument=[], cwd=None, stdout=None, + stderr=None, log=None, **opts): + """launch one job and wait for it""" + print("__CUDACPP_DEBUG: MultiCore.launch_and_wait starting '" + prog + "'") + cudacpp_start = time.perf_counter() + if isinstance(stdout, str): + stdout = open(stdout, 'w') + if isinstance(stderr, str): + stdout = open(stderr, 'w') + out = misc.call([prog] + argument, stdout=stdout, stderr=stderr, cwd=cwd) + cudacpp_end = time.perf_counter() + cudacpp_length = cudacpp_end - cudacpp_start + print("__CUDACPP_DEBUG: MultiCore.launch_and_wait finished '" + prog + "' in %.4f seconds"%cudacpp_length) + return out + + def remove(self, error=None): + """Ensure that all thread are killed""" + + # ensure the worker to stop + self.stoprequest.set() + if error and not self.fail_msg: + self.fail_msg = error + + # cleaning the queue done_pid_queue and move them to done_pid + while not self.done_pid_queue.empty(): + pid = self.done_pid_queue.get() + self.done_pid.append(pid) +# self.done_pid_queue.task_done() + + while not self.pids.empty(): + pid = self.pids.get() + self.pids.task_done() + if isinstance(pid, tuple): + continue + if pid in self.done_pid: + continue + out = os.system('CPIDS=$(pgrep -P %(pid)s); kill -15 $CPIDS > /dev/null 2>&1' \ + % {'pid':pid} ) + out = os.system('kill -15 %(pid)s > /dev/null 2>&1' % {'pid':pid} ) + + + + def wait(self, me_dir, update_status, update_first=None): + """Waiting that all the jobs are done. This function also control that + the submission by packet are handle correctly (i.e. submit the function)""" + + import six.moves.queue + import threading + + try: # to catch KeyBoardInterupt to see which kind of error to display + last_status = (0, 0, 0) + sleep_time = 1 + use_lock = True + first = True + while True: + force_one_more_loop = False # some security + + # Loop over the job tagged as done to check if some packet of jobs + # are finished in case, put the associate function in the queue + while self.done.qsize(): + try: + tag = self.done.get(True, 1) + except six.moves.queue.Empty: + pass + else: + if self.id_to_packet and tuple(tag) in self.id_to_packet: + packet = self.id_to_packet[tuple(tag)] + remaining = packet.remove_one() + if remaining == 0: + # fully ensure that the packet is finished (thread safe) + packet.queue.join() + self.submit(packet.fct, packet.args) + force_one_more_loop = True + self.nb_done += 1 + self.done.task_done() + + # Get from the various queue the Idle/Done/Running information + # Those variable should be thread safe but approximate. + Idle = self.queue.qsize() + Done = self.nb_done + self.done.qsize() + Running = max(0, self.submitted.qsize() - Idle - Done) + + if Idle + Running <= 0 and not force_one_more_loop: + update_status(Idle, Running, Done) + # Going the quit since everything is done + # Fully Ensure that everything is indeed done. + self.queue.join() + break + + if (Idle, Running, Done) != last_status: + if first and update_first: + update_first(Idle, Running, Done) + first = False + else: + update_status(Idle, Running, Done) + last_status = (Idle, Running, Done) + + # cleaning the queue done_pid_queue and move them to done_pid + while not self.done_pid_queue.empty(): + pid = self.done_pid_queue.get() + self.done_pid.append(pid) + self.done_pid_queue.task_done() + + + # Define how to wait for the next iteration + if use_lock: + # simply wait that a worker release the lock + use_lock = self.lock.wait(300) + self.lock.clear() + if not use_lock and Idle > 0: + use_lock = True + else: + # to be sure that we will never fully lock at the end pass to + # a simple time.sleep() + time.sleep(sleep_time) + sleep_time = min(sleep_time + 2, 180) + if update_first: + update_first(Idle, Running, Done) + + if self.stoprequest.isSet(): + if isinstance(self.fail_msg, Exception): + raise self.fail_msg + elif isinstance(self.fail_msg, str): + raise Exception(self.fail_msg) + elif self.fail_msg: + # can happend that stoprequest is set bu not fail if no job have been resubmitted + six.reraise(self.fail_msg[0], self.fail_msg[1], self.fail_msg[2]) + # self.fail_msg is None can happen when no job was submitted -> ignore + + # reset variable for next submission + try: + self.lock.clear() + except Exception: + pass + self.done = six.moves.queue.Queue() + self.done_pid = [] + self.done_pid_queue = six.moves.queue.Queue() + self.nb_done = 0 + self.submitted = six.moves.queue.Queue() + self.pids = six.moves.queue.Queue() + self.stoprequest.clear() + self.id_to_packet = {} + + + except KeyboardInterrupt: + # if one of the node fails -> return that error + if isinstance(self.fail_msg, Exception): + raise self.fail_msg + elif isinstance(self.fail_msg, str): + raise Exception(self.fail_msg) + elif self.fail_msg: + six.reraise(self.fail_msg[0], self.fail_msg[1], self.fail_msg[2]) + # else return orignal error + raise + + if not self.keep_thread: + self.stoprequest.set() + +class CondorCluster(Cluster): + """Basic class for dealing with cluster submission""" + + name = 'condor' + job_id = 'CONDOR_ID' + + + + @multiple_try() + def submit(self, prog, argument=[], cwd=None, stdout=None, stderr=None, log=None, + required_output=[], nb_submit=0): + """Submit a job prog to a Condor cluster""" + + text = """Executable = %(prog)s + output = %(stdout)s + error = %(stderr)s + log = %(log)s + %(argument)s + environment = CONDOR_ID=$(Cluster).$(Process) + Universe = vanilla + notification = Error + Initialdir = %(cwd)s + %(requirement)s + getenv=True + queue 1 + """ + + if self.cluster_queue not in ['None', None]: + requirement = 'Requirements = %s=?=True' % self.cluster_queue + else: + requirement = '' + + if cwd is None: + cwd = os.getcwd() + if stdout is None: + stdout = '/dev/null' + if stderr is None: + stderr = '/dev/null' + if log is None: + log = '/dev/null' + if not os.path.exists(prog): + prog = os.path.join(cwd, prog) + if argument: + argument = 'Arguments = %s' % ' '.join(argument) + else: + argument = '' + + + dico = {'prog': prog, 'cwd': cwd, 'stdout': stdout, + 'stderr': stderr,'log': log,'argument': argument, + 'requirement': requirement} + + #open('submit_condor','w').write(text % dico) + a = misc.Popen(['condor_submit'], stdout=subprocess.PIPE, + stdin=subprocess.PIPE) + output, _ = a.communicate((text % dico).encode()) + #output = a.stdout.read() + #Submitting job(s). + #Logging submit event(s). + #1 job(s) submitted to cluster 2253622. + pat = re.compile("submitted to cluster (\d*)",re.MULTILINE) + output = output.decode(errors='ignore') + try: + id = pat.search(output).groups()[0] + except: + raise ClusterManagmentError('fail to submit to the cluster: \n%s' \ + % output) + self.submitted += 1 + self.submitted_ids.append(id) + return id + + @store_input() + @multiple_try() + def submit2(self, prog, argument=[], cwd=None, stdout=None, stderr=None, + log=None, input_files=[], output_files=[], required_output=[], + nb_submit=0): + """Submit the job on the cluster NO SHARE DISK + input/output file should be give relative to cwd + """ + + if not required_output and output_files: + required_output = output_files + + if (input_files == [] == output_files): + return self.submit(prog, argument, cwd, stdout, stderr, log, + required_output=required_output, nb_submit=nb_submit) + + text = """Executable = %(prog)s + output = %(stdout)s + error = %(stderr)s + log = %(log)s + %(argument)s + should_transfer_files = YES + when_to_transfer_output = ON_EXIT + transfer_input_files = %(input_files)s + %(output_files)s + Universe = vanilla + notification = Error + Initialdir = %(cwd)s + %(requirement)s + getenv=True + queue 1 + """ + + if self.cluster_queue not in ['None', None]: + requirement = 'Requirements = %s=?=True' % self.cluster_queue + else: + requirement = '' + + if cwd is None: + cwd = os.getcwd() + if stdout is None: + stdout = '/dev/null' + if stderr is None: + stderr = '/dev/null' + if log is None: + log = '/dev/null' + if not os.path.exists(prog): + prog = os.path.join(cwd, prog) + if argument: + argument = 'Arguments = %s' % ' '.join([str(a) for a in argument]) + else: + argument = '' + # input/output file treatment + if input_files: + input_files = ','.join(input_files) + else: + input_files = '' + if output_files: + output_files = 'transfer_output_files = %s' % ','.join(output_files) + else: + output_files = '' + + + + dico = {'prog': prog, 'cwd': cwd, 'stdout': stdout, + 'stderr': stderr,'log': log,'argument': argument, + 'requirement': requirement, 'input_files':input_files, + 'output_files':output_files} + + #open('submit_condor','w').write(text % dico) + a = subprocess.Popen(['condor_submit'], stdout=subprocess.PIPE, + stdin=subprocess.PIPE) + output, _ = a.communicate((text % dico).encode()) + #output = a.stdout.read() + #Submitting job(s). + #Logging submit event(s). + #1 job(s) submitted to cluster 2253622. + output = output.decode(errors='ignore') + pat = re.compile("submitted to cluster (\d*)",re.MULTILINE) + try: + id = pat.search(output).groups()[0] + except: + raise ClusterManagmentError('fail to submit to the cluster: \n%s' \ + % output) + self.submitted += 1 + self.submitted_ids.append(id) + return id + + + + + + @multiple_try(nb_try=10, sleep=10) + def control_one_job(self, id): + """ control the status of a single job with it's cluster id """ + cmd = 'condor_q '+str(id)+" -format \'%-2s \\n\' \'ifThenElse(JobStatus==0,\"U\",ifThenElse(JobStatus==1,\"I\",ifThenElse(JobStatus==2,\"R\",ifThenElse(JobStatus==3,\"X\",ifThenElse(JobStatus==4,\"C\",ifThenElse(JobStatus==5,\"H\",ifThenElse(JobStatus==6,\"E\",string(JobStatus))))))))\'" + status = misc.Popen([cmd], shell=True, stdout=subprocess.PIPE, + stderr=subprocess.PIPE) + + error = status.stderr.read().decode(errors='ignore') + if status.returncode and error: + raise ClusterManagmentError('condor_q returns error: %s' % error) + elif status.returncode: + raise ClusterManagmentError('condor_q fails with status code: %s' % status.returncode) + elif error: + sys.stderr.write("condor_q error (returncode was 0): %s" % error) + + return status.stdout.readline().decode(errors='ignore').strip() + + jobstatus = {'0':'U', '1':'I','2':'R','3':'X','4':'C','5':'H','6':'E'} + @check_interupt() + @multiple_try(nb_try=10, sleep=10) + def control(self, me_dir): + """ control the status of a single job with it's cluster id """ + + if not self.submitted_ids: + return 0, 0, 0, 0 + + packet = 15000 + idle, run, fail = 0, 0, 0 + ongoing = [] + for i in range(1+(len(self.submitted_ids)-1)//packet): + start = i * packet + stop = (i+1) * packet + cmd = "condor_q " + ' '.join(self.submitted_ids[start:stop]) + \ + " -format \"%d \" ClusterId " + \ + " -format \"%d\\n\" JobStatus " + + status = misc.Popen([cmd], shell=True, stdout=subprocess.PIPE, + stderr=subprocess.PIPE) + error = status.stderr.read().decode(errors='ignore') + if status.returncode and error: + raise ClusterManagmentError('condor_q returns error: %s' % error) + elif status.returncode: + raise ClusterManagmentError('condor_q fails with status code: %s' % status.returncode) + elif error: + sys.stderr.write("condor_q error (returncode was 0): %s" % error) + + for line in status.stdout: + id, status = line.decode(errors='ignore').strip().split() + status = self.jobstatus[status] + ongoing.append(id) + if status in ['I','U']: + idle += 1 + elif status == 'R': + run += 1 + elif status != 'C': + fail += 1 + + for id in list(self.submitted_ids): + if id not in ongoing: + status = self.check_termination(id) + if status == 'wait': + run += 1 + elif status == 'resubmit': + idle += 1 + + return idle, run, self.submitted - (idle+run+fail), fail + + @multiple_try() + def remove(self, *args, **opts): + """Clean the jobson the cluster""" + + if not self.submitted_ids: + return + cmd = "condor_rm %s" % ' '.join(self.submitted_ids) + + status = misc.Popen([cmd], shell=True, stdout=open(os.devnull,'w')) + self.submitted_ids = [] + +class PBSCluster(Cluster): + """Basic class for dealing with cluster submission""" + + name = 'pbs' + job_id = 'PBS_JOBID' + idle_tag = ['Q'] + running_tag = ['T','E','R'] + complete_tag = ['C'] + + maximum_submited_jobs = 2500 + + @multiple_try() + def submit(self, prog, argument=[], cwd=None, stdout=None, stderr=None, log=None, + required_output=[], nb_submit=0): + """Submit a job prog to a PBS cluster""" + + me_dir = self.get_jobs_identifier(cwd, prog) + + if len(self.submitted_ids) > self.maximum_submited_jobs: + fct = lambda idle, run, finish: logger.info('Waiting for free slot: %s %s %s' % (idle, run, finish)) + self.wait(me_dir, fct, self.maximum_submited_jobs) + + + text = "" + if cwd is None: + cwd = os.getcwd() + else: + text = " cd %s;" % cwd + if stdout is None: + stdout = '/dev/null' + if stderr is None: + stderr = '/dev/null' + elif stderr == -2: # -2 is subprocess.STDOUT + stderr = stdout + if log is None: + log = '/dev/null' + + if not os.path.isabs(prog): + text += "./%s" % prog + else: + text+= prog + + if argument: + text += ' ' + ' '.join(argument) + + command = ['qsub','-o', stdout, + '-N', me_dir, + '-e', stderr, + '-V'] + + if self.cluster_queue and self.cluster_queue != 'None': + command.extend(['-q', self.cluster_queue]) + + a = misc.Popen(command, stdout=subprocess.PIPE, + stderr=subprocess.STDOUT, + stdin=subprocess.PIPE, cwd=cwd) + + output = a.communicate(text.encode())[0].decode(errors='ignore') + id = output.split('.')[0] + if not id.isdigit() or a.returncode !=0: + raise ClusterManagmentError('fail to submit to the cluster: \n%s' \ + % output) + + self.submitted += 1 + self.submitted_ids.append(id) + return id + + @multiple_try() + def control_one_job(self, id): + """ control the status of a single job with it's cluster id """ + cmd = 'qstat '+str(id) + status = misc.Popen([cmd], shell=True, stdout=subprocess.PIPE, + stderr=subprocess.STDOUT) + + for line in status.stdout: + line = line.decode(errors='ignore').strip() + if 'cannot connect to server' in line or 'cannot read reply' in line: + raise ClusterManagmentError('server disconnected') + if 'Unknown' in line: + return 'F' + elif line.startswith(str(id)): + jobstatus = line.split()[4] + else: + jobstatus="" + + if status.returncode != 0 and status.returncode is not None: + raise ClusterManagmentError('server fails in someway (errorcode %s)' % status.returncode) + if jobstatus in self.idle_tag: + return 'I' + elif jobstatus in self.running_tag: + return 'R' + return 'F' + + + @multiple_try() + def control(self, me_dir): + """ control the status of a single job with it's cluster id """ + cmd = "qstat" + status = misc.Popen([cmd], stdout=subprocess.PIPE) + + me_dir = self.get_jobs_identifier(me_dir) + + ongoing = [] + + idle, run, fail = 0, 0, 0 + for line in status.stdout: + line = line.decode(errors='ignore') + if 'cannot connect to server' in line or 'cannot read reply' in line: + raise ClusterManagmentError('server disconnected') + if me_dir in line: + ongoing.append(line.split()[0].split('.')[0]) + status2 = line.split()[4] + if status2 in self.idle_tag: + idle += 1 + elif status2 in self.running_tag: + run += 1 + elif status2 in self.complete_tag: + if not self.check_termination(line.split()[0].split('.')[0]): + idle += 1 + else: + fail += 1 + + if status.returncode != 0 and status.returncode is not None: + raise ClusterManagmentError('server fails in someway (errorcode %s)' % status.returncode) + + for id in list(self.submitted_ids): + if id not in ongoing: + status2 = self.check_termination(id) + if status2 == 'wait': + run += 1 + elif status2 == 'resubmit': + idle += 1 + + return idle, run, self.submitted - (idle+run+fail), fail + + @multiple_try() + def remove(self, *args, **opts): + """Clean the jobs on the cluster""" + + if not self.submitted_ids: + return + cmd = "qdel %s" % ' '.join(self.submitted_ids) + status = misc.Popen([cmd], shell=True, stdout=open(os.devnull,'w')) + self.submitted_ids = [] + + +class SGECluster(Cluster): + """Basic class for dealing with cluster submission""" + # Class written by Arian Abrahantes. + + name = 'sge' + job_id = 'JOB_ID' + idle_tag = ['qw', 'hqw','hRqw','w'] + running_tag = ['r','t','Rr','Rt'] + identifier_length = 10 + + def def_get_path(self,location): + """replace string for path issues""" + location = os.path.realpath(location) + homePath = os.getenv("HOME") + if homePath: + location = location.replace(homePath,'$HOME') + return location + + @multiple_try() + def submit(self, prog, argument=[], cwd=None, stdout=None, stderr=None, log=None, + required_output=[], nb_submit=0): + """Submit a job prog to an SGE cluster""" + + me_dir = self.get_jobs_identifier(cwd, prog) + + + if cwd is None: + #cwd = os.getcwd() + cwd = self.def_get_path(os.getcwd()) + cwd1 = self.def_get_path(cwd) + text = " cd %s;" % cwd1 + if stdout is None: + stdout = '/dev/null' + else: + stdout = self.def_get_path(stdout) + if stderr is None: + stderr = '/dev/null' + elif stderr == -2: # -2 is subprocess.STDOUT + stderr = stdout + else: + stderr = self.def_get_path(stderr) + + if log is None: + log = '/dev/null' + else: + log = self.def_get_path(log) + + text += prog + if argument: + text += ' ' + ' '.join(argument) + + #if anything slips through argument + #print "!=== inteded change ",text.replace('/srv/nfs','') + #text = text.replace('/srv/nfs','') + homePath = os.getenv("HOME") + if homePath: + text = text.replace(homePath,'$HOME') + + logger.debug("!=== input %s" % text) + logger.debug("!=== output %s" % stdout) + logger.debug("!=== error %s" % stderr) + logger.debug("!=== logs %s" % log) + + command = ['qsub','-o', stdout, + '-N', me_dir, + '-e', stderr, + '-V'] + + if self.cluster_queue and self.cluster_queue != 'None': + command.extend(['-q', self.cluster_queue]) + + a = misc.Popen(command, stdout=subprocess.PIPE, + stderr=subprocess.STDOUT, + stdin=subprocess.PIPE, cwd=cwd) + + output = a.communicate(text.encode())[0].decode(errors='ignore') + id = output.split(' ')[2] + if not id.isdigit(): + raise ClusterManagmentError('fail to submit to the cluster: \n%s' \ + % output) + self.submitted += 1 + self.submitted_ids.append(id) + logger.debug(output) + + return id + + @multiple_try() + def control_one_job(self, id): + """ control the status of a single job with it's cluster id """ + #cmd = 'qstat '+str(id) + cmd = 'qstat ' + status = misc.Popen([cmd], shell=True, stdout=subprocess.PIPE) + for line in status.stdout: + line = line.decode(errors='ignore') + #print "!==",line + #line = line.strip() + #if 'Unknown' in line: + # return 'F' + #elif line.startswith(str(id)): + # status = line.split()[4] + if str(id) in line: + status = line.split()[4] + #print "!=status", status + if status in self.idle_tag: + return 'I' + elif status in self.running_tag: + return 'R' + return 'F' + + @multiple_try() + def control(self, me_dir): + """ control the status of a single job with it's cluster id """ + cmd = "qstat " + status = misc.Popen([cmd], shell=True, stdout=subprocess.PIPE) + + me_dir = self.get_jobs_identifier(me_dir) + + finished = list(self.submitted_ids) + + idle, run, fail = 0, 0, 0 + for line in status.stdout: + line = line.decode(errors='ignore') + if me_dir in line: + id,_,_,_,status = line.split()[:5] + if status in self.idle_tag: + idle += 1 + finished.remove(id) + elif status in self.running_tag: + run += 1 + finished.remove(id) + else: + logger.debug(line) + fail += 1 + finished.remove(id) + + for id in finished: + self.check_termination(id) + + return idle, run, self.submitted - (idle+run+fail), fail + + + + @multiple_try() + def remove(self, *args, **opts): + """Clean the jobs on the cluster""" + + if not self.submitted_ids: + return + cmd = "qdel %s" % ' '.join(self.submitted_ids) + status = misc.Popen([cmd], shell=True, stdout=open(os.devnull,'w')) + self.submitted_ids = [] + + +class LSFCluster(Cluster): + """Basic class for dealing with cluster submission""" + + name = 'lsf' + job_id = 'LSB_JOBID' + + @multiple_try() + def submit(self, prog, argument=[], cwd=None, stdout=None, stderr=None, log=None, + required_output=[], nb_submit=0): + """Submit the job prog to an LSF cluster""" + + + me_dir = self.get_jobs_identifier(cwd, prog) + + text = "" + command = ['bsub', '-C0', '-J', me_dir] + if cwd is None: + cwd = os.getcwd() + else: + text = " cd %s;" % cwd + if stdout and isinstance(stdout, str): + command.extend(['-o', stdout]) + if stderr and isinstance(stdout, str): + command.extend(['-e', stderr]) + elif stderr == -2: # -2 is subprocess.STDOUT + pass + if log is None: + log = '/dev/null' + + text += prog + if argument: + text += ' ' + ' '.join(argument) + + if self.cluster_queue and self.cluster_queue != 'None': + command.extend(['-q', self.cluster_queue]) + + a = misc.Popen(command, stdout=subprocess.PIPE, + stderr=subprocess.STDOUT, + stdin=subprocess.PIPE, cwd=cwd) + + output = a.communicate(text.encode())[0].decode(errors='ignore') + #Job is submitted to default queue . + try: + id = output.split('>',1)[0].split('<')[1] + except: + raise ClusterManagmentError('fail to submit to the cluster: \n%s' \ + % output) + if not id.isdigit(): + raise ClusterManagmentError('fail to submit to the cluster: \n%s' \ + % output) + self.submitted += 1 + self.submitted_ids.append(id) + return id + + + @multiple_try() + def control_one_job(self, id): + """ control the status of a single job with it's cluster id """ + + cmd = 'bjobs '+str(id) + status = misc.Popen([cmd], shell=True, stdout=subprocess.PIPE) + + for line in status.stdout: + line = line.decode(errors='ignore').strip().upper() + if 'JOBID' in line: + continue + elif str(id) not in line: + continue + status = line.split()[2] + if status == 'RUN': + return 'R' + elif status == 'PEND': + return 'I' + elif status == 'DONE': + return 'F' + else: + return 'H' + return 'F' + + @multiple_try() + def control(self, me_dir): + """ control the status of a single job with it's cluster id """ + + if not self.submitted_ids: + return 0, 0, 0, 0 + + cmd = "bjobs " + ' '.join(self.submitted_ids) + status = misc.Popen([cmd], shell=True, stdout=subprocess.PIPE) + + jobstatus = {} + for line in status.stdout: + line = line.decode(errors='ignore').strip() + if 'JOBID' in line: + continue + splitline = line.split() + id = splitline[0] + if id not in self.submitted_ids: + continue + jobstatus[id] = splitline[2] + + idle, run, fail = 0, 0, 0 + for id in self.submitted_ids[:]: + if id in jobstatus: + status = jobstatus[id] + else: + status = 'MISSING' + if status == 'RUN': + run += 1 + elif status == 'PEND': + idle += 1 + else: + status = self.check_termination(id) + if status == 'wait': + run += 1 + elif status == 'resubmit': + idle += 1 + + return idle, run, self.submitted - (idle+run+fail), fail + + @multiple_try() + def remove(self, *args,**opts): + """Clean the jobs on the cluster""" + + if not self.submitted_ids: + return + cmd = "bkill %s" % ' '.join(self.submitted_ids) + status = misc.Popen([cmd], shell=True, stdout=open(os.devnull,'w')) + self.submitted_ids = [] + +class GECluster(Cluster): + """Class for dealing with cluster submission on a GE cluster""" + + name = 'ge' + job_id = 'JOB_ID' + idle_tag = ['qw'] + running_tag = ['r'] + + @multiple_try() + def submit(self, prog, argument=[], cwd=None, stdout=None, stderr=None, log=None, + required_output=[], nb_submit=0): + """Submit a job prog to a GE cluster""" + + text = "" + if cwd is None: + cwd = os.getcwd() + else: + text = " cd %s; bash " % cwd + if stdout is None: + stdout = os.path.join(cwd, "log.%s" % prog.split('/')[-1]) + if stderr is None: + stderr = os.path.join(cwd, "err.%s" % prog.split('/')[-1]) + elif stderr == -2: # -2 is subprocess.STDOUT + stderr = stdout + if log is None: + log = '/dev/null' + + text += prog + if argument: + text += ' ' + ' '.join(argument) + text += '\n' + tmp_submit = os.path.join(cwd, 'tmp_submit') + open(tmp_submit,'w').write(text) + + a = misc.Popen(['qsub','-o', stdout, + '-e', stderr, + tmp_submit], + stdout=subprocess.PIPE, + stderr=subprocess.STDOUT, + stdin=subprocess.PIPE, cwd=cwd) + + output = a.communicate()[0].decode(errors='ignore') + #Your job 874511 ("test.sh") has been submitted + pat = re.compile("Your job (\d*) \(",re.MULTILINE) + try: + id = pat.search(output).groups()[0] + except: + raise ClusterManagmentError('fail to submit to the cluster: \n%s' \ + % output) + self.submitted += 1 + self.submitted_ids.append(id) + return id + + @multiple_try() + def control_one_job(self, id): + """ control the status of a single job with it's cluster id """ + cmd = 'qstat | grep '+str(id) + status = misc.Popen([cmd], shell=True, stdout=subprocess.PIPE) + if not status: + return 'F' + #874516 0.00000 test.sh alwall qw 03/04/2012 22:30:35 1 + pat = re.compile("^(\d+)\s+[\d\.]+\s+[\w\d\.]+\s+[\w\d\.]+\s+(\w+)\s") + stat = '' + for line in status.stdout.read().decode(errors='ignore').split('\n'): + if not line: + continue + line = line.strip() + try: + groups = pat.search(line).groups() + except: + raise ClusterManagmentError('bad syntax for stat: \n\"%s\"' % line) + if groups[0] != id: continue + stat = groups[1] + if not stat: + return 'F' + if stat in self.idle_tag: + return 'I' + if stat in self.running_tag: + return 'R' + + @multiple_try() + def control(self, me_dir=None): + """Check the status of job associated to directory me_dir. return (idle, run, finish, fail)""" + if not self.submitted_ids: + return 0, 0, 0, 0 + idle, run, fail = 0, 0, 0 + ongoing = [] + for statusflag in ['p', 'r', 'sh']: + cmd = 'qstat -s %s' % statusflag + status = misc.Popen([cmd], shell=True, stdout=subprocess.PIPE) + #874516 0.00000 test.sh alwall qw 03/04/2012 22:30:35 1 + pat = re.compile("^(\d+)") + for line in status.stdout.read().decode(errors='ignore').split('\n'): + line = line.strip() + try: + id = pat.search(line).groups()[0] + except Exception: + pass + else: + if id not in self.submitted_ids: + continue + ongoing.append(id) + if statusflag == 'p': + idle += 1 + if statusflag == 'r': + run += 1 + if statusflag == 'sh': + fail += 1 + for id in list(self.submitted_ids): + if id not in ongoing: + self.check_termination(id) + #self.submitted_ids = ongoing + + return idle, run, self.submitted - idle - run - fail, fail + + @multiple_try() + def remove(self, *args, **opts): + """Clean the jobs on the cluster""" + + if not self.submitted_ids: + return + cmd = "qdel %s" % ' '.join(self.submitted_ids) + status = misc.Popen([cmd], shell=True, stdout=open(os.devnull,'w')) + self.submitted_ids = [] + +def asyncrone_launch(exe, cwd=None, stdout=None, argument = [], **opt): + """start a computation and not wait for it to finish. + this fonction returns a lock which is locked as long as the job is + running.""" + + mc = MultiCore(1) + mc.submit(exe, argument, cwd, stdout, **opt) + mc.need_waiting = True + return mc.lock + + +class SLURMCluster(Cluster): + """Basic class for dealing with cluster submission""" + + name = 'slurm' + job_id = 'SLURM_JOBID' + idle_tag = ['Q','PD','S','CF'] + running_tag = ['R', 'CG'] + complete_tag = ['C'] + identifier_length = 8 + + @multiple_try() + def submit(self, prog, argument=[], cwd=None, stdout=None, stderr=None, log=None, + required_output=[], nb_submit=0): + """Submit a job prog to a SLURM cluster""" + + me_dir = self.get_jobs_identifier(cwd, prog) + import sys + if prog == sys.executable: + argument.insert(0, prog) + if MADEVENT: + prog = pjoin(LOCALDIR,'bin','internal','eval.sh') + else: + prog = pjoin(MG5DIR, 'Template','Common','bin','internal','eval.sh') + + if cwd is None: + cwd = os.getcwd() + if stdout is None: + stdout = '/dev/null' + if stderr is None: + stderr = '/dev/null' + elif stderr == -2: # -2 is subprocess.STDOUT + stderr = stdout + if log is None: + log = '/dev/null' + + command = ['sbatch', '-o', stdout, + '-J', me_dir, + '-e', stderr, prog] + argument + + + + if self.cluster_queue and self.cluster_queue != 'None': + command.insert(1, '-p') + command.insert(2, self.cluster_queue) + + + a = misc.Popen(command, stdout=subprocess.PIPE, + stderr=subprocess.STDOUT, + stdin=subprocess.PIPE, cwd=cwd) + + output = a.communicate() + output_arr = output[0].decode(errors='ignore').split(' ') + id = output_arr[3].rstrip() + + if not id.isdigit(): + id = re.findall('Submitted batch job ([\d\.]+)', ' '.join(output_arr)) + + if not id or len(id)>1: + raise ClusterManagmentError( 'fail to submit to the cluster: \n%s' \ + % ('stdout: %s\nstderr %s' %(output[0],output[1]))) + id = id[0] + + + self.submitted += 1 + self.submitted_ids.append(id) + return id + + @multiple_try() + def control_one_job(self, id): + """ control the status of a single job with it's cluster id """ + cmd = 'squeue j'+str(id) + # Remove incompatible squeue formats + env = os.environ.copy() + if "SQUEUE_FORMAT" in env: + del env["SQUEUE_FORMAT"] + status = misc.Popen([cmd], shell=True, stdout=subprocess.PIPE, + stderr=open(os.devnull,'w'),env=env) + + for line in status.stdout: + line = line.decode(errors='ignore').strip() + if 'Invalid' in line: + return 'F' + elif line.startswith(str(id)): + status = line.split()[4] + if status in self.idle_tag: + return 'I' + elif status in self.running_tag: + return 'R' + return 'F' + + @multiple_try() + def control(self, me_dir): + """ control the status of a single job with it's cluster id """ + cmd = "squeue" + # Remove incompatible squeue formats + env = os.environ.copy() + if "SQUEUE_FORMAT" in env: + del env["SQUEUE_FORMAT"] + pstatus = misc.Popen([cmd], stdout=subprocess.PIPE,env=env) + + me_dir = self.get_jobs_identifier(me_dir) + + idle, run, fail = 0, 0, 0 + ongoing=[] + for line in pstatus.stdout: + line = line.decode(errors='ignore') + if me_dir in line: + id, _, _,_ , status,_ = line.split(None,5) + ongoing.append(id) + if status in self.idle_tag: + idle += 1 + elif status in self.running_tag: + run += 1 + elif status in self.complete_tag: + status = self.check_termination(id) + if status == 'wait': + run += 1 + elif status == 'resubmit': + idle += 1 + else: + fail += 1 + + #control other finished job + for id in list(self.submitted_ids): + if id not in ongoing: + status = self.check_termination(id) + if status == 'wait': + run += 1 + elif status == 'resubmit': + idle += 1 + + + return idle, run, self.submitted - (idle+run+fail), fail + + @multiple_try() + def remove(self, *args, **opts): + """Clean the jobs on the cluster""" + + if not self.submitted_ids: + return + cmd = "scancel %s" % ' '.join(self.submitted_ids) + status = misc.Popen([cmd], shell=True, stdout=open(os.devnull,'w')) + self.submitted_ids = [] + +class HTCaaSCluster(Cluster): + """Class for dealing with cluster submission on a HTCaaS cluster using GPFS """ + + name= 'htcaas' + job_id = 'HTCAAS_JOBID' + idle_tag = ['waiting'] + running_tag = ['preparing','running'] + complete_tag = ['done'] + + @store_input() + @multiple_try() + def submit2(self, prog, argument=[], cwd=None, stdout=None, stderr=None, + log=None, input_files=[], output_files=[], required_output=[], + nb_submit=0): + """Submit the HTCaaS job on the cluster with NO SHARE DISK + input/output file should be given as relative to CWd + """ + # To make workspace name(temp) + cur_usr = os.getenv('USER') + + if cwd is None: + cwd = os.getcwd() + + cwd_cp = cwd.rsplit("/",2) + + if not stdout is None: + print("stdout: %s" % stdout) + + if not os.path.exists(prog): + prog = os.path.join(cwd, prog) + + if not required_output and output_files: + required_output = output_files + + logger.debug(prog) + if 'combine' not in prog and 'pythia' not in prog and 'shower' not in prog : + cwd_arg = cwd+"/arguments" + temp = ' '.join([str(a) for a in argument]) + arg_cmd="echo '"+temp+"' > " + cwd_arg + command = ['htcaas-mgjob-submit','-d',cwd,'-e',os.path.basename(prog)] + if argument : + command.extend(['-a ', '='.join([str(a) for a in argument])]) + a = misc.Popen(command, stdout=subprocess.PIPE, stderr=subprocess.PIPE, stdin=subprocess.PIPE, cwd=cwd) + id = a.stdout.read().strip() + + else: + cwd_arg = cwd+"/arguments" + temp = ' '.join([str(a) for a in argument]) + temp_file_name = "sub." + os.path.basename(prog) + text = """#!/bin/bash + MYPWD=%(cwd)s + cd $MYPWD + input_files=(%(input_files)s ) + for i in ${input_files[@]} + do + chmod -f +x $i + done + /bin/bash %(prog)s %(arguments)s > %(stdout)s + """ + dico = {'cwd':cwd, 'input_files': ' '.join(input_files + [prog]), 'stdout': stdout, 'prog':prog, + 'arguments': ' '.join([str(a) for a in argument]), + 'program': ' ' if '.py' in prog else 'bash'} + + # writing a new script for the submission + new_prog = pjoin(cwd, temp_file_name) + open(new_prog, 'w').write(text % dico) + misc.Popen(['chmod','+x',new_prog],cwd=cwd) + command = ['htcaas-mgjob-submit','-d',cwd,'-e',temp_file_name] + a = misc.Popen(command, stdout=subprocess.PIPE, stderr=subprocess.PIPE, stdin=subprocess.PIPE, cwd=cwd) + id = a.stdout.read().strip() + logger.debug(id) + + nb_try=0 + nb_limit=5 + if not id.isdigit() : + print("[ID is not digit]:" + id) + + while not id.isdigit() : + nb_try+=1 + print("[fail_retry]:"+ nb_try) + a=misc.Popen(command, stdout=subprocess.PIPE, stderr=subprocess.PIPE, stdin=subprocess.PIPE, cwd=cwd) + id = a.stdout.read().strip() + if nb_try > nb_limit : + raise ClusterManagementError('fail to submit to the HTCaaS cluster: \n %s' % id) + break + + self.submitted += 1 + self.submitted_ids.append(id) + + return id + + @multiple_try(nb_try=10, sleep=5) + def control_one_job(self, id): + """ control the status of a single job with it's cluster id """ + + if id == 0 : + status_out ='C' + else : + cmd = 'htcaas-job-status -m '+str(id)+ " -s | grep Status " + status = misc.Popen([cmd], shell=True,stdout=subprocess.PIPE, + stderr=subprocess.PIPE) + error = status.stderr.read().decode(errors='ignore') + if status.returncode or error: + raise ClusterManagmentError('htcaas-job-submit returns error: %s' % error) + status_out= status.stdout.read().decode(errors='ignore').strip() + status_out= status_out.split(":",1)[1] + if status_out == 'waiting': + status_out='I' + elif status_out == 'preparing' or status_out == 'running': + status_out = 'R' + elif status_out != 'done': + status_out = 'F' + elif status_out == 'done': + status_out = 'C' + + return status_out + + @multiple_try() + def control(self, me_dir): + """ control the status of a single job with it's cluster id """ + if not self.submitted_ids: + logger.debug("self.submitted_ids not exists") + return 0, 0, 0, 0 + + ongoing = [] + idle, run, fail = 0, 0, 0 + + start = self.submitted_ids[0] + end = self.submitted_ids[-1] + + cmd = "htcaas-job-status -c "+str(start)+"-"+str(end)#+" -ac" + status = misc.Popen([cmd], shell=True, stdout=subprocess.PIPE) + + for line in status.stdout: + #ongoing.append(line.split()[0].strip()) + status2 = line.decode(errors='ignore').split()[-1] + if status2 != 'null' or line.split()[0].strip() != '0': + ongoing.append(line.split()[0].strip()) + logger.debug("["+line.split()[0].strip()+"]"+status2) + if status2 != 'null' or line.split()[0].strip() != '0': + idle += 1 + elif status2 in self.idle_tag: + idle += 1 + elif status2 in self.running_tag: + run += 1 + elif status2 in self.complete_tag: + if not self.check_termination(line.split()[0]): + idle +=1 + else: + fail += 1 + + return idle, run, self.submitted - (idle+run+fail), fail + + @multiple_try() + def remove(self, *args, **opts): + """Clean the jobson the cluster""" + + if not self.submitted_ids: + return + for i in range(len(self.submitted_ids)): + cmd = "htcaas-job-cancel -m %s" % self.submitted_ids[i] + status = misc.Popen([cmd], shell=True, stdout=open(os.devnull,'w')) + +class HTCaaS2Cluster(Cluster): + """Class for dealing with cluster submission on a HTCaaS cluster without GPFS """ + + name= 'htcaas2' + job_id = 'HTCAAS2_JOBID' + idle_tag = ['waiting'] + running_tag = ['preparing','running'] + complete_tag = ['done'] + + @store_input() + @multiple_try() + def submit2(self, prog, argument=[], cwd=None, stdout=None, stderr=None, + log=None, input_files=[], output_files=[], required_output=[], + nb_submit=0): + + """Submit the HTCaaS job on the cluster with NO SHARE DISK + input/output file should be given as relative to CWD + """ + if cwd is None: + cwd = os.getcwd() + + if not os.path.exists(prog): + prog = os.path.join(cwd, prog) + + if 'combine' not in prog and 'pythia' not in prog and 'shower' not in prog : + if cwd or prog : + self.submitted_dirs.append(cwd) + self.submitted_exes.append(prog) + else: + logger.debug("cwd and prog not exist->"+cwd+" / "+ os.path.basename(prog)) + + if argument : + self.submitted_args.append('='.join([str(a) for a in argument])) + + if cwd or prog : + self.submitted += 1 + id = self.submitted + self.submitted_ids.append(id) + else: + logger.debug("cwd and prog are not exist! ") + id = 0 + + else: + temp_file_name = "sub."+ os.path.basename(prog) + text = """#!/bin/bash + MYPWD=%(cwd)s + cd $MYPWD + input_files=(%(input_files)s ) + for i in ${input_files[@]} + do + chmod -f +x $i + done + /bin/bash %(prog)s %(arguments)s > %(stdout)s + """ + dico = {'cwd':cwd, 'input_files': ' '.join(input_files + [prog]), 'stdout': stdout, 'prog':prog, + 'arguments': ' '.join([str(a) for a in argument]), + 'program': ' ' if '.py' in prog else 'bash'} + # writing a new script for the submission + new_prog = pjoin(cwd, temp_file_name) + open(new_prog, 'w').write(text % dico) + misc.Popen(['chmod','+x',new_prog],cwd=cwd) + command = ['htcaas-mgjob-submit','-d',cwd,'-e',new_prog] + a = misc.Popen(command, stdout=subprocess.PIPE, stderr=subprocess.PIPE, stdin=subprocess.PIPE, cwd=cwd) + id = a.stdout.read().strip() + logger.debug("[mode2]-["+str(id)+"]") + if cwd and prog : + self.submitted += 1 + self.submitted_ids.append(id) + else: + logger.debug("cwd and prog are not exist! ") + id = 0 + + return id + + @multiple_try() + def metasubmit(self, me_dir=None): + if self.submitted > 1100 and self.submitted == len(self.submitted_ids): + tmp_leng= len(self.submitted_ids)/2 + tmp_dirs1= self.submitted_dirs[0:tmp_leng] + tmp_dirs2= self.submitted_dirs[tmp_leng:] + tmp_exes1= self.submitted_exes[0:tmp_leng] + tmp_exes2= self.submitted_exes[tmp_leng:] + command1 = ['htcaas-mgjob-submit','-d',":".join([str(a) for a in tmp_dirs1 if a and a != ' ']), + '-e', ":".join([str(a) for a in tmp_exes1 if a and a != ' '])] + command2 = ['htcaas-mgjob-submit','-d',":".join([str(a) for a in tmp_dirs2 if a and a != ' ']), + '-e', ":".join([str(a) for a in tmp_exes2 if a and a != ' '])] + if len(self.submitted_args) > 0 : + tmp_args1= self.submitted_args[0:tmp_leng] + tmp_args2= self.submitted_args[tmp_leng:] + command1.extend(['-a', ':'.join([str(a) for a in tmp_args1])]) + command2.extend(['-a', ':'.join([str(a) for a in tmp_args2])]) + result1 = misc.Popen(command1, stdout=subprocess.PIPE, stderr=subprocess.PIPE, stdin=subprocess.PIPE) + result2 = misc.Popen(command2, stdout=subprocess.PIPE, stderr=subprocess.PIPE, stdin=subprocess.PIPE) + me_dir = str(result1.stdout.read().strip())+ "//" + str(result2.stdout.read().strip()) + + elif self.submitted > 0 and self.submitted == self.submitted_ids[-1]: + command = ['htcaas-mgjob-submit','-d',":".join([str(a) for a in self.submitted_dirs if a and a != ' ']), + '-e', ":".join([str(a) for a in self.submitted_exes if a and a != ' '])] + if len(self.submitted_args) > 0 : + command.extend(['-a', ':'.join([str(a) for a in self.submitted_args])]) + if self.submitted_dirs[0] or self.submitted_exes[0] : + result = misc.Popen(command, stdout=subprocess.PIPE, stderr=subprocess.PIPE, stdin=subprocess.PIPE) + me_dir = result.stdout.read().strip() + self.submitted_ids[0]=me_dir + else: + me_dir = self.submitted_ids[-1] + elif self.submitted > 0 and self.submitted != self.submitted_ids[-1]: + me_dir = self.submitted_ids[0] + else: + me_dir = -1 + + logger.debug("[" + str(me_dir) + "]") + + self.submitted_dirs = [] + self.submitted_exes = [] + self.submitted_args = [] + + return me_dir + + + @multiple_try(nb_try=10, sleep=5) + def control_one_job(self, id): + """ control the status of a single job with it's cluster id """ + #logger.debug("CONTROL ONE JOB MODE") + if self.submitted == self.submitted_ids[-1] : + id = self.metasubmit(self) + tempid = self.submitted_ids[-1] + self.submitted_ids.remove(self.submitted_ids[-1]) + self.submitted_ids.append(id) + logger.debug(str(id)+" // "+str(self.submitted_ids[-1])) + + if id == 0 : + status_out ='C' + else: + cmd = 'htcaas-job-status -m '+ str(id) + " -s | grep Status " + status = misc.Popen([cmd],shell=True,stdout=subprocess.PIPE, + stderr=subprocess.PIPE) + error = status.stderr.read().decode(errors='ignore') + if status.returncode or error: + raise ClusterManagmentError('htcaas-job-status returns error: %s' % error) + status_out= status.stdout.read().decode(errors='ignore').strip() + status_out= status_out.split(":",1)[1] + logger.debug("[["+str(id)+"]]"+status_out) + if status_out == 'waiting': + status_out='I' + elif status_out == 'preparing' or status_out == 'running': + status_out = 'R' + elif status_out != 'done': + status_out = 'F' + elif status_out == 'done': + status_out = 'C' + self.submitted -= 1 + + return status_out + + @multiple_try() + def control(self, me_dir): + """ control the status of a single job with it's cluster id """ + if not self.submitted_ids: + logger.debug("self.submitted_ids not exists") + return 0, 0, 0, 0 + + if "//" in me_dir : + if int(me_dir.split("//")[0]) < int(me_dir.split("//")[1]) : + start = me_dir.split("//")[0] + end = me_dir.split("//")[1] + else : + start = me_dir.split("//")[1] + end = me_dir.split("//")[0] + elif "/" in me_dir : # update + start = 0 + end = 0 + elif me_dir.isdigit(): + start = me_dir + end = me_dir + elif not me_dir.isdigit(): + me_dir = self.submitted_ids[0] + logger.debug("Meta_ID is not digit(control), self.submitted_ids[0]: "+str(me_dir) ) + + ongoing = [] + idle, run, fail, done = 0, 0, 0, 0 + + cmd = "htcaas-job-status -c "+str(start)+"-"+str(end) +" -ac" + status = misc.Popen([cmd], shell=True, stdout=subprocess.PIPE) + + for line in status.stdout: + line = line.decode(errors='ignore') + status2 = line.split()[-1] + if status2 != 'null' or line.split()[0].strip() != '0': + ongoing.append(str(line.split()[0].strip())+"-"+str(line.split()[1].strip())) + logger.debug("["+line.split()[0].strip()+"-"+line.split()[1].strip()+"]"+status2) + + if status2 == 'null' or line.split()[0].strip() == '0': + idle += 1 + elif status2 in self.idle_tag: + idle += 1 + elif status2 in self.running_tag: + run += 1 + elif status2 in self.complete_tag: + done += 1 + self.submitted -= 1 + if not self.check_termination(line.split()[1]): + idle +=1 + else: + fail += 1 + + return idle, run, self.submitted - (idle+run+fail), fail + + @multiple_try() + def remove(self, *args, **opts): + """Clean the jobson the cluster""" + + if not self.submitted_ids: + return + id = self.submitted_ids[0] + if id: + cmd = "htcaas-job-cancel -m %s" % str(id) + status = misc.Popen([cmd], shell=True, stdout=open(os.devnull,'w')) + +from_name = {'condor':CondorCluster, 'pbs': PBSCluster, 'sge': SGECluster, + 'lsf': LSFCluster, 'ge':GECluster, 'slurm': SLURMCluster, + 'htcaas':HTCaaSCluster, 'htcaas2':HTCaaS2Cluster} + +onecore=MultiCore(1) # create a thread to run simple bash job without having to + #fork the main process diff --git a/epochX/cudacpp/tlau/gridpacks/MG5aMC_patches/gen_ximprove.py b/epochX/cudacpp/tlau/gridpacks/MG5aMC_patches/gen_ximprove.py new file mode 100755 index 0000000000..d2a5fe20ba --- /dev/null +++ b/epochX/cudacpp/tlau/gridpacks/MG5aMC_patches/gen_ximprove.py @@ -0,0 +1,2029 @@ +################################################################################ +# +# Copyright (c) 2014 The MadGraph5_aMC@NLO Development team and Contributors +# +# This file is a part of the MadGraph5_aMC@NLO project, an application which +# automatically generates Feynman diagrams and matrix elements for arbitrary +# high-energy processes in the Standard Model and beyond. +# +# It is subject to the MadGraph5_aMC@NLO license which should accompany this +# distribution. +# +# For more information, visit madgraph.phys.ucl.ac.be and amcatnlo.web.cern.ch +# +################################################################################ +""" A python file to replace the fortran script gen_ximprove. + This script analyses the result of the survey/ previous refine and + creates the jobs for the following script. +""" +from __future__ import division + +from __future__ import absolute_import +import collections +import os +import glob +import logging +import math +import re +import subprocess +import shutil +import stat +import sys +import six +import time # for __CUDACPP_DEBUG +from six.moves import range +from six.moves import zip + +try: + import madgraph +except ImportError: + MADEVENT = True + import internal.sum_html as sum_html + import internal.banner as bannermod + import internal.misc as misc + import internal.files as files + import internal.cluster as cluster + import internal.combine_grid as combine_grid + import internal.combine_runs as combine_runs + import internal.lhe_parser as lhe_parser + if six.PY3: + import internal.hel_recycle as hel_recycle +else: + MADEVENT= False + import madgraph.madevent.sum_html as sum_html + import madgraph.various.banner as bannermod + import madgraph.various.misc as misc + import madgraph.iolibs.files as files + import madgraph.various.cluster as cluster + import madgraph.madevent.combine_grid as combine_grid + import madgraph.madevent.combine_runs as combine_runs + import madgraph.various.lhe_parser as lhe_parser + if six.PY3: + import madgraph.madevent.hel_recycle as hel_recycle + +logger = logging.getLogger('madgraph.madevent.gen_ximprove') +pjoin = os.path.join + +class gensym(object): + """a class to call the fortran gensym executable and handle it's output + in order to create the various job that are needed for the survey""" + + #convenient shortcut for the formatting of variable + @ staticmethod + def format_variable(*args): + return bannermod.ConfigFile.format_variable(*args) + + combining_job = 2 # number of channel by ajob + splitted_grid = False + min_iterations = 3 + mode= "survey" + + + def __init__(self, cmd, opt=None): + + try: + super(gensym, self).__init__(cmd, opt) + except TypeError: + pass + + # Run statistics, a dictionary of RunStatistics(), with + self.run_statistics = {} + + self.cmd = cmd + self.run_card = cmd.run_card + self.me_dir = cmd.me_dir + + + # dictionary to keep track of the precision when combining iteration + self.cross = collections.defaultdict(int) + self.abscross = collections.defaultdict(int) + self.sigma = collections.defaultdict(int) + self.chi2 = collections.defaultdict(int) + + self.splitted_grid = False + if self.cmd.proc_characteristics['loop_induced']: + nexternal = self.cmd.proc_characteristics['nexternal'] + self.splitted_grid = max(2, (nexternal-2)**2) + if hasattr(self.cmd, "opts") and self.cmd.opts['accuracy'] == 0.1: + self.cmd.opts['accuracy'] = 0.02 + + if isinstance(cmd.cluster, cluster.MultiCore) and self.splitted_grid > 1: + self.splitted_grid = int(cmd.cluster.nb_core**0.5) + if self.splitted_grid == 1 and cmd.cluster.nb_core >1: + self.splitted_grid = 2 + + #if the user defines it in the run_card: + if self.run_card['survey_splitting'] != -1: + self.splitted_grid = self.run_card['survey_splitting'] + if self.run_card['survey_nchannel_per_job'] != 1 and 'survey_nchannel_per_job' in self.run_card.user_set: + self.combining_job = self.run_card['survey_nchannel_per_job'] + elif self.run_card['hard_survey'] > 1: + self.combining_job = 1 + + + self.splitted_Pdir = {} + self.splitted_for_dir = lambda x,y: self.splitted_grid + self.combining_job_for_Pdir = lambda x: self.combining_job + self.lastoffset = {} + + done_warning_zero_coupling = False + def get_helicity(self, to_submit=True, clean=True): + """launch a single call to madevent to get the list of non zero helicity""" + + self.subproc = [l.strip() for l in open(pjoin(self.me_dir,'SubProcesses', + 'subproc.mg'))] + subproc = self.subproc + P_zero_result = [] + nb_tot_proc = len(subproc) + job_list = {} + + + for nb_proc,subdir in enumerate(subproc): + self.cmd.update_status('Compiling for process %s/%s.' % \ + (nb_proc+1,nb_tot_proc), level=None) + + subdir = subdir.strip() + Pdir = pjoin(self.me_dir, 'SubProcesses',subdir) + logger.info(' %s ' % subdir) + + #compile gensym + self.cmd.compile(['gensym'], cwd=Pdir) + if not os.path.exists(pjoin(Pdir, 'gensym')): + raise Exception('Error make gensym not successful') + + # Launch gensym + p = misc.Popen(['./gensym'], stdout=subprocess.PIPE, + stderr=subprocess.STDOUT, cwd=Pdir) + #sym_input = "%(points)d %(iterations)d %(accuracy)f \n" % self.opts + + (stdout, _) = p.communicate(''.encode()) + stdout = stdout.decode('ascii',errors='ignore') + if stdout: + nb_channel = max([math.floor(float(d)) for d in stdout.split()]) + else: + for matrix_file in misc.glob('matrix*orig.f', Pdir): + files.cp(matrix_file, matrix_file.replace('orig','optim')) + P_zero_result.append(Pdir) + if os.path.exists(pjoin(self.me_dir, 'error')): + os.remove(pjoin(self.me_dir, 'error')) + continue # bypass bad process + + self.cmd.compile(['madevent_forhel'], cwd=Pdir) + if not os.path.exists(pjoin(Pdir, 'madevent_forhel')): + raise Exception('Error make madevent_forhel not successful') + + if not os.path.exists(pjoin(Pdir, 'Hel')): + os.mkdir(pjoin(Pdir, 'Hel')) + ff = open(pjoin(Pdir, 'Hel', 'input_app.txt'),'w') + ff.write('1000 1 1 \n 0.1 \n 2\n 0\n -1\n -%s\n' % nb_channel) + ff.close() + else: + try: + os.remove(pjoin(Pdir, 'Hel','results.dat')) + except Exception: + pass + # Launch gensym + p = misc.Popen(['../madevent_forhel < input_app.txt'], stdout=subprocess.PIPE, + stderr=subprocess.STDOUT, cwd=pjoin(Pdir,'Hel'), shell=True) + #sym_input = "%(points)d %(iterations)d %(accuracy)f \n" % self.opts + (stdout, _) = p.communicate(" ".encode()) + stdout = stdout.decode('ascii',errors='ignore') + if os.path.exists(pjoin(self.me_dir, 'error')): + raise Exception(pjoin(self.me_dir,'error')) + # note a continue is not enough here, we have in top to link + # the matrixX_optim.f to matrixX_orig.f to let the code to work + # after this error. + # for matrix_file in misc.glob('matrix*orig.f', Pdir): + # files.cp(matrix_file, matrix_file.replace('orig','optim')) + + if 'no events passed cuts' in stdout: + raise Exception + + all_zamp = set() + all_hel = set() + zero_gc = list() + all_zampperhel = set() + all_bad_amps_perhel = set() + + for line in stdout.splitlines(): + if "=" not in line and ":" not in line: + continue + if ' GC_' in line: + lsplit = line.split() + if float(lsplit[2]) ==0 == float(lsplit[3]): + zero_gc.append(lsplit[0]) + if 'Matrix Element/Good Helicity:' in line: + all_hel.add(tuple(line.split()[3:5])) + if 'Amplitude/ZEROAMP:' in line: + all_zamp.add(tuple(line.split()[1:3])) + if 'HEL/ZEROAMP:' in line: + nb_mat, nb_hel, nb_amp = line.split()[1:4] + if (nb_mat, nb_hel) not in all_hel: + continue + if (nb_mat,nb_amp) in all_zamp: + continue + all_zampperhel.add(tuple(line.split()[1:4])) + + if zero_gc and not gensym.done_warning_zero_coupling: + gensym.done_warning_zero_coupling = True + logger.warning("The optimizer detects that you have coupling evaluated to zero: \n"+\ + "%s\n" % (' '.join(zero_gc)) +\ + "This will slow down the computation. Please consider using restricted model:\n" +\ + "https://answers.launchpad.net/mg5amcnlo/+faq/2312") + + + all_good_hels = collections.defaultdict(list) + for me_index, hel in all_hel: + all_good_hels[me_index].append(int(hel)) + + #print(all_hel) + if self.run_card['hel_zeroamp']: + all_bad_amps = collections.defaultdict(list) + for me_index, amp in all_zamp: + all_bad_amps[me_index].append(int(amp)) + + all_bad_amps_perhel = collections.defaultdict(list) + for me_index, hel, amp in all_zampperhel: + all_bad_amps_perhel[me_index].append((int(hel),int(amp))) + + elif all_zamp: + nb_zero = sum(int(a[1]) for a in all_zamp) + if zero_gc: + logger.warning("Those zero couplings lead to %s Feynman diagram evaluated to zero (on 10 PS point),\n" % nb_zero +\ + "This part can optimize if you set the flag hel_zeroamp to True in the run_card."+\ + "Note that restricted model will be more optimal.") + else: + logger.warning("The optimization detected that you have %i zero matrix-element for this SubProcess: %s.\n" % nb_zero +\ + "This part can optimize if you set the flag hel_zeroamp to True in the run_card.") + + #check if we need to do something and write associate information" + data = [all_hel, all_zamp, all_bad_amps_perhel] + if not self.run_card['hel_zeroamp']: + data[1] = '' + if not self.run_card['hel_filtering']: + data[0] = '' + data = str(data) + if os.path.exists(pjoin(Pdir,'Hel','selection')): + old_data = open(pjoin(Pdir,'Hel','selection')).read() + if old_data == data: + continue + + + with open(pjoin(Pdir,'Hel','selection'),'w') as fsock: + fsock.write(data) + + + for matrix_file in misc.glob('matrix*orig.f', Pdir): + + split_file = matrix_file.split('/') + me_index = split_file[-1][len('matrix'):-len('_orig.f')] + + basename = split_file[-1].replace('orig', 'optim') + split_out = split_file[:-1] + [basename] + out_file = pjoin('/', '/'.join(split_out)) + + basename = 'template_%s' % split_file[-1].replace("_orig", "") + split_templ = split_file[:-1] + [basename] + templ_file = pjoin('/', '/'.join(split_templ)) + + # Convert to sorted list for reproducibility + #good_hels = sorted(list(good_hels)) + good_hels = [str(x) for x in sorted(all_good_hels[me_index])] + if self.run_card['hel_zeroamp']: + + bad_amps = [str(x) for x in sorted(all_bad_amps[me_index])] + bad_amps_perhel = [x for x in sorted(all_bad_amps_perhel[me_index])] + else: + bad_amps = [] + bad_amps_perhel = [] + if __debug__: + mtext = open(matrix_file).read() + nb_amp = int(re.findall('PARAMETER \(NGRAPHS=(\d+)\)', mtext)[0]) + logger.debug('nb_hel: %s zero amp: %s bad_amps_hel: %s/%s', len(good_hels),len(bad_amps),len(bad_amps_perhel), len(good_hels)*nb_amp ) + if len(good_hels) == 1: + files.cp(matrix_file, matrix_file.replace('orig','optim')) + continue # avoid optimization if onlye one helicity + recycler = hel_recycle.HelicityRecycler(good_hels, bad_amps, bad_amps_perhel) + # In case of bugs you can play around with these: + recycler.hel_filt = self.run_card['hel_filtering'] + recycler.amp_splt = self.run_card['hel_splitamp'] + recycler.amp_filt = self.run_card['hel_zeroamp'] + + recycler.set_input(matrix_file) + recycler.set_output(out_file) + recycler.set_template(templ_file) + recycler.generate_output_file() + del recycler + + # with misc.chdir(): + # pass + + #files.ln(pjoin(Pdir, 'madevent_forhel'), Pdir, name='madevent') ##to be removed + + return {}, P_zero_result + + + def launch(self, to_submit=True, clean=True): + """ """ + + if not hasattr(self, 'subproc'): + self.subproc = [l.strip() for l in open(pjoin(self.me_dir,'SubProcesses', + 'subproc.mg'))] + subproc = self.subproc + + P_zero_result = [] # check the number of times where they are no phase-space + + nb_tot_proc = len(subproc) + job_list = {} + for nb_proc,subdir in enumerate(subproc): + self.cmd.update_status('Compiling for process %s/%s.
(previous processes already running)' % \ + (nb_proc+1,nb_tot_proc), level=None) + + subdir = subdir.strip() + Pdir = pjoin(self.me_dir, 'SubProcesses',subdir) + logger.info(' %s ' % subdir) + + # clean previous run + if clean: + for match in misc.glob('*ajob*', Pdir): + if os.path.basename(match)[:4] in ['ajob', 'wait', 'run.', 'done']: + os.remove(match) + for match in misc.glob('G*', Pdir): + if os.path.exists(pjoin(match,'results.dat')): + os.remove(pjoin(match, 'results.dat')) + if os.path.exists(pjoin(match, 'ftn25')): + os.remove(pjoin(match, 'ftn25')) + + #compile gensym + self.cmd.compile(['gensym'], cwd=Pdir) + if not os.path.exists(pjoin(Pdir, 'gensym')): + raise Exception('Error make gensym not successful') + + # Launch gensym + p = misc.Popen(['./gensym'], stdout=subprocess.PIPE, + stderr=subprocess.STDOUT, cwd=Pdir) + #sym_input = "%(points)d %(iterations)d %(accuracy)f \n" % self.opts + (stdout, _) = p.communicate(''.encode()) + stdout = stdout.decode('ascii',errors='ignore') + if os.path.exists(pjoin(self.me_dir,'error')): + files.mv(pjoin(self.me_dir,'error'), pjoin(Pdir,'ajob.no_ps.log')) + P_zero_result.append(subdir) + continue + + jobs = stdout.split() + job_list[Pdir] = jobs + try: + # check that all input are valid + [float(s) for s in jobs] + except Exception: + logger.debug("unformated string found in gensym. Please check:\n %s" % stdout) + done=False + job_list[Pdir] = [] + lines = stdout.split('\n') + for l in lines: + try: + [float(s) for s in l.split()] + except: + continue + else: + if done: + raise Exception('Parsing error in gensym: %s' % stdout) + job_list[Pdir] = l.split() + done = True + if not done: + raise Exception('Parsing error in gensym: %s' % stdout) + + self.cmd.compile(['madevent'], cwd=Pdir) + if to_submit: + self.submit_to_cluster(job_list) + job_list = {} + + return job_list, P_zero_result + + def resubmit(self, min_precision=1.0, resubmit_zero=False): + """collect the result of the current run and relaunch each channel + not completed or optionally a completed one with a precision worse than + a threshold (and/or the zero result channel)""" + + job_list, P_zero_result = self.launch(to_submit=False, clean=False) + + for P , jobs in dict(job_list).items(): + misc.sprint(jobs) + to_resub = [] + for job in jobs: + if os.path.exists(pjoin(P, 'G%s' % job)) and os.path.exists(pjoin(P, 'G%s' % job, 'results.dat')): + one_result = sum_html.OneResult(job) + try: + one_result.read_results(pjoin(P, 'G%s' % job, 'results.dat')) + except: + to_resub.append(job) + if one_result.xsec == 0: + if resubmit_zero: + to_resub.append(job) + elif max(one_result.xerru, one_result.xerrc)/one_result.xsec > min_precision: + to_resub.append(job) + else: + to_resub.append(job) + if to_resub: + for G in to_resub: + try: + shutil.rmtree(pjoin(P, 'G%s' % G)) + except Exception as error: + misc.sprint(error) + pass + misc.sprint(to_resub) + self.submit_to_cluster({P: to_resub}) + + + + + + + + + + + + def submit_to_cluster(self, job_list): + """ """ + + if self.run_card['job_strategy'] > 0: + if len(job_list) >1: + for path, dirs in job_list.items(): + self.submit_to_cluster({path:dirs}) + return + path, value = list(job_list.items())[0] + nexternal = self.cmd.proc_characteristics['nexternal'] + current = open(pjoin(path, "nexternal.inc")).read() + ext = re.search(r"PARAMETER \(NEXTERNAL=(\d+)\)", current).group(1) + + if self.run_card['job_strategy'] == 2: + self.splitted_grid = 2 + if nexternal == int(ext): + to_split = 2 + else: + to_split = 0 + if hasattr(self, 'splitted_Pdir'): + self.splitted_Pdir[path] = to_split + else: + self.splitted_Pdir = {path: to_split} + self.splitted_for_dir = lambda x,y : self.splitted_Pdir[x] + elif self.run_card['job_strategy'] == 1: + if nexternal == int(ext): + combine = 1 + else: + combine = self.combining_job + if hasattr(self, 'splitted_Pdir'): + self.splitted_Pdir[path] = combine + else: + self.splitted_Pdir = {path: combine} + self.combining_job_for_Pdir = lambda x : self.splitted_Pdir[x] + + if not self.splitted_grid: + return self.submit_to_cluster_no_splitting(job_list) + elif self.cmd.cluster_mode == 0: + return self.submit_to_cluster_no_splitting(job_list) + elif self.cmd.cluster_mode == 2 and self.cmd.options['nb_core'] == 1: + return self.submit_to_cluster_no_splitting(job_list) + else: + return self.submit_to_cluster_splitted(job_list) + + + def submit_to_cluster_no_splitting(self, job_list): + """submit the survey without the parralelization. + This is the old mode which is still usefull in single core""" + + # write the template file for the parameter file + self.write_parameter(parralelization=False, Pdirs=list(job_list.keys())) + + + # launch the job with the appropriate grouping + for Pdir, jobs in job_list.items(): + jobs = list(jobs) + i=0 + while jobs: + i+=1 + to_submit = ['0'] # the first entry is actually the offset + for _ in range(self.combining_job_for_Pdir(Pdir)): + if jobs: + to_submit.append(jobs.pop(0)) + + self.cmd.launch_job(pjoin(self.me_dir, 'SubProcesses', 'survey.sh'), + argument=to_submit, + cwd=pjoin(self.me_dir,'SubProcesses' , Pdir)) + + + def create_resubmit_one_iter(self, Pdir, G, submit_ps, nb_job, step=0): + """prepare the input_file for submitting the channel""" + + + if 'SubProcesses' not in Pdir: + Pdir = pjoin(self.me_dir, 'SubProcesses', Pdir) + + #keep track of how many job are sended + self.splitted_Pdir[(Pdir, G)] = int(nb_job) + + + # 1. write the new input_app.txt + run_card = self.cmd.run_card + options = {'event' : submit_ps, + 'maxiter': 1, + 'miniter': 1, + 'accuracy': self.cmd.opts['accuracy'], + 'helicity': run_card['nhel_survey'] if 'nhel_survey' in run_card \ + else run_card['nhel'], + 'gridmode': -2, + 'channel' : G + } + + Gdir = pjoin(Pdir, 'G%s' % G) + self.write_parameter_file(pjoin(Gdir, 'input_app.txt'), options) + + # 2. check that ftn25 exists. + assert os.path.exists(pjoin(Gdir, "ftn25")) + + + # 3. Submit the new jobs + #call back function + packet = cluster.Packet((Pdir, G, step+1), + self.combine_iteration, + (Pdir, G, step+1)) + + if step ==0: + self.lastoffset[(Pdir, G)] = 0 + + # resubmit the new jobs + for i in range(int(nb_job)): + name = "G%s_%s" % (G,i+1) + self.lastoffset[(Pdir, G)] += 1 + offset = self.lastoffset[(Pdir, G)] + self.cmd.launch_job(pjoin(self.me_dir, 'SubProcesses', 'refine_splitted.sh'), + argument=[name, 'G%s'%G, offset], + cwd= Pdir, + packet_member=packet) + + + def submit_to_cluster_splitted(self, job_list): + """ submit the version of the survey with splitted grid creation + """ + + #if self.splitted_grid <= 1: + # return self.submit_to_cluster_no_splitting(job_list) + + for Pdir, jobs in job_list.items(): + if not jobs: + continue + if self.splitted_for_dir(Pdir, jobs[0]) <= 1: + return self.submit_to_cluster_no_splitting({Pdir:jobs}) + + self.write_parameter(parralelization=True, Pdirs=[Pdir]) + # launch the job with the appropriate grouping + + for job in jobs: + packet = cluster.Packet((Pdir, job, 1), self.combine_iteration, (Pdir, job, 1)) + for i in range(self.splitted_for_dir(Pdir, job)): + self.cmd.launch_job(pjoin(self.me_dir, 'SubProcesses', 'survey.sh'), + argument=[i+1, job], + cwd=pjoin(self.me_dir,'SubProcesses' , Pdir), + packet_member=packet) + + def combine_iteration(self, Pdir, G, step): + + grid_calculator, cross, error = self.combine_grid(Pdir, G, step) + + # Compute the number of events used for this run. + nb_events = grid_calculator.target_evt + + Gdirs = [] #build the the list of directory + for i in range(self.splitted_for_dir(Pdir, G)): + path = pjoin(Pdir, "G%s_%s" % (G, i+1)) + Gdirs.append(path) + + # 4. make the submission of the next iteration + # Three cases - less than 3 iteration -> continue + # - more than 3 and less than 5 -> check error + # - more than 5 -> prepare info for refine + need_submit = False + if step < self.min_iterations and cross != 0: + if step == 1: + need_submit = True + else: + across = self.abscross[(Pdir,G)]/(self.sigma[(Pdir,G)]+1e-99) + tot_across = self.get_current_axsec() + if across / tot_across < 1e-6: + need_submit = False + elif error < self.cmd.opts['accuracy'] / 100: + need_submit = False + else: + need_submit = True + + elif step >= self.cmd.opts['iterations']: + need_submit = False + elif self.cmd.opts['accuracy'] < 0: + #check for luminosity + raise Exception("Not Implemented") + elif self.abscross[(Pdir,G)] == 0: + need_submit = False + else: + across = self.abscross[(Pdir,G)]/(self.sigma[(Pdir,G)]+1e-99) + tot_across = self.get_current_axsec() + if across == 0: + need_submit = False + elif across / tot_across < 1e-5: + need_submit = False + elif error > self.cmd.opts['accuracy']: + need_submit = True + else: + need_submit = False + + + if cross: + grid_calculator.write_grid_for_submission(Pdir,G, + self.splitted_for_dir(Pdir, G), + nb_events,mode=self.mode, + conservative_factor=5.0) + + xsec_format = '.%ig'%(max(3,int(math.log10(1.0/float(error)))+2) + if float(cross)!=0.0 and float(error)!=0.0 else 8) + if need_submit: + message = "%%s/G%%s is at %%%s +- %%.3g pb. Now submitting iteration #%s."%(xsec_format, step+1) + logger.info(message%\ + (os.path.basename(Pdir), G, float(cross), + float(error)*float(cross))) + self.resubmit_survey(Pdir,G, Gdirs, step) + elif cross: + logger.info("Survey finished for %s/G%s at %s"%( + os.path.basename(Pdir),G,('%%%s +- %%.3g pb'%xsec_format))% + (float(cross), float(error)*float(cross))) + # prepare information for refine + newGpath = pjoin(self.me_dir,'SubProcesses' , Pdir, 'G%s' % G) + if not os.path.exists(newGpath): + os.mkdir(newGpath) + + # copy the new grid: + files.cp(pjoin(Gdirs[0], 'ftn25'), + pjoin(self.me_dir,'SubProcesses' , Pdir, 'G%s' % G, 'ftn26')) + + # copy the events + fsock = open(pjoin(newGpath, 'events.lhe'), 'w') + for Gdir in Gdirs: + fsock.write(open(pjoin(Gdir, 'events.lhe')).read()) + + # copy one log + files.cp(pjoin(Gdirs[0], 'log.txt'), + pjoin(self.me_dir,'SubProcesses' , Pdir, 'G%s' % G)) + + + # create the appropriate results.dat + self.write_results(grid_calculator, cross, error, Pdir, G, step) + else: + logger.info("Survey finished for %s/G%s [0 cross]", os.path.basename(Pdir),G) + + Gdir = pjoin(self.me_dir,'SubProcesses' , Pdir, 'G%s' % G) + if not os.path.exists(Gdir): + os.mkdir(Gdir) + # copy one log + files.cp(pjoin(Gdirs[0], 'log.txt'), Gdir) + # create the appropriate results.dat + self.write_results(grid_calculator, cross, error, Pdir, G, step) + + return 0 + + def combine_grid(self, Pdir, G, step, exclude_sub_jobs=[]): + """ exclude_sub_jobs is to remove some of the subjobs if a numerical + issue is detected in one of them. Warning is issue when this occurs. + """ + + # 1. create an object to combine the grid information and fill it + grid_calculator = combine_grid.grid_information(self.run_card['nhel']) + + for i in range(self.splitted_for_dir(Pdir, G)): + if i in exclude_sub_jobs: + continue + path = pjoin(Pdir, "G%s_%s" % (G, i+1)) + fsock = misc.mult_try_open(pjoin(path, 'results.dat')) + one_result = grid_calculator.add_results_information(fsock) + fsock.close() + if one_result.axsec == 0: + grid_calculator.onefail = True + continue # grid_information might not exists + fsock = misc.mult_try_open(pjoin(path, 'grid_information')) + grid_calculator.add_one_grid_information(fsock) + fsock.close() + os.remove(pjoin(path, 'results.dat')) + #os.remove(pjoin(path, 'grid_information')) + + + + #2. combine the information about the total crossection / error + # start by keep the interation in memory + cross, across, sigma = grid_calculator.get_cross_section() + + #3. Try to avoid one single PS point which ruins the integration + # Should be related to loop evaluation instability. + maxwgt = grid_calculator.get_max_wgt(0.01) + if maxwgt: + nunwgt = grid_calculator.get_nunwgt(maxwgt) + # Make sure not to apply the security below during the first step of the + # survey. Also, disregard channels with a contribution relative to the + # total cross-section smaller than 1e-8 since in this case it is unlikely + # that this channel will need more than 1 event anyway. + apply_instability_security = False + rel_contrib = 0.0 + if (self.__class__ != gensym or step > 1): + Pdir_across = 0.0 + Gdir_across = 0.0 + for (mPdir,mG) in self.abscross.keys(): + if mPdir == Pdir: + Pdir_across += (self.abscross[(mPdir,mG)]/ + (self.sigma[(mPdir,mG)]+1e-99)) + if mG == G: + Gdir_across += (self.abscross[(mPdir,mG)]/ + (self.sigma[(mPdir,mG)]+1e-99)) + rel_contrib = abs(Gdir_across/(Pdir_across+1e-99)) + if rel_contrib > (1.0e-8) and \ + nunwgt < 2 and len(grid_calculator.results) > 1: + apply_instability_security = True + + if apply_instability_security: + # check the ratio between the different submit + th_maxwgt = [(r.th_maxwgt,i) for i,r in enumerate(grid_calculator.results)] + th_maxwgt.sort() + ratio = th_maxwgt[-1][0]/th_maxwgt[-2][0] + if ratio > 1e4: + logger.warning( +""""One Event with large weight have been found (ratio = %.3g) in channel G%s (with rel.contrib=%.3g). +This is likely due to numerical instabilities. The associated job is discarded to recover. +For offline investigation, the problematic discarded events are stored in: +%s"""%(ratio,G,rel_contrib,pjoin(Pdir,'DiscardedUnstableEvents'))) + exclude_sub_jobs = list(exclude_sub_jobs) + exclude_sub_jobs.append(th_maxwgt[-1][1]) + grid_calculator.results.run_statistics['skipped_subchannel'] += 1 + + # Add some monitoring of the problematic events + gPath = pjoin(Pdir, "G%s_%s" % (G, th_maxwgt[-1][1]+1)) + if os.path.isfile(pjoin(gPath,'events.lhe')): + lhe_file = lhe_parser.EventFile(pjoin(gPath,'events.lhe')) + discardedPath = pjoin(Pdir,'DiscardedUnstableEvents') + if not os.path.exists(discardedPath): + os.mkdir(discardedPath) + if os.path.isdir(discardedPath): + # Keep only the event with a maximum weight, as it surely + # is the problematic one. + evtRecord = open(pjoin(discardedPath,'discarded_G%s.dat'%G),'a') + lhe_file.seek(0) #rewind the file + try: + evtRecord.write('\n'+str(max(lhe_file,key=lambda evt:abs(evt.wgt)))) + except Exception: + #something wrong write the full file. + lhe_file.close() + evtRecord.write(pjoin(gPath,'events.lhe').read()) + evtRecord.close() + + return self.combine_grid(Pdir, G, step, exclude_sub_jobs) + + + if across !=0: + if sigma != 0: + self.cross[(Pdir,G)] += cross**3/sigma**2 + self.abscross[(Pdir,G)] += across * cross**2/sigma**2 + self.sigma[(Pdir,G)] += cross**2/ sigma**2 + self.chi2[(Pdir,G)] += cross**4/sigma**2 + # and use those iteration to get the current estimator + cross = self.cross[(Pdir,G)]/self.sigma[(Pdir,G)] + if step > 1: + error = math.sqrt(abs((self.chi2[(Pdir,G)]/cross**2 - \ + self.sigma[(Pdir,G)])/(step-1))/self.sigma[(Pdir,G)]) + else: + error = sigma/cross + else: + self.cross[(Pdir,G)] = cross + self.abscross[(Pdir,G)] = across + self.sigma[(Pdir,G)] = 0 + self.chi2[(Pdir,G)] = 0 + cross = self.cross[(Pdir,G)] + error = 0 + + else: + error = 0 + + grid_calculator.results.compute_values(update_statistics=True) + if (str(os.path.basename(Pdir)), G) in self.run_statistics: + self.run_statistics[(str(os.path.basename(Pdir)), G)]\ + .aggregate_statistics(grid_calculator.results.run_statistics) + else: + self.run_statistics[(str(os.path.basename(Pdir)), G)] = \ + grid_calculator.results.run_statistics + + self.warnings_from_statistics(G, grid_calculator.results.run_statistics) + stats_msg = grid_calculator.results.run_statistics.nice_output( + '/'.join([os.path.basename(Pdir),'G%s'%G])) + + if stats_msg: + logger.log(5, stats_msg) + + # Clean up grid_information to avoid border effects in case of a crash + for i in range(self.splitted_for_dir(Pdir, G)): + path = pjoin(Pdir, "G%s_%s" % (G, i+1)) + try: + os.remove(pjoin(path, 'grid_information')) + except OSError as oneerror: + if oneerror.errno != 2: + raise + return grid_calculator, cross, error + + def warnings_from_statistics(self,G,stats): + """Possible warn user for worrying MadLoop stats for this channel.""" + + if stats['n_madloop_calls']==0: + return + + EPS_fraction = float(stats['exceptional_points'])/stats['n_madloop_calls'] + + msg = "Channel %s has encountered a fraction of %.3g\n"+ \ + "of numerically unstable loop matrix element computations\n"+\ + "(which could not be rescued using quadruple precision).\n"+\ + "The results might not be trusted." + + if 0.01 > EPS_fraction > 0.001: + logger.warning(msg%(G,EPS_fraction)) + elif EPS_fraction > 0.01: + logger.critical((msg%(G,EPS_fraction)).replace('might', 'can')) + raise Exception((msg%(G,EPS_fraction)).replace('might', 'can')) + + def get_current_axsec(self): + + across = 0 + for (Pdir,G) in self.abscross: + across += self.abscross[(Pdir,G)]/(self.sigma[(Pdir,G)]+1e-99) + return across + + def write_results(self, grid_calculator, cross, error, Pdir, G, step): + + #compute the value + if cross == 0: + abscross,nw, luminosity = 0, 0, 0 + wgt, maxit,nunwgt, wgt, nevents = 0,0,0,0,0 + maxwgt = 0 + error = 0 + else: + grid_calculator.results.compute_values() + abscross = self.abscross[(Pdir,G)]/self.sigma[(Pdir,G)] + nw = grid_calculator.results.nw + wgt = grid_calculator.results.wgt + maxit = step + wgt = 0 + nevents = grid_calculator.results.nevents + maxwgt = grid_calculator.get_max_wgt() + nunwgt = grid_calculator.get_nunwgt() + luminosity = nunwgt/cross + + #format the results.dat + def fstr(nb): + data = '%E' % nb + nb, power = data.split('E') + nb = float(nb) /10 + power = int(power) + 1 + return '%.5fE%+03i' %(nb,power) + line = '%s %s %s %i %i %i %i %s %s %s %s 0.0 0\n' % \ + (fstr(cross), fstr(error*cross), fstr(error*cross), + nevents, nw, maxit,nunwgt, + fstr(luminosity), fstr(wgt), fstr(abscross), fstr(maxwgt)) + + fsock = open(pjoin(self.me_dir,'SubProcesses' , Pdir, 'G%s' % G, + 'results.dat'),'w') + fsock.writelines(line) + fsock.close() + + def resubmit_survey(self, Pdir, G, Gdirs, step): + """submit the next iteration of the survey""" + + # 1. write the new input_app.txt to double the number of points + run_card = self.cmd.run_card + options = {'event' : 2**(step) * self.cmd.opts['points'] / self.splitted_grid, + 'maxiter': 1, + 'miniter': 1, + 'accuracy': self.cmd.opts['accuracy'], + 'helicity': run_card['nhel_survey'] if 'nhel_survey' in run_card \ + else run_card['nhel'], + 'gridmode': -2, + 'channel' : '' + } + + if int(options['helicity']) == 1: + options['event'] = options['event'] * 2**(self.cmd.proc_characteristics['nexternal']//3) + + for Gdir in Gdirs: + self.write_parameter_file(pjoin(Gdir, 'input_app.txt'), options) + + + #2. resubmit the new jobs + packet = cluster.Packet((Pdir, G, step+1), self.combine_iteration, \ + (Pdir, G, step+1)) + nb_step = len(Gdirs) * (step+1) + for i,subdir in enumerate(Gdirs): + subdir = subdir.rsplit('_',1)[1] + subdir = int(subdir) + offset = nb_step+i+1 + offset=str(offset) + tag = "%s.%s" % (subdir, offset) + + self.cmd.launch_job(pjoin(self.me_dir, 'SubProcesses', 'survey.sh'), + argument=[tag, G], + cwd=pjoin(self.me_dir,'SubProcesses' , Pdir), + packet_member=packet) + + + + + def write_parameter_file(self, path, options): + """ """ + + template =""" 8192 1 1 !Number of events and max and min iterations + %(accuracy)s !Accuracy + %(gridmode)s !Grid Adjustment 0=none, 2=adjust + 1 !Suppress Amplitude 1=yes + %(helicity)s !Helicity Sum/event 0=exact + %(channel)s """ + options['event'] = int(options['event']) + open(path, 'w').write(template % options) + + + + def write_parameter(self, parralelization, Pdirs=None): + """Write the parameter of the survey run""" + + run_card = self.cmd.run_card + + options = {'event' : self.cmd.opts['points'], + 'maxiter': self.cmd.opts['iterations'], + 'miniter': self.min_iterations, + 'accuracy': self.cmd.opts['accuracy'], + 'helicity': run_card['nhel_survey'] if 'nhel_survey' in run_card \ + else run_card['nhel'], + 'gridmode': 2, + 'channel': '' + } + + if int(options['helicity'])== 1: + options['event'] = options['event'] * 2**(self.cmd.proc_characteristics['nexternal']//3) + + if parralelization: + options['gridmode'] = -2 + options['maxiter'] = 1 #this is automatic in dsample anyway + options['miniter'] = 1 #this is automatic in dsample anyway + options['event'] /= self.splitted_grid + + if not Pdirs: + Pdirs = self.subproc + + for Pdir in Pdirs: + path =pjoin(Pdir, 'input_app.txt') + self.write_parameter_file(path, options) + + + +class gen_ximprove(object): + + + # some hardcoded value which impact the generation + gen_events_security = 1.2 # multiply the number of requested event by this number for security + combining_job = 0 # allow to run multiple channel in sequence + max_request_event = 1000 # split jobs if a channel if it needs more than that + max_event_in_iter = 5000 + min_event_in_iter = 1000 + max_splitting = 130 # maximum duplication of a given channel + min_iter = 3 + max_iter = 9 + keep_grid_for_refine = False # only apply if needed to split the job + + #convenient shortcut for the formatting of variable + @ staticmethod + def format_variable(*args): + return bannermod.ConfigFile.format_variable(*args) + + + def __new__(cls, cmd, opt): + """Choose in which type of refine we want to be""" + + if hasattr(cls, 'force_class'): + if cls.force_class == 'gridpack': + return super(gen_ximprove, cls).__new__(gen_ximprove_gridpack) + elif cls.force_class == 'loop_induced': + return super(gen_ximprove, cls).__new__(gen_ximprove_share) + + if cmd.proc_characteristics['loop_induced']: + return super(gen_ximprove, cls).__new__(gen_ximprove_share) + elif gen_ximprove.format_variable(cmd.run_card['gridpack'], bool): + return super(gen_ximprove, cls).__new__(gen_ximprove_gridpack) + elif cmd.run_card["job_strategy"] == 2: + return super(gen_ximprove, cls).__new__(gen_ximprove_share) + else: + return super(gen_ximprove, cls).__new__(gen_ximprove_v4) + + + def __init__(self, cmd, opt=None): + + try: + super(gen_ximprove, self).__init__(cmd, opt) + except TypeError: + pass + + self.run_statistics = {} + self.cmd = cmd + self.run_card = cmd.run_card + run_card = self.run_card + self.me_dir = cmd.me_dir + + #extract from the run_card the information that we need. + self.gridpack = run_card['gridpack'] + self.nhel = run_card['nhel'] + if "nhel_refine" in run_card: + self.nhel = run_card["nhel_refine"] + + if self.run_card['refine_evt_by_job'] != -1: + self.max_request_event = run_card['refine_evt_by_job'] + + + # Default option for the run + self.gen_events = True + self.parralel = False + # parameter which was input for the normal gen_ximprove run + self.err_goal = 0.01 + self.max_np = 9 + self.split_channels = False + # parameter for the gridpack run + self.nreq = 2000 + self.iseed = 4321 + + # placeholder for information + self.results = 0 #updated in launch/update_html + + if isinstance(opt, dict): + self.configure(opt) + elif isinstance(opt, bannermod.GridpackCard): + self.configure_gridpack(opt) + + def __call__(self): + return self.launch() + + def launch(self): + """running """ + print("__CUDACPP_DEBUG: gen_ximprove.launch starting") + cudacpp_start = time.perf_counter() + #start the run + self.handle_seed() + self.results = sum_html.collect_result(self.cmd, + main_dir=pjoin(self.cmd.me_dir,'SubProcesses')) #main_dir is for gridpack readonly mode + if self.gen_events: + # We run to provide a given number of events + self.get_job_for_event() + else: + # We run to achieve a given precision + self.get_job_for_precision() + cudacpp_end = time.perf_counter() + cudacpp_length = cudacpp_end - cudacpp_start + print("__CUDACPP_DEBUG: gen_ximprove.launch finished in %.4f seconds"%cudacpp_length) + + + def configure(self, opt): + """Defines some parameter of the run""" + + for key, value in opt.items(): + if key in self.__dict__: + targettype = type(getattr(self, key)) + setattr(self, key, self.format_variable(value, targettype, key)) + else: + raise Exception('%s not define' % key) + + + # special treatment always do outside the loop to avoid side effect + if 'err_goal' in opt: + if self.err_goal < 1: + logger.info("running for accuracy %s%%" % (self.err_goal*100)) + self.gen_events = False + elif self.err_goal >= 1: + logger.info("Generating %s unweighted events." % self.err_goal) + self.gen_events = True + self.err_goal = self.err_goal * self.gen_events_security # security + + def handle_seed(self): + """not needed but for gridpack --which is not handle here for the moment""" + return + + + def find_job_for_event(self): + """return the list of channel that need to be improved""" + + assert self.err_goal >=1 + self.err_goal = int(self.err_goal) + + goal_lum = self.err_goal/(self.results.axsec+1e-99) #pb^-1 + logger.info('Effective Luminosity %s pb^-1', goal_lum) + + all_channels = sum([list(P) for P in self.results],[]) + all_channels.sort(key= lambda x:x.get('luminosity'), reverse=True) + + to_refine = [] + for C in all_channels: + if C.get('axsec') == 0: + continue + if goal_lum/(C.get('luminosity')+1e-99) >= 1 + (self.gen_events_security-1)/2: + logger.debug("channel %s need to improve by %.2f (xsec=%s pb, iter=%s)", C.name, goal_lum/(C.get('luminosity')+1e-99), C.get('xsec'), int(C.get('maxit'))) + to_refine.append(C) + elif C.get('xerr') > max(C.get('axsec'), + (1/(100*math.sqrt(self.err_goal)))*all_channels[-1].get('axsec')): + to_refine.append(C) + + logger.info('need to improve %s channels' % len(to_refine)) + return goal_lum, to_refine + + def update_html(self): + """update the html from this object since it contains all the information""" + + + run = self.cmd.results.current['run_name'] + if not os.path.exists(pjoin(self.cmd.me_dir, 'HTML', run)): + os.mkdir(pjoin(self.cmd.me_dir, 'HTML', run)) + + unit = self.cmd.results.unit + P_text = "" + if self.results: + Presults = self.results + else: + self.results = sum_html.collect_result(self.cmd, None) + Presults = self.results + + for P_comb in Presults: + P_text += P_comb.get_html(run, unit, self.cmd.me_dir) + + Presults.write_results_dat(pjoin(self.cmd.me_dir,'SubProcesses', 'results.dat')) + + fsock = open(pjoin(self.cmd.me_dir, 'HTML', run, 'results.html'),'w') + fsock.write(sum_html.results_header) + fsock.write('%s
' % Presults.get_html(run, unit, self.cmd.me_dir)) + fsock.write('%s
' % P_text) + + self.cmd.results.add_detail('cross', Presults.xsec) + self.cmd.results.add_detail('error', Presults.xerru) + + return Presults.xsec, Presults.xerru + + +class gen_ximprove_v4(gen_ximprove): + + # some hardcoded value which impact the generation + gen_events_security = 1.2 # multiply the number of requested event by this number for security + combining_job = 0 # allow to run multiple channel in sequence + max_request_event = 1000 # split jobs if a channel if it needs more than that + max_event_in_iter = 5000 + min_event_in_iter = 1000 + max_splitting = 130 # maximum duplication of a given channel + min_iter = 3 + max_iter = 9 + keep_grid_for_refine = False # only apply if needed to split the job + + + + def __init__(self, cmd, opt=None): + + super(gen_ximprove_v4, self).__init__(cmd, opt) + + if cmd.opts['accuracy'] < cmd._survey_options['accuracy'][1]: + self.increase_precision(cmd._survey_options['accuracy'][1]/cmd.opts['accuracy']) + + def reset_multijob(self): + + for path in misc.glob(pjoin('*', '*','multijob.dat'), pjoin(self.me_dir, 'SubProcesses')): + open(path,'w').write('0\n') + + def write_multijob(self, Channel, nb_split): + """ """ + if nb_split <=1: + return + f = open(pjoin(self.me_dir, 'SubProcesses', Channel.get('name'), 'multijob.dat'), 'w') + f.write('%i\n' % nb_split) + f.close() + + def increase_precision(self, rate=3): + #misc.sprint(rate) + if rate < 3: + self.max_event_in_iter = 20000 + self.min_events = 7500 + self.gen_events_security = 1.3 + else: + rate = rate -2 + self.max_event_in_iter = int((rate+1) * 10000) + self.min_events = int(rate+2) * 2500 + self.gen_events_security = 1 + 0.1 * (rate+2) + + if int(self.nhel) == 1: + self.min_event_in_iter *= 2**(self.cmd.proc_characteristics['nexternal']//3) + self.max_event_in_iter *= 2**(self.cmd.proc_characteristics['nexternal']//2) + + + + alphabet = "abcdefghijklmnopqrstuvwxyz" + def get_job_for_event(self): + """generate the script in order to generate a given number of event""" + # correspond to write_gen in the fortran version + + + goal_lum, to_refine = self.find_job_for_event() + + #reset the potential multijob of previous run + self.reset_multijob() + + jobs = [] # list of the refine if some job are split is list of + # dict with the parameter of the run. + + # try to have a smart load on the cluster (not really important actually) + if self.combining_job >1: + # add a nice ordering for the jobs + new_order = [] + if self.combining_job % 2 == 0: + for i in range(len(to_refine) //2): + new_order.append(to_refine[i]) + new_order.append(to_refine[-i-1]) + if len(to_refine) % 2: + new_order.append(to_refine[i+1]) + else: + for i in range(len(to_refine) //3): + new_order.append(to_refine[i]) + new_order.append(to_refine[-2*i-1]) + new_order.append(to_refine[-2*i-2]) + if len(to_refine) % 3 == 1: + new_order.append(to_refine[i+1]) + elif len(to_refine) % 3 == 2: + new_order.append(to_refine[i+2]) + #ensure that the reordering is done nicely + assert set([id(C) for C in to_refine]) == set([id(C) for C in new_order]) + to_refine = new_order + + + # loop over the channel to refine + for C in to_refine: + #1. Compute the number of points are needed to reach target + needed_event = goal_lum*C.get('axsec') + nb_split = int(max(1,((needed_event-1)// self.max_request_event) +1)) + if not self.split_channels: + nb_split = 1 + if nb_split > self.max_splitting: + nb_split = self.max_splitting + nb_split=max(1, nb_split) + + + #2. estimate how many points we need in each iteration + if C.get('nunwgt') > 0: + nevents = needed_event / nb_split * (C.get('nevents') / C.get('nunwgt')) + #split by iter + nevents = int(nevents / (2**self.min_iter-1)) + else: + nevents = self.max_event_in_iter + + if nevents < self.min_event_in_iter: + nb_split = int(nb_split * nevents / self.min_event_in_iter) + 1 + nevents = self.min_event_in_iter + # + # forbid too low/too large value + nevents = max(self.min_event_in_iter, min(self.max_event_in_iter, nevents)) + logger.debug("%s : need %s event. Need %s split job of %s points", C.name, needed_event, nb_split, nevents) + + + # write the multi-job information + self.write_multijob(C, nb_split) + + packet = cluster.Packet((C.parent_name, C.name), + combine_runs.CombineRuns, + (pjoin(self.me_dir, 'SubProcesses', C.parent_name)), + {"subproc": C.name, "nb_split":nb_split}) + + + #create the info dict assume no splitting for the default + info = {'name': self.cmd.results.current['run_name'], + 'script_name': 'unknown', + 'directory': C.name, # need to be change for splitted job + 'P_dir': C.parent_name, + #'Ppath': pjoin(self.cmd.me_dir, 'SubProcesses', C.parent_name), + 'offset': 1, # need to be change for splitted job + 'nevents': nevents, + 'maxiter': self.max_iter, + 'miniter': self.min_iter, + 'precision': -goal_lum/nb_split, + 'nhel': self.run_card['nhel'], + 'channel': C.name.replace('G',''), + 'grid_refinment' : 0, #no refinment of the grid + 'base_directory': '', #should be change in splitted job if want to keep the grid + 'packet': packet, + } + + if nb_split == 1: + jobs.append(info) + else: + for i in range(nb_split): + new_info = dict(info) + new_info['offset'] = i+1 + new_info['directory'] += self.alphabet[i % 26] + str((i+1)//26) + if self.keep_grid_for_refine: + new_info['base_directory'] = info['directory'] + jobs.append(new_info) + + self.create_ajob(pjoin(self.me_dir, 'SubProcesses', 'refine.sh'), jobs) + + + def create_ajob(self, template, jobs, write_dir=None): + """create the ajob""" + + if not jobs: + return + + if not write_dir: + write_dir = pjoin(self.me_dir, 'SubProcesses') + + #filter the job according to their SubProcess directory # no mix submition + P2job= collections.defaultdict(list) + for j in jobs: + P2job[j['P_dir']].append(j) + if len(P2job) >1: + for P in P2job.values(): + self.create_ajob(template, P, write_dir) + return + + + #Here we can assume that all job are for the same directory. + path = pjoin(write_dir, jobs[0]['P_dir']) + + template_text = open(template, 'r').read() + # special treatment if needed to combine the script + # computes how many submition miss one job + if self.combining_job > 1: + skip1=0 + n_channels = len(jobs) + nb_sub = n_channels // self.combining_job + nb_job_in_last = n_channels % self.combining_job + if nb_sub == 0: + nb_sub = 1 + nb_job_in_last =0 + if nb_job_in_last: + nb_sub +=1 + skip1 = self.combining_job - nb_job_in_last + if skip1 > nb_sub: + self.combining_job -=1 + return self.create_ajob(template, jobs, write_dir) + combining_job = self.combining_job + else: + #define the variable for combining jobs even in not combine mode + #such that we can use the same routine + skip1=0 + combining_job =1 + nb_sub = len(jobs) + + + nb_use = 0 + for i in range(nb_sub): + script_number = i+1 + if i < skip1: + nb_job = combining_job -1 + else: + nb_job = min(combining_job, len(jobs)) + fsock = open(pjoin(path, 'ajob%i' % script_number), 'w') + for j in range(nb_use, nb_use + nb_job): + if j> len(jobs): + break + info = jobs[j] + info['script_name'] = 'ajob%i' % script_number + ###info['keeplog'] = 'false' + info['keeplog'] = 'true' # for __CUDACPP_DEBUG + if "base_directory" not in info: + info["base_directory"] = "./" + fsock.write(template_text % info) + nb_use += nb_job + + fsock.close() + return script_number + + def get_job_for_precision(self): + """create the ajob to achieve a give precision on the total cross-section""" + + + assert self.err_goal <=1 + xtot = abs(self.results.xsec) + logger.info("Working on precision: %s %%" %(100*self.err_goal)) + all_channels = sum([list(P) for P in self.results if P.mfactor],[]) + limit = self.err_goal * xtot / len(all_channels) + to_refine = [] + rerr = 0 #error of the job not directly selected + for C in all_channels: + cerr = C.mfactor*(C.xerru + len(all_channels)*C.xerrc) + if cerr > abs(limit): + to_refine.append(C) + else: + rerr += cerr + rerr *=rerr + if not len(to_refine): + return + + # change limit since most don't contribute + limit = math.sqrt((self.err_goal * xtot)**2 - rerr/math.sqrt(len(to_refine))) + for C in to_refine[:]: + cerr = C.mfactor*(C.xerru + len(to_refine)*C.xerrc) + if cerr < limit: + to_refine.remove(C) + + # all the channel are now selected. create the channel information + logger.info('need to improve %s channels' % len(to_refine)) + + + jobs = [] # list of the refine if some job are split is list of + # dict with the parameter of the run. + + # loop over the channel to refine + for C in to_refine: + + #1. Determine how many events we need in each iteration + yerr = C.mfactor*(C.xerru+len(to_refine)*C.xerrc) + nevents = 0.2*C.nevents*(yerr/limit)**2 + + nb_split = int((nevents*(C.nunwgt/C.nevents)/self.max_request_event/ (2**self.min_iter-1))**(2/3)) + nb_split = max(nb_split, 1) + # **(2/3) to slow down the increase in number of jobs + if nb_split > self.max_splitting: + nb_split = self.max_splitting + + if nb_split >1: + nevents = nevents / nb_split + self.write_multijob(C, nb_split) + # forbid too low/too large value + nevents = min(self.min_event_in_iter, max(self.max_event_in_iter, nevents)) + + + #create the info dict assume no splitting for the default + info = {'name': self.cmd.results.current['run_name'], + 'script_name': 'unknown', + 'directory': C.name, # need to be change for splitted job + 'P_dir': C.parent_name, + #'Ppath': pjoin(self.cmd.me_dir, 'SubProcesses', C.parent_name), + 'offset': 1, # need to be change for splitted job + 'nevents': nevents, + 'maxiter': self.max_iter, + 'miniter': self.min_iter, + 'precision': yerr/math.sqrt(nb_split)/(C.get('xsec')+ yerr), + 'nhel': self.run_card['nhel'], + 'channel': C.name.replace('G',''), + 'grid_refinment' : 1 + } + + if nb_split == 1: + jobs.append(info) + else: + for i in range(nb_split): + new_info = dict(info) + new_info['offset'] = i+1 + new_info['directory'] += self.alphabet[i % 26] + str((i+1)//26) + jobs.append(new_info) + self.create_ajob(pjoin(self.me_dir, 'SubProcesses', 'refine.sh'), jobs) + + def update_html(self): + """update the html from this object since it contains all the information""" + + + run = self.cmd.results.current['run_name'] + if not os.path.exists(pjoin(self.cmd.me_dir, 'HTML', run)): + os.mkdir(pjoin(self.cmd.me_dir, 'HTML', run)) + + unit = self.cmd.results.unit + P_text = "" + if self.results: + Presults = self.results + else: + self.results = sum_html.collect_result(self.cmd, None) + Presults = self.results + + for P_comb in Presults: + P_text += P_comb.get_html(run, unit, self.cmd.me_dir) + + Presults.write_results_dat(pjoin(self.cmd.me_dir,'SubProcesses', 'results.dat')) + + fsock = open(pjoin(self.cmd.me_dir, 'HTML', run, 'results.html'),'w') + fsock.write(sum_html.results_header) + fsock.write('%s
' % Presults.get_html(run, unit, self.cmd.me_dir)) + fsock.write('%s
' % P_text) + + self.cmd.results.add_detail('cross', Presults.xsec) + self.cmd.results.add_detail('error', Presults.xerru) + + return Presults.xsec, Presults.xerru + + + + +class gen_ximprove_v4_nogridupdate(gen_ximprove_v4): + + # some hardcoded value which impact the generation + gen_events_security = 1.1 # multiply the number of requested event by this number for security + combining_job = 0 # allow to run multiple channel in sequence + max_request_event = 400 # split jobs if a channel if it needs more than that + max_event_in_iter = 500 + min_event_in_iter = 250 + max_splitting = 260 # maximum duplication of a given channel + min_iter = 2 + max_iter = 6 + keep_grid_for_refine = True + + + def __init__(self, cmd, opt=None): + + gen_ximprove.__init__(cmd, opt) + + if cmd.proc_characteristics['loopinduced'] and \ + cmd.proc_characteristics['nexternal'] > 2: + self.increase_parralelization(cmd.proc_characteristics['nexternal']) + + def increase_parralelization(self, nexternal): + + self.max_splitting = 1000 + + if self.run_card['refine_evt_by_job'] != -1: + pass + elif nexternal == 3: + self.max_request_event = 200 + elif nexternal == 4: + self.max_request_event = 100 + elif nexternal >= 5: + self.max_request_event = 50 + self.min_event_in_iter = 125 + self.max_iter = 5 + +class gen_ximprove_share(gen_ximprove, gensym): + """Doing the refine in multicore. Each core handle a couple of PS point.""" + + nb_ps_by_job = 2000 + mode = "refine" + gen_events_security = 1.15 + # Note the real security is lower since we stop the jobs if they are at 96% + # of this target. + + def __init__(self, *args, **opts): + + super(gen_ximprove_share, self).__init__(*args, **opts) + self.generated_events = {} + self.splitted_for_dir = lambda x,y : self.splitted_Pdir[(x,y)] + + + def get_job_for_event(self): + """generate the script in order to generate a given number of event""" + # correspond to write_gen in the fortran version + + + goal_lum, to_refine = self.find_job_for_event() + self.goal_lum = goal_lum + + # loop over the channel to refine to find the number of PS point to launch + total_ps_points = 0 + channel_to_ps_point = [] + for C in to_refine: + #0. remove previous events files + try: + os.remove(pjoin(self.me_dir, "SubProcesses",C.parent_name, C.name, "events.lhe")) + except: + pass + + #1. Compute the number of points are needed to reach target + needed_event = goal_lum*C.get('axsec') + if needed_event == 0: + continue + #2. estimate how many points we need in each iteration + if C.get('nunwgt') > 0: + nevents = needed_event * (C.get('nevents') / C.get('nunwgt')) + #split by iter + nevents = int(nevents / (2**self.min_iter-1)) + else: + nb_split = int(max(1,((needed_event-1)// self.max_request_event) +1)) + if not self.split_channels: + nb_split = 1 + if nb_split > self.max_splitting: + nb_split = self.max_splitting + nevents = self.max_event_in_iter * self.max_splitting + else: + nevents = self.max_event_in_iter * nb_split + + if nevents > self.max_splitting*self.max_event_in_iter: + logger.warning("Channel %s/%s has a very low efficiency of unweighting. Might not be possible to reach target" % \ + (C.name, C.parent_name)) + nevents = self.max_event_in_iter * self.max_splitting + + total_ps_points += nevents + channel_to_ps_point.append((C, nevents)) + + if self.cmd.options["run_mode"] == 1: + if self.cmd.options["cluster_size"]: + nb_ps_by_job = total_ps_points /int(self.cmd.options["cluster_size"]) + else: + nb_ps_by_job = self.nb_ps_by_job + elif self.cmd.options["run_mode"] == 2: + remain = total_ps_points % self.cmd.options["nb_core"] + if remain: + nb_ps_by_job = 1 + (total_ps_points - remain) / self.cmd.options["nb_core"] + else: + nb_ps_by_job = total_ps_points / self.cmd.options["nb_core"] + else: + nb_ps_by_job = self.nb_ps_by_job + + nb_ps_by_job = int(max(nb_ps_by_job, 500)) + + for C, nevents in channel_to_ps_point: + if nevents % nb_ps_by_job: + nb_job = 1 + int(nevents // nb_ps_by_job) + else: + nb_job = int(nevents // nb_ps_by_job) + submit_ps = min(nevents, nb_ps_by_job) + if nb_job == 1: + submit_ps = max(submit_ps, self.min_event_in_iter) + self.create_resubmit_one_iter(C.parent_name, C.name[1:], submit_ps, nb_job, step=0) + needed_event = goal_lum*C.get('xsec') + logger.debug("%s/%s : need %s event. Need %s split job of %s points", C.parent_name, C.name, needed_event, nb_job, submit_ps) + + + def combine_iteration(self, Pdir, G, step): + + grid_calculator, cross, error = self.combine_grid(Pdir, G, step) + + # collect all the generated_event + Gdirs = [] #build the the list of directory + for i in range(self.splitted_for_dir(Pdir, G)): + path = pjoin(Pdir, "G%s_%s" % (G, i+1)) + Gdirs.append(path) + assert len(grid_calculator.results) == len(Gdirs) == self.splitted_for_dir(Pdir, G) + + + # Check how many events are going to be kept after un-weighting. + needed_event = cross * self.goal_lum + if needed_event == 0: + return 0 + # check that the number of events requested is not higher than the actual + # total number of events to generate. + if self.err_goal >=1: + if needed_event > self.gen_events_security * self.err_goal: + needed_event = int(self.gen_events_security * self.err_goal) + + if (Pdir, G) in self.generated_events: + old_nunwgt, old_maxwgt = self.generated_events[(Pdir, G)] + else: + old_nunwgt, old_maxwgt = 0, 0 + + if old_nunwgt == 0 and os.path.exists(pjoin(Pdir,"G%s" % G, "events.lhe")): + # possible for second refine. + lhe = lhe_parser.EventFile(pjoin(Pdir,"G%s" % G, "events.lhe")) + old_nunwgt = lhe.unweight(None, trunc_error=0.005, log_level=0) + old_maxwgt = lhe.max_wgt + + + + maxwgt = max(grid_calculator.get_max_wgt(), old_maxwgt) + new_evt = grid_calculator.get_nunwgt(maxwgt) + efficiency = new_evt / sum([R.nevents for R in grid_calculator.results]) + nunwgt = old_nunwgt * old_maxwgt / maxwgt + nunwgt += new_evt + + # check the number of event for this iteration alone + one_iter_nb_event = max(grid_calculator.get_nunwgt(),1) + drop_previous_iteration = False + # compare the number of events to generate if we discard the previous iteration + n_target_one_iter = (needed_event-one_iter_nb_event) / ( one_iter_nb_event/ sum([R.nevents for R in grid_calculator.results])) + n_target_combined = (needed_event-nunwgt) / efficiency + if n_target_one_iter < n_target_combined: + # the last iteration alone has more event that the combine iteration. + # it is therefore interesting to drop previous iteration. + drop_previous_iteration = True + nunwgt = one_iter_nb_event + maxwgt = grid_calculator.get_max_wgt() + new_evt = nunwgt + efficiency = ( one_iter_nb_event/ sum([R.nevents for R in grid_calculator.results])) + + try: + if drop_previous_iteration: + raise IOError + output_file = open(pjoin(Pdir,"G%s" % G, "events.lhe"), 'a') + except IOError: + output_file = open(pjoin(Pdir,"G%s" % G, "events.lhe"), 'w') + + misc.call(["cat"] + [pjoin(d, "events.lhe") for d in Gdirs], + stdout=output_file) + output_file.close() + # For large number of iteration. check the number of event by doing the + # real unweighting. + if nunwgt < 0.6 * needed_event and step > self.min_iter: + lhe = lhe_parser.EventFile(output_file.name) + old_nunwgt =nunwgt + nunwgt = lhe.unweight(None, trunc_error=0.01, log_level=0) + + + self.generated_events[(Pdir, G)] = (nunwgt, maxwgt) + + # misc.sprint("Adding %s event to %s. Currently at %s" % (new_evt, G, nunwgt)) + # check what to do + if nunwgt >= int(0.96*needed_event)+1: # 0.96*1.15=1.10 =real security + # We did it. + logger.info("found enough event for %s/G%s" % (os.path.basename(Pdir), G)) + self.write_results(grid_calculator, cross, error, Pdir, G, step, efficiency) + return 0 + elif step >= self.max_iter: + logger.debug("fail to find enough event") + self.write_results(grid_calculator, cross, error, Pdir, G, step, efficiency) + return 0 + + nb_split_before = len(grid_calculator.results) + nevents = grid_calculator.results[0].nevents + if nevents == 0: # possible if some integral returns 0 + nevents = max(g.nevents for g in grid_calculator.results) + + need_ps_point = (needed_event - nunwgt)/(efficiency+1e-99) + need_job = need_ps_point // nevents + 1 + + if step < self.min_iter: + # This is normal but check if we are on the good track + job_at_first_iter = nb_split_before/2**(step-1) + expected_total_job = job_at_first_iter * (2**self.min_iter-1) + done_job = job_at_first_iter * (2**step-1) + expected_remaining_job = expected_total_job - done_job + + logger.debug("efficiency status (smaller is better): %s", need_job/expected_remaining_job) + # increase if needed but not too much + need_job = min(need_job, expected_remaining_job*1.25) + + nb_job = (need_job-0.5)//(2**(self.min_iter-step)-1) + 1 + nb_job = max(1, nb_job) + grid_calculator.write_grid_for_submission(Pdir,G, + self.splitted_for_dir(Pdir, G), nb_job*nevents ,mode=self.mode, + conservative_factor=self.max_iter) + logger.info("%s/G%s is at %i/%i (%.2g%%) event. Resubmit %i job at iteration %i." \ + % (os.path.basename(Pdir), G, int(nunwgt),int(needed_event)+1, + (float(nunwgt)/needed_event)*100.0 if needed_event>0.0 else 0.0, + nb_job, step)) + self.create_resubmit_one_iter(Pdir, G, nevents, nb_job, step) + #self.create_job(Pdir, G, nb_job, nevents, step) + + elif step < self.max_iter: + if step + 1 == self.max_iter: + need_job = 1.20 * need_job # avoid to have just too few event. + + nb_job = int(min(need_job, nb_split_before*1.5)) + grid_calculator.write_grid_for_submission(Pdir,G, + self.splitted_for_dir(Pdir, G), nb_job*nevents ,mode=self.mode, + conservative_factor=self.max_iter) + + + logger.info("%s/G%s is at %i/%i ('%.2g%%') event. Resubmit %i job at iteration %i." \ + % (os.path.basename(Pdir), G, int(nunwgt),int(needed_event)+1, + (float(nunwgt)/needed_event)*100.0 if needed_event>0.0 else 0.0, + nb_job, step)) + self.create_resubmit_one_iter(Pdir, G, nevents, nb_job, step) + + + + return 0 + + + def write_results(self, grid_calculator, cross, error, Pdir, G, step, efficiency): + + #compute the value + if cross == 0: + abscross,nw, luminosity = 0, 0, 0 + wgt, maxit,nunwgt, wgt, nevents = 0,0,0,0,0 + error = 0 + else: + grid_calculator.results.compute_values() + abscross = self.abscross[(Pdir,G)]/self.sigma[(Pdir,G)] + nunwgt, wgt = self.generated_events[(Pdir, G)] + nw = int(nunwgt / efficiency) + nunwgt = int(nunwgt) + maxit = step + nevents = nunwgt + # make the unweighting to compute the number of events: + luminosity = nunwgt/cross + + #format the results.dat + def fstr(nb): + data = '%E' % nb + nb, power = data.split('E') + nb = float(nb) /10 + power = int(power) + 1 + return '%.5fE%+03i' %(nb,power) + line = '%s %s %s %i %i %i %i %s %s %s 0.0 0.0 0\n' % \ + (fstr(cross), fstr(error*cross), fstr(error*cross), + nevents, nw, maxit,nunwgt, + fstr(luminosity), fstr(wgt), fstr(abscross)) + + fsock = open(pjoin(self.me_dir,'SubProcesses' , Pdir, 'G%s' % G, + 'results.dat'),'w') + fsock.writelines(line) + fsock.close() + + + + +class gen_ximprove_gridpack(gen_ximprove_v4): + min_iter = 1 + max_iter = 13 + max_request_event = 1e12 # split jobs if a channel if it needs more than that + max_event_in_iter = 4000 + min_event_in_iter = 500 + combining_job = sys.maxsize + gen_events_security = 1.00 + + def __new__(cls, *args, **opts): + + cls.force_class = 'gridpack' + return super(gen_ximprove_gridpack, cls).__new__(cls, *args, **opts) + + def __init__(self, *args, **opts): + + self.ngran = -1 + self.gscalefact = {} + self.readonly = False + if 'ngran' in opts: + self.gran = opts['ngran'] +# del opts['ngran'] + if 'readonly' in opts: + self.readonly = opts['readonly'] + super(gen_ximprove_gridpack,self).__init__(*args, **opts) + if self.ngran == -1: + self.ngran = 1 + + def find_job_for_event(self): + """return the list of channel that need to be improved""" + import random + + assert self.err_goal >=1 + self.err_goal = int(self.err_goal) + self.gscalefact = {} + + xtot = self.results.axsec + goal_lum = self.err_goal/(xtot+1e-99) #pb^-1 +# logger.info('Effective Luminosity %s pb^-1', goal_lum) + + all_channels = sum([list(P) for P in self.results],[]) + all_channels.sort(key=lambda x : x.get('luminosity'), reverse=True) + + to_refine = [] + for C in all_channels: + tag = C.get('name') + self.gscalefact[tag] = 0 + R = random.random() + if C.get('axsec') == 0: + continue + if (goal_lum * C.get('axsec') < R*self.ngran ): + continue # no event to generate events + self.gscalefact[tag] = max(1, 1/(goal_lum * C.get('axsec')/ self.ngran)) + #need to generate events + logger.debug('request events for ', C.get('name'), 'cross=', + C.get('axsec'), 'needed events = ', goal_lum * C.get('axsec')) + to_refine.append(C) + + logger.info('need to improve %s channels' % len(to_refine)) + return goal_lum, to_refine + + def get_job_for_event(self): + """generate the script in order to generate a given number of event""" + # correspond to write_gen in the fortran version + print("__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event starting") + cudacpp_start = time.perf_counter() + + goal_lum, to_refine = self.find_job_for_event() + + jobs = [] # list of the refine if some job are split is list of + # dict with the parameter of the run. + + # loop over the channel to refine + for C in to_refine: + #1. Compute the number of points are needed to reach target + needed_event = max(goal_lum*C.get('axsec'), self.ngran) + nb_split = 1 + + #2. estimate how many points we need in each iteration + if C.get('nunwgt') > 0: + nevents = needed_event / nb_split * (C.get('nevents') / C.get('nunwgt')) + #split by iter + nevents = int(nevents / (2**self.min_iter-1)) + else: + nevents = self.max_event_in_iter + + if nevents < self.min_event_in_iter: + nevents = self.min_event_in_iter + # + # forbid too low/too large value + nevents = max(self.min_event_in_iter, min(self.max_event_in_iter, nevents)) + logger.debug("%s : need %s event. Need %s split job of %s points", C.name, needed_event, nb_split, nevents) + + + #create the info dict assume no splitting for the default + info = {'name': self.cmd.results.current['run_name'], + 'script_name': 'unknown', + 'directory': C.name, # need to be change for splitted job + 'P_dir': os.path.basename(C.parent_name), + 'offset': 1, # need to be change for splitted job + #'Ppath': pjoin(self.cmd.me_dir, 'SubProcesses', C.parent_name), + 'nevents': nevents, #int(nevents*self.gen_events_security)+1, + 'maxiter': self.max_iter, + 'miniter': self.min_iter, + 'precision': -1*int(needed_event)/C.get('axsec'), + 'requested_event': needed_event, + 'nhel': self.run_card['nhel'], + 'channel': C.name.replace('G',''), + 'grid_refinment' : 0, #no refinment of the grid + 'base_directory': '', #should be change in splitted job if want to keep the grid + 'packet': None, + } + + if self.readonly: + basedir = pjoin(os.path.dirname(__file__), '..','..','SubProcesses', info['P_dir'], info['directory']) + info['base_directory'] = basedir + + jobs.append(info) + + + write_dir = '.' if self.readonly else None + self.create_ajob(pjoin(self.me_dir, 'SubProcesses', 'refine.sh'), jobs, write_dir) + + done = [] + for j in jobs: + if j['P_dir'] in done: + continue + done.append(j['P_dir']) + # set the working directory path. + pwd = pjoin(os.getcwd(),j['P_dir']) if self.readonly else pjoin(self.me_dir, 'SubProcesses', j['P_dir']) + exe = pjoin(pwd, 'ajob1') + st = os.stat(exe) + os.chmod(exe, st.st_mode | stat.S_IEXEC) + + # run the code\ + print("__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event will call launch_and_wait '" + exe + "'") + cluster.onecore.launch_and_wait(exe, cwd=pwd, packet_member=j['packet']) + print("__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event back from launch_and_wait '" + exe + "'") + for log in glob.glob(pwd+'/G*/GridRun_21_app.log'): + print(log) + with open(log, 'r') as f: + for line in f.readlines(): + if 'COUNTERS' in line: + print(line, end='') # line already contains '\n' + print("__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished dumping logs from '" + exe + "'") + write_dir = '.' if self.readonly else pjoin(self.me_dir, 'SubProcesses') + + self.check_events(goal_lum, to_refine, jobs, write_dir) + cudacpp_end = time.perf_counter() + cudacpp_length = cudacpp_end - cudacpp_start + print("__CUDACPP_DEBUG: gen_ximprove_gridpack.get_job_for_event finished in %.4f seconds"%cudacpp_length) + + def check_events(self, goal_lum, to_refine, jobs, Sdir): + """check that we get the number of requested events if not resubmit.""" + + new_jobs = [] + + for C, job_info in zip(to_refine, jobs): + P = job_info['P_dir'] + G = job_info['channel'] + axsec = C.get('axsec') + requested_events= job_info['requested_event'] + + + new_results = sum_html.OneResult((P,G)) + new_results.read_results(pjoin(Sdir,P, 'G%s'%G, 'results.dat')) + + # need to resubmit? + if new_results.get('nunwgt') < requested_events: + pwd = pjoin(os.getcwd(),job_info['P_dir'],'G%s'%G) if self.readonly else \ + pjoin(self.me_dir, 'SubProcesses', job_info['P_dir'],'G%s'%G) + job_info['requested_event'] -= new_results.get('nunwgt') + job_info['precision'] -= -1*job_info['requested_event']/axsec + job_info['offset'] += 1 + new_jobs.append(job_info) + files.mv(pjoin(pwd, 'events.lhe'), pjoin(pwd, 'events.lhe.previous')) + + if new_jobs: + self.create_ajob(pjoin(self.me_dir, 'SubProcesses', 'refine.sh'), new_jobs, Sdir) + + done = [] + for j in new_jobs: + if j['P_dir'] in done: + continue + G = j['channel'] + # set the working directory path. + pwd = pjoin(os.getcwd(),j['P_dir']) if self.readonly \ + else pjoin(self.me_dir, 'SubProcesses', j['P_dir']) + exe = pjoin(pwd, 'ajob1') + st = os.stat(exe) + os.chmod(exe, st.st_mode | stat.S_IEXEC) + + # run the code + cluster.onecore.launch_and_wait(exe, cwd=pwd, packet_member=j['packet']) + pwd = pjoin(pwd, 'G%s'%G) + # concatanate with old events file + files.put_at_end(pjoin(pwd, 'events.lhe'),pjoin(pwd, 'events.lhe.previous')) + + return self.check_events(goal_lum, to_refine, new_jobs, Sdir) + + + + + + diff --git a/epochX/cudacpp/tlau/gridpacks/MG5aMC_patches/madevent_interface.py b/epochX/cudacpp/tlau/gridpacks/MG5aMC_patches/madevent_interface.py new file mode 100755 index 0000000000..9236ac6b12 --- /dev/null +++ b/epochX/cudacpp/tlau/gridpacks/MG5aMC_patches/madevent_interface.py @@ -0,0 +1,7472 @@ +############################################################################### +# +# Copyright (c) 2011 The MadGraph5_aMC@NLO Development team and Contributors +# +# This file is a part of the MadGraph5_aMC@NLO project, an application which +# automatically generates Feynman diagrams and matrix elements for arbitrary +# high-energy processes in the Standard Model and beyond. +# +# It is subject to the MadGraph5_aMC@NLO license which should accompany this +# distribution. +# +# For more information, visit madgraph.phys.ucl.ac.be and amcatnlo.web.cern.ch +# +################################################################################ +"""A user friendly command line interface to access MadGraph5_aMC@NLO features. + Uses the cmd package for command interpretation and tab completion. +""" +from __future__ import division + +from __future__ import absolute_import +import collections +import itertools +import glob +import logging +import math +import os +import random +import re + +import stat +import subprocess +import sys +import time +import tarfile +import shutil +import copy +from six.moves import range +import six +StringIO = six +try: + import readline + GNU_SPLITTING = ('GNU' in readline.__doc__) +except: + GNU_SPLITTING = True + +root_path = os.path.split(os.path.dirname(os.path.realpath( __file__ )))[0] +root_path = os.path.split(root_path)[0] +if __name__ == '__main__': + sys.path.insert(0, os.path.join(root_path,'bin')) + +# usefull shortcut +pjoin = os.path.join +# Special logger for the Cmd Interface +logger = logging.getLogger('madevent.stdout') # -> stdout +logger_stderr = logging.getLogger('madevent.stderr') # ->stderr + +try: + import madgraph +except ImportError as error: + # import from madevent directory + MADEVENT = True + import internal.extended_cmd as cmd + import internal.common_run_interface as common_run + import internal.banner as banner_mod + import internal.misc as misc + from internal import InvalidCmd, MadGraph5Error, ReadWrite + import internal.files as files + import internal.gen_crossxhtml as gen_crossxhtml + import internal.gen_ximprove as gen_ximprove + import internal.save_load_object as save_load_object + import internal.cluster as cluster + import internal.check_param_card as check_param_card + import internal.sum_html as sum_html + import internal.combine_runs as combine_runs + import internal.lhe_parser as lhe_parser +# import internal.histograms as histograms # imported later to not slow down the loading of the code + from internal.files import ln +else: + # import from madgraph directory + MADEVENT = False + import madgraph.interface.extended_cmd as cmd + import madgraph.interface.common_run_interface as common_run + import madgraph.iolibs.files as files + import madgraph.iolibs.save_load_object as save_load_object + import madgraph.madevent.gen_crossxhtml as gen_crossxhtml + import madgraph.madevent.gen_ximprove as gen_ximprove + import madgraph.madevent.sum_html as sum_html + import madgraph.various.banner as banner_mod + import madgraph.various.cluster as cluster + import madgraph.various.misc as misc + import madgraph.madevent.combine_runs as combine_runs + import madgraph.various.lhe_parser as lhe_parser +# import madgraph.various.histograms as histograms # imported later to not slow down the loading of the code + import models.check_param_card as check_param_card + from madgraph.iolibs.files import ln + from madgraph import InvalidCmd, MadGraph5Error, MG5DIR, ReadWrite + + + +class MadEventError(Exception): pass +ZeroResult = common_run.ZeroResult +class SysCalcError(InvalidCmd): pass + +MadEventAlreadyRunning = common_run.MadEventAlreadyRunning + +#=============================================================================== +# CmdExtended +#=============================================================================== +class CmdExtended(common_run.CommonRunCmd): + """Particularisation of the cmd command for MadEvent""" + + #suggested list of command + next_possibility = { + 'start': [], + } + + debug_output = 'ME5_debug' + error_debug = 'Please report this bug on https://bugs.launchpad.net/mg5amcnlo\n' + error_debug += 'More information is found in \'%(debug)s\'.\n' + error_debug += 'Please attach this file to your report.' + + config_debug = 'If you need help with this issue please contact us on https://answers.launchpad.net/mg5amcnlo\n' + + + keyboard_stop_msg = """stopping all operation + in order to quit MadGraph5_aMC@NLO please enter exit""" + + # Define the Error + InvalidCmd = InvalidCmd + ConfigurationError = MadGraph5Error + + def __init__(self, me_dir, options, *arg, **opt): + """Init history and line continuation""" + + # Tag allowing/forbiding question + self.force = False + + # If possible, build an info line with current version number + # and date, from the VERSION text file + info = misc.get_pkg_info() + info_line = "" + if info and 'version' in info and 'date' in info: + len_version = len(info['version']) + len_date = len(info['date']) + if len_version + len_date < 30: + info_line = "#* VERSION %s %s %s *\n" % \ + (info['version'], + (30 - len_version - len_date) * ' ', + info['date']) + else: + version = open(pjoin(root_path,'MGMEVersion.txt')).readline().strip() + info_line = "#* VERSION %s %s *\n" % \ + (version, (24 - len(version)) * ' ') + + # Create a header for the history file. + # Remember to fill in time at writeout time! + self.history_header = \ + '#************************************************************\n' + \ + '#* MadGraph5_aMC@NLO/MadEvent *\n' + \ + '#* *\n' + \ + "#* * * *\n" + \ + "#* * * * * *\n" + \ + "#* * * * * 5 * * * * *\n" + \ + "#* * * * * *\n" + \ + "#* * * *\n" + \ + "#* *\n" + \ + "#* *\n" + \ + info_line + \ + "#* *\n" + \ + "#* The MadGraph5_aMC@NLO Development Team - Find us at *\n" + \ + "#* https://server06.fynu.ucl.ac.be/projects/madgraph *\n" + \ + '#* *\n' + \ + '#************************************************************\n' + \ + '#* *\n' + \ + '#* Command File for MadEvent *\n' + \ + '#* *\n' + \ + '#* run as ./bin/madevent.py filename *\n' + \ + '#* *\n' + \ + '#************************************************************\n' + + if info_line: + info_line = info_line[1:] + + logger.info(\ + "************************************************************\n" + \ + "* *\n" + \ + "* W E L C O M E to *\n" + \ + "* M A D G R A P H 5 _ a M C @ N L O *\n" + \ + "* M A D E V E N T *\n" + \ + "* *\n" + \ + "* * * *\n" + \ + "* * * * * *\n" + \ + "* * * * * 5 * * * * *\n" + \ + "* * * * * *\n" + \ + "* * * *\n" + \ + "* *\n" + \ + info_line + \ + "* *\n" + \ + "* The MadGraph5_aMC@NLO Development Team - Find us at *\n" + \ + "* https://server06.fynu.ucl.ac.be/projects/madgraph *\n" + \ + "* *\n" + \ + "* Type 'help' for in-line help. *\n" + \ + "* *\n" + \ + "************************************************************") + super(CmdExtended, self).__init__(me_dir, options, *arg, **opt) + + def get_history_header(self): + """return the history header""" + return self.history_header % misc.get_time_info() + + def stop_on_keyboard_stop(self): + """action to perform to close nicely on a keyboard interupt""" + try: + if hasattr(self, 'cluster'): + logger.info('rm jobs on queue') + self.cluster.remove() + if hasattr(self, 'results'): + self.update_status('Stop by the user', level=None, makehtml=False, error=True) + self.add_error_log_in_html(KeyboardInterrupt) + except: + pass + + def postcmd(self, stop, line): + """ Update the status of the run for finishing interactive command """ + + stop = super(CmdExtended, self).postcmd(stop, line) + # relaxing the tag forbidding question + self.force = False + + if not self.use_rawinput: + return stop + + if self.results and not self.results.current: + return stop + + arg = line.split() + if len(arg) == 0: + return stop + if isinstance(self.results.status, str) and self.results.status.startswith('Error'): + return stop + if isinstance(self.results.status, str) and self.results.status == 'Stop by the user': + self.update_status('%s Stop by the user' % arg[0], level=None, error=True) + return stop + elif not self.results.status: + return stop + elif str(arg[0]) in ['exit','quit','EOF']: + return stop + + try: + self.update_status('Command \'%s\' done.
Waiting for instruction.' % arg[0], + level=None, error=True) + except Exception: + misc.sprint('update_status fails') + pass + + + def nice_user_error(self, error, line): + """If a ME run is currently running add a link in the html output""" + + self.add_error_log_in_html() + return cmd.Cmd.nice_user_error(self, error, line) + + def nice_config_error(self, error, line): + """If a ME run is currently running add a link in the html output""" + + self.add_error_log_in_html() + stop = cmd.Cmd.nice_config_error(self, error, line) + + + try: + debug_file = open(self.debug_output, 'a') + debug_file.write(open(pjoin(self.me_dir,'Cards','proc_card_mg5.dat'))) + debug_file.close() + except: + pass + return stop + + + def nice_error_handling(self, error, line): + """If a ME run is currently running add a link in the html output""" + + if isinstance(error, ZeroResult): + self.add_error_log_in_html(error) + logger.warning('Zero result detected: %s' % error) + # create a banner if needed + try: + if not self.banner: + self.banner = banner_mod.Banner() + if 'slha' not in self.banner: + self.banner.add(pjoin(self.me_dir,'Cards','param_card.dat')) + if 'mgruncard' not in self.banner: + self.banner.add(pjoin(self.me_dir,'Cards','run_card.dat')) + if 'mg5proccard' not in self.banner: + proc_card = pjoin(self.me_dir,'Cards','proc_card_mg5.dat') + if os.path.exists(proc_card): + self.banner.add(proc_card) + + out_dir = pjoin(self.me_dir, 'Events', self.run_name) + if not os.path.isdir(out_dir): + os.mkdir(out_dir) + output_path = pjoin(out_dir, '%s_%s_banner.txt' % \ + (self.run_name, self.run_tag)) + self.banner.write(output_path) + except Exception: + if __debug__: + raise + else: + pass + else: + self.add_error_log_in_html() + stop = cmd.Cmd.nice_error_handling(self, error, line) + try: + debug_file = open(self.debug_output, 'a') + debug_file.write(open(pjoin(self.me_dir,'Cards','proc_card_mg5.dat'))) + debug_file.close() + except: + pass + return stop + + +#=============================================================================== +# HelpToCmd +#=============================================================================== +class HelpToCmd(object): + """ The Series of help routine for the MadEventCmd""" + + def help_pythia(self): + logger.info("syntax: pythia [RUN] [--run_options]") + logger.info("-- run pythia on RUN (current one by default)") + self.run_options_help([('-f','answer all question by default'), + ('--tag=', 'define the tag for the pythia run'), + ('--no_default', 'not run if pythia_card not present')]) + + def help_pythia8(self): + logger.info("syntax: pythia8 [RUN] [--run_options]") + logger.info("-- run pythia8 on RUN (current one by default)") + self.run_options_help([('-f','answer all question by default'), + ('--tag=', 'define the tag for the pythia8 run'), + ('--no_default', 'not run if pythia8_card not present')]) + + def help_rivet(self): + logger.info("syntax: rivet [RUN] [--run_options]") + logger.info("-- run rivet on RUN (current one by default)") + self.run_options_help([('-f','answer all question by default'), + ('--tag=', 'define the tag for the rivet run'), + ('--no_default', 'not run if rivet_card not present')]) + + def help_banner_run(self): + logger.info("syntax: banner_run Path|RUN [--run_options]") + logger.info("-- Reproduce a run following a given banner") + logger.info(" One of the following argument is require:") + logger.info(" Path should be the path of a valid banner.") + logger.info(" RUN should be the name of a run of the current directory") + self.run_options_help([('-f','answer all question by default'), + ('--name=X', 'Define the name associated with the new run')]) + + def help_open(self): + logger.info("syntax: open FILE ") + logger.info("-- open a file with the appropriate editor.") + logger.info(' If FILE belongs to index.html, param_card.dat, run_card.dat') + logger.info(' the path to the last created/used directory is used') + logger.info(' The program used to open those files can be chosen in the') + logger.info(' configuration file ./input/mg5_configuration.txt') + + + def run_options_help(self, data): + if data: + logger.info('-- local options:') + for name, info in data: + logger.info(' %s : %s' % (name, info)) + + logger.info("-- session options:") + logger.info(" Note that those options will be kept for the current session") + logger.info(" --cluster : Submit to the cluster. Current cluster: %s" % self.options['cluster_type']) + logger.info(" --multicore : Run in multi-core configuration") + logger.info(" --nb_core=X : limit the number of core to use to X.") + + + def help_generate_events(self): + logger.info("syntax: generate_events [run_name] [options]",) + logger.info("-- Launch the full chain of script for the generation of events") + logger.info(" Including possible plotting, shower and detector resolution.") + logger.info(" Those steps are performed if the related program are installed") + logger.info(" and if the related card are present in the Cards directory.") + self.run_options_help([('-f', 'Use default for all questions.'), + ('--laststep=', 'argument might be parton/pythia/pgs/delphes and indicate the last level to be run.'), + ('-M', 'in order to add MadSpin'), + ('-R', 'in order to add the reweighting module')]) + + def help_initMadLoop(self): + logger.info("syntax: initMadLoop [options]",'$MG:color:GREEN') + logger.info( +"""-- Command only useful when MadEvent simulates loop-induced processes. This command compiles and run + the MadLoop output for the matrix element computation so as to initialize the filter for analytically + zero helicity configurations and loop topologies. If you suspect that a change you made in the model + parameters can have affected these filters, this command allows you to automatically refresh them. """) + logger.info(" The available options are:",'$MG:color:BLUE') + logger.info(" -f : Bypass the edition of MadLoopParams.dat.",'$MG:color:BLUE') + logger.info(" -r : Refresh of the existing filters (erasing them if already present).",'$MG:color:BLUE') + logger.info(" --nPS= : Specify how many phase-space points should be tried to set up the filters.",'$MG:color:BLUE') + + + + def help_calculate_decay_widths(self): + + if self.ninitial != 1: + logger.warning("This command is only valid for processes of type A > B C.") + logger.warning("This command can not be run in current context.") + logger.warning("") + + logger.info("syntax: calculate_decay_widths [run_name] [options])") + logger.info("-- Calculate decay widths and enter widths and BRs in param_card") + logger.info(" for a series of processes of type A > B C ...") + self.run_options_help([('-f', 'Use default for all questions.'), + ('--accuracy=', 'accuracy (for each partial decay width).'\ + + ' Default is 0.01.')]) + + def help_multi_run(self): + logger.info("syntax: multi_run NB_RUN [run_name] [--run_options])") + logger.info("-- Launch the full chain of script for the generation of events") + logger.info(" NB_RUN times. This chains includes possible plotting, shower") + logger.info(" and detector resolution.") + self.run_options_help([('-f', 'Use default for all questions.'), + ('--laststep=', 'argument might be parton/pythia/pgs/delphes and indicate the last level to be run.')]) + + def help_survey(self): + logger.info("syntax: survey [run_name] [--run_options])") + logger.info("-- evaluate the different channel associate to the process") + self.run_options_help([("--" + key,value[-1]) for (key,value) in \ + self._survey_options.items()]) + + + def help_restart_gridpack(self): + logger.info("syntax: restart_gridpack --precision= --restart_zero") + + + def help_launch(self): + """exec generate_events for 2>N and calculate_width for 1>N""" + logger.info("syntax: launch [run_name] [options])") + logger.info(" --alias for either generate_events/calculate_decay_widths") + logger.info(" depending of the number of particles in the initial state.") + + if self.ninitial == 1: + logger.info("For this directory this is equivalent to calculate_decay_widths") + self.help_calculate_decay_widths() + else: + logger.info("For this directory this is equivalent to $generate_events") + self.help_generate_events() + + def help_refine(self): + logger.info("syntax: refine require_precision [max_channel] [--run_options]") + logger.info("-- refine the LAST run to achieve a given precision.") + logger.info(" require_precision: can be either the targeted number of events") + logger.info(' or the required relative error') + logger.info(' max_channel:[5] maximal number of channel per job') + self.run_options_help([]) + + def help_combine_events(self): + """ """ + logger.info("syntax: combine_events [run_name] [--tag=tag_name] [--run_options]") + logger.info("-- Combine the last run in order to write the number of events") + logger.info(" asked in the run_card.") + self.run_options_help([]) + + def help_store_events(self): + """ """ + logger.info("syntax: store_events [--run_options]") + logger.info("-- Write physically the events in the files.") + logger.info(" should be launch after \'combine_events\'") + self.run_options_help([]) + + def help_create_gridpack(self): + """ """ + logger.info("syntax: create_gridpack [--run_options]") + logger.info("-- create the gridpack. ") + logger.info(" should be launch after \'store_events\'") + self.run_options_help([]) + + def help_import(self): + """ """ + logger.info("syntax: import command PATH") + logger.info("-- Execute the command present in the file") + self.run_options_help([]) + + def help_syscalc(self): + logger.info("syntax: syscalc [RUN] [%s] [-f | --tag=]" % '|'.join(self._plot_mode)) + logger.info("-- calculate systematics information for the RUN (current run by default)") + logger.info(" at different stages of the event generation for scale/pdf/...") + + def help_remove(self): + logger.info("syntax: remove RUN [all|parton|pythia|pgs|delphes|banner] [-f] [--tag=]") + logger.info("-- Remove all the files linked to previous run RUN") + logger.info(" if RUN is 'all', then all run will be cleaned.") + logger.info(" The optional argument precise which part should be cleaned.") + logger.info(" By default we clean all the related files but the banners.") + logger.info(" the optional '-f' allows to by-pass all security question") + logger.info(" The banner can be remove only if all files are removed first.") + + +class AskRun(cmd.ControlSwitch): + """a class for the question on what to do on a madevent run""" + + to_control = [('shower', 'Choose the shower/hadronization program'), + ('detector', 'Choose the detector simulation program'), + ('analysis', 'Choose an analysis package (plot/convert)'), + ('madspin', 'Decay onshell particles'), + ('reweight', 'Add weights to events for new hypp.') + ] + + def __init__(self, question, line_args=[], mode=None, force=False, + *args, **opt): + + self.check_available_module(opt['mother_interface'].options) + self.me_dir = opt['mother_interface'].me_dir + super(AskRun,self).__init__(self.to_control, opt['mother_interface'], + *args, **opt) + + + def check_available_module(self, options): + + self.available_module = set() + if options['pythia-pgs_path']: + self.available_module.add('PY6') + self.available_module.add('PGS') + if options['pythia8_path']: + self.available_module.add('PY8') + if options['madanalysis_path']: + self.available_module.add('MA4') + if options['madanalysis5_path']: + self.available_module.add('MA5') + if options['exrootanalysis_path']: + self.available_module.add('ExRoot') + if options['delphes_path']: + if 'PY6' in self.available_module or 'PY8' in self.available_module: + self.available_module.add('Delphes') + else: + logger.warning("Delphes program installed but no parton shower module detected.\n Please install pythia8") + if options['rivet_path']: + if 'PY8' in self.available_module: + self.available_module.add('Rivet') + else: + logger.warning("Rivet program installed but no parton shower with hepmc output detected.\n Please install pythia8") + + if not MADEVENT or ('mg5_path' in options and options['mg5_path']): + self.available_module.add('MadSpin') + if misc.has_f2py() or options['f2py_compiler']: + self.available_module.add('reweight') + +# old mode to activate the shower + def ans_parton(self, value=None): + """None: means that the user type 'pythia' + value: means that the user type pythia=value""" + + if value is None: + self.set_all_off() + else: + logger.warning('Invalid command: parton=%s' % value) + + +# +# HANDLING SHOWER +# + def get_allowed_shower(self): + """return valid entry for the shower switch""" + + if hasattr(self, 'allowed_shower'): + return self.allowed_shower + + self.allowed_shower = [] + if 'PY6' in self.available_module: + self.allowed_shower.append('Pythia6') + if 'PY8' in self.available_module: + self.allowed_shower.append('Pythia8') + if self.allowed_shower: + self.allowed_shower.append('OFF') + return self.allowed_shower + + def set_default_shower(self): + + if 'PY6' in self.available_module and\ + os.path.exists(pjoin(self.me_dir,'Cards','pythia_card.dat')): + self.switch['shower'] = 'Pythia6' + elif 'PY8' in self.available_module and\ + os.path.exists(pjoin(self.me_dir,'Cards','pythia8_card.dat')): + self.switch['shower'] = 'Pythia8' + elif self.get_allowed_shower(): + self.switch['shower'] = 'OFF' + else: + self.switch['shower'] = 'Not Avail.' + + def check_value_shower(self, value): + """check an entry is valid. return the valid entry in case of shortcut""" + + if value in self.get_allowed_shower(): + return True + + value =value.lower() + if value in ['py6','p6','pythia_6'] and 'PY6' in self.available_module: + return 'Pythia6' + elif value in ['py8','p8','pythia_8'] and 'PY8' in self.available_module: + return 'Pythia8' + else: + return False + + +# old mode to activate the shower + def ans_pythia(self, value=None): + """None: means that the user type 'pythia' + value: means that the user type pythia=value""" + + if 'PY6' not in self.available_module: + logger.info('pythia-pgs not available. Ignore commmand') + return + + if value is None: + self.set_all_off() + self.switch['shower'] = 'Pythia6' + elif value == 'on': + self.switch['shower'] = 'Pythia6' + elif value == 'off': + self.set_switch('shower', 'OFF') + else: + logger.warning('Invalid command: pythia=%s' % value) + + + def consistency_shower_detector(self, vshower, vdetector): + """consistency_XX_YY(val_XX, val_YY) + -> XX is the new key set by the user to a new value val_XX + -> YY is another key + -> return value should be None or "replace_YY" + """ + + if vshower == 'OFF': + if self.check_value('detector', vdetector) and vdetector!= 'OFF': + return 'OFF' + if vshower == 'Pythia8' and vdetector == 'PGS': + return 'OFF' + + return None + + + +# +# HANDLING DETECTOR +# + def get_allowed_detector(self): + """return valid entry for the switch""" + + if hasattr(self, 'allowed_detector'): + return self.allowed_detector + + self.allowed_detector = [] + if 'PGS' in self.available_module: + self.allowed_detector.append('PGS') + if 'Delphes' in self.available_module: + self.allowed_detector.append('Delphes') + + + if self.allowed_detector: + self.allowed_detector.append('OFF') + return self.allowed_detector + + def set_default_detector(self): + + self.set_default_shower() #ensure that this one is called first! + + if 'PGS' in self.available_module and self.switch['shower'] == 'Pythia6'\ + and os.path.exists(pjoin(self.me_dir,'Cards','pgs_card.dat')): + self.switch['detector'] = 'PGS' + elif 'Delphes' in self.available_module and self.switch['shower'] != 'OFF'\ + and os.path.exists(pjoin(self.me_dir,'Cards','delphes_card.dat')): + self.switch['detector'] = 'Delphes' + elif self.get_allowed_detector(): + self.switch['detector'] = 'OFF' + else: + self.switch['detector'] = 'Not Avail.' + +# old mode to activate pgs + def ans_pgs(self, value=None): + """None: means that the user type 'pgs' + value: means that the user type pgs=value""" + + if 'PGS' not in self.available_module: + logger.info('pythia-pgs not available. Ignore commmand') + return + + if value is None: + self.set_all_off() + self.switch['shower'] = 'Pythia6' + self.switch['detector'] = 'PGS' + elif value == 'on': + self.switch['shower'] = 'Pythia6' + self.switch['detector'] = 'PGS' + elif value == 'off': + self.set_switch('detector', 'OFF') + else: + logger.warning('Invalid command: pgs=%s' % value) + + +# old mode to activate Delphes + def ans_delphes(self, value=None): + """None: means that the user type 'delphes' + value: means that the user type delphes=value""" + + if 'Delphes' not in self.available_module: + logger.warning('Delphes not available. Ignore commmand') + return + + if value is None: + self.set_all_off() + if 'PY6' in self.available_module: + self.switch['shower'] = 'Pythia6' + else: + self.switch['shower'] = 'Pythia8' + self.switch['detector'] = 'Delphes' + elif value == 'on': + return self.ans_delphes(None) + elif value == 'off': + self.set_switch('detector', 'OFF') + else: + logger.warning('Invalid command: pgs=%s' % value) + + def consistency_detector_shower(self,vdetector, vshower): + """consistency_XX_YY(val_XX, val_YY) + -> XX is the new key set by the user to a new value val_XX + -> YY is another key + -> return value should be None or "replace_YY" + """ + + if vdetector == 'PGS' and vshower != 'Pythia6': + return 'Pythia6' + if vdetector == 'Delphes' and vshower not in ['Pythia6', 'Pythia8']: + if 'PY8' in self.available_module: + return 'Pythia8' + elif 'PY6' in self.available_module: + return 'Pythia6' + else: + raise Exception + return None + + +# +# HANDLING ANALYSIS +# + def get_allowed_analysis(self): + """return valid entry for the shower switch""" + + if hasattr(self, 'allowed_analysis'): + return self.allowed_analysis + + self.allowed_analysis = [] + if 'ExRoot' in self.available_module: + self.allowed_analysis.append('ExRoot') + if 'MA4' in self.available_module: + self.allowed_analysis.append('MadAnalysis4') + if 'MA5' in self.available_module: + self.allowed_analysis.append('MadAnalysis5') + if 'Rivet' in self.available_module: + self.allowed_analysis.append('Rivet') + + if self.allowed_analysis: + self.allowed_analysis.append('OFF') + + return self.allowed_analysis + + def check_analysis(self, value): + """check an entry is valid. return the valid entry in case of shortcut""" + + if value in self.get_allowed_analysis(): + return True + if value.lower() in ['ma4', 'madanalysis4', 'madanalysis_4','4']: + return 'MadAnalysis4' + if value.lower() in ['ma5', 'madanalysis5', 'madanalysis_5','5']: + return 'MadAnalysis5' + if value.lower() in ['ma', 'madanalysis']: + if 'MA5' in self.available_module: + return 'MadAnalysis5' + elif 'MA4' in self.available_module: + return 'MadAnalysis4' + else: + return False + else: + return False + + def consistency_shower_analysis(self, vshower, vanalysis): + """consistency_XX_YY(val_XX, val_YY) + -> XX is the new key set by the user to a new value val_XX + -> YY is another key + -> return value should be None or "replace_YY" + """ + + if vshower != 'Pythia8' and vanalysis == 'Rivet': + return 'OFF' #new value for analysis + + return None + + def consistency_analysis_shower(self, vanalysis, vshower): + """consistency_XX_YY(val_XX, val_YY) + -> XX is the new key set by the user to a new value val_XX + -> YY is another key + -> return value should be None or "replace_YY" + """ + + if vshower != 'Pythia8' and vanalysis == 'Rivet': + return 'Pythia8' #new value for analysis + + return None + + + def set_default_analysis(self): + """initialise the switch for analysis""" + + if 'MA4' in self.available_module and \ + os.path.exists(pjoin(self.me_dir,'Cards','plot_card.dat')): + self.switch['analysis'] = 'MadAnalysis4' + elif 'MA5' in self.available_module and\ + (os.path.exists(pjoin(self.me_dir,'Cards','madanalysis5_parton_card.dat'))\ + or os.path.exists(pjoin(self.me_dir,'Cards', 'madanalysis5_hadron_card.dat'))): + self.switch['analysis'] = 'MadAnalysis5' + elif 'ExRoot' in self.available_module: + self.switch['analysis'] = 'ExRoot' + elif self.get_allowed_analysis(): + self.switch['analysis'] = 'OFF' + else: + self.switch['analysis'] = 'Not Avail.' + +# +# MADSPIN handling +# + def get_allowed_madspin(self): + """ ON|OFF|onshell """ + + if hasattr(self, 'allowed_madspin'): + return self.allowed_madspin + + self.allowed_madspin = [] + if 'MadSpin' in self.available_module: + self.allowed_madspin = ['OFF',"ON",'onshell',"full"] + return self.allowed_madspin + + def check_value_madspin(self, value): + """handle alias and valid option not present in get_allowed_madspin""" + + if value.upper() in self.get_allowed_madspin(): + return True + elif value.lower() in self.get_allowed_madspin(): + return True + + if 'MadSpin' not in self.available_module: + return False + + if value.lower() in ['madspin', 'full']: + return 'full' + elif value.lower() in ['none']: + return 'none' + + + def set_default_madspin(self): + """initialise the switch for madspin""" + + if 'MadSpin' in self.available_module: + if os.path.exists(pjoin(self.me_dir,'Cards','madspin_card.dat')): + self.switch['madspin'] = 'ON' + else: + self.switch['madspin'] = 'OFF' + else: + self.switch['madspin'] = 'Not Avail.' + + def get_cardcmd_for_madspin(self, value): + """set some command to run before allowing the user to modify the cards.""" + + if value == 'onshell': + return ["edit madspin_card --replace_line='set spinmode' --before_line='decay' set spinmode onshell"] + elif value in ['full', 'madspin']: + return ["edit madspin_card --replace_line='set spinmode' --before_line='decay' set spinmode full"] + elif value == 'none': + return ["edit madspin_card --replace_line='set spinmode' --before_line='decay' set spinmode none"] + else: + return [] + +# +# ReWeight handling +# + def get_allowed_reweight(self): + """ return the list of valid option for reweight=XXX """ + + if hasattr(self, 'allowed_reweight'): + return getattr(self, 'allowed_reweight') + + if 'reweight' not in self.available_module: + self.allowed_reweight = [] + return + self.allowed_reweight = ['OFF', 'ON'] + + # check for plugin mode + plugin_path = self.mother_interface.plugin_path + opts = misc.from_plugin_import(plugin_path, 'new_reweight', warning=False) + self.allowed_reweight += opts + + def set_default_reweight(self): + """initialise the switch for reweight""" + + if 'reweight' in self.available_module: + if os.path.exists(pjoin(self.me_dir,'Cards','reweight_card.dat')): + self.switch['reweight'] = 'ON' + else: + self.switch['reweight'] = 'OFF' + else: + self.switch['reweight'] = 'Not Avail.' + +#=============================================================================== +# CheckValidForCmd +#=============================================================================== +class CheckValidForCmd(object): + """ The Series of check routine for the MadEventCmd""" + + def check_banner_run(self, args): + """check the validity of line""" + + if len(args) == 0: + self.help_banner_run() + raise self.InvalidCmd('banner_run requires at least one argument.') + + tag = [a[6:] for a in args if a.startswith('--tag=')] + + + if os.path.exists(args[0]): + type ='banner' + format = self.detect_card_type(args[0]) + if format != 'banner': + raise self.InvalidCmd('The file is not a valid banner.') + elif tag: + args[0] = pjoin(self.me_dir,'Events', args[0], '%s_%s_banner.txt' % \ + (args[0], tag)) + if not os.path.exists(args[0]): + raise self.InvalidCmd('No banner associates to this name and tag.') + else: + name = args[0] + type = 'run' + banners = misc.glob('*_banner.txt', pjoin(self.me_dir,'Events', args[0])) + if not banners: + raise self.InvalidCmd('No banner associates to this name.') + elif len(banners) == 1: + args[0] = banners[0] + else: + #list the tag and propose those to the user + tags = [os.path.basename(p)[len(args[0])+1:-11] for p in banners] + tag = self.ask('which tag do you want to use?', tags[0], tags) + args[0] = pjoin(self.me_dir,'Events', args[0], '%s_%s_banner.txt' % \ + (args[0], tag)) + + run_name = [arg[7:] for arg in args if arg.startswith('--name=')] + if run_name: + try: + self.exec_cmd('remove %s all banner -f' % run_name) + except Exception: + pass + self.set_run_name(args[0], tag=None, level='parton', reload_card=True) + elif type == 'banner': + self.set_run_name(self.find_available_run_name(self.me_dir)) + elif type == 'run': + if not self.results[name].is_empty(): + run_name = self.find_available_run_name(self.me_dir) + logger.info('Run %s is not empty so will use run_name: %s' % \ + (name, run_name)) + self.set_run_name(run_name) + else: + try: + self.exec_cmd('remove %s all banner -f' % run_name) + except Exception: + pass + self.set_run_name(name) + + def check_history(self, args): + """check the validity of line""" + + if len(args) > 1: + self.help_history() + raise self.InvalidCmd('\"history\" command takes at most one argument') + + if not len(args): + return + elif args[0] != 'clean': + dirpath = os.path.dirname(args[0]) + if dirpath and not os.path.exists(dirpath) or \ + os.path.isdir(args[0]): + raise self.InvalidCmd("invalid path %s " % dirpath) + + def check_save(self, args): + """ check the validity of the line""" + + if len(args) == 0: + args.append('options') + + if args[0] not in self._save_opts: + raise self.InvalidCmd('wrong \"save\" format') + + if args[0] != 'options' and len(args) != 2: + self.help_save() + raise self.InvalidCmd('wrong \"save\" format') + elif args[0] != 'options' and len(args) == 2: + basename = os.path.dirname(args[1]) + if not os.path.exists(basename): + raise self.InvalidCmd('%s is not a valid path, please retry' % \ + args[1]) + + if args[0] == 'options': + has_path = None + for arg in args[1:]: + if arg in ['--auto', '--all']: + continue + elif arg.startswith('--'): + raise self.InvalidCmd('unknow command for \'save options\'') + else: + basename = os.path.dirname(arg) + if not os.path.exists(basename): + raise self.InvalidCmd('%s is not a valid path, please retry' % \ + arg) + elif has_path: + raise self.InvalidCmd('only one path is allowed') + else: + args.remove(arg) + args.insert(1, arg) + has_path = True + if not has_path: + if '--auto' in arg and self.options['mg5_path']: + args.insert(1, pjoin(self.options['mg5_path'],'input','mg5_configuration.txt')) + else: + args.insert(1, pjoin(self.me_dir,'Cards','me5_configuration.txt')) + + def check_set(self, args): + """ check the validity of the line""" + + if len(args) < 2: + self.help_set() + raise self.InvalidCmd('set needs an option and an argument') + + if args[0] not in self._set_options + list(self.options.keys()): + self.help_set() + raise self.InvalidCmd('Possible options for set are %s' % \ + self._set_options) + + if args[0] in ['stdout_level']: + if args[1] not in ['DEBUG','INFO','WARNING','ERROR','CRITICAL'] \ + and not args[1].isdigit(): + raise self.InvalidCmd('output_level needs ' + \ + 'a valid level') + + if args[0] in ['timeout']: + if not args[1].isdigit(): + raise self.InvalidCmd('timeout values should be a integer') + + def check_open(self, args): + """ check the validity of the line """ + + if len(args) != 1: + self.help_open() + raise self.InvalidCmd('OPEN command requires exactly one argument') + + if args[0].startswith('./'): + if not os.path.isfile(args[0]): + raise self.InvalidCmd('%s: not such file' % args[0]) + return True + + # if special : create the path. + if not self.me_dir: + if not os.path.isfile(args[0]): + self.help_open() + raise self.InvalidCmd('No MadEvent path defined. Unable to associate this name to a file') + else: + return True + + path = self.me_dir + if os.path.isfile(os.path.join(path,args[0])): + args[0] = os.path.join(path,args[0]) + elif os.path.isfile(os.path.join(path,'Cards',args[0])): + args[0] = os.path.join(path,'Cards',args[0]) + elif os.path.isfile(os.path.join(path,'HTML',args[0])): + args[0] = os.path.join(path,'HTML',args[0]) + # special for card with _default define: copy the default and open it + elif '_card.dat' in args[0]: + name = args[0].replace('_card.dat','_card_default.dat') + if os.path.isfile(os.path.join(path,'Cards', name)): + files.cp(os.path.join(path,'Cards', name), os.path.join(path,'Cards', args[0])) + args[0] = os.path.join(path,'Cards', args[0]) + else: + raise self.InvalidCmd('No default path for this file') + elif not os.path.isfile(args[0]): + raise self.InvalidCmd('No default path for this file') + + def check_initMadLoop(self, args): + """ check initMadLoop command arguments are valid.""" + + opt = {'refresh': False, 'nPS': None, 'force': False} + + for arg in args: + if arg in ['-r','--refresh']: + opt['refresh'] = True + if arg in ['-f','--force']: + opt['force'] = True + elif arg.startswith('--nPS='): + n_attempts = arg.split('=')[1] + try: + opt['nPS'] = int(n_attempts) + except ValueError: + raise InvalidCmd("The number of attempts specified "+ + "'%s' is not a valid integer."%n_attempts) + + return opt + + def check_treatcards(self, args): + """check that treatcards arguments are valid + [param|run|all] [--output_dir=] [--param_card=] [--run_card=] + """ + + opt = {'output_dir':pjoin(self.me_dir,'Source'), + 'param_card':pjoin(self.me_dir,'Cards','param_card.dat'), + 'run_card':pjoin(self.me_dir,'Cards','run_card.dat'), + 'forbid_MadLoopInit': False} + mode = 'all' + for arg in args: + if arg.startswith('--') and '=' in arg: + key,value =arg[2:].split('=',1) + if not key in opt: + self.help_treatcards() + raise self.InvalidCmd('Invalid option for treatcards command:%s ' \ + % key) + if key in ['param_card', 'run_card']: + if os.path.isfile(value): + card_name = self.detect_card_type(value) + if card_name != key: + raise self.InvalidCmd('Format for input file detected as %s while expecting %s' + % (card_name, key)) + opt[key] = value + elif os.path.isfile(pjoin(self.me_dir,value)): + card_name = self.detect_card_type(pjoin(self.me_dir,value)) + if card_name != key: + raise self.InvalidCmd('Format for input file detected as %s while expecting %s' + % (card_name, key)) + opt[key] = value + else: + raise self.InvalidCmd('No such file: %s ' % value) + elif key in ['output_dir']: + if os.path.isdir(value): + opt[key] = value + elif os.path.isdir(pjoin(self.me_dir,value)): + opt[key] = pjoin(self.me_dir, value) + else: + raise self.InvalidCmd('No such directory: %s' % value) + elif arg in ['loop','param','run','all']: + mode = arg + elif arg == '--no_MadLoopInit': + opt['forbid_MadLoopInit'] = True + else: + self.help_treatcards() + raise self.InvalidCmd('Unvalid argument %s' % arg) + + return mode, opt + + + def check_survey(self, args, cmd='survey'): + """check that the argument for survey are valid""" + + + self.opts = dict([(key,value[1]) for (key,value) in \ + self._survey_options.items()]) + + # Treat any arguments starting with '--' + while args and args[-1].startswith('--'): + arg = args.pop(-1) + try: + for opt,value in self._survey_options.items(): + if arg.startswith('--%s=' % opt): + exec('self.opts[\'%s\'] = %s(arg.split(\'=\')[-1])' % \ + (opt, value[0])) + arg = "" + if arg != "": raise Exception + except Exception: + self.help_survey() + raise self.InvalidCmd('invalid %s argument'% arg) + + if len(args) > 1: + self.help_survey() + raise self.InvalidCmd('Too many argument for %s command' % cmd) + elif not args: + # No run name assigned -> assigned one automaticaly + self.set_run_name(self.find_available_run_name(self.me_dir)) + else: + self.set_run_name(args[0], None,'parton', True) + args.pop(0) + + return True + + def check_generate_events(self, args): + """check that the argument for generate_events are valid""" + + run = None + if args and args[-1].startswith('--laststep='): + run = args[-1].split('=')[-1] + if run not in ['auto','parton', 'pythia', 'pgs', 'delphes']: + self.help_generate_events() + raise self.InvalidCmd('invalid %s argument'% args[-1]) + if run != 'parton' and not self.options['pythia-pgs_path']: + raise self.InvalidCmd('''pythia-pgs not install. Please install this package first. + To do so type: \'install pythia-pgs\' in the mg5 interface''') + if run == 'delphes' and not self.options['delphes_path']: + raise self.InvalidCmd('''delphes not install. Please install this package first. + To do so type: \'install Delphes\' in the mg5 interface''') + del args[-1] + + + #if len(args) > 1: + # self.help_generate_events() + # raise self.InvalidCmd('Too many argument for generate_events command: %s' % cmd) + + return run + + def check_calculate_decay_widths(self, args): + """check that the argument for calculate_decay_widths are valid""" + + if self.ninitial != 1: + raise self.InvalidCmd('Can only calculate decay widths for decay processes A > B C ...') + + accuracy = 0.01 + run = None + if args and args[-1].startswith('--accuracy='): + try: + accuracy = float(args[-1].split('=')[-1]) + except Exception: + raise self.InvalidCmd('Argument error in calculate_decay_widths command') + del args[-1] + if len(args) > 1: + self.help_calculate_decay_widths() + raise self.InvalidCmd('Too many argument for calculate_decay_widths command: %s' % cmd) + + return accuracy + + + + def check_multi_run(self, args): + """check that the argument for survey are valid""" + + run = None + + if not len(args): + self.help_multi_run() + raise self.InvalidCmd("""multi_run command requires at least one argument for + the number of times that it call generate_events command""") + + if args[-1].startswith('--laststep='): + run = args[-1].split('=')[-1] + if run not in ['parton', 'pythia', 'pgs', 'delphes']: + self.help_multi_run() + raise self.InvalidCmd('invalid %s argument'% args[-1]) + if run != 'parton' and not self.options['pythia-pgs_path']: + raise self.InvalidCmd('''pythia-pgs not install. Please install this package first. + To do so type: \'install pythia-pgs\' in the mg5 interface''') + if run == 'delphes' and not self.options['delphes_path']: + raise self.InvalidCmd('''delphes not install. Please install this package first. + To do so type: \'install Delphes\' in the mg5 interface''') + del args[-1] + + + elif not args[0].isdigit(): + self.help_multi_run() + raise self.InvalidCmd("The first argument of multi_run should be a integer.") + #pass nb run to an integer + nb_run = args.pop(0) + args.insert(0, int(nb_run)) + + + return run + + def check_refine(self, args): + """check that the argument for survey are valid""" + + # if last argument is not a number -> it's the run_name (Not allow anymore) + try: + float(args[-1]) + except ValueError: + self.help_refine() + raise self.InvalidCmd('Not valid arguments') + except IndexError: + self.help_refine() + raise self.InvalidCmd('require_precision argument is require for refine cmd') + + + if not self.run_name: + if self.results.lastrun: + self.set_run_name(self.results.lastrun) + else: + raise self.InvalidCmd('No run_name currently define. Unable to run refine') + + if len(args) > 2: + raise self.InvalidCmd('Too many argument for refine command') + else: + try: + [float(arg) for arg in args] + except ValueError: + self.help_refine() + raise self.InvalidCmd('refine arguments are suppose to be number') + + return True + + def check_combine_events(self, arg): + """ Check the argument for the combine events command """ + + tag = [a for a in arg if a.startswith('--tag=')] + if tag: + arg.remove(tag[0]) + tag = tag[0][6:] + elif not self.run_tag: + tag = 'tag_1' + else: + tag = self.run_tag + self.run_tag = tag + + if len(arg) > 1: + self.help_combine_events() + raise self.InvalidCmd('Too many argument for combine_events command') + + if len(arg) == 1: + self.set_run_name(arg[0], self.run_tag, 'parton', True) + + if not self.run_name: + if not self.results.lastrun: + raise self.InvalidCmd('No run_name currently define. Unable to run combine') + else: + self.set_run_name(self.results.lastrun) + + return True + + def check_pythia(self, args): + """Check the argument for pythia command + syntax: pythia [NAME] + Note that other option are already removed at this point + """ + + mode = None + laststep = [arg for arg in args if arg.startswith('--laststep=')] + if laststep and len(laststep)==1: + mode = laststep[0].split('=')[-1] + if mode not in ['auto', 'pythia', 'pgs', 'delphes']: + self.help_pythia() + raise self.InvalidCmd('invalid %s argument'% args[-1]) + elif laststep: + raise self.InvalidCmd('only one laststep argument is allowed') + + if not self.options['pythia-pgs_path']: + logger.info('Retry to read configuration file to find pythia-pgs path') + self.set_configuration() + + if not self.options['pythia-pgs_path'] or not \ + os.path.exists(pjoin(self.options['pythia-pgs_path'],'src')): + error_msg = 'No valid pythia-pgs path set.\n' + error_msg += 'Please use the set command to define the path and retry.\n' + error_msg += 'You can also define it in the configuration file.\n' + raise self.InvalidCmd(error_msg) + + + + tag = [a for a in args if a.startswith('--tag=')] + if tag: + args.remove(tag[0]) + tag = tag[0][6:] + + if len(args) == 0 and not self.run_name: + if self.results.lastrun: + args.insert(0, self.results.lastrun) + else: + raise self.InvalidCmd('No run name currently define. Please add this information.') + + if len(args) >= 1: + if args[0] != self.run_name and\ + not os.path.exists(pjoin(self.me_dir,'Events',args[0], 'unweighted_events.lhe.gz')): + raise self.InvalidCmd('No events file corresponding to %s run. '% args[0]) + self.set_run_name(args[0], tag, 'pythia') + else: + if tag: + self.run_card['run_tag'] = tag + self.set_run_name(self.run_name, tag, 'pythia') + + input_file = pjoin(self.me_dir,'Events',self.run_name, 'unweighted_events.lhe') + output_file = pjoin(self.me_dir, 'Events', 'unweighted_events.lhe') + if not os.path.exists('%s.gz' % input_file): + if not os.path.exists(input_file): + raise self.InvalidCmd('No events file corresponding to %s run. '% self.run_name) + files.ln(input_file, os.path.dirname(output_file)) + else: + misc.gunzip(input_file, keep=True, stdout=output_file) + + args.append(mode) + + def check_pythia8(self, args): + """Check the argument for pythia command + syntax: pythia8 [NAME] + Note that other option are already removed at this point + """ + mode = None + laststep = [arg for arg in args if arg.startswith('--laststep=')] + if laststep and len(laststep)==1: + mode = laststep[0].split('=')[-1] + if mode not in ['auto', 'pythia','pythia8','delphes']: + self.help_pythia8() + raise self.InvalidCmd('invalid %s argument'% args[-1]) + elif laststep: + raise self.InvalidCmd('only one laststep argument is allowed') + + # If not pythia-pgs path + if not self.options['pythia8_path']: + logger.info('Retry reading configuration file to find pythia8 path') + self.set_configuration() + + if not self.options['pythia8_path'] or not \ + os.path.exists(pjoin(self.options['pythia8_path'],'bin','pythia8-config')): + error_msg = 'No valid pythia8 path set.\n' + error_msg += 'Please use the set command to define the path and retry.\n' + error_msg += 'You can also define it in the configuration file.\n' + error_msg += 'Finally, it can be installed automatically using the' + error_msg += ' install command.\n' + raise self.InvalidCmd(error_msg) + + tag = [a for a in args if a.startswith('--tag=')] + if tag: + args.remove(tag[0]) + tag = tag[0][6:] + + if len(args) == 0 and not self.run_name: + if self.results.lastrun: + args.insert(0, self.results.lastrun) + else: + raise self.InvalidCmd('No run name currently define. '+ + 'Please add this information.') + + if len(args) >= 1: + if args[0] != self.run_name and\ + not os.path.exists(pjoin(self.me_dir,'Events',args[0], + 'unweighted_events.lhe.gz')): + raise self.InvalidCmd('No events file corresponding to %s run. ' + % args[0]) + self.set_run_name(args[0], tag, 'pythia8') + else: + if tag: + self.run_card['run_tag'] = tag + self.set_run_name(self.run_name, tag, 'pythia8') + + input_file = pjoin(self.me_dir,'Events',self.run_name, 'unweighted_events.lhe') + if not os.path.exists('%s.gz'%input_file): + if os.path.exists(input_file): + misc.gzip(input_file, stdout='%s.gz'%input_file) + else: + raise self.InvalidCmd('No event file corresponding to %s run. ' + % self.run_name) + + args.append(mode) + + def check_remove(self, args): + """Check that the remove command is valid""" + + tmp_args = args[:] + + tag = [a[6:] for a in tmp_args if a.startswith('--tag=')] + if tag: + tag = tag[0] + tmp_args.remove('--tag=%s' % tag) + + + if len(tmp_args) == 0: + self.help_remove() + raise self.InvalidCmd('clean command require the name of the run to clean') + elif len(tmp_args) == 1: + return tmp_args[0], tag, ['all'] + else: + for arg in tmp_args[1:]: + if arg not in self._clean_mode: + self.help_remove() + raise self.InvalidCmd('%s is not a valid options for clean command'\ + % arg) + return tmp_args[0], tag, tmp_args[1:] + + def check_plot(self, args): + """Check the argument for the plot command + plot run_name modes""" + + madir = self.options['madanalysis_path'] + td = self.options['td_path'] + + if not madir or not td: + logger.info('Retry to read configuration file to find madanalysis/td') + self.set_configuration() + + madir = self.options['madanalysis_path'] + td = self.options['td_path'] + + if not madir: + error_msg = 'No valid MadAnalysis path set.\n' + error_msg += 'Please use the set command to define the path and retry.\n' + error_msg += 'You can also define it in the configuration file.\n' + raise self.InvalidCmd(error_msg) + if not td: + error_msg = 'No valid td path set.\n' + error_msg += 'Please use the set command to define the path and retry.\n' + error_msg += 'You can also define it in the configuration file.\n' + raise self.InvalidCmd(error_msg) + + if len(args) == 0: + if not hasattr(self, 'run_name') or not self.run_name: + self.help_plot() + raise self.InvalidCmd('No run name currently define. Please add this information.') + args.append('all') + return + + + if args[0] not in self._plot_mode: + self.set_run_name(args[0], level='plot') + del args[0] + if len(args) == 0: + args.append('all') + elif not self.run_name: + self.help_plot() + raise self.InvalidCmd('No run name currently define. Please add this information.') + + for arg in args: + if arg not in self._plot_mode and arg != self.run_name: + self.help_plot() + raise self.InvalidCmd('unknown options %s' % arg) + + def check_syscalc(self, args): + """Check the argument for the syscalc command + syscalc run_name modes""" + + scdir = self.options['syscalc_path'] + + if not scdir: + logger.info('Retry to read configuration file to find SysCalc') + self.set_configuration() + + scdir = self.options['syscalc_path'] + + if not scdir: + error_msg = 'No valid SysCalc path set.\n' + error_msg += 'Please use the set command to define the path and retry.\n' + error_msg += 'You can also define it in the configuration file.\n' + error_msg += 'Please note that you need to compile SysCalc first.' + raise self.InvalidCmd(error_msg) + + if len(args) == 0: + if not hasattr(self, 'run_name') or not self.run_name: + self.help_syscalc() + raise self.InvalidCmd('No run name currently defined. Please add this information.') + args.append('all') + return + + #deal options + tag = [a for a in args if a.startswith('--tag=')] + if tag: + args.remove(tag[0]) + tag = tag[0][6:] + + if args[0] not in self._syscalc_mode: + self.set_run_name(args[0], tag=tag, level='syscalc') + del args[0] + if len(args) == 0: + args.append('all') + elif not self.run_name: + self.help_syscalc() + raise self.InvalidCmd('No run name currently defined. Please add this information.') + elif tag and tag != self.run_tag: + self.set_run_name(self.run_name, tag=tag, level='syscalc') + + for arg in args: + if arg not in self._syscalc_mode and arg != self.run_name: + self.help_syscalc() + raise self.InvalidCmd('unknown options %s' % arg) + + if self.run_card['use_syst'] not in self.true: + raise self.InvalidCmd('Run %s does not include ' % self.run_name + \ + 'systematics information needed for syscalc.') + + + def check_pgs(self, arg, no_default=False): + """Check the argument for pythia command + syntax is "pgs [NAME]" + Note that other option are already remove at this point + """ + + # If not pythia-pgs path + if not self.options['pythia-pgs_path']: + logger.info('Retry to read configuration file to find pythia-pgs path') + self.set_configuration() + + if not self.options['pythia-pgs_path'] or not \ + os.path.exists(pjoin(self.options['pythia-pgs_path'],'src')): + error_msg = 'No valid pythia-pgs path set.\n' + error_msg += 'Please use the set command to define the path and retry.\n' + error_msg += 'You can also define it in the configuration file.\n' + raise self.InvalidCmd(error_msg) + + tag = [a for a in arg if a.startswith('--tag=')] + if tag: + arg.remove(tag[0]) + tag = tag[0][6:] + + + if len(arg) == 0 and not self.run_name: + if self.results.lastrun: + arg.insert(0, self.results.lastrun) + else: + raise self.InvalidCmd('No run name currently define. Please add this information.') + + if len(arg) == 1 and self.run_name == arg[0]: + arg.pop(0) + + if not len(arg) and \ + not os.path.exists(pjoin(self.me_dir,'Events','pythia_events.hep')): + if not no_default: + self.help_pgs() + raise self.InvalidCmd('''No file file pythia_events.hep currently available + Please specify a valid run_name''') + + lock = None + if len(arg) == 1: + prev_tag = self.set_run_name(arg[0], tag, 'pgs') + if not os.path.exists(pjoin(self.me_dir,'Events',self.run_name,'%s_pythia_events.hep.gz' % prev_tag)): + raise self.InvalidCmd('No events file corresponding to %s run with tag %s. '% (self.run_name, prev_tag)) + else: + input_file = pjoin(self.me_dir,'Events', self.run_name, '%s_pythia_events.hep.gz' % prev_tag) + output_file = pjoin(self.me_dir, 'Events', 'pythia_events.hep') + lock = cluster.asyncrone_launch('gunzip',stdout=open(output_file,'w'), + argument=['-c', input_file]) + + else: + if tag: + self.run_card['run_tag'] = tag + self.set_run_name(self.run_name, tag, 'pgs') + + return lock + + def check_display(self, args): + """check the validity of line + syntax is "display XXXXX" + """ + + if len(args) < 1 or args[0] not in self._display_opts: + self.help_display() + raise self.InvalidCmd + + if args[0] == 'variable' and len(args) !=2: + raise self.InvalidCmd('variable need a variable name') + + + + + + def check_import(self, args): + """check the validity of line""" + + if not args: + self.help_import() + raise self.InvalidCmd('wrong \"import\" format') + + if args[0] != 'command': + args.insert(0,'command') + + + if not len(args) == 2 or not os.path.exists(args[1]): + raise self.InvalidCmd('PATH is mandatory for import command\n') + + +#=============================================================================== +# CompleteForCmd +#=============================================================================== +class CompleteForCmd(CheckValidForCmd): + """ The Series of help routine for the MadGraphCmd""" + + + def complete_banner_run(self, text, line, begidx, endidx, formatting=True): + "Complete the banner run command" + try: + + + args = self.split_arg(line[0:begidx], error=False) + + if args[-1].endswith(os.path.sep): + return self.path_completion(text, + os.path.join('.',*[a for a in args \ + if a.endswith(os.path.sep)])) + + + if len(args) > 1: + # only options are possible + tags = misc.glob('%s_*_banner.txt' % args[1], pjoin(self.me_dir, 'Events' , args[1])) + tags = ['%s' % os.path.basename(t)[len(args[1])+1:-11] for t in tags] + + if args[-1] != '--tag=': + tags = ['--tag=%s' % t for t in tags] + else: + return self.list_completion(text, tags) + return self.list_completion(text, tags +['--name=','-f'], line) + + # First argument + possibilites = {} + + comp = self.path_completion(text, os.path.join('.',*[a for a in args \ + if a.endswith(os.path.sep)])) + if os.path.sep in line: + return comp + else: + possibilites['Path from ./'] = comp + + run_list = misc.glob(pjoin('*','*_banner.txt'), pjoin(self.me_dir, 'Events')) + run_list = [n.rsplit('/',2)[1] for n in run_list] + possibilites['RUN Name'] = self.list_completion(text, run_list) + + return self.deal_multiple_categories(possibilites, formatting) + + + except Exception as error: + print(error) + + + def complete_history(self, text, line, begidx, endidx): + "Complete the history command" + + args = self.split_arg(line[0:begidx], error=False) + + # Directory continuation + if args[-1].endswith(os.path.sep): + return self.path_completion(text, + os.path.join('.',*[a for a in args \ + if a.endswith(os.path.sep)])) + + if len(args) == 1: + return self.path_completion(text) + + def complete_open(self, text, line, begidx, endidx): + """ complete the open command """ + + args = self.split_arg(line[0:begidx]) + + # Directory continuation + if os.path.sep in args[-1] + text: + return self.path_completion(text, + os.path.join('.',*[a for a in args if \ + a.endswith(os.path.sep)])) + + possibility = [] + if self.me_dir: + path = self.me_dir + possibility = ['index.html'] + if os.path.isfile(os.path.join(path,'README')): + possibility.append('README') + if os.path.isdir(os.path.join(path,'Cards')): + possibility += [f for f in os.listdir(os.path.join(path,'Cards')) + if f.endswith('.dat')] + if os.path.isdir(os.path.join(path,'HTML')): + possibility += [f for f in os.listdir(os.path.join(path,'HTML')) + if f.endswith('.html') and 'default' not in f] + else: + possibility.extend(['./','../']) + if os.path.exists('ME5_debug'): + possibility.append('ME5_debug') + if os.path.exists('MG5_debug'): + possibility.append('MG5_debug') + return self.list_completion(text, possibility) + + def complete_set(self, text, line, begidx, endidx): + "Complete the set command" + + args = self.split_arg(line[0:begidx]) + + # Format + if len(args) == 1: + return self.list_completion(text, self._set_options + list(self.options.keys()) ) + + if len(args) == 2: + if args[1] == 'stdout_level': + return self.list_completion(text, ['DEBUG','INFO','WARNING','ERROR','CRITICAL']) + else: + first_set = ['None','True','False'] + # directory names + second_set = [name for name in self.path_completion(text, '.', only_dirs = True)] + return self.list_completion(text, first_set + second_set) + elif len(args) >2 and args[-1].endswith(os.path.sep): + return self.path_completion(text, + os.path.join('.',*[a for a in args if a.endswith(os.path.sep)]), + only_dirs = True) + + def complete_survey(self, text, line, begidx, endidx): + """ Complete the survey command """ + + if line.endswith('nb_core=') and not text: + import multiprocessing + max = multiprocessing.cpu_count() + return [str(i) for i in range(2,max+1)] + + return self.list_completion(text, self._run_options, line) + + complete_refine = complete_survey + complete_combine_events = complete_survey + complite_store = complete_survey + complete_generate_events = complete_survey + complete_create_gridpack = complete_survey + + def complete_generate_events(self, text, line, begidx, endidx): + """ Complete the generate events""" + + if line.endswith('nb_core=') and not text: + import multiprocessing + max = multiprocessing.cpu_count() + return [str(i) for i in range(2,max+1)] + if line.endswith('laststep=') and not text: + return ['parton','pythia','pgs','delphes'] + elif '--laststep=' in line.split()[-1] and line and line[-1] != ' ': + return self.list_completion(text,['parton','pythia','pgs','delphes'],line) + + opts = self._run_options + self._generate_options + return self.list_completion(text, opts, line) + + + def complete_initMadLoop(self, text, line, begidx, endidx): + "Complete the initMadLoop command" + + numbers = [str(i) for i in range(10)] + opts = ['-f','-r','--nPS='] + + args = self.split_arg(line[0:begidx], error=False) + if len(line) >=6 and line[begidx-6:begidx]=='--nPS=': + return self.list_completion(text, numbers, line) + else: + return self.list_completion(text, [opt for opt in opts if not opt in + line], line) + + def complete_launch(self, *args, **opts): + + if self.ninitial == 1: + return self.complete_calculate_decay_widths(*args, **opts) + else: + return self.complete_generate_events(*args, **opts) + + def complete_calculate_decay_widths(self, text, line, begidx, endidx): + """ Complete the calculate_decay_widths command""" + + if line.endswith('nb_core=') and not text: + import multiprocessing + max = multiprocessing.cpu_count() + return [str(i) for i in range(2,max+1)] + + opts = self._run_options + self._calculate_decay_options + return self.list_completion(text, opts, line) + + def complete_display(self, text, line, begidx, endidx): + """ Complete the display command""" + + args = self.split_arg(line[0:begidx], error=False) + if len(args) >= 2 and args[1] =='results': + start = line.find('results') + return self.complete_print_results(text, 'print_results '+line[start+7:], begidx+2+start, endidx+2+start) + return super(CompleteForCmd, self).complete_display(text, line, begidx, endidx) + + def complete_multi_run(self, text, line, begidx, endidx): + """complete multi run command""" + + args = self.split_arg(line[0:begidx], error=False) + if len(args) == 1: + data = [str(i) for i in range(0,20)] + return self.list_completion(text, data, line) + + if line.endswith('run=') and not text: + return ['parton','pythia','pgs','delphes'] + elif '--laststep=' in line.split()[-1] and line and line[-1] != ' ': + return self.list_completion(text,['parton','pythia','pgs','delphes'],line) + + opts = self._run_options + self._generate_options + return self.list_completion(text, opts, line) + + + + if line.endswith('nb_core=') and not text: + import multiprocessing + max = multiprocessing.cpu_count() + return [str(i) for i in range(2,max+1)] + opts = self._run_options + self._generate_options + return self.list_completion(text, opts, line) + + def complete_plot(self, text, line, begidx, endidx): + """ Complete the plot command """ + + args = self.split_arg(line[0:begidx], error=False) + if len(args) > 1: + return self.list_completion(text, self._plot_mode) + else: + return self.list_completion(text, self._plot_mode + list(self.results.keys())) + + def complete_syscalc(self, text, line, begidx, endidx, formatting=True): + """ Complete the syscalc command """ + + output = {} + args = self.split_arg(line[0:begidx], error=False) + + if len(args) <=1: + output['RUN_NAME'] = self.list_completion(list(self.results.keys())) + output['MODE'] = self.list_completion(text, self._syscalc_mode) + output['options'] = ['-f'] + if len(args) > 1 and (text.startswith('--t')): + run = args[1] + if run in self.results: + tags = ['--tag=%s' % tag['tag'] for tag in self.results[run]] + output['options'] += tags + + return self.deal_multiple_categories(output, formatting) + + def complete_remove(self, text, line, begidx, endidx): + """Complete the remove command """ + + args = self.split_arg(line[0:begidx], error=False) + if len(args) > 1 and (text.startswith('--t')): + run = args[1] + tags = ['--tag=%s' % tag['tag'] for tag in self.results[run]] + return self.list_completion(text, tags) + elif len(args) > 1 and '--' == args[-1]: + run = args[1] + tags = ['tag=%s' % tag['tag'] for tag in self.results[run]] + return self.list_completion(text, tags) + elif len(args) > 1 and '--tag=' == args[-1]: + run = args[1] + tags = [tag['tag'] for tag in self.results[run]] + return self.list_completion(text, tags) + elif len(args) > 1: + return self.list_completion(text, self._clean_mode + ['-f','--tag=']) + else: + data = misc.glob(pjoin('*','*_banner.txt'), pjoin(self.me_dir, 'Events')) + data = [n.rsplit('/',2)[1] for n in data] + return self.list_completion(text, ['all'] + data) + + + def complete_shower(self,text, line, begidx, endidx): + "Complete the shower command" + args = self.split_arg(line[0:begidx], error=False) + if len(args) == 1: + return self.list_completion(text, self._interfaced_showers) + elif len(args)>1 and args[1] in self._interfaced_showers: + return getattr(self, 'complete_%s' % text)\ + (text, args[1],line.replace(args[0]+' ',''), + begidx-len(args[0])-1, endidx-len(args[0])-1) + + def complete_pythia8(self,text, line, begidx, endidx): + "Complete the pythia8 command" + args = self.split_arg(line[0:begidx], error=False) + if len(args) == 1: + #return valid run_name + data = misc.glob(pjoin('*','unweighted_events.lhe.gz'),pjoin(self.me_dir, 'Events')) + data = [n.rsplit('/',2)[1] for n in data] + tmp1 = self.list_completion(text, data) + if not self.run_name: + return tmp1 + else: + tmp2 = self.list_completion(text, self._run_options + ['-f', + '--no_default', '--tag='], line) + return tmp1 + tmp2 + elif line[-1] != '=': + return self.list_completion(text, self._run_options + ['-f', + '--no_default','--tag='], line) + + def complete_madanalysis5_parton(self,text, line, begidx, endidx): + "Complete the madanalysis5 command" + args = self.split_arg(line[0:begidx], error=False) + if len(args) == 1: + #return valid run_name + data = [] + for name in ['unweighted_events.lhe']: + data += misc.glob(pjoin('*','%s'%name), pjoin(self.me_dir, 'Events')) + data += misc.glob(pjoin('*','%s.gz'%name), pjoin(self.me_dir, 'Events')) + data = [n.rsplit('/',2)[1] for n in data] + tmp1 = self.list_completion(text, data) + if not self.run_name: + return tmp1 + else: + tmp2 = self.list_completion(text, ['-f', + '--MA5_stdout_lvl=','--no_default','--tag='], line) + return tmp1 + tmp2 + elif '--MA5_stdout_lvl=' in line and not any(arg.startswith( + '--MA5_stdout_lvl=') for arg in args): + return self.list_completion(text, + ['--MA5_stdout_lvl=%s'%opt for opt in + ['logging.INFO','logging.DEBUG','logging.WARNING', + 'logging.CRITICAL','90']], line) + else: + return self.list_completion(text, ['-f', + '--MA5_stdout_lvl=','--no_default','--tag='], line) + + def complete_pythia(self,text, line, begidx, endidx): + "Complete the pythia command" + args = self.split_arg(line[0:begidx], error=False) + + if len(args) == 1: + #return valid run_name + data = misc.glob(pjoin('*','unweighted_events.lhe.gz'), pjoin(self.me_dir, 'Events')) + data = [n.rsplit('/',2)[1] for n in data] + tmp1 = self.list_completion(text, data) + if not self.run_name: + return tmp1 + else: + tmp2 = self.list_completion(text, self._run_options + ['-f', + '--no_default', '--tag='], line) + return tmp1 + tmp2 + elif line[-1] != '=': + return self.list_completion(text, self._run_options + ['-f', + '--no_default','--tag='], line) + + def complete_pgs(self,text, line, begidx, endidx): + "Complete the pythia command" + args = self.split_arg(line[0:begidx], error=False) + if len(args) == 1: + #return valid run_name + data = misc.glob(pjoin('*', '*_pythia_events.hep.gz'), pjoin(self.me_dir, 'Events')) + data = [n.rsplit('/',2)[1] for n in data] + tmp1 = self.list_completion(text, data) + if not self.run_name: + return tmp1 + else: + tmp2 = self.list_completion(text, self._run_options + ['-f', + '--tag=' ,'--no_default'], line) + return tmp1 + tmp2 + else: + return self.list_completion(text, self._run_options + ['-f', + '--tag=','--no_default'], line) + + complete_delphes = complete_pgs + complete_rivet = complete_pgs + +#=============================================================================== +# MadEventCmd +#=============================================================================== +class MadEventCmd(CompleteForCmd, CmdExtended, HelpToCmd, common_run.CommonRunCmd): + + """The command line processor of Mad Graph""" + + + LO = True + # Truth values + true = ['T','.true.',True,'true'] + # Options and formats available + _run_options = ['--cluster','--multicore','--nb_core=','--nb_core=2', '-c', '-m'] + _generate_options = ['-f', '--laststep=parton', '--laststep=pythia', '--laststep=pgs', '--laststep=delphes'] + _calculate_decay_options = ['-f', '--accuracy=0.'] + _interfaced_showers = ['pythia','pythia8'] + _set_options = ['stdout_level','fortran_compiler','timeout'] + _plot_mode = ['all', 'parton','pythia','pgs','delphes','channel', 'banner'] + _syscalc_mode = ['all', 'parton','pythia'] + _clean_mode = _plot_mode + _display_opts = ['run_name', 'options', 'variable', 'results'] + _save_opts = ['options'] + _initMadLoop_opts = ['-f','-r','--nPS='] + # survey options, dict from name to type, default value, and help text + _survey_options = {'points':('int', 8192,'Number of points for first iteration'), + 'iterations':('int', 1, 'Number of iterations'), + 'accuracy':('float', 0.1, 'Required accuracy'), + 'gridpack':('str', '.false.', 'Gridpack generation')} + # Variables to store object information + true = ['T','.true.',True,'true', 1, '1'] + web = False + cluster_mode = 0 + queue = 'madgraph' + nb_core = None + + next_possibility = { + 'start': ['generate_events [OPTIONS]', 'multi_run [OPTIONS]', + 'calculate_decay_widths [OPTIONS]', + 'help generate_events'], + 'generate_events': ['generate_events [OPTIONS]', 'multi_run [OPTIONS]', 'pythia', 'pgs','delphes'], + 'calculate_decay_widths': ['calculate_decay_widths [OPTIONS]', + 'generate_events [OPTIONS]'], + 'multi_run': ['generate_events [OPTIONS]', 'multi_run [OPTIONS]'], + 'survey': ['refine'], + 'refine': ['combine_events'], + 'combine_events': ['store'], + 'store': ['pythia'], + 'pythia': ['pgs', 'delphes'], + 'pgs': ['generate_events [OPTIONS]', 'multi_run [OPTIONS]'], + 'delphes' : ['generate_events [OPTIONS]', 'multi_run [OPTIONS]'] + } + + asking_for_run = AskRun + + ############################################################################ + def __init__(self, me_dir = None, options={}, *completekey, **stdin): + """ add information to the cmd """ + + CmdExtended.__init__(self, me_dir, options, *completekey, **stdin) + #common_run.CommonRunCmd.__init__(self, me_dir, options) + + self.mode = 'madevent' + self.nb_refine=0 + if self.web: + os.system('touch %s' % pjoin(self.me_dir,'Online')) + + self.load_results_db() + self.results.def_web_mode(self.web) + + self.Gdirs = None + + self.prompt = "%s>"%os.path.basename(pjoin(self.me_dir)) + self.configured = 0 # time for reading the card + self._options = {} # for compatibility with extended_cmd + + + def pass_in_web_mode(self): + """configure web data""" + self.web = True + self.results.def_web_mode(True) + self.force = True + if os.environ['MADGRAPH_BASE']: + self.options['mg5_path'] = pjoin(os.environ['MADGRAPH_BASE'],'MG5') + + ############################################################################ + def check_output_type(self, path): + """ Check that the output path is a valid madevent directory """ + + bin_path = os.path.join(path,'bin') + if os.path.isfile(os.path.join(bin_path,'generate_events')): + return True + else: + return False + + ############################################################################ + def set_configuration(self, amcatnlo=False, final=True, **opt): + """assign all configuration variable from file + loop over the different config file if config_file not define """ + + super(MadEventCmd,self).set_configuration(amcatnlo=amcatnlo, + final=final, **opt) + + if not final: + return self.options # the return is usefull for unittest + + + # Treat each expected input + # delphes/pythia/... path + # ONLY the ONE LINKED TO Madevent ONLY!!! + for key in (k for k in self.options if k.endswith('path')): + path = self.options[key] + if path is None or key.startswith("cluster"): + continue + if not os.path.isdir(path): + path = pjoin(self.me_dir, self.options[key]) + if os.path.isdir(path): + self.options[key] = None + if key == "pythia-pgs_path": + if not os.path.exists(pjoin(path, 'src','pythia')): + logger.info("No valid pythia-pgs path found") + continue + elif key == "delphes_path": + if not os.path.exists(pjoin(path, 'Delphes')) and not\ + os.path.exists(pjoin(path, 'DelphesSTDHEP')): + logger.info("No valid Delphes path found") + continue + elif key == "madanalysis_path": + if not os.path.exists(pjoin(path, 'plot_events')): + logger.info("No valid MadAnalysis path found") + continue + elif key == "rivet_path": + if not os.path.exists(pjoin(path, 'bin', 'rivet')): + logger.info("No valid rivet path found") + continue + elif key == "td_path": + if not os.path.exists(pjoin(path, 'td')): + logger.info("No valid td path found") + continue + elif key == "syscalc_path": + if not os.path.exists(pjoin(path, 'sys_calc')): + logger.info("No valid SysCalc path found") + continue + # No else since the next line reinitialize the option to the + #previous value anyway + self.options[key] = os.path.realpath(path) + continue + else: + self.options[key] = None + + + return self.options + + ############################################################################ + def do_banner_run(self, line): + """Make a run from the banner file""" + + args = self.split_arg(line) + #check the validity of the arguments + self.check_banner_run(args) + + # Remove previous cards + for name in ['delphes_trigger.dat', 'delphes_card.dat', + 'pgs_card.dat', 'pythia_card.dat', 'madspin_card.dat', + 'reweight_card.dat']: + try: + os.remove(pjoin(self.me_dir, 'Cards', name)) + except Exception: + pass + + banner_mod.split_banner(args[0], self.me_dir, proc_card=False) + + # Check if we want to modify the run + if not self.force: + ans = self.ask('Do you want to modify the Cards?', 'n', ['y','n']) + if ans == 'n': + self.force = True + + # Call Generate events + self.exec_cmd('generate_events %s %s' % (self.run_name, self.force and '-f' or '')) + + + + ############################################################################ + def do_display(self, line, output=sys.stdout): + """Display current internal status""" + + args = self.split_arg(line) + #check the validity of the arguments + self.check_display(args) + + if args[0] == 'run_name': + #return valid run_name + data = misc.glob(pjoin('*','*_banner.txt'), pjoin(self.me_dir, 'Events')) + data = [n.rsplit('/',2)[1:] for n in data] + + if data: + out = {} + for name, tag in data: + tag = tag[len(name)+1:-11] + if name in out: + out[name].append(tag) + else: + out[name] = [tag] + print('the runs available are:') + for run_name, tags in out.items(): + print(' run: %s' % run_name) + print(' tags: ', end=' ') + print(', '.join(tags)) + else: + print('No run detected.') + + elif args[0] == 'options': + outstr = " Run Options \n" + outstr += " ----------- \n" + for key, default in self.options_madgraph.items(): + value = self.options[key] + if value == default: + outstr += " %25s \t:\t%s\n" % (key,value) + else: + outstr += " %25s \t:\t%s (user set)\n" % (key,value) + outstr += "\n" + outstr += " MadEvent Options \n" + outstr += " ---------------- \n" + for key, default in self.options_madevent.items(): + if key in self.options: + value = self.options[key] + else: + default = '' + if value == default: + outstr += " %25s \t:\t%s\n" % (key,value) + else: + outstr += " %25s \t:\t%s (user set)\n" % (key,value) + outstr += "\n" + outstr += " Configuration Options \n" + outstr += " --------------------- \n" + for key, default in self.options_configuration.items(): + value = self.options[key] + if value == default: + outstr += " %25s \t:\t%s\n" % (key,value) + else: + outstr += " %25s \t:\t%s (user set)\n" % (key,value) + output.write(outstr) + elif args[0] == 'results': + self.do_print_results(' '.join(args[1:])) + else: + super(MadEventCmd, self).do_display(line, output) + + def do_save(self, line, check=True, to_keep={}): + """Not in help: Save information to file""" + + args = self.split_arg(line) + # Check argument validity + if check: + self.check_save(args) + + if args[0] == 'options': + # First look at options which should be put in MG5DIR/input + to_define = {} + for key, default in self.options_configuration.items(): + if self.options[key] != self.options_configuration[key]: + to_define[key] = self.options[key] + + if not '--auto' in args: + for key, default in self.options_madevent.items(): + if self.options[key] != self.options_madevent[key]: + to_define[key] = self.options[key] + + if '--all' in args: + for key, default in self.options_madgraph.items(): + if self.options[key] != self.options_madgraph[key]: + to_define[key] = self.options[key] + elif not '--auto' in args: + for key, default in self.options_madgraph.items(): + if self.options[key] != self.options_madgraph[key]: + logger.info('The option %s is modified [%s] but will not be written in the configuration files.' \ + % (key,self.options_madgraph[key]) ) + logger.info('If you want to make this value the default for future session, you can run \'save options --all\'') + if len(args) >1 and not args[1].startswith('--'): + filepath = args[1] + else: + filepath = pjoin(self.me_dir, 'Cards', 'me5_configuration.txt') + basefile = pjoin(self.me_dir, 'Cards', 'me5_configuration.txt') + basedir = self.me_dir + + if to_keep: + to_define = to_keep + self.write_configuration(filepath, basefile, basedir, to_define) + + + + + def do_edit_cards(self, line): + """Advanced commands: Basic edition of the cards""" + args = self.split_arg(line) + # Check argument's validity + mode = self.check_generate_events(args) + self.ask_run_configuration(mode) + + return + + ############################################################################ + + ############################################################################ + def do_restart_gridpack(self, line): + """ syntax restart_gridpack --precision=1.0 --restart_zero + collect the result of the current run and relaunch each channel + not completed or optionally a completed one with a precision worse than + a threshold (and/or the zero result channel)""" + + + args = self.split_arg(line) + # Check argument's validity + self.check_survey(args) + + # initialize / remove lhapdf mode + #self.run_card = banner_mod.RunCard(pjoin(self.me_dir, 'Cards', 'run_card.dat')) + #self.configure_directory() + + gensym = gen_ximprove.gensym(self) + + min_precision = 1.0 + resubmit_zero=False + if '--precision=' in line: + s = line.index('--precision=') + len('--precision=') + arg=line[s:].split(1)[0] + min_precision = float(arg) + + if '--restart_zero' in line: + resubmit_zero = True + + + gensym.resubmit(min_precision, resubmit_zero) + self.monitor(run_type='All jobs submitted for gridpack', html=True) + + #will be done during the refine (more precisely in gen_ximprove) + cross, error = sum_html.make_all_html_results(self) + self.results.add_detail('cross', cross) + self.results.add_detail('error', error) + self.exec_cmd("print_results %s" % self.run_name, + errorhandling=False, printcmd=False, precmd=False, postcmd=False) + + self.results.add_detail('run_statistics', dict(gensym.run_statistics)) + + + #self.exec_cmd('combine_events', postcmd=False) + #self.exec_cmd('store_events', postcmd=False) + self.exec_cmd('decay_events -from_cards', postcmd=False) + self.exec_cmd('create_gridpack', postcmd=False) + + + + ############################################################################ + + ############################################################################ + def do_generate_events(self, line): + """Main Commands: launch the full chain """ + + self.banner = None + self.Gdirs = None + + args = self.split_arg(line) + # Check argument's validity + mode = self.check_generate_events(args) + switch_mode = self.ask_run_configuration(mode, args) + if not args: + # No run name assigned -> assigned one automaticaly + self.set_run_name(self.find_available_run_name(self.me_dir), None, 'parton') + else: + self.set_run_name(args[0], None, 'parton', True) + args.pop(0) + + self.run_generate_events(switch_mode, args) + + self.postprocessing() + + + # postprocessing : runs after all the event generation has been done + # even for the 'scan' mode, madevent->pythia->madevent->pythia->...->POSTPROCESSING + def postprocessing(self): + + # Run Rivet postprocessor + cmd_do_rivet = common_run.CommonRunCmd.do_rivet(self,"--no_default",True) + if cmd_do_rivet: + rivet_config = cmd_do_rivet[0] + postprocess_RIVET = cmd_do_rivet[1] + postprocess_CONTUR = cmd_do_rivet[2] + if postprocess_RIVET or postprocess_CONTUR: + self.rivet_postprocessing(rivet_config, postprocess_RIVET, postprocess_CONTUR) + + def rivet_postprocessing(self, rivet_config, postprocess_RIVET, postprocess_CONTUR): + + # Check number of Rivet jobs to run + run_dirs = [pjoin(self.me_dir, 'Events',run_name) + for run_name in self.postprocessing_dirs] + + nb_rivet = len(run_dirs) + + if postprocess_RIVET: + + # Submit Rivet jobs + for i_rivet in range(nb_rivet): + self.cluster.submit2(pjoin(run_dirs[i_rivet], "run_rivet.sh"), argument=[str(i_rivet)]) + + startRivet = time.time() + + def wait_monitoring(Idle, Running, Done): + if Idle+Running+Done == 0: + return + logger.info('Rivet analysis jobs: %d Idle, %d Running, %d Done [%s]'\ + %(Idle, Running, Done, misc.format_time(time.time() - startRivet))) + self.cluster.wait(pjoin(self.me_dir, 'Events'),wait_monitoring) + + self.update_status("postprocessing rivet done", level="rivet") + + if postprocess_CONTUR: + + self.update_status("Starting postprocess contur", level="rivet") + + set_env = "#!{0}\n".format(misc.which('bash' if misc.get_shell_type() in ['bash',None] else 'tcsh')) + rivet_path = self.options['rivet_path'] + yoda_path = self.options['yoda_path'] + set_env = set_env + "# RIVET/YODA PATH SETUP\n" + set_env = set_env + "export PATH={0}:{1}:$PATH\n"\ + .format(pjoin(rivet_path, 'bin'),\ + pjoin(yoda_path, 'bin')) + set_env = set_env + "export LD_LIBRARY_PATH={0}:{1}:{2}:{3}:$LD_LIBRARY_PATH\n"\ + .format(pjoin(rivet_path, 'lib'),\ + pjoin(rivet_path, 'lib64'),\ + pjoin(yoda_path, 'lib'),\ + pjoin(yoda_path, 'lib64')) + major, minor = sys.version_info[0:2] + set_env = set_env + "export PYTHONPATH={0}:{1}:{2}:{3}:$PYTHONPATH\n\n"\ + .format(pjoin(rivet_path, 'lib', 'python%s.%s' %(major,minor), 'site-packages'),\ + pjoin(rivet_path, 'lib64', 'python%s.%s' %(major,minor), 'site-packages'),\ + pjoin(yoda_path, 'lib', 'python%s.%s' %(major,minor), 'site-packages'),\ + pjoin(yoda_path, 'lib64', 'python%s.%s' %(major,minor), 'site-packages')) + + contur_path = self.options['contur_path'] + set_env = set_env + "# CONTUR PATH SETUP\n" + set_env = set_env + "export PATH={0}:$PATH\n".format(pjoin(contur_path, 'python%s.%s' %(major,minor), 'bin')) + set_env = set_env + "export PYTHONPATH={0}:$PYTHONPATH\n".format(pjoin(contur_path, 'python%s.%s' %(major,minor))) + + set_env = set_env + "source {0} >> contur.log 2>&1\n\n".format(pjoin(contur_path, "contur", "setupContur.sh")) + + os.system("mkdir -p {0}".format(pjoin(self.me_dir, 'Analysis', 'contur'))) + + if nb_rivet == 1: + this_yoda_file = pjoin(run_dirs[0], "rivet_result.yoda") + os.system("ln -s {0} {1}".format(this_yoda_file, pjoin(self.me_dir, 'Analysis', 'contur', 'rivet_result.yoda'))) + if not rivet_config["weight_name"] == "None": + contur_cmd = 'contur --wn "{0}" {1}\n'.format(rivet_config["weight_name"], pjoin(self.me_dir, 'Analysis', 'contur', 'rivet_result.yoda')) + else: + contur_cmd = 'contur {0}\n'.format(pjoin(self.me_dir, 'Analysis', 'contur', 'rivet_result.yoda')) + else: + # Link yoda and params files inside analysis/contur/scan directory + scan_subdirs = [] + for i_rivet in range(nb_rivet): + this_scan_dir = pjoin(self.me_dir, 'Analysis', 'contur', 'scan', rivet_config["contur_ra"]) + os.system("mkdir -p {0}".format(this_scan_dir)) + + this_scan_subdir = pjoin(this_scan_dir, str(i_rivet+1).zfill(4)) + scan_subdirs.append(this_scan_subdir) + os.mkdir(this_scan_subdir) + + this_yoda_file = pjoin(run_dirs[i_rivet], "rivet_result.yoda") + this_param_file = pjoin(run_dirs[i_rivet], "params.dat") + os.system("ln -s {0} {1}".format(this_yoda_file, pjoin(this_scan_subdir, "runpoint_"+str(i_rivet+1).zfill(4)+".yoda"))) + os.system("ln -s {0} {1}".format(this_param_file, pjoin(this_scan_subdir, "params.dat"))) + + if rivet_config['xaxis_relvar'] or rivet_config['yaxis_relvar']: + f_params = open(pjoin(run_dirs[i_rivet], "params.dat")) + f_relparams = open(pjoin(run_dirs[i_rivet], "params_replace.dat"), "w") + rivet_config.setRelevantParamCard(f_params=f_params,f_relparams=f_relparams) + f_params.close() + f_relparams.close() + + files.mv(pjoin(run_dirs[i_rivet], "params_replace.dat"), pjoin(run_dirs[i_rivet], "params.dat")) + + contur_add = "" + if not (rivet_config["contur_add"] == "default" or rivet_config["contur_add"] == None): + contur_add = " " + rivet_config["contur_add"] + + if rivet_config["weight_name"] == "None": + contur_cmd = 'contur --nomultip -g scan >> contur.log 2>&1\n' + else: + contur_cmd = 'contur --nomultip -g scan --wn "{0}" >> contur.log 2>&1\n'.format(rivet_config["weight_name"] + contur_add) + + if rivet_config["draw_contur_heatmap"]: + + axis_log = "" + if rivet_config["xaxis_log"]: + axis_log = axis_log + " --xlog" + if rivet_config["yaxis_log"]: + axis_log = axis_log + " --ylog" + + axis_label = "" + if rivet_config["xaxis_label"]: + axis_label = axis_label + " -x " + rivet_config["xaxis_label"] + if rivet_config["yaxis_label"]: + axis_label = axis_label + " -y " + rivet_config["yaxis_label"] + + if rivet_config["xaxis_relvar"]: + if rivet_config["xaxis_label"]: + xaxis_var = rivet_config["xaxis_label"] + else: + xaxis_var = "xaxis_relvar" + else: + xaxis_var = rivet_config["xaxis_var"] + if rivet_config["yaxis_relvar"]: + if rivet_config["yaxis_label"]: + yaxis_var = rivet_config["yaxis_label"] + else: + yaxis_var = "yaxis_relvar" + else: + yaxis_var = rivet_config["yaxis_var"] + + contur_cmd = contur_cmd + 'contur-plot ANALYSIS/contur.map {0} {1} {2} {3}' \ + .format(xaxis_var, yaxis_var,axis_label, axis_log) + + wrapper = open(pjoin(self.me_dir, "Analysis", "contur", "run_contur.sh"), "w") + wrapper.write(set_env) + + wrapper.write('{0}\n'.format(contur_cmd)) + wrapper.close() + + misc.call(["run_contur.sh"], cwd=(pjoin(self.me_dir, "Analysis", "contur"))) + + logger.info("Contur outputs are stored in {0}".format(pjoin(self.me_dir, "Analysis", "contur","conturPlot"))) + self.update_status("postprocessing contur done", level="rivet") + + # this decorator handle the loop related to scan. + @common_run.scanparamcardhandling() + def run_generate_events(self, switch_mode, args): + + if self.proc_characteristics['loop_induced'] and self.options['run_mode']==0: + # Also the single core mode is not supported for loop-induced. + # We therefore emulate it with multi-core mode with one core + logger.warning( +"""Single-core mode not supported for loop-induced processes. +Beware that MG5aMC now changes your runtime options to a multi-core mode with only one active core.""") + self.do_set('run_mode 2') + self.do_set('nb_core 1') + + if self.run_card['gridpack'] in self.true: + # Running gridpack warmup + gridpack_opts=[('accuracy', 0.01), + ('points', 2000), + ('iterations',8), + ('gridpack','.true.')] + logger.info('Generating gridpack with run name %s' % self.run_name) + self.exec_cmd('survey %s %s' % \ + (self.run_name, + " ".join(['--' + opt + '=' + str(val) for (opt,val) \ + in gridpack_opts])), + postcmd=False) + self.exec_cmd('combine_events', postcmd=False) + self.exec_cmd('store_events', postcmd=False) + with misc.TMP_variable(self, 'run_name', self.run_name): + self.exec_cmd('decay_events -from_cards', postcmd=False) + self.exec_cmd('create_gridpack', postcmd=False) + else: + # Regular run mode + logger.info('Generating %s events with run name %s' % + (self.run_card['nevents'], self.run_name)) + + self.exec_cmd('survey %s %s' % (self.run_name,' '.join(args)), + postcmd=False) + nb_event = self.run_card['nevents'] + bypass_run=False + self.exec_cmd('refine %s' % nb_event, postcmd=False) + if not float(self.results.current['cross']): + # Zero cross-section. Try to guess why + text = '''Survey return zero cross section. + Typical reasons are the following: + 1) A massive s-channel particle has a width set to zero. + 2) The pdf are zero for at least one of the initial state particles + or you are using maxjetflavor=4 for initial state b:s. + 3) The cuts are too strong. + Please check/correct your param_card and/or your run_card.''' + logger_stderr.critical(text) + if not self.param_card_iterator: + raise ZeroResult('See https://cp3.irmp.ucl.ac.be/projects/madgraph/wiki/FAQ-General-14') + else: + bypass_run = True + + #we can bypass the following if scan and first result is zero + if not bypass_run: + self.exec_cmd('refine %s --treshold=%s' % (nb_event,self.run_card['second_refine_treshold']) + , postcmd=False) + + self.exec_cmd('combine_events', postcmd=False,printcmd=False) + self.print_results_in_shell(self.results.current) + + if self.run_card['use_syst']: + if self.run_card['systematics_program'] == 'auto': + scdir = self.options['syscalc_path'] + if not scdir or not os.path.exists(scdir): + to_use = 'systematics' + else: + to_use = 'syscalc' + elif self.run_card['systematics_program'].lower() in ['systematics','syscalc', 'none']: + to_use = self.run_card['systematics_program'] + else: + logger.critical('Unvalid options for systematics_program: bypass computation of systematics variations.') + to_use = 'none' + + if to_use == 'systematics': + if self.run_card['systematics_arguments'] != ['']: + self.exec_cmd('systematics %s %s ' % (self.run_name, + ' '.join(self.run_card['systematics_arguments'])), + postcmd=False, printcmd=False) + else: + self.exec_cmd('systematics %s --from_card' % self.run_name, + postcmd=False,printcmd=False) + elif to_use == 'syscalc': + self.run_syscalc('parton') + + + self.create_plot('parton') + self.exec_cmd('store_events', postcmd=False) + if self.run_card['boost_event'].strip() and self.run_card['boost_event'] != 'False': + self.boost_events() + + + self.exec_cmd('reweight -from_cards', postcmd=False) + self.exec_cmd('decay_events -from_cards', postcmd=False) + if self.run_card['time_of_flight']>=0: + self.exec_cmd("add_time_of_flight --threshold=%s" % self.run_card['time_of_flight'] ,postcmd=False) + + if switch_mode['analysis'] == 'ExRoot': + input = pjoin(self.me_dir, 'Events', self.run_name,'unweighted_events.lhe.gz') + output = pjoin(self.me_dir, 'Events', self.run_name, 'unweighted_events.root') + self.create_root_file(input , output) + + self.exec_cmd('madanalysis5_parton --no_default', postcmd=False, printcmd=False) + # shower launches pgs/delphes if needed + self.exec_cmd('shower --no_default', postcmd=False, printcmd=False) + self.exec_cmd('madanalysis5_hadron --no_default', postcmd=False, printcmd=False) + self.exec_cmd('rivet --no_default', postcmd=False, printcmd=False) + self.store_result() + + if self.allow_notification_center: + misc.system_notify('Run %s finished' % os.path.basename(self.me_dir), + '%s: %s +- %s ' % (self.results.current['run_name'], + self.results.current['cross'], + self.results.current['error'])) + + def boost_events(self): + + if not self.run_card['boost_event']: + return + + if self.run_card['boost_event'].startswith('lambda'): + if not isinstance(self, cmd.CmdShell): + raise Exception("boost not allowed online") + filter = eval(self.run_card['boost_event']) + else: + raise Exception + + path = [pjoin(self.me_dir, 'Events', self.run_name, 'unweighted_events.lhe.gz'), + pjoin(self.me_dir, 'Events', self.run_name, 'unweighted_events.lhe'), + pjoin(self.me_dir, 'Events', self.run_name, 'events.lhe.gz'), + pjoin(self.me_dir, 'Events', self.run_name, 'events.lhe')] + + for p in path: + if os.path.exists(p): + event_path = p + break + else: + raise Exception("fail to find event file for the boost") + + + lhe = lhe_parser.EventFile(event_path) + with misc.TMP_directory() as tmp_dir: + output = lhe_parser.EventFile(pjoin(tmp_dir, os.path.basename(event_path)), 'w') + #write the banner to the output file + output.write(lhe.banner) + # Loop over all events + for event in lhe: + event.boost(filter) + #write this modify event + output.write(str(event)) + output.write('\n') + lhe.close() + files.mv(pjoin(tmp_dir, os.path.basename(event_path)), event_path) + + + + + + def do_initMadLoop(self,line): + """Compile and run MadLoop for a certain number of PS point so as to + initialize MadLoop (setup the zero helicity and loop filter.)""" + + args = line.split() + # Check argument's validity + options = self.check_initMadLoop(args) + + if not options['force']: + self.ask_edit_cards(['MadLoopParams.dat'], mode='fixed', plot=False) + self.exec_cmd('treatcards loop --no_MadLoopInit') + + if options['refresh']: + for filter in misc.glob('*Filter*', + pjoin(self.me_dir,'SubProcesses','MadLoop5_resources')): + logger.debug("Resetting filter '%s'."%os.path.basename(filter)) + os.remove(filter) + + MLCard = banner_mod.MadLoopParam(pjoin(self.me_dir, + 'Cards','MadLoopParams.dat')) + if options['nPS'] is None: + options['nPS'] = MLCard['CheckCycle']+2 + elif options['nPS'] < MLCard['CheckCycle']+2: + new_n_PS = MLCard['CheckCycle']+2 + logger.debug('Hard-setting user-defined n_PS (%d) to %d, because '\ + %(options['nPS'],new_n_PS)+"of the 'CheckCycle' value (%d) "%MLCard['CheckCycle']+\ + "specified in the ML param card.") + options['nPS'] = new_n_PS + + MadLoopInitializer.init_MadLoop(self.me_dir,n_PS=options['nPS'], + subproc_prefix='PV', MG_options=self.options, interface=self) + + def do_launch(self, line, *args, **opt): + """Main Commands: exec generate_events for 2>N and calculate_width for 1>N""" + + if self.ninitial == 1: + logger.info("Note that since 2.3. The launch for 1>N pass in event generation\n"+ + " To have the previous behavior use the calculate_decay_widths function") + # self.do_calculate_decay_widths(line, *args, **opt) + #else: + self.do_generate_events(line, *args, **opt) + + def print_results_in_shell(self, data): + """Have a nice results prints in the shell, + data should be of type: gen_crossxhtml.OneTagResults""" + + if not data: + return + + if data['run_statistics']: + globalstat = sum_html.RunStatistics() + + logger.info(" " ) + logger.debug(" === Run statistics summary ===") + for key, value in data['run_statistics'].items(): + globalstat.aggregate_statistics(value) + level = 5 + if value.has_warning(): + level = 10 + logger.log(level, value.nice_output(str('/'.join([key[0],'G%s'%key[1]]))).\ + replace(' statistics','')) + logger.info(" " ) + logger.debug(globalstat.nice_output('combined', no_warning=True)) + if globalstat.has_warning(): + logger.warning(globalstat.get_warning_text()) + logger.info(" ") + + + logger.info(" === Results Summary for run: %s tag: %s ===\n" % (data['run_name'],data['tag'])) + + total_time = int(sum(_['cumulative_timing'] for _ in data['run_statistics'].values())) + if total_time > 0: + logger.info(" Cumulative sequential time for this run: %s"%misc.format_time(total_time)) + + if self.ninitial == 1: + logger.info(" Width : %.4g +- %.4g GeV" % (data['cross'], data['error'])) + else: + logger.info(" Cross-section : %.4g +- %.4g pb" % (data['cross'], data['error'])) + logger.info(" Nb of events : %s" % data['nb_event'] ) + + if data['run_mode']=='madevent': + if data['cross_pythia'] and data['nb_event_pythia']: + if data['cross_pythia'] == -1: + path = pjoin(self.me_dir, 'Events', self.run_name, '%s_merged_xsecs.txt' % self.run_tag) + cross_sections = {} + if os.path.exists(path): + for line in open(path): + split = line.split() + if len(split)!=3: + continue + scale, cross, error = split + cross_sections[float(scale)] = (float(cross), float(error)) + if len(cross_sections)>0: + logger.info(' Pythia8 merged cross-sections are:') + for scale in sorted(cross_sections.keys()): + logger.info(' > Merging scale = %-6.4g : %-11.5g +/- %-7.2g [pb]'%\ + (scale,cross_sections[scale][0],cross_sections[scale][1])) + + else: + if self.ninitial == 1: + logger.info(" Matched width : %.4g +- %.4g GeV" % (data['cross_pythia'], data['error_pythia'])) + else: + logger.info(" Matched cross-section : %.4g +- %.4g pb" % (data['cross_pythia'], data['error_pythia'])) + logger.info(" Nb of events after matching/merging : %d" % int(data['nb_event_pythia'])) + if self.run_card['use_syst'] in self.true and \ + (int(self.run_card['ickkw'])==1 or self.run_card['ktdurham']>0.0 + or self.run_card['ptlund']>0.0): + logger.info(" Notice that because Systematics computation is turned on, the merging did not veto events but modified their weights instead.\n"+\ + " The resulting hepmc/stdhep file should therefore be use with those weights.") + else: + logger.info(" Nb of events after merging : %s" % data['nb_event_pythia']) + + logger.info(" " ) + + def print_results_in_file(self, data, path, mode='w', format='full'): + """Have a nice results prints in the shell, + data should be of type: gen_crossxhtml.OneTagResults""" + if not data: + return + + fsock = open(path, mode) + + if data['run_statistics']: + logger.debug(" === Run statistics summary ===") + for key, value in data['run_statistics'].items(): + logger.debug(value.nice_output(str('/'.join([key[0],'G%s'%key[1]]))).\ + replace(' statistics','')) + logger.info(" " ) + + if format == "full": + fsock.write(" === Results Summary for run: %s tag: %s process: %s ===\n" % \ + (data['run_name'],data['tag'], os.path.basename(self.me_dir))) + + if self.ninitial == 1: + fsock.write(" Width : %.4g +- %.4g GeV\n" % (data['cross'], data['error'])) + else: + fsock.write(" Cross-section : %.4g +- %.4g pb\n" % (data['cross'], data['error'])) + fsock.write(" Nb of events : %s\n" % data['nb_event'] ) + if data['cross_pythia'] and data['nb_event_pythia']: + if self.ninitial == 1: + fsock.write(" Matched Width : %.4g +- %.4g GeV\n" % (data['cross_pythia'], data['error_pythia'])) + else: + fsock.write(" Matched Cross-section : %.4g +- %.4g pb\n" % (data['cross_pythia'], data['error_pythia'])) + fsock.write(" Nb of events after Matching : %s\n" % data['nb_event_pythia']) + fsock.write(" \n" ) + elif format == "short": + if mode == "w": + fsock.write("# run_name tag cross error Nb_event cross_after_matching nb_event_after matching\n") + + if data['cross_pythia'] and data['nb_event_pythia']: + text = "%(run_name)s %(tag)s %(cross)s %(error)s %(nb_event)s %(cross_pythia)s %(nb_event_pythia)s\n" + else: + text = "%(run_name)s %(tag)s %(cross)s %(error)s %(nb_event)s\n" + fsock.write(text % data) + + ############################################################################ + def do_calculate_decay_widths(self, line): + """Main Commands: launch decay width calculation and automatic inclusion of + calculated widths and BRs in the param_card.""" + + args = self.split_arg(line) + # Check argument's validity + accuracy = self.check_calculate_decay_widths(args) + self.ask_run_configuration('parton') + self.banner = None + self.Gdirs = None + + if not args: + # No run name assigned -> assigned one automaticaly + self.set_run_name(self.find_available_run_name(self.me_dir)) + else: + self.set_run_name(args[0], reload_card=True) + args.pop(0) + + self.configure_directory() + + # Running gridpack warmup + opts=[('accuracy', accuracy), # default 0.01 + ('points', 1000), + ('iterations',9)] + + logger.info('Calculating decay widths with run name %s' % self.run_name) + + self.exec_cmd('survey %s %s' % \ + (self.run_name, + " ".join(['--' + opt + '=' + str(val) for (opt,val) \ + in opts])), + postcmd=False) + self.refine_mode = "old" # specify how to combine event + self.exec_cmd('combine_events', postcmd=False) + self.exec_cmd('store_events', postcmd=False) + + self.collect_decay_widths() + self.print_results_in_shell(self.results.current) + self.update_status('calculate_decay_widths done', + level='parton', makehtml=False) + + + ############################################################################ + def collect_decay_widths(self): + """ Collect the decay widths and calculate BRs for all particles, and put + in param_card form. + """ + + particle_dict = {} # store the results + run_name = self.run_name + + # Looping over the Subprocesses + for P_path in SubProcesses.get_subP(self.me_dir): + ids = SubProcesses.get_subP_ids(P_path) + # due to grouping we need to compute the ratio factor for the + # ungroup resutls (that we need here). Note that initial particles + # grouping are not at the same stage as final particle grouping + nb_output = len(ids) / (len(set([p[0] for p in ids]))) + results = open(pjoin(P_path, run_name + '_results.dat')).read().split('\n')[0] + result = float(results.strip().split(' ')[0]) + for particles in ids: + try: + particle_dict[particles[0]].append([particles[1:], result/nb_output]) + except KeyError: + particle_dict[particles[0]] = [[particles[1:], result/nb_output]] + + self.update_width_in_param_card(particle_dict, + initial = pjoin(self.me_dir, 'Cards', 'param_card.dat'), + output=pjoin(self.me_dir, 'Events', run_name, "param_card.dat")) + + @staticmethod + def update_width_in_param_card(decay_info, initial=None, output=None): + # Open the param_card.dat and insert the calculated decays and BRs + + if not output: + output = initial + + param_card_file = open(initial) + param_card = param_card_file.read().split('\n') + param_card_file.close() + + decay_lines = [] + line_number = 0 + # Read and remove all decays from the param_card + while line_number < len(param_card): + line = param_card[line_number] + if line.lower().startswith('decay'): + # Read decay if particle in decay_info + # DECAY 6 1.455100e+00 + line = param_card.pop(line_number) + line = line.split() + particle = 0 + if int(line[1]) not in decay_info: + try: # If formatting is wrong, don't want this particle + particle = int(line[1]) + width = float(line[2]) + except Exception: + particle = 0 + # Read BRs for this decay + line = param_card[line_number] + while re.search('^(#|\s|\d)', line): + line = param_card.pop(line_number) + if not particle or line.startswith('#'): + line=param_card[line_number] + continue + # 6.668201e-01 3 5 2 -1 + line = line.split() + try: # Remove BR if formatting is wrong + partial_width = float(line[0])*width + decay_products = [int(p) for p in line[2:2+int(line[1])]] + except Exception: + line=param_card[line_number] + continue + try: + decay_info[particle].append([decay_products, partial_width]) + except KeyError: + decay_info[particle] = [[decay_products, partial_width]] + if line_number == len(param_card): + break + line=param_card[line_number] + if particle and particle not in decay_info: + # No decays given, only total width + decay_info[particle] = [[[], width]] + else: # Not decay + line_number += 1 + # Clean out possible remaining comments at the end of the card + while not param_card[-1] or param_card[-1].startswith('#'): + param_card.pop(-1) + + # Append calculated and read decays to the param_card + param_card.append("#\n#*************************") + param_card.append("# Decay widths *") + param_card.append("#*************************") + for key in sorted(decay_info.keys()): + width = sum([r for p,r in decay_info[key]]) + param_card.append("#\n# PDG Width") + param_card.append("DECAY %i %e" % (key, width.real)) + if not width: + continue + if decay_info[key][0][0]: + param_card.append("# BR NDA ID1 ID2 ...") + brs = [[(val[1]/width).real, val[0]] for val in decay_info[key] if val[1]] + for val in sorted(brs, reverse=True): + param_card.append(" %e %i %s # %s" % + (val[0].real, len(val[1]), + " ".join([str(v) for v in val[1]]), + val[0] * width + )) + decay_table = open(output, 'w') + decay_table.write("\n".join(param_card) + "\n") + decay_table.close() + logger.info("Results written to %s" % output) + + + ############################################################################ + def do_multi_run(self, line): + + args = self.split_arg(line) + # Check argument's validity + mode = self.check_multi_run(args) + nb_run = args.pop(0) + if nb_run == 1: + logger.warn("'multi_run 1' command is not optimal. Think of using generate_events instead") + self.ask_run_configuration(mode) + + self.check_survey(args, cmd='multi_run') + main_name = self.run_name + # check if the param_card requires a scan over parameter. + path=pjoin(self.me_dir, 'Cards', 'param_card.dat') + self.check_param_card(path, run=False) + #store it locally to avoid relaunch + param_card_iterator, self.param_card_iterator = self.param_card_iterator, [] + + crossoversig = 0 + inv_sq_err = 0 + nb_event = 0 + for i in range(nb_run): + self.nb_refine = 0 + self.exec_cmd('generate_events %s_%s -f' % (main_name, i), postcmd=False) + # Update collected value + nb_event += int(self.results[self.run_name][-1]['nb_event']) + self.results.add_detail('nb_event', nb_event , run=main_name) + cross = self.results[self.run_name][-1]['cross'] + error = self.results[self.run_name][-1]['error'] + 1e-99 + crossoversig+=cross/error**2 + inv_sq_err+=1.0/error**2 + self.results[main_name][-1]['cross'] = crossoversig/inv_sq_err + self.results[main_name][-1]['error'] = math.sqrt(1.0/inv_sq_err) + self.results.def_current(main_name) + self.run_name = main_name + self.update_status("Merging LHE files", level='parton') + try: + os.mkdir(pjoin(self.me_dir,'Events', self.run_name)) + except Exception: + pass + os.system('%(bin)s/merge.pl %(event)s/%(name)s_*/unweighted_events.lhe.gz %(event)s/%(name)s/unweighted_events.lhe.gz %(event)s/%(name)s_banner.txt' + % {'bin': self.dirbin, 'event': pjoin(self.me_dir,'Events'), + 'name': self.run_name}) + + eradir = self.options['exrootanalysis_path'] + if eradir and misc.is_executable(pjoin(eradir,'ExRootLHEFConverter')): + self.update_status("Create Root file", level='parton') + path = '%s/%s/unweighted_events.lhe.gz' % (pjoin(self.me_dir,'Events'), self.run_name) + + if os.path.exists(path): + misc.gunzip(path) + + self.create_root_file('%s/unweighted_events.lhe' % self.run_name, + '%s/unweighted_events.root' % self.run_name) + + path = pjoin(self.me_dir, "Events", self.run_name, "unweighted_events.lhe") + self.create_plot('parton', path, + pjoin(self.me_dir, 'HTML',self.run_name, 'plots_parton.html') + ) + + + if not os.path.exists('%s.gz' % path): + misc.gzip(path) + + self.update_status('', level='parton') + self.print_results_in_shell(self.results.current) + + cpath = pjoin(self.me_dir,'Cards','param_card.dat') + if param_card_iterator: + + param_card_iterator.store_entry(self.run_name, self.results.current['cross'],param_card_path=cpath) + #check if the param_card defines a scan. + orig_name=self.run_name + for card in param_card_iterator: + card.write(cpath) + self.exec_cmd("multi_run %s -f " % nb_run ,precmd=True, postcmd=True,errorhandling=False) + param_card_iterator.store_entry(self.run_name, self.results.current['cross'], param_card_path=cpath) + param_card_iterator.write(pjoin(self.me_dir,'Cards','param_card.dat')) + scan_name = misc.get_scan_name(orig_name, self.run_name) + path = pjoin(self.me_dir, 'Events','scan_%s.txt' % scan_name) + logger.info("write all cross-section results in %s" % path, '$MG:BOLD') + param_card_iterator.write_summary(path) + + + ############################################################################ + def do_treatcards(self, line, mode=None, opt=None): + """Advanced commands: create .inc files from param_card.dat/run_card.dat""" + + if not mode and not opt: + args = self.split_arg(line) + mode, opt = self.check_treatcards(args) + + # To decide whether to refresh MadLoop's helicity filters, it is necessary + # to check if the model parameters where modified or not, before doing + # anything else. + need_MadLoopFilterUpdate = False + # Just to record what triggered the reinitialization of MadLoop for a + # nice debug message. + type_of_change = '' + if not opt['forbid_MadLoopInit'] and self.proc_characteristics['loop_induced'] \ + and mode in ['loop', 'all']: + paramDat = pjoin(self.me_dir, 'Cards','param_card.dat') + paramInc = pjoin(opt['output_dir'], 'param_card.inc') + if (not os.path.isfile(paramDat)) or (not os.path.isfile(paramInc)) or \ + (os.path.getmtime(paramDat)-os.path.getmtime(paramInc)) > 0.0: + need_MadLoopFilterUpdate = True + type_of_change = 'model' + + ML_in = pjoin(self.me_dir, 'Cards', 'MadLoopParams.dat') + ML_out = pjoin(self.me_dir,"SubProcesses", + "MadLoop5_resources", "MadLoopParams.dat") + if (not os.path.isfile(ML_in)) or (not os.path.isfile(ML_out)) or \ + (os.path.getmtime(ML_in)-os.path.getmtime(ML_out)) > 0.0: + need_MadLoopFilterUpdate = True + type_of_change = 'MadLoop' + + #check if no 'Auto' are present in the file + self.check_param_card(pjoin(self.me_dir, 'Cards','param_card.dat')) + + if mode in ['param', 'all']: + model = self.find_model_name() + tmp_model = os.path.basename(model) + if tmp_model == 'mssm' or tmp_model.startswith('mssm-'): + if not '--param_card=' in line: + param_card = pjoin(self.me_dir, 'Cards','param_card.dat') + mg5_param = pjoin(self.me_dir, 'Source', 'MODEL', 'MG5_param.dat') + check_param_card.convert_to_mg5card(param_card, mg5_param) + check_param_card.check_valid_param_card(mg5_param) + opt['param_card'] = pjoin(self.me_dir, 'Source', 'MODEL', 'MG5_param.dat') + else: + check_param_card.check_valid_param_card(opt['param_card']) + + logger.debug('write compile file for card: %s' % opt['param_card']) + param_card = check_param_card.ParamCard(opt['param_card']) + outfile = pjoin(opt['output_dir'], 'param_card.inc') + ident_card = pjoin(self.me_dir,'Cards','ident_card.dat') + if os.path.isfile(pjoin(self.me_dir,'bin','internal','ufomodel','restrict_default.dat')): + default = pjoin(self.me_dir,'bin','internal','ufomodel','restrict_default.dat') + elif os.path.isfile(pjoin(self.me_dir,'bin','internal','ufomodel','param_card.dat')): + default = pjoin(self.me_dir,'bin','internal','ufomodel','param_card.dat') + elif not os.path.exists(pjoin(self.me_dir,'bin','internal','ufomodel')): + fsock = open(pjoin(self.me_dir,'Source','param_card.inc'),'w') + fsock.write(' ') + fsock.close() + if mode == 'all': + self.do_treatcards('', 'run', opt) + return + else: + devnull = open(os.devnull,'w') + subprocess.call([sys.executable, 'write_param_card.py'], + cwd=pjoin(self.me_dir,'bin','internal','ufomodel'), + stdout=devnull) + devnull.close() + default = pjoin(self.me_dir,'bin','internal','ufomodel','param_card.dat') + + need_mp = self.proc_characteristics['loop_induced'] + param_card.write_inc_file(outfile, ident_card, default, need_mp=need_mp) + + + if mode in ['run', 'all']: + if not hasattr(self, 'run_card'): + run_card = banner_mod.RunCard(opt['run_card'], path=pjoin(self.me_dir, 'Cards', 'run_card.dat')) + else: + run_card = self.run_card + self.run_card = run_card + if self.cluster: + self.cluster.modify_interface(self) + if self.ninitial == 1: + run_card['lpp1'] = 0 + run_card['lpp2'] = 0 + run_card['ebeam1'] = 0 + run_card['ebeam2'] = 0 + + # Ensure that the bias parameters has all the required input from the + # run_card + if run_card['bias_module'].lower() not in ['dummy','none']: + # Using basename here means that the module will not be overwritten if already existing. + bias_module_path = pjoin(self.me_dir,'Source','BIAS', + os.path.basename(run_card['bias_module'])) + if not os.path.isdir(bias_module_path): + if not os.path.isdir(run_card['bias_module']): + raise InvalidCmd("The bias module at '%s' cannot be found."%run_card['bias_module']) + else: + for mandatory_file in ['makefile','%s.f'%os.path.basename(run_card['bias_module'])]: + if not os.path.isfile(pjoin(run_card['bias_module'],mandatory_file)): + raise InvalidCmd("Could not find the mandatory file '%s' in bias module '%s'."%( + mandatory_file,run_card['bias_module'])) + misc.copytree(run_card['bias_module'], pjoin(self.me_dir,'Source','BIAS', + os.path.basename(run_card['bias_module']))) + + #check expected parameters for the module. + default_bias_parameters = {} + start, last = False,False + for line in open(pjoin(bias_module_path,'%s.f'%os.path.basename(bias_module_path))): + if start and last: + break + if not start and not re.search('c\s*parameters\s*=\s*{',line, re.I): + continue + start = True + if not line.startswith('C'): + continue + line = line[1:] + if '{' in line: + line = line.split('{')[-1] + # split for } ! # + split_result = re.split('(\}|!|\#)', line,1, re.M) + line = split_result[0] + sep = split_result[1] if len(split_result)>1 else None + if sep == '}': + last = True + if ',' in line: + for pair in line.split(','): + if not pair.strip(): + continue + x,y =pair.split(':') + x=x.strip() + if x.startswith(('"',"'")) and x.endswith(x[0]): + x = x[1:-1] + default_bias_parameters[x] = y + elif ':' in line: + x,y = line.split(':') + x = x.strip() + if x.startswith(('"',"'")) and x.endswith(x[0]): + x = x[1:-1] + default_bias_parameters[x] = y + for key,value in run_card['bias_parameters'].items(): + if key not in default_bias_parameters: + logger.warning('%s not supported by the bias module. We discard this entry.', key) + else: + default_bias_parameters[key] = value + run_card['bias_parameters'] = default_bias_parameters + + + # Finally write the include file + run_card.write_include_file(opt['output_dir']) + + + if self.proc_characteristics['loop_induced'] and mode in ['loop', 'all']: + self.MadLoopparam = banner_mod.MadLoopParam(pjoin(self.me_dir, + 'Cards', 'MadLoopParams.dat')) + # The writing out of MadLoop filter is potentially dangerous + # when running in multi-core with a central disk. So it is turned + # off here. If these filters were not initialized then they will + # have to be re-computed at the beginning of each run. + if 'WriteOutFilters' in self.MadLoopparam.user_set and \ + self.MadLoopparam.get('WriteOutFilters'): + logger.info( +"""You chose to have MadLoop writing out filters. +Beware that this can be dangerous for local multicore runs.""") + self.MadLoopparam.set('WriteOutFilters',False, changeifuserset=False) + + # The conservative settings below for 'CTModeInit' and 'ZeroThres' + # help adress issues for processes like g g > h z, and g g > h g + # where there are some helicity configuration heavily suppressed + # (by several orders of magnitude) so that the helicity filter + # needs high numerical accuracy to correctly handle this spread in + # magnitude. Also, because one cannot use the Born as a reference + # scale, it is better to force quadruple precision *for the + # initialization points only*. This avoids numerical accuracy issues + # when setting up the helicity filters and does not significantly + # slow down the run. +# self.MadLoopparam.set('CTModeInit',4, changeifuserset=False) + # Consequently, we can allow for a finer threshold for vanishing + # helicity configuration +# self.MadLoopparam.set('ZeroThres',1.0e-11, changeifuserset=False) + +# It is a bit superficial to use the level 2 which tries to numerically +# map matching helicities (because of CP symmetry typically) together. +# It is useless in the context of MC over helicities and it can +# potentially make the helicity double checking fail. + self.MadLoopparam.set('HelicityFilterLevel',1, changeifuserset=False) + +# To be on the safe side however, we ask for 4 consecutive matching +# helicity filters. + self.MadLoopparam.set('CheckCycle',4, changeifuserset=False) + + # For now it is tricky to have each channel performing the helicity + # double check. What we will end up doing is probably some kind + # of new initialization round at the beginning of each launch + # command, to reset the filters. + self.MadLoopparam.set('DoubleCheckHelicityFilter',False, + changeifuserset=False) + + # Thanks to TIR recycling, TIR is typically much faster for Loop-induced + # processes when not doing MC over helicities, so that we place OPP last. + if not hasattr(self, 'run_card'): + run_card = banner_mod.RunCard(opt['run_card']) + else: + run_card = self.run_card + if run_card['nhel'] == 0: + if 'MLReductionLib' in self.MadLoopparam.user_set and \ + (self.MadLoopparam.get('MLReductionLib').startswith('1') or + self.MadLoopparam.get('MLReductionLib').startswith('6')): + logger.warning( + """You chose to set the preferred reduction technique in MadLoop to be OPP (see parameter MLReductionLib). + Beware that this can bring significant slowdown; the optimal choice --when not MC over helicity-- being to first start with TIR reduction.""") + # We do not include GOLEM for now since it cannot recycle TIR coefs yet. + self.MadLoopparam.set('MLReductionLib','7|6|1', changeifuserset=False) + else: + if 'MLReductionLib' in self.MadLoopparam.user_set and \ + not (self.MadLoopparam.get('MLReductionLib').startswith('1') or + self.MadLoopparam.get('MLReductionLib').startswith('6')): + logger.warning( + """You chose to set the preferred reduction technique in MadLoop to be different than OPP (see parameter MLReductionLib). + Beware that this can bring significant slowdown; the optimal choice --when MC over helicity-- being to first start with OPP reduction.""") + self.MadLoopparam.set('MLReductionLib','6|7|1', changeifuserset=False) + + # Also TIR cache will only work when NRotations_DP=0 (but only matters + # when not MC-ing over helicities) so it will be hard-reset by MadLoop + # to zero when not MC-ing over helicities, unless the parameter + # Force_ML_Helicity_Sum is set to True in the matrix.f codes. + if run_card['nhel'] == 0: + if ('NRotations_DP' in self.MadLoopparam.user_set and \ + self.MadLoopparam.get('NRotations_DP')!=0) or \ + ('NRotations_QP' in self.MadLoopparam.user_set and \ + self.MadLoopparam.get('NRotations_QP')!=0): + logger.warning( + """You chose to also use a lorentz rotation for stability tests (see parameter NRotations_[DP|QP]). + Beware that, for optimization purposes, MadEvent uses manual TIR cache clearing which is not compatible + with the lorentz rotation stability test. The number of these rotations to be used will be reset to + zero by MadLoop. You can avoid this by changing the parameter 'FORCE_ML_HELICITY_SUM' int he matrix.f + files to be .TRUE. so that the sum over helicity configurations is performed within MadLoop (in which case + the helicity of final state particles cannot be speicfied in the LHE file.""") + self.MadLoopparam.set('NRotations_DP',0,changeifuserset=False) + self.MadLoopparam.set('NRotations_QP',0,changeifuserset=False) + else: + # When MC-ing over helicities, the manual TIR cache clearing is + # not necessary, so that one can use the lorentz check + # Using NRotations_DP=1 slows down the code by close to 100% + # but it is typicaly safer. + # self.MadLoopparam.set('NRotations_DP',0,changeifuserset=False) + # Revert to the above to be slightly less robust but twice faster. + self.MadLoopparam.set('NRotations_DP',1,changeifuserset=False) + self.MadLoopparam.set('NRotations_QP',0,changeifuserset=False) + + # Finally, the stability tests are slightly less reliable for process + # with less or equal than 4 final state particles because the + # accessible kinematic is very limited (i.e. lorentz rotations don't + # shuffle invariants numerics much). In these cases, we therefore + # increase the required accuracy to 10^-7. + # This is important for getting g g > z z [QCD] working with a + # ptheavy cut as low as 1 GeV. + if self.proc_characteristics['nexternal']<=4: + if ('MLStabThres' in self.MadLoopparam.user_set and \ + self.MadLoopparam.get('MLStabThres')>1.0e-7): + logger.warning( + """You chose to increase the default value of the MadLoop parameter 'MLStabThres' above 1.0e-7. + Stability tests can be less reliable on the limited kinematic of processes with less or equal + than four external legs, so this is not recommended (especially not for g g > z z).""") + self.MadLoopparam.set('MLStabThres',1.0e-7,changeifuserset=False) + else: + self.MadLoopparam.set('MLStabThres',1.0e-4,changeifuserset=False) + + #write the output file + self.MadLoopparam.write(pjoin(self.me_dir,"SubProcesses","MadLoop5_resources", + "MadLoopParams.dat")) + + if self.proc_characteristics['loop_induced'] and mode in ['loop', 'all']: + # Now Update MadLoop filters if necessary (if modifications were made to + # the model parameters). + if need_MadLoopFilterUpdate: + logger.debug('Changes to the %s parameters'%type_of_change+\ + ' have been detected. Madevent will then now reinitialize'+\ + ' MadLoop filters.') + self.exec_cmd('initMadLoop -r -f') + # The need_MadLoopInit condition is just there so as to avoid useless + # printout if there is not initialization to be performed. But even + # without it, and because we call 'initMadLoop' without the '-r' option + # no time would be wasted anyway, since the existing filters would not + # be overwritten. + elif not opt['forbid_MadLoopInit'] and \ + MadLoopInitializer.need_MadLoopInit(self.me_dir): + self.exec_cmd('initMadLoop -f') + + ############################################################################ + def do_survey(self, line): + """Advanced commands: launch survey for the current process """ + + + args = self.split_arg(line) + # Check argument's validity + self.check_survey(args) + # initialize / remove lhapdf mode + + if os.path.exists(pjoin(self.me_dir,'error')): + os.remove(pjoin(self.me_dir,'error')) + + self.configure_directory() + # Save original random number + self.random_orig = self.random + logger.info("Using random number seed offset = %s" % self.random) + # Update random number + self.update_random() + self.save_random() + self.update_status('Running Survey', level=None) + if self.cluster_mode: + logger.info('Creating Jobs') + + self.total_jobs = 0 + subproc = [l.strip() for l in open(pjoin(self.me_dir, + 'SubProcesses', 'subproc.mg'))] + + P_zero_result = [] # check the number of times where they are no phase-space + + # File for the loop (for loop induced) + if os.path.exists(pjoin(self.me_dir,'SubProcesses', + 'MadLoop5_resources')) and cluster.need_transfer(self.options): + tf=tarfile.open(pjoin(self.me_dir, 'SubProcesses', + 'MadLoop5_resources.tar.gz'), 'w:gz', dereference=True) + tf.add(pjoin(self.me_dir,'SubProcesses','MadLoop5_resources'), + arcname='MadLoop5_resources') + tf.close() + + logger.info('Working on SubProcesses') + ajobcreator = gen_ximprove.gensym(self) + + #check difficult PS case + if float(self.run_card['mmjj']) > 0.01 * (float(self.run_card['ebeam1'])+float(self.run_card['ebeam2'])): + self.pass_in_difficult_integration_mode() + elif self.run_card['hard_survey']: + self.pass_in_difficult_integration_mode(self.run_card['hard_survey']) + + if self.proc_characteristics['hel_recycling'] and self.run_card['hel_recycling']: + jobs, P_zero_result = ajobcreator.get_helicity() + else: + for p in subproc: + for f in misc.glob('matrix*_orig.f', pjoin(self.me_dir, 'SubProcesses', p)): + new_file = f.replace('_orig','_optim') + files.cp(f, f.replace('_orig','_optim')) + f = '%s.o' % f[:-2] + if os.path.exists(f): + files.cp(f, f.replace('_orig','_optim')) + try: + os.remove(pjoin(self.me_dir, 'SubProcesses', p, 'Hel', 'selection')) + except Exception as error: + logger.debug(error) + pass + + jobs, P_zero_result = ajobcreator.launch() + # Check if all or only some fails + if P_zero_result: + if len(P_zero_result) == len(subproc): + Pdir = pjoin(self.me_dir, 'SubProcesses',subproc[0].strip()) + raise ZeroResult('%s' % \ + open(pjoin(Pdir,'ajob.no_ps.log')).read()) + else: + logger.warning(''' %s SubProcesses doesn\'t have available phase-space. + Please check mass spectrum.''' % ','.join(P_zero_result)) + self.get_Gdir() + for P in P_zero_result: + self.Gdirs[0][pjoin(self.me_dir,'SubProcesses',P)] = [] + + self.monitor(run_type='All jobs submitted for survey', html=True) + if not self.history or 'survey' in self.history[-1] or self.ninitial ==1 or \ + self.run_card['gridpack']: + #will be done during the refine (more precisely in gen_ximprove) + cross, error = self.make_make_all_html_results() + self.results.add_detail('cross', cross) + self.results.add_detail('error', error) + self.exec_cmd("print_results %s" % self.run_name, + errorhandling=False, printcmd=False, precmd=False, postcmd=False) + + self.results.add_detail('run_statistics', dict(ajobcreator.run_statistics)) + self.update_status('End survey', 'parton', makehtml=False) + + ############################################################################ + def pass_in_difficult_integration_mode(self, rate=1): + """be more secure for the integration to not miss it due to strong cut""" + + # improve survey options if default + if self.opts['points'] == self._survey_options['points'][1]: + self.opts['points'] = (rate+2) * self._survey_options['points'][1] + if self.opts['iterations'] == self._survey_options['iterations'][1]: + self.opts['iterations'] = 1 + rate + self._survey_options['iterations'][1] + if self.opts['accuracy'] == self._survey_options['accuracy'][1]: + self.opts['accuracy'] = self._survey_options['accuracy'][1]/(rate+2) + + # Modify run_config.inc in order to improve the refine + conf_path = pjoin(self.me_dir, 'Source','run_config.inc') + files.cp(conf_path, conf_path + '.bk') + # + text = open(conf_path).read() + min_evt, max_evt = 2500 *(2+rate), 10000*(rate+1) + + text = re.sub('''\(min_events = \d+\)''', '(min_events = %i )' % min_evt, text) + text = re.sub('''\(max_events = \d+\)''', '(max_events = %i )' % max_evt, text) + fsock = open(conf_path, 'w') + fsock.write(text) + fsock.close() + + # Compile + for name in ['../bin/internal/gen_ximprove', 'all']: + self.compile(arg=[name], cwd=os.path.join(self.me_dir, 'Source')) + + + ############################################################################ + def do_refine(self, line): + """Advanced commands: launch survey for the current process """ + devnull = open(os.devnull, 'w') + self.nb_refine += 1 + args = self.split_arg(line) + treshold=None + + + + for a in args: + if a.startswith('--treshold='): + treshold = float(a.split('=',1)[1]) + old_xsec = self.results.current['prev_cross'] + new_xsec = self.results.current['cross'] + if old_xsec > new_xsec * treshold: + logger.info('No need for second refine due to stability of cross-section') + return + else: + args.remove(a) + break + # Check argument's validity + self.check_refine(args) + + refine_opt = {'err_goal': args[0], 'split_channels': True} + precision = args[0] + if len(args) == 2: + refine_opt['max_process']= args[1] + + # initialize / remove lhapdf mode + self.configure_directory() + + # Update random number + self.update_random() + self.save_random() + + if self.cluster_mode: + logger.info('Creating Jobs') + self.update_status('Refine results to %s' % precision, level=None) + + self.total_jobs = 0 + subproc = [l.strip() for l in open(pjoin(self.me_dir,'SubProcesses', + 'subproc.mg'))] + + # cleanning the previous job + for nb_proc,subdir in enumerate(subproc): + subdir = subdir.strip() + Pdir = pjoin(self.me_dir, 'SubProcesses', subdir) + for match in misc.glob('*ajob*', Pdir): + if os.path.basename(match)[:4] in ['ajob', 'wait', 'run.', 'done']: + os.remove(match) + + x_improve = gen_ximprove.gen_ximprove(self, refine_opt) + # Load the run statistics from the survey + survey_statistics = dict(self.results.get_detail('run_statistics')) + # Printout survey statistics + if __debug__ and survey_statistics: + globalstat = sum_html.RunStatistics() + logger.debug(" === Survey statistics summary ===") + for key, value in survey_statistics.items(): + globalstat.aggregate_statistics(value) + level = 5 + if value.has_warning(): + level = 10 + logger.log(level, + value.nice_output(str('/'.join([key[0],'G%s'%key[1]]))). + replace(' statistics','')) + logger.debug(globalstat.nice_output('combined', no_warning=True)) + + if survey_statistics: + x_improve.run_statistics = survey_statistics + + x_improve.launch() # create the ajob for the refinment. + if not self.history or 'refine' not in self.history[-1]: + cross, error = x_improve.update_html() #update html results for survey + if cross == 0: + return + logger.info("- Current estimate of cross-section: %s +- %s" % (cross, error)) + if isinstance(x_improve, gen_ximprove.gen_ximprove_v4): + # Non splitted mode is based on writting ajob so need to track them + # Splitted mode handle the cluster submition internally. + for nb_proc,subdir in enumerate(subproc): + subdir = subdir.strip() + Pdir = pjoin(self.me_dir, 'SubProcesses',subdir) + bindir = pjoin(os.path.relpath(self.dirbin, Pdir)) + + logger.info(' %s ' % subdir) + + if os.path.exists(pjoin(Pdir, 'ajob1')): + self.compile(['madevent'], cwd=Pdir) + + alljobs = misc.glob('ajob*', Pdir) + + #remove associated results.dat (ensure to not mix with all data) + Gre = re.compile("\s*j=(G[\d\.\w]+)") + for job in alljobs: + Gdirs = Gre.findall(open(job).read()) + for Gdir in Gdirs: + if os.path.exists(pjoin(Pdir, Gdir, 'results.dat')): + os.remove(pjoin(Pdir, Gdir,'results.dat')) + + nb_tot = len(alljobs) + self.total_jobs += nb_tot + for i, job in enumerate(alljobs): + job = os.path.basename(job) + self.launch_job('%s' % job, cwd=Pdir, remaining=(nb_tot-i-1), + run_type='Refine number %s on %s (%s/%s)' % + (self.nb_refine, subdir, nb_proc+1, len(subproc))) + + + self.monitor(run_type='All job submitted for refine number %s' % self.nb_refine, + html=True) + + self.update_status("Combining runs", level='parton') + try: + os.remove(pjoin(Pdir, 'combine_runs.log')) + except Exception: + pass + + if isinstance(x_improve, gen_ximprove.gen_ximprove_v4): + # the merge of the events.lhe is handle in the x_improve class + # for splitted runs. (and partly in store_events). + combine_runs.CombineRuns(self.me_dir) + self.refine_mode = "old" + else: + self.refine_mode = "new" + + cross, error = self.make_make_all_html_results() + self.results.add_detail('cross', cross) + self.results.add_detail('error', error) + + self.results.add_detail('run_statistics', + dict(self.results.get_detail('run_statistics'))) + + self.update_status('finish refine', 'parton', makehtml=False) + devnull.close() + + ############################################################################ + def do_comine_iteration(self, line): + """Not in help: Combine a given iteration combine_iteration Pdir Gdir S|R step + S is for survey + R is for refine + step is the iteration number (not very critical)""" + + self.set_run_name("tmp") + self.configure_directory(html_opening=False) + Pdir, Gdir, mode, step = self.split_arg(line) + if Gdir.startswith("G"): + Gdir = Gdir[1:] + if "SubProcesses" not in Pdir: + Pdir = pjoin(self.me_dir, "SubProcesses", Pdir) + if mode == "S": + self.opts = dict([(key,value[1]) for (key,value) in \ + self._survey_options.items()]) + gensym = gen_ximprove.gensym(self) + gensym.combine_iteration(Pdir, Gdir, int(step)) + elif mode == "R": + refine = gen_ximprove.gen_ximprove_share(self) + refine.combine_iteration(Pdir, Gdir, int(step)) + + + + + ############################################################################ + def do_combine_events(self, line): + """Advanced commands: Launch combine events""" + start=time.time() + args = self.split_arg(line) + start = time.time() + # Check argument's validity + self.check_combine_events(args) + self.update_status('Combining Events', level='parton') + + + if self.run_card['gridpack'] and isinstance(self, GridPackCmd): + return GridPackCmd.do_combine_events(self, line) + + + # Define The Banner + tag = self.run_card['run_tag'] + # Update the banner with the pythia card + if not self.banner: + self.banner = banner_mod.recover_banner(self.results, 'parton') + self.banner.load_basic(self.me_dir) + # Add cross-section/event information + self.banner.add_generation_info(self.results.current['cross'], self.run_card['nevents']) + if not hasattr(self, 'random_orig'): self.random_orig = 0 + self.banner.change_seed(self.random_orig) + if not os.path.exists(pjoin(self.me_dir, 'Events', self.run_name)): + os.mkdir(pjoin(self.me_dir, 'Events', self.run_name)) + self.banner.write(pjoin(self.me_dir, 'Events', self.run_name, + '%s_%s_banner.txt' % (self.run_name, tag))) + + + get_wgt = lambda event: event.wgt + AllEvent = lhe_parser.MultiEventFile() + AllEvent.banner = self.banner + + partials = 0 # if too many file make some partial unweighting + sum_xsec, sum_xerru, sum_axsec = 0,[],0 + Gdirs = self.get_Gdir() + Gdirs.sort() + for Gdir in Gdirs: + if os.path.exists(pjoin(Gdir, 'events.lhe')): + result = sum_html.OneResult('') + result.read_results(pjoin(Gdir, 'results.dat')) + sum_xsec += result.get('xsec') + sum_xerru.append(result.get('xerru')) + sum_axsec += result.get('axsec') + + if self.run_card['gridpack'] or self.run_card['nevents']==0: + os.remove(pjoin(Gdir, 'events.lhe')) + continue + + AllEvent.add(pjoin(Gdir, 'events.lhe'), + result.get('xsec'), + result.get('xerru'), + result.get('axsec') + ) + + if len(AllEvent) >= 80: #perform a partial unweighting + AllEvent.unweight(pjoin(self.me_dir, "Events", self.run_name, "partials%s.lhe.gz" % partials), + get_wgt, log_level=5, trunc_error=1e-2, event_target=self.run_card['nevents']) + AllEvent = lhe_parser.MultiEventFile() + AllEvent.banner = self.banner + AllEvent.add(pjoin(self.me_dir, "Events", self.run_name, "partials%s.lhe.gz" % partials), + sum_xsec, + math.sqrt(sum(x**2 for x in sum_xerru)), + sum_axsec) + partials +=1 + + if not hasattr(self,'proc_characteristic'): + self.proc_characteristic = self.get_characteristics() + if len(AllEvent) == 0: + nb_event = 0 + else: + nb_event = AllEvent.unweight(pjoin(self.me_dir, "Events", self.run_name, "unweighted_events.lhe.gz"), + get_wgt, trunc_error=1e-2, event_target=self.run_card['nevents'], + log_level=logging.DEBUG, normalization=self.run_card['event_norm'], + proc_charac=self.proc_characteristic) + + if nb_event < self.run_card['nevents']: + logger.warning("failed to generate enough events. Please follow one of the following suggestions to fix the issue:") + logger.warning(" - set in the run_card.dat 'sde_strategy' to %s", 1 + self.run_card['sde_strategy'] % 2) + logger.warning(" - set in the run_card.dat 'hard_survey' to 1 or 2.") + logger.warning(" - reduce the number of requested events (if set too high)") + logger.warning(" - check that you do not have -integrable- singularity in your amplitude.") + + if partials: + for i in range(partials): + try: + os.remove(pjoin(self.me_dir, "Events", self.run_name, "partials%s.lhe.gz" % i)) + except Exception: + os.remove(pjoin(self.me_dir, "Events", self.run_name, "partials%s.lhe" % i)) + + self.results.add_detail('nb_event', nb_event) + + if self.run_card['bias_module'].lower() not in ['dummy', 'none'] and nb_event: + self.correct_bias() + elif self.run_card['custom_fcts']: + self.correct_bias() + logger.info("combination of events done in %s s ", time.time()-start) + + self.to_store.append('event') + + ############################################################################ + def correct_bias(self): + """check the first event and correct the weight by the bias + and correct the cross-section. + If the event do not have the bias tag it means that the bias is + one modifying the cross-section/shape so we have nothing to do + """ + + lhe = lhe_parser.EventFile(pjoin(self.me_dir, 'Events', self.run_name, 'unweighted_events.lhe.gz')) + init = False + cross = collections.defaultdict(float) + nb_event = 0 + for event in lhe: + rwgt_info = event.parse_reweight() + if not init: + if 'bias' in rwgt_info: + output = lhe_parser.EventFile(pjoin(self.me_dir, 'Events', self.run_name, '.unweighted_events.lhe.tmp.gz'),'w') + #output.write(lhe.banner) + init = True + else: + return + #change the weight + event.wgt /= rwgt_info['bias'] + #remove the bias info + del event.reweight_data['bias'] + # compute the new cross-section + cross[event.ievent] += event.wgt + nb_event +=1 + output.write(str(event)) + output.write('') + output.close() + lhe.close() + + # MODIFY THE BANNER i.e. INIT BLOCK + # ensure information compatible with normalisation choice + total_cross = sum(cross[key] for key in cross) + if 'event_norm' in self.run_card: # if not this is "sum" + if self.run_card['event_norm'] == 'average': + total_cross = total_cross / nb_event + for key in cross: + cross[key] /= nb_event + elif self.run_card['event_norm'] == 'unity': + total_cross = self.results.current['cross'] * total_cross / nb_event + for key in cross: + cross[key] *= total_cross / nb_event + + bannerfile = lhe_parser.EventFile(pjoin(self.me_dir, 'Events', self.run_name, '.banner.tmp.gz'),'w') + banner = banner_mod.Banner(lhe.banner) + banner.modify_init_cross(cross) + banner.set_lha_strategy(-4) + banner.write(bannerfile, close_tag=False) + bannerfile.close() + # replace the lhe file by the new one + if lhe.name.endswith('.gz'): + os.system('cat %s %s > %s' %(bannerfile.name, output.name, lhe.name)) + else: + os.system('cat %s %s > %s.gz' %(bannerfile.name, output.name, lhe.name)) + os.remove(lhe.name) + os.remove(bannerfile.name) + os.remove(output.name) + + + self.results.current['cross'] = total_cross + self.results.current['error'] = 0 + + ############################################################################ + def do_store_events(self, line): + """Advanced commands: Launch store events""" + + args = self.split_arg(line) + # Check argument's validity + self.check_combine_events(args) + self.update_status('Storing parton level results', level='parton') + + run = self.run_name + tag = self.run_card['run_tag'] + devnull = open(os.devnull, 'w') + + if not os.path.exists(pjoin(self.me_dir, 'Events', run)): + os.mkdir(pjoin(self.me_dir, 'Events', run)) + if not os.path.exists(pjoin(self.me_dir, 'HTML', run)): + os.mkdir(pjoin(self.me_dir, 'HTML', run)) + + # 1) Store overall process information + #input = pjoin(self.me_dir, 'SubProcesses', 'results.dat') + #output = pjoin(self.me_dir, 'SubProcesses', '%s_results.dat' % run) + #files.cp(input, output) + + + # 2) Treat the files present in the P directory + # Ensure that the number of events is different of 0 + if self.results.current['nb_event'] == 0 and not self.run_card['gridpack']: + logger.warning("No event detected. No cleaning performed! This should allow to run:\n" + + " cd Subprocesses; ../bin/internal/combine_events\n"+ + " to have your events if those one are missing.") + else: + for G_path in self.get_Gdir(): + try: + # Remove events file (if present) + if os.path.exists(pjoin(G_path, 'events.lhe')): + os.remove(pjoin(G_path, 'events.lhe')) + except Exception: + continue + #try: + # # Store results.dat + # if os.path.exists(pjoin(G_path, 'results.dat')): + # input = pjoin(G_path, 'results.dat') + # output = pjoin(G_path, '%s_results.dat' % run) + # files.cp(input, output) + #except Exception: + # continue + # Store log + try: + if os.path.exists(pjoin(G_path, 'log.txt')): + input = pjoin(G_path, 'log.txt') + output = pjoin(G_path, '%s_log.txt' % run) + files.mv(input, output) + except Exception: + continue + #try: + # # Grid + # for name in ['ftn26']: + # if os.path.exists(pjoin(G_path, name)): + # if os.path.exists(pjoin(G_path, '%s_%s.gz'%(run,name))): + # os.remove(pjoin(G_path, '%s_%s.gz'%(run,name))) + # input = pjoin(G_path, name) + # output = pjoin(G_path, '%s_%s' % (run,name)) + # files.mv(input, output) + # misc.gzip(pjoin(G_path, output), error=None) + #except Exception: + # continue + # Delete ftn25 to ensure reproducible runs + if os.path.exists(pjoin(G_path, 'ftn25')): + os.remove(pjoin(G_path, 'ftn25')) + + # 3) Update the index.html + self.gen_card_html() + + + # 4) Move the Files present in Events directory + E_path = pjoin(self.me_dir, 'Events') + O_path = pjoin(self.me_dir, 'Events', run) + + # The events file + for name in ['events.lhe', 'unweighted_events.lhe']: + finput = pjoin(E_path, name) + foutput = pjoin(O_path, name) + if os.path.exists(finput): + logger.debug("File %s exists BAAAAD. Not move anymore!" % pjoin(E_path, name)) + if os.path.exists(foutput): + if os.path.exists("%s.gz" % foutput): + os.remove(foutput) + else: + misc.gzip(foutput, stdout="%s.gz" % foutput, error=False) + # if os.path.exists(pjoin(O_path, '%s.gz' % name)): + # os.remove(pjoin(O_path, '%s.gz' % name)) + # input = pjoin(E_path, name) + ## output = pjoin(O_path, name) + + + self.update_status('End Parton', level='parton', makehtml=False) + devnull.close() + + + ############################################################################ + def do_create_gridpack(self, line): + """Advanced commands: Create gridpack from present run""" + + self.update_status('Creating gridpack', level='parton') + # compile gen_ximprove + misc.compile(['../bin/internal/gen_ximprove'], cwd=pjoin(self.me_dir, "Source")) + + Gdir = self.get_Gdir() + Pdir = set([os.path.dirname(G) for G in Gdir]) + for P in Pdir: + allG = misc.glob('G*', path=P) + for G in allG: + # avoid case where some file starts with G (madgraph5/madgraph4gpu#947) + if not os.path.isdir(G): + continue + if pjoin(P, G) not in Gdir: + logger.debug('removing %s', pjoin(P,G)) + shutil.rmtree(pjoin(P,G)) + + + args = self.split_arg(line) + self.check_combine_events(args) + if not self.run_tag: self.run_tag = 'tag_1' + os.system("sed -i.bak \"s/ *.false.*=.*GridRun/ .true. = GridRun/g\" %s/Cards/grid_card.dat" \ + % self.me_dir) + misc.call(['./bin/internal/restore_data', self.run_name], + cwd=self.me_dir) + misc.call(['./bin/internal/store4grid', + self.run_name, self.run_tag], + cwd=self.me_dir) + misc.call(['./bin/internal/clean'], cwd=self.me_dir) + misc.call(['./bin/internal/make_gridpack'], cwd=self.me_dir) + files.mv(pjoin(self.me_dir, 'gridpack.tar.gz'), + pjoin(self.me_dir, '%s_gridpack.tar.gz' % self.run_name)) + os.system("sed -i.bak \"s/\s*.true.*=.*GridRun/ .false. = GridRun/g\" %s/Cards/grid_card.dat" \ + % self.me_dir) + self.update_status('gridpack created', level='gridpack') + + ############################################################################ + def do_shower(self, line): + """launch the shower""" + + args = self.split_arg(line) + if len(args)>1 and args[0] in self._interfaced_showers: + chosen_showers = [args.pop(0)] + elif '--no_default' in line: + # If '--no_default' was specified in the arguments, then only one + # shower will be run, depending on which card is present. + # but we each of them are called. (each of them check if the file exists) + chosen_showers = list(self._interfaced_showers) + else: + chosen_showers = list(self._interfaced_showers) + # It is preferable to run only one shower, even if several are available and no + # specific one has been selected + shower_priority = ['pythia8','pythia'] + chosen_showers = [sorted(chosen_showers,key=lambda sh: + shower_priority.index(sh) if sh in shower_priority else len(shower_priority)+1)[0]] + + for shower in chosen_showers: + self.exec_cmd('%s %s'%(shower,' '.join(args)), + postcmd=False, printcmd=False) + + def do_madanalysis5_parton(self, line): + """launch MadAnalysis5 at the parton level.""" + return self.run_madanalysis5(line,mode='parton') + + #=============================================================================== + # Return a warning (if applicable) on the consistency of the current Pythia8 + # and MG5_aMC version specified. It is placed here because it should be accessible + # from both madgraph5_interface and madevent_interface + #=============================================================================== + @staticmethod + def mg5amc_py8_interface_consistency_warning(options): + """ Check the consistency of the mg5amc_py8_interface installed with + the current MG5 and Pythia8 versions. """ + + # All this is only relevant is Pythia8 is interfaced to MG5 + if not options['pythia8_path']: + return None + + if not options['mg5amc_py8_interface_path']: + return \ + """ + A Pythia8 path is specified via the option 'pythia8_path' but no path for option + 'mg5amc_py8_interface_path' is specified. This means that Pythia8 cannot be used + leading order simulations with MadEvent. + Consider installing the MG5_aMC-PY8 interface with the following command: + MG5_aMC>install mg5amc_py8_interface + """ + + mg5amc_py8_interface_path = options['mg5amc_py8_interface_path'] + py8_path = options['pythia8_path'] + # If the specified interface path is relative, make it absolut w.r.t MGDIR if + # avaialble. + if not MADEVENT: + mg5amc_py8_interface_path = pjoin(MG5DIR,mg5amc_py8_interface_path) + py8_path = pjoin(MG5DIR,py8_path) + + # Retrieve all the on-install and current versions + fsock = open(pjoin(mg5amc_py8_interface_path, 'MG5AMC_VERSION_ON_INSTALL')) + MG5_version_on_install = fsock.read().replace('\n','') + fsock.close() + if MG5_version_on_install == 'UNSPECIFIED': + MG5_version_on_install = None + fsock = open(pjoin(mg5amc_py8_interface_path, 'PYTHIA8_VERSION_ON_INSTALL')) + PY8_version_on_install = fsock.read().replace('\n','') + fsock.close() + MG5_curr_version =misc.get_pkg_info()['version'] + try: + p = subprocess.Popen(['./get_pythia8_version.py',py8_path], + stdout=subprocess.PIPE, stderr=subprocess.PIPE, + cwd=mg5amc_py8_interface_path) + (out, err) = p.communicate() + out = out.decode(errors='ignore').replace('\n','') + PY8_curr_version = out + # In order to test that the version is correctly formed, we try to cast + # it to a float + float(out) + except: + PY8_curr_version = None + + if not MG5_version_on_install is None and not MG5_curr_version is None: + if MG5_version_on_install != MG5_curr_version: + return \ + """ + The current version of MG5_aMC (v%s) is different than the one active when + installing the 'mg5amc_py8_interface_path' (which was MG5aMC v%s). + Please consider refreshing the installation of this interface with the command: + MG5_aMC>install mg5amc_py8_interface + """%(MG5_curr_version, MG5_version_on_install) + + if not PY8_version_on_install is None and not PY8_curr_version is None: + if PY8_version_on_install != PY8_curr_version: + return \ + """ + The current version of Pythia8 (v%s) is different than the one active when + installing the 'mg5amc_py8_interface' tool (which was Pythia8 v%s). + Please consider refreshing the installation of this interface with the command: + MG5_aMC>install mg5amc_py8_interface + """%(PY8_curr_version,PY8_version_on_install) + + return None + + def setup_Pythia8RunAndCard(self, PY8_Card, run_type): + """ Setup the Pythia8 Run environment and card. In particular all the process and run specific parameters + of the card are automatically set here. This function returns the path where HEPMC events will be output, + if any.""" + + HepMC_event_output = None + tag = self.run_tag + + PY8_Card.subruns[0].systemSet('Beams:LHEF',"unweighted_events.lhe.gz") + + hepmc_format = PY8_Card['HEPMCoutput:file'].lower() + if hepmc_format == "auto": + hepmc_format = "hepmc.gz" + elif hepmc_format == "autoremove": + hepmc_format = "hepmcremove" + + # output format : hepmc/fifo + if hepmc_format.startswith("hepmc"): + + hepmc_specs = hepmc_format.split('@') + hepmc_path = pjoin(self.me_dir,'Events', self.run_name, '%s_pythia8_events.hepmc'%tag) + + # In case @ is given (output path) + if len(hepmc_specs) > 1: + if os.path.isabs(hepmc_specs[1]): + if os.path.exists(hepmc_specs[1]): + os.mkdir(pjoin(hepmc_specs[1], self.run_name)) + self.to_store.append("moveHEPMC@" + pjoin(hepmc_specs[1], self.run_name)) + else: + logger.warning("%s does not exist, using default output path"%hepmc_specs[1]) + else: + self.to_store.append("moveHEPMC@" + pjoin(self.me_dir, 'Events', hepmc_specs[1], self.run_name)) + os.mkdir(pjoin(self.me_dir, 'Events', hepmc_specs[1], self.run_name)) + + # Compress if .gz is given + if hepmc_specs[0].endswith(".gz"): + if not 'compressHEPMC' in self.to_store: + self.to_store.append('compressHEPMC') + else: + if 'compressHEPMC' in self.to_store: + self.to_store.remove('compressHEPMC') + + # Remove if remove is given + if hepmc_specs[0].endswith("remove"): + if not 'removeHEPMC' in self.to_store: + self.to_store.append('removeHEPMC') + else: + if 'removeHEPMC' in self.to_store: + self.to_store.remove('removeHEPMC') + + HepMC_event_output=hepmc_path + PY8_Card.MadGraphSet('HEPMCoutput:file','%s_pythia8_events.hepmc'%tag, force=True) + + elif hepmc_format.startswith('fifo'): + fifo_specs = hepmc_format.split('@') + fifo_path = None + if len(fifo_specs)<=1: + fifo_path = pjoin(self.me_dir,'Events', self.run_name,'PY8.hepmc.fifo') + if os.path.exists(fifo_path): + os.remove(fifo_path) + misc.mkfifo(fifo_path) + # Use defaultSet not to overwrite the current userSet status + PY8_Card.defaultSet('HEPMCoutput:file','PY8.hepmc.fifo') + else: + fifo_path = fifo_specs[1] + if os.path.exists(fifo_path): + if stat.S_ISFIFO(os.stat(fifo_path).st_mode): + logger.warning('PY8 will be reusing already existing '+ + 'custom fifo file at:\n %s'%fifo_path) + else: + raise InvalidCmd( +"""The fifo path speficied for the PY8 parameter 'HEPMCoutput:file': + %s +already exists and is not a fifo file."""%fifo_path) + else: + misc.mkfifo(fifo_path) + # Use defaultSet not to overwrite the current userSet status + PY8_Card.defaultSet('HEPMCoutput:file',fifo_path) + HepMC_event_output=fifo_path + elif hepmc_format in ['','/dev/null','None']: + logger.warning('User disabled the HepMC output of Pythia8.') + HepMC_event_output = None + else: + raise InvalidCmd("Unknow HEPMCoutput:file setting, hepmc/hepmc.gz/hepmcremove/fifo") + + # We specify by hand all necessary parameters, so that there is no + # need to read parameters from the Banner. + PY8_Card.MadGraphSet('JetMatching:setMad', False) + if run_type=='MLM': + # When running MLM make sure that we do not write out the parameter + # Merging:xxx as this can interfere with the MLM merging in older + # versions of the driver. + PY8_Card.vetoParamWriteOut('Merging:TMS') + PY8_Card.vetoParamWriteOut('Merging:Process') + PY8_Card.vetoParamWriteOut('Merging:nJetMax') + # MadGraphSet sets the corresponding value (in system mode) + # only if it is not already user_set. + if PY8_Card['JetMatching:qCut']==-1.0: + PY8_Card.MadGraphSet('JetMatching:qCut',1.5*self.run_card['xqcut'], force=True) + + if PY8_Card['JetMatching:qCut']<(1.5*self.run_card['xqcut']): + logger.error( + 'The MLM merging qCut parameter you chose (%f) is less than '%PY8_Card['JetMatching:qCut']+ + '1.5*xqcut, with xqcut your run_card parameter (=%f).\n'%self.run_card['xqcut']+ + 'It would be better/safer to use a larger qCut or a smaller xqcut.') + + # Also make sure to use the shower starting scales specified in the LHE + # unless the user specified it + PY8_Card.systemSet('Beams:setProductionScalesFromLHEF',True) + + # Automatically set qWeed to xqcut if not defined by the user. + if PY8_Card['SysCalc:qWeed']==-1.0: + PY8_Card.MadGraphSet('SysCalc:qWeed',self.run_card['xqcut'], force=True) + + if PY8_Card['SysCalc:qCutList']=='auto': + if self.run_card['use_syst']: + if self.run_card['sys_matchscale']=='auto': + qcut = PY8_Card['JetMatching:qCut'] + value = [factor*qcut for factor in [0.5,0.75,1.0,1.5,2.0] if\ + factor*qcut> 1.5*self.run_card['xqcut'] ] + PY8_Card.MadGraphSet('SysCalc:qCutList', value, force=True) + else: + qCutList = [float(qc) for qc in self.run_card['sys_matchscale'].split()] + if PY8_Card['JetMatching:qCut'] not in qCutList: + qCutList.append(PY8_Card['JetMatching:qCut']) + PY8_Card.MadGraphSet('SysCalc:qCutList', qCutList, force=True) + + + if PY8_Card['SysCalc:qCutList']!='auto': + for scale in PY8_Card['SysCalc:qCutList']: + if scale<(1.5*self.run_card['xqcut']): + logger.error( + 'One of the MLM merging qCut parameter you chose (%f) in the variation list'%scale+\ + " (either via 'SysCalc:qCutList' in the PY8 shower card or "+\ + "'sys_matchscale' in the run_card) is less than 1.5*xqcut, where xqcut is"+ + ' the run_card parameter (=%f)\n'%self.run_card['xqcut']+ + 'It would be better/safer to use a larger qCut or a smaller xqcut.') + + # Specific MLM settings + # PY8 should not implement the MLM veto since the driver should do it + # if merging scale variation is turned on + if self.run_card['use_syst']: + # We do no force it here, but it is clear that the user should know what + # he's doing if he were to force it to True. + PY8_Card.MadGraphSet('JetMatching:doVeto',False) + PY8_Card.MadGraphSet('JetMatching:merge',True) + PY8_Card.MadGraphSet('JetMatching:scheme',1) + # Use the parameter maxjetflavor for JetMatching:nQmatch which specifies + # up to which parton must be matched.Merging:nQuarksMerge + PY8_Card.MadGraphSet('JetMatching:nQmatch',self.run_card['maxjetflavor']) + # For MLM, a cone radius of 1.0 is to be prefered. + PY8_Card.MadGraphSet('JetMatching:coneRadius',1.0) + # And the value of etaj_max is already infinity by default. + # PY8_Card.MadGraphSet('JetMatching:etaJetMax',1000.0) + if not hasattr(self,'proc_characteristic'): + self.proc_characteristic = self.get_characteristics() + nJetMax = self.proc_characteristic['max_n_matched_jets'] + if PY8_Card['JetMatching:nJetMax'.lower()] == -1: + logger.info("No user-defined value for Pythia8 parameter "+ + "'JetMatching:nJetMax'. Setting it automatically to %d."%nJetMax) + PY8_Card.MadGraphSet('JetMatching:nJetMax',nJetMax, force=True) + # We use the positivity of 'ktdurham' cut as a CKKWl marker. + elif run_type=='CKKW': + + # Make sure the user correctly filled in the lowest order process to be considered + if PY8_Card['Merging:Process']=='': + raise self.InvalidCmd('When running CKKWl merging, the user must'+ + " specifiy the option 'Merging:Process' in pythia8_card.dat.\n"+ + "Read section 'Defining the hard process' of "+\ + "http://home.thep.lu.se/~torbjorn/pythia81html/CKKWLMerging.html for more information.") + + # When running CKKWL make sure that we do not write out the parameter + # JetMatching:xxx as this can interfere with the MLM merging in older + # versions of the driver. + PY8_Card.vetoParamWriteOut('JetMatching:qCut') + PY8_Card.vetoParamWriteOut('JetMatching:doShowerKt') + PY8_Card.vetoParamWriteOut('JetMatching:nJetMax') + + CKKW_cut = None + # Specific CKKW settings + if self.run_card['ptlund']<=0.0 and self.run_card['ktdurham']>0.0: + PY8_Card.subruns[0].MadGraphSet('Merging:doKTMerging',True) + PY8_Card.subruns[0].MadGraphSet('Merging:Dparameter', + self.run_card['dparameter']) + CKKW_cut = 'ktdurham' + elif self.run_card['ptlund']>0.0 and self.run_card['ktdurham']<=0.0: + PY8_Card.subruns[0].MadGraphSet('Merging:doPTLundMerging',True) + CKKW_cut = 'ptlund' + else: + raise InvalidCmd("*Either* the 'ptlund' or 'ktdurham' cut in "+\ + " the run_card must be turned on to activate CKKW(L) merging"+ + " with Pythia8, but *both* cuts cannot be turned on at the same time."+ + "\n ptlund=%f, ktdurham=%f."%(self.run_card['ptlund'],self.run_card['ktdurham'])) + + + # Automatically set qWeed to the CKKWL cut if not defined by the user. + if PY8_Card['SysCalc:qWeed']==-1.0: + PY8_Card.MadGraphSet('SysCalc:qWeed',self.run_card[CKKW_cut], force=True) + + # MadGraphSet sets the corresponding value (in system mode) + # only if it is not already user_set. + if PY8_Card['Merging:TMS']==-1.0: + if self.run_card[CKKW_cut]>0.0: + PY8_Card.MadGraphSet('Merging:TMS',self.run_card[CKKW_cut], force=True) + else: + raise self.InvalidCmd('When running CKKWl merging, the user'+\ + " select a '%s' cut larger than 0.0 in the run_card."%CKKW_cut) + if PY8_Card['Merging:TMS'] self.run_card[CKKW_cut]] + PY8_Card.MadGraphSet('SysCalc:tmsList', value, force=True) + else: + tmsList = [float(tms) for tms in self.run_card['sys_matchscale'].split()] + if PY8_Card['Merging:TMS'] not in tmsList: + tmsList.append(PY8_Card['Merging:TMS']) + PY8_Card.MadGraphSet('SysCalc:tmsList', tmsList, force=True) + #else: + # PY8_Card.MadGraphSet('SysCalc:tmsList', [], force=True) + if PY8_Card['SysCalc:tmsList']!='auto': + for scale in PY8_Card['SysCalc:tmsList']: + if float(scale) install mg5amc_py8_interface_path""") + else: + pythia_main = pjoin(self.options['mg5amc_py8_interface_path'], + 'MG5aMC_PY8_interface') + warnings = MadEventCmd.mg5amc_py8_interface_consistency_warning(self.options) + if warnings: + logger.warning(warnings) + + self.results.add_detail('run_mode', 'madevent') + + # Again here 'pythia' is just a keyword for the simulation level. + self.update_status('\033[92mRunning Pythia8 [arXiv:1410.3012]\033[0m', 'pythia8') + + tag = self.run_tag + # Now write Pythia8 card + # Start by reading, starting from the default one so that the 'user_set' + # tag are correctly set. + PY8_Card = banner_mod.PY8Card(pjoin(self.me_dir, 'Cards', + 'pythia8_card_default.dat')) + PY8_Card.read(pjoin(self.me_dir, 'Cards', 'pythia8_card.dat'), + setter='user') + + run_type = 'default' + merged_run_types = ['MLM','CKKW'] + if int(self.run_card['ickkw'])==1: + run_type = 'MLM' + elif int(self.run_card['ickkw'])==2 or \ + self.run_card['ktdurham']>0.0 or self.run_card['ptlund']>0.0: + run_type = 'CKKW' + + # Edit the card and run environment according to the run specification + HepMC_event_output = self.setup_Pythia8RunAndCard(PY8_Card, run_type) + + # Now write the card. + pythia_cmd_card = pjoin(self.me_dir, 'Events', self.run_name , + '%s_pythia8.cmd' % tag) + cmd_card = StringIO.StringIO() + PY8_Card.write(cmd_card,pjoin(self.me_dir,'Cards','pythia8_card_default.dat'), + direct_pythia_input=True) + + # Now setup the preamble to make sure that everything will use the locally + # installed tools (if present) even if the user did not add it to its + # environment variables. + if 'heptools_install_dir' in self.options: + preamble = misc.get_HEPTools_location_setter( + self.options['heptools_install_dir'],'lib') + else: + if MADEVENT: + preamble = misc.get_HEPTools_location_setter( + pjoin(self.options['mg5amc_py8_interface_path'],os.pardir),'lib') + else: + preamble = misc.get_HEPTools_location_setter( + pjoin(MG5DIR,'HEPTools'),'lib') + preamble += "\n unset PYTHIA8DATA\n" + + open(pythia_cmd_card,'w').write("""! +! It is possible to run this card manually with: +! %s %s +! +"""%(preamble+pythia_main,os.path.basename(pythia_cmd_card))+cmd_card.getvalue()) + + # launch pythia8 + pythia_log = pjoin(self.me_dir , 'Events', self.run_name , + '%s_pythia8.log' % tag) + + # Write a bash wrapper to run the shower with custom environment variables + wrapper_path = pjoin(self.me_dir,'Events',self.run_name,'run_shower.sh') + wrapper = open(wrapper_path,'w') + shell = 'bash' if misc.get_shell_type() in ['bash',None] else 'tcsh' + shell_exe = None + if os.path.exists('/usr/bin/env'): + shell_exe = '/usr/bin/env %s'%shell + else: + shell_exe = misc.which(shell) + if not shell_exe: + raise self.InvalidCmd('No s hell could be found in your environment.\n'+ + "Make sure that either '%s' is in your path or that the"%shell+\ + " command '/usr/bin/env %s' exists and returns a valid path."%shell) + + exe_cmd = "#!%s\n%s"%(shell_exe,' '.join( + [preamble+pythia_main, + os.path.basename(pythia_cmd_card)])) + + wrapper.write(exe_cmd) + wrapper.close() + + # Set it as executable + st = os.stat(wrapper_path) + os.chmod(wrapper_path, st.st_mode | stat.S_IEXEC) + + # If the target HEPMC output file is a fifo, don't hang MG5_aMC and let + # it proceed. + is_HepMC_output_fifo = False if not HepMC_event_output else \ + ( os.path.exists(HepMC_event_output) and \ + stat.S_ISFIFO(os.stat(HepMC_event_output).st_mode)) + startPY8timer = time.time() + + # Information that will be extracted from this PY8 run + PY8_extracted_information={ 'sigma_m':None, 'Nacc':None, 'Ntry':None, + 'cross_sections':{} } + + if is_HepMC_output_fifo: + logger.info( +"""Pythia8 is set to output HEPMC events to to a fifo file. +You can follow PY8 run with the following command (in a separate terminal): + tail -f %s"""%pythia_log ) + py8_log = open( pythia_log,'w') + py8_bkgrd_proc = misc.Popen([wrapper_path], + stdout=py8_log,stderr=py8_log, + cwd=pjoin(self.me_dir,'Events',self.run_name)) + # Now directly return to madevent interactive interface if we are piping PY8 + if not no_default: + logger.info('You can now run a tool that reads the following fifo file:'+\ + '\n %s\nwhere PY8 outputs HEPMC events (e.g. MadAnalysis5).' + %HepMC_event_output,'$MG:color:GREEN') + return + else: + if self.options ['run_mode']!=0: + # Start a parallelization instance (stored in self.cluster) + self.configure_run_mode(self.options['run_mode']) + if self.options['run_mode']==1: + n_cores = max(int(self.options['cluster_size']),1) + elif self.options['run_mode']==2: + n_cores = max(int(self.cluster.nb_core),1) + + lhe_file_name = os.path.basename(PY8_Card.subruns[0]['Beams:LHEF']) + lhe_file = lhe_parser.EventFile(pjoin(self.me_dir,'Events', + self.run_name,PY8_Card.subruns[0]['Beams:LHEF'])) + n_available_events = len(lhe_file) + if PY8_Card['Main:numberOfEvents']==-1: + n_events = n_available_events + else: + n_events = PY8_Card['Main:numberOfEvents'] + if n_events > n_available_events: + raise self.InvalidCmd('You specified more events (%d) in the PY8 parameter'%n_events+\ + "'Main:numberOfEvents' than the total number of events available (%d)"%n_available_events+\ + ' in the event file:\n %s'%pjoin(self.me_dir,'Events',self.run_name,PY8_Card.subruns[0]['Beams:LHEF'])) + + # Implement a security to insure a minimum numbe of events per job + if self.options['run_mode']==2: + min_n_events_per_job = 100 + elif self.options['run_mode']==1: + min_n_events_per_job = 1000 + min_n_core = n_events//min_n_events_per_job + n_cores = max(min(min_n_core,n_cores),1) + + if self.options['run_mode']==0 or (self.options['run_mode']==2 and self.options['nb_core']==1): + # No need for parallelization anymore + self.cluster = None + logger.info('Follow Pythia8 shower by running the '+ + 'following command (in a separate terminal):\n tail -f %s'%pythia_log) + + if self.options['run_mode']==2 and self.options['nb_core']>1: + ret_code = self.cluster.launch_and_wait(wrapper_path, + argument= [], stdout= pythia_log, stderr=subprocess.STDOUT, + cwd=pjoin(self.me_dir,'Events',self.run_name)) + else: + stdout = open(pythia_log,'w') + ret_code = misc.call(wrapper_path, stdout=stdout, stderr=subprocess.STDOUT, + cwd=pjoin(self.me_dir,'Events',self.run_name)) + stdout.close() + if ret_code != 0: + raise self.InvalidCmd('Pythia8 shower interrupted with return'+\ + ' code %d.\n'%ret_code+\ + 'You can find more information in this log file:\n%s'%pythia_log) + else: + if self.run_card['event_norm']=='sum': + logger.error("") + logger.error("Either run in single core or change event_norm to 'average'.") + raise InvalidCmd("Pythia8 parallelization with event_norm set to 'sum' is not supported." + "Either run in single core or change event_norm to 'average'.") + + # Create the parallelization folder + parallelization_dir = pjoin(self.me_dir,'Events',self.run_name,'PY8_parallelization') + if os.path.isdir(parallelization_dir): + shutil.rmtree(parallelization_dir) + os.mkdir(parallelization_dir) + # Copy what should be the now standalone executable for PY8 + shutil.copy(pythia_main,parallelization_dir) + # Add a safe card in parallelization + ParallelPY8Card = copy.copy(PY8_Card) + assert ParallelPY8Card['JetMatching:nJetMax'] == PY8_Card['JetMatching:nJetMax'] + + # Normalize the name of the HEPMCouput and lhe input + if HepMC_event_output: + ParallelPY8Card['HEPMCoutput:file']='events.hepmc' + else: + ParallelPY8Card['HEPMCoutput:file']='/dev/null' + + ParallelPY8Card.subruns[0].systemSet('Beams:LHEF','events.lhe.gz') + ParallelPY8Card.write(pjoin(parallelization_dir,'PY8Card.dat'), + pjoin(self.me_dir,'Cards','pythia8_card_default.dat'), + direct_pythia_input=True) + # Write the wrapper + wrapper_path = pjoin(parallelization_dir,'run_PY8.sh') + wrapper = open(wrapper_path,'w') + if self.options['cluster_temp_path'] is None: + exe_cmd = \ +"""#!%s +./%s PY8Card.dat >& PY8_log.txt +""" + else: + exe_cmd = \ +"""#!%s +ln -s ./events_$1.lhe.gz ./events.lhe.gz +./%s PY8Card_$1.dat >& PY8_log.txt +mkdir split_$1 +if [ -f ./events.hepmc ]; +then + mv ./events.hepmc ./split_$1/ +fi +if [ -f ./pts.dat ]; +then + mv ./pts.dat ./split_$1/ +fi +if [ -f ./djrs.dat ]; +then + mv ./djrs.dat ./split_$1/ +fi +if [ -f ./PY8_log.txt ]; +then + mv ./PY8_log.txt ./split_$1/ +fi +tar -czf split_$1.tar.gz split_$1 +""" + exe_cmd = exe_cmd%(shell_exe,os.path.basename(pythia_main)) + wrapper.write(exe_cmd) + wrapper.close() + # Set it as executable + st = os.stat(wrapper_path) + os.chmod(wrapper_path, st.st_mode | stat.S_IEXEC) + + # Split the .lhe event file, create event partition + partition=[n_available_events//n_cores]*n_cores + for i in range(n_available_events%n_cores): + partition[i] += 1 + + # Splitting according to the total number of events requested by the user + # Will be used to determine the number of events to indicate in the PY8 split cards. + partition_for_PY8=[n_events//n_cores]*n_cores + for i in range(n_events%n_cores): + partition_for_PY8[i] += 1 + + logger.info('Splitting .lhe event file for PY8 parallelization...') + n_splits = lhe_file.split(partition=partition, cwd=parallelization_dir, zip=True) + + if n_splits!=len(partition): + raise MadGraph5Error('Error during lhe file splitting. Expected %d files but obtained %d.' + %(len(partition),n_splits)) + # Distribute the split events + split_files = [] + split_dirs = [] + for split_id in range(n_splits): + split_files.append('events_%s.lhe.gz'%split_id) + split_dirs.append(pjoin(parallelization_dir,'split_%d'%split_id)) + # Add the necessary run content + shutil.move(pjoin(parallelization_dir,lhe_file.name+'_%d.lhe.gz'%split_id), + pjoin(parallelization_dir,split_files[-1])) + + logger.info('Submitting Pythia8 jobs...') + for i, split_file in enumerate(split_files): + # We must write a PY8Card tailored for each split so as to correct the normalization + # HEPMCoutput:scaling of each weight since the lhe showered will not longer contain the + # same original number of events + split_PY8_Card = banner_mod.PY8Card(pjoin(parallelization_dir,'PY8Card.dat')) + # Make sure to sure the number of split_events determined during the splitting. + split_PY8_Card.systemSet('Main:numberOfEvents',partition_for_PY8[i]) + split_PY8_Card.systemSet('HEPMCoutput:scaling',split_PY8_Card['HEPMCoutput:scaling']* + (float(partition_for_PY8[i])/float(n_events))) + # Add_missing set to False so as to be sure not to add any additional parameter w.r.t + # the ones in the original PY8 param_card copied. + split_PY8_Card.write(pjoin(parallelization_dir,'PY8Card_%d.dat'%i), + pjoin(parallelization_dir,'PY8Card.dat'), add_missing=False) + in_files = [pjoin(parallelization_dir,os.path.basename(pythia_main)), + pjoin(parallelization_dir,'PY8Card_%d.dat'%i), + pjoin(parallelization_dir,split_file)] + if self.options['cluster_temp_path'] is None: + out_files = [] + os.mkdir(pjoin(parallelization_dir,'split_%d'%i)) + selected_cwd = pjoin(parallelization_dir,'split_%d'%i) + for in_file in in_files+[pjoin(parallelization_dir,'run_PY8.sh')]: + # Make sure to rename the split_file link from events_.lhe.gz to events.lhe.gz + # and similarly for PY8Card + if os.path.basename(in_file)==split_file: + ln(in_file,selected_cwd,name='events.lhe.gz') + elif os.path.basename(in_file).startswith('PY8Card'): + ln(in_file,selected_cwd,name='PY8Card.dat') + else: + ln(in_file,selected_cwd) + in_files = [] + wrapper_path = os.path.basename(wrapper_path) + else: + out_files = ['split_%d.tar.gz'%i] + selected_cwd = parallelization_dir + + self.cluster.submit2(wrapper_path, + argument=[str(i)], cwd=selected_cwd, + input_files=in_files, + output_files=out_files, + required_output=out_files) + + def wait_monitoring(Idle, Running, Done): + if Idle+Running+Done == 0: + return + logger.info('Pythia8 shower jobs: %d Idle, %d Running, %d Done [%s]'\ + %(Idle, Running, Done, misc.format_time(time.time() - startPY8timer))) + self.cluster.wait(parallelization_dir,wait_monitoring) + + logger.info('Merging results from the split PY8 runs...') + if self.options['cluster_temp_path']: + # Decompressing the output + for i, split_file in enumerate(split_files): + misc.call(['tar','-xzf','split_%d.tar.gz'%i],cwd=parallelization_dir) + os.remove(pjoin(parallelization_dir,'split_%d.tar.gz'%i)) + + # Now merge logs + pythia_log_file = open(pythia_log,'w') + n_added = 0 + for split_dir in split_dirs: + log_file = pjoin(split_dir,'PY8_log.txt') + pythia_log_file.write('='*35+'\n') + pythia_log_file.write(' -> Pythia8 log file for run %d <-'%i+'\n') + pythia_log_file.write('='*35+'\n') + pythia_log_file.write(open(log_file,'r').read()+'\n') + if run_type in merged_run_types: + sigma_m, Nacc, Ntry = self.parse_PY8_log_file(log_file) + if any(elem is None for elem in [sigma_m, Nacc, Ntry]): + continue + n_added += 1 + if PY8_extracted_information['sigma_m'] is None: + PY8_extracted_information['sigma_m'] = sigma_m + else: + PY8_extracted_information['sigma_m'] += sigma_m + if PY8_extracted_information['Nacc'] is None: + PY8_extracted_information['Nacc'] = Nacc + else: + PY8_extracted_information['Nacc'] += Nacc + if PY8_extracted_information['Ntry'] is None: + PY8_extracted_information['Ntry'] = Ntry + else: + PY8_extracted_information['Ntry'] += Ntry + + # Normalize the values added + if n_added>0: + PY8_extracted_information['sigma_m'] /= float(n_added) + pythia_log_file.close() + + # djr plots + djr_HwU = None + n_added = 0 + for split_dir in split_dirs: + djr_file = pjoin(split_dir,'djrs.dat') + if not os.path.isfile(djr_file): + continue + xsecs = self.extract_cross_sections_from_DJR(djr_file) + if len(xsecs)>0: + n_added += 1 + if len(PY8_extracted_information['cross_sections'])==0: + PY8_extracted_information['cross_sections'] = xsecs + # Square the error term + for key in PY8_extracted_information['cross_sections']: + PY8_extracted_information['cross_sections'][key][1] = \ + PY8_extracted_information['cross_sections'][key][1]**2 + else: + for key, value in xsecs.items(): + PY8_extracted_information['cross_sections'][key][0] += value[0] + # Add error in quadrature + PY8_extracted_information['cross_sections'][key][1] += value[1]**2 + new_djr_HwU = histograms.HwUList(djr_file,run_id=0) + if djr_HwU is None: + djr_HwU = new_djr_HwU + else: + for i, hist in enumerate(djr_HwU): + djr_HwU[i] = hist + new_djr_HwU[i] + + + if not djr_HwU is None: + djr_HwU.output(pjoin(self.me_dir,'Events',self.run_name,'djrs'),format='HwU') + shutil.move(pjoin(self.me_dir,'Events',self.run_name,'djrs.HwU'), + pjoin(self.me_dir,'Events',self.run_name,'%s_djrs.dat'%tag)) + + if n_added>0: + for key in PY8_extracted_information['cross_sections']: + # The cross-sections in the DJR are normalized for the original number of events, so we should not + # divide by n_added anymore for the cross-section value + # PY8_extracted_information['cross_sections'][key][0] /= float(n_added) + PY8_extracted_information['cross_sections'][key][1] = \ + math.sqrt(PY8_extracted_information['cross_sections'][key][1]) / float(n_added) + + # pts plots + pts_HwU = None + for split_dir in split_dirs: + pts_file = pjoin(split_dir,'pts.dat') + if not os.path.isfile(pts_file): + continue + new_pts_HwU = histograms.HwUList(pts_file,run_id=0) + if pts_HwU is None: + pts_HwU = new_pts_HwU + else: + for i, hist in enumerate(pts_HwU): + pts_HwU[i] = hist + new_pts_HwU[i] + if not pts_HwU is None: + pts_HwU.output(pjoin(self.me_dir,'Events',self.run_name,'pts'),format='HwU') + shutil.move(pjoin(self.me_dir,'Events',self.run_name,'pts.HwU'), + pjoin(self.me_dir,'Events',self.run_name,'%s_pts.dat'%tag)) + + # HepMC events now. + all_hepmc_files = [] + for split_dir in split_dirs: + hepmc_file = pjoin(split_dir,'events.hepmc') + if not os.path.isfile(hepmc_file): + continue + all_hepmc_files.append(hepmc_file) + + if len(all_hepmc_files)>0: + hepmc_output = pjoin(self.me_dir,'Events',self.run_name,HepMC_event_output) + with misc.TMP_directory() as tmp_dir: + # Use system calls to quickly put these together + header = open(pjoin(tmp_dir,'header.hepmc'),'w') + n_head = 0 + for line in open(all_hepmc_files[0],'r'): + if not line.startswith('E'): + n_head += 1 + header.write(line) + else: + break + header.close() + tail = open(pjoin(tmp_dir,'tail.hepmc'),'w') + n_tail = 0 + + for line in misc.reverse_readline(all_hepmc_files[-1]): + if line.startswith('HepMC::'): + n_tail += 1 + tail.write(line) + else: + break + tail.close() + if n_tail>1: + raise MadGraph5Error('HEPMC files should only have one trailing command.') + ###################################################################### + # This is the most efficient way of putting together HEPMC's, *BUT* # + # WARNING: NEED TO RENDER THE CODE BELOW SAFE TOWARDS INJECTION # + ###################################################################### + for hepmc_file in all_hepmc_files: + # Remove in an efficient way the starting and trailing HEPMC tags + # check for support of negative argument in head + devnull = open(os.path.devnull, 'w') + pid = misc.call(['head','-n', '-1', __file__], stdout=devnull, stderr=devnull) + devnull.close() + if pid == 0: + misc.call('head -n -1 %s | tail -n +%d > %s/tmpfile' % + (hepmc_file, n_head+1, os.path.dirname(hepmc_file)), shell=True) + misc.call(['mv', 'tmpfile', os.path.basename(hepmc_file)], cwd=os.path.dirname(hepmc_file)) + elif sys.platform == 'darwin': + # sed on MAC has slightly different synthax than on + os.system(' '.join(['sed','-i',"''","'%s;$d'"% + (';'.join('%id'%(i+1) for i in range(n_head))),hepmc_file])) + else: + # other UNIX systems + os.system(' '.join(['sed','-i']+["-e '%id'"%(i+1) for i in range(n_head)]+ + ["-e '$d'",hepmc_file])) + + os.system(' '.join(['cat',pjoin(tmp_dir,'header.hepmc')]+all_hepmc_files+ + [pjoin(tmp_dir,'tail.hepmc'),'>',hepmc_output])) + + # We are done with the parallelization directory. Clean it. + if os.path.isdir(parallelization_dir): + shutil.rmtree(parallelization_dir) + + # Properly rename the djr and pts output if present. + djr_output = pjoin(self.me_dir,'Events', self.run_name, 'djrs.dat') + if os.path.isfile(djr_output): + shutil.move(djr_output, pjoin(self.me_dir,'Events', + self.run_name, '%s_djrs.dat' % tag)) + pt_output = pjoin(self.me_dir,'Events', self.run_name, 'pts.dat') + if os.path.isfile(pt_output): + shutil.move(pt_output, pjoin(self.me_dir,'Events', + self.run_name, '%s_pts.dat' % tag)) + + if not os.path.isfile(pythia_log) or \ + 'Inclusive cross section:' not in '\n'.join(open(pythia_log,'r').readlines()[-20:]): + logger.warning('Fail to produce a pythia8 output. More info in \n %s'%pythia_log) + return + + # Plot for Pythia8 + successful = self.create_plot('Pythia8') + if not successful: + logger.warning('Failed to produce Pythia8 merging plots.') + + self.to_store.append('pythia8') + + # Study matched cross-sections + if run_type in merged_run_types: + # From the log file + if all(PY8_extracted_information[_] is None for _ in ['sigma_m','Nacc','Ntry']): + # When parallelization is enable we shouldn't have cannot look in the log in this way + if self.options['run_mode']==0 or (self.options['run_mode']==2 and self.options['nb_core']==1): + PY8_extracted_information['sigma_m'],PY8_extracted_information['Nacc'],\ + PY8_extracted_information['Ntry'] = self.parse_PY8_log_file( + pjoin(self.me_dir,'Events', self.run_name,'%s_pythia8.log' % tag)) + else: + logger.warning('Pythia8 cross-section could not be retreived.\n'+ + 'Try turning parallelization off by setting the option nb_core to 1. YYYYY') + + if not any(PY8_extracted_information[_] is None for _ in ['sigma_m','Nacc','Ntry']): + self.results.add_detail('cross_pythia', PY8_extracted_information['sigma_m']) + self.results.add_detail('nb_event_pythia', PY8_extracted_information['Nacc']) + # Shorthands + Nacc = PY8_extracted_information['Nacc'] + Ntry = PY8_extracted_information['Ntry'] + sigma_m = PY8_extracted_information['sigma_m'] + # Compute pythia error + error = self.results[self.run_name].return_tag(self.run_tag)['error'] + try: + error_m = math.sqrt((error * Nacc/Ntry)**2 + sigma_m**2 *(1-Nacc/Ntry)/Nacc) + except ZeroDivisionError: + # Cannot compute error + error_m = -1.0 + # works both for fixed number of generated events and fixed accepted events + self.results.add_detail('error_pythia', error_m) + + if self.run_card['use_syst']: + self.results.add_detail('cross_pythia', -1) + self.results.add_detail('error_pythia', 0) + + # From the djr file generated + djr_output = pjoin(self.me_dir,'Events',self.run_name,'%s_djrs.dat'%tag) + if os.path.isfile(djr_output) and len(PY8_extracted_information['cross_sections'])==0: + # When parallelization is enable we shouldn't have cannot look in the log in this way + if self.options['run_mode']==0 or (self.options['run_mode']==2 and self.options['nb_core']==1): + PY8_extracted_information['cross_sections'] = self.extract_cross_sections_from_DJR(djr_output) + else: + logger.warning('Pythia8 merged cross-sections could not be retreived.\n'+ + 'Try turning parallelization off by setting the option nb_core to 1.XXXXX') + PY8_extracted_information['cross_sections'] = {} + + cross_sections = PY8_extracted_information['cross_sections'] + if cross_sections: + # Filter the cross_sections specified an keep only the ones + # with central parameters and a different merging scale + a_float_re = '[\+|-]?\d+(\.\d*)?([EeDd][\+|-]?\d+)?' + central_merging_re = re.compile( + '^\s*Weight_MERGING\s*=\s*(?P%s)\s*$'%a_float_re, + re.IGNORECASE) + cross_sections = dict( + (float(central_merging_re.match(xsec).group('merging')),value) + for xsec, value in cross_sections.items() if not + central_merging_re.match(xsec) is None) + central_scale = PY8_Card['JetMatching:qCut'] if \ + int(self.run_card['ickkw'])==1 else PY8_Card['Merging:TMS'] + if central_scale in cross_sections: + self.results.add_detail('cross_pythia8', cross_sections[central_scale][0]) + self.results.add_detail('error_pythia8', cross_sections[central_scale][1]) + + #logger.info('Pythia8 merged cross-sections are:') + #for scale in sorted(cross_sections.keys()): + # logger.info(' > Merging scale = %-6.4g : %-11.5g +/- %-7.2g [pb]'%\ + # (scale,cross_sections[scale][0],cross_sections[scale][1])) + + xsecs_file = open(pjoin(self.me_dir,'Events',self.run_name, + '%s_merged_xsecs.txt'%tag),'w') + if cross_sections: + xsecs_file.write('%-20s%-20s%-20s\n'%('Merging scale', + 'Cross-section [pb]','MC uncertainty [pb]')) + for scale in sorted(cross_sections.keys()): + xsecs_file.write('%-20.4g%-20.6e%-20.2e\n'% + (scale,cross_sections[scale][0],cross_sections[scale][1])) + else: + xsecs_file.write('Cross-sections could not be read from the'+\ + "XML node 'xsection' of the .dat file produced by Pythia8.") + xsecs_file.close() + + #Update the banner + # We add directly the pythia command card because it has the full + # information + self.banner.add(pythia_cmd_card) + + if int(self.run_card['ickkw']): + # Add the matched cross-section + if 'MGGenerationInfo' in self.banner: + self.banner['MGGenerationInfo'] += '# Matched Integrated weight (pb) : %s\n' % self.results.current['cross_pythia'] + else: + self.banner['MGGenerationInfo'] = '# Matched Integrated weight (pb) : %s\n' % self.results.current['cross_pythia'] + banner_path = pjoin(self.me_dir, 'Events', self.run_name, '%s_%s_banner.txt' % (self.run_name, tag)) + self.banner.write(banner_path) + + self.update_status('Pythia8 shower finished after %s.'%misc.format_time(time.time() - startPY8timer), level='pythia8') + if self.options['delphes_path']: + self.exec_cmd('delphes --no_default', postcmd=False, printcmd=False) + self.print_results_in_shell(self.results.current) + + def parse_PY8_log_file(self, log_file_path): + """ Parse a log file to extract number of event and cross-section. """ + pythiare = re.compile("Les Houches User Process\(es\)\s*\d+\s*\|\s*(?P\d+)\s*(?P\d+)\s*(?P\d+)\s*\|\s*(?P[\d\.e\-\+]+)\s*(?P[\d\.e\-\+]+)") + pythia_xsec_re = re.compile("Inclusive cross section\s*:\s*(?P[\d\.e\-\+]+)\s*(?P[\d\.e\-\+]+)") + sigma_m, Nacc, Ntry = None, None, None + for line in misc.BackRead(log_file_path): + info = pythiare.search(line) + if not info: + # Also try to obtain the cross-section and error from the final xsec line of pythia8 log + # which is more reliable, in general for example when there is merging and the last event + # is skipped. + final_PY8_xsec = pythia_xsec_re.search(line) + if not final_PY8_xsec: + continue + else: + sigma_m = float(final_PY8_xsec.group('xsec')) *1e9 + continue + else: + try: + # Pythia cross section in mb, we want pb + if sigma_m is None: + sigma_m = float(info.group('xsec')) *1e9 + if Nacc is None: + Nacc = int(info.group('generated')) + if Ntry is None: + Ntry = int(info.group('tried')) + if Nacc==0: + raise self.InvalidCmd('Pythia8 shower failed since it'+\ + ' did not accept any event from the MG5aMC event file.') + return sigma_m, Nacc, Ntry + except ValueError: + return None,None,None + + raise self.InvalidCmd("Could not find cross-section and event number information "+\ + "in Pythia8 log\n '%s'."%log_file_path) + + def extract_cross_sections_from_DJR(self,djr_output): + """Extract cross-sections from a djr XML output.""" + import xml.dom.minidom as minidom + run_nodes = minidom.parse(djr_output).getElementsByTagName("run") + all_nodes = dict((int(node.getAttribute('id')),node) for + node in run_nodes) + try: + selected_run_node = all_nodes[0] + except: + return {} + xsections = selected_run_node.getElementsByTagName("xsection") + # In the DJR, the conversion to pb is already performed + return dict((xsec.getAttribute('name'), + [float(xsec.childNodes[0].data.split()[0]), + float(xsec.childNodes[0].data.split()[1])]) + for xsec in xsections) + + def do_pythia(self, line): + """launch pythia""" + + + # Check argument's validity + args = self.split_arg(line) + if '--no_default' in args: + if not os.path.exists(pjoin(self.me_dir, 'Cards', 'pythia_card.dat')): + return + no_default = True + args.remove('--no_default') + else: + no_default = False + + if not self.run_name: + self.check_pythia(args) + self.configure_directory(html_opening =False) + else: + # initialize / remove lhapdf mode + self.configure_directory(html_opening =False) + self.check_pythia(args) + + if self.run_card['event_norm'] != 'sum': + logger.error('pythia-pgs require event_norm to be on sum. Do not run pythia6') + return + + # the args are modify and the last arg is always the mode + if not no_default: + self.ask_pythia_run_configuration(args[-1]) + if self.options['automatic_html_opening']: + misc.open_file(os.path.join(self.me_dir, 'crossx.html')) + self.options['automatic_html_opening'] = False + + # Update the banner with the pythia card + if not self.banner or len(self.banner) <=1: + self.banner = banner_mod.recover_banner(self.results, 'pythia') + + pythia_src = pjoin(self.options['pythia-pgs_path'],'src') + + self.results.add_detail('run_mode', 'madevent') + + self.update_status('Running Pythia', 'pythia') + try: + os.remove(pjoin(self.me_dir,'Events','pythia.done')) + except Exception: + pass + + ## LAUNCHING PYTHIA + # check that LHAPATH is define. + if not re.search(r'^\s*LHAPATH=%s/PDFsets' % pythia_src, + open(pjoin(self.me_dir,'Cards','pythia_card.dat')).read(), + re.M): + f = open(pjoin(self.me_dir,'Cards','pythia_card.dat'),'a') + f.write('\n LHAPATH=%s/PDFsets' % pythia_src) + f.close() + tag = self.run_tag + pythia_log = pjoin(self.me_dir, 'Events', self.run_name , '%s_pythia.log' % tag) + #self.cluster.launch_and_wait('../bin/internal/run_pythia', + # argument= [pythia_src], stdout= pythia_log, + # stderr=subprocess.STDOUT, + # cwd=pjoin(self.me_dir,'Events')) + output_files = ['pythia_events.hep'] + if self.run_card['use_syst']: + output_files.append('syst.dat') + if self.run_card['ickkw'] == 1: + output_files += ['beforeveto.tree', 'xsecs.tree', 'events.tree'] + + os.environ['PDG_MASS_TBL'] = pjoin(pythia_src,'mass_width_2004.mc') + self.cluster.launch_and_wait(pjoin(pythia_src, 'pythia'), + input_files=[pjoin(self.me_dir, "Events", "unweighted_events.lhe"), + pjoin(self.me_dir,'Cards','pythia_card.dat'), + pjoin(pythia_src,'mass_width_2004.mc')], + output_files=output_files, + stdout= pythia_log, + stderr=subprocess.STDOUT, + cwd=pjoin(self.me_dir,'Events')) + + + os.remove(pjoin(self.me_dir, "Events", "unweighted_events.lhe")) + + if not os.path.exists(pjoin(self.me_dir,'Events','pythia_events.hep')): + logger.warning('Fail to produce pythia output. More info in \n %s' % pythia_log) + return + + self.to_store.append('pythia') + + # Find the matched cross-section + if int(self.run_card['ickkw']): + # read the line from the bottom of the file + #pythia_log = misc.BackRead(pjoin(self.me_dir,'Events', self.run_name, + # '%s_pythia.log' % tag)) + pythiare = re.compile("\s*I\s+0 All included subprocesses\s+I\s+(?P\d+)\s+(?P\d+)\s+I\s+(?P[\d\.D\-+]+)\s+I") + for line in misc.reverse_readline(pjoin(self.me_dir,'Events', self.run_name, + '%s_pythia.log' % tag)): + info = pythiare.search(line) + if not info: + continue + try: + # Pythia cross section in mb, we want pb + sigma_m = float(info.group('xsec').replace('D','E')) *1e9 + Nacc = int(info.group('generated')) + Ntry = int(info.group('tried')) + except ValueError: + # xsec is not float - this should not happen + self.results.add_detail('cross_pythia', 0) + self.results.add_detail('nb_event_pythia', 0) + self.results.add_detail('error_pythia', 0) + else: + self.results.add_detail('cross_pythia', sigma_m) + self.results.add_detail('nb_event_pythia', Nacc) + #compute pythia error + error = self.results[self.run_name].return_tag(self.run_tag)['error'] + if Nacc: + error_m = math.sqrt((error * Nacc/Ntry)**2 + sigma_m**2 *(1-Nacc/Ntry)/Nacc) + else: + error_m = 10000 * sigma_m + # works both for fixed number of generated events and fixed accepted events + self.results.add_detail('error_pythia', error_m) + break + + #pythia_log.close() + + pydir = pjoin(self.options['pythia-pgs_path'], 'src') + eradir = self.options['exrootanalysis_path'] + madir = self.options['madanalysis_path'] + td = self.options['td_path'] + + #Update the banner + self.banner.add(pjoin(self.me_dir, 'Cards','pythia_card.dat')) + if int(self.run_card['ickkw']): + # Add the matched cross-section + if 'MGGenerationInfo' in self.banner: + self.banner['MGGenerationInfo'] += '# Matched Integrated weight (pb) : %s\n' % self.results.current['cross_pythia'] + else: + self.banner['MGGenerationInfo'] = '# Matched Integrated weight (pb) : %s\n' % self.results.current['cross_pythia'] + banner_path = pjoin(self.me_dir, 'Events', self.run_name, '%s_%s_banner.txt' % (self.run_name, tag)) + self.banner.write(banner_path) + + # Creating LHE file + self.run_hep2lhe(banner_path) + + if int(self.run_card['ickkw']): + misc.gzip(pjoin(self.me_dir,'Events','beforeveto.tree'), + stdout=pjoin(self.me_dir,'Events',self.run_name, tag+'_pythia_beforeveto.tree.gz')) + + + if self.run_card['use_syst'] in self.true: + # Calculate syscalc info based on syst.dat + try: + self.run_syscalc('Pythia') + except SysCalcError as error: + logger.error(str(error)) + else: + if os.path.exists(pjoin(self.me_dir,'Events', 'syst.dat')): + # Store syst.dat + misc.gzip(pjoin(self.me_dir,'Events', 'syst.dat'), + stdout=pjoin(self.me_dir,'Events',self.run_name, tag + '_pythia_syst.dat.gz')) + + # Store syscalc.dat + if os.path.exists(pjoin(self.me_dir, 'Events', 'syscalc.dat')): + filename = pjoin(self.me_dir, 'Events' ,self.run_name, + '%s_syscalc.dat' % self.run_tag) + misc.gzip(pjoin(self.me_dir, 'Events','syscalc.dat'), + stdout = "%s.gz" % filename) + + # Plot for pythia + self.create_plot('Pythia') + + if os.path.exists(pjoin(self.me_dir,'Events','pythia_events.lhe')): + misc.gzip(pjoin(self.me_dir,'Events','pythia_events.lhe'), + stdout=pjoin(self.me_dir,'Events', self.run_name,'%s_pythia_events.lhe.gz' % tag)) + + self.update_status('finish', level='pythia', makehtml=False) + self.exec_cmd('pgs --no_default', postcmd=False, printcmd=False) + if self.options['delphes_path']: + self.exec_cmd('delphes --no_default', postcmd=False, printcmd=False) + self.print_results_in_shell(self.results.current) + + + ################################################################################ + def do_remove(self, line): + """Remove one/all run or only part of it""" + + args = self.split_arg(line) + run, tag, mode = self.check_remove(args) + if 'banner' in mode: + mode.append('all') + + + if run == 'all': + # Check first if they are not a run with a name run. + if os.path.exists(pjoin(self.me_dir, 'Events', 'all')): + logger.warning('A run with name all exists. So we will not supress all processes.') + else: + for match in misc.glob(pjoin('*','*_banner.txt'), pjoin(self.me_dir, 'Events')): + run = match.rsplit(os.path.sep,2)[1] + if self.force: + args.append('-f') + try: + self.exec_cmd('remove %s %s' % (run, ' '.join(args[1:]) ) ) + except self.InvalidCmd as error: + logger.info(error) + pass # run already clear + return + + # Check that run exists + if not os.path.exists(pjoin(self.me_dir, 'Events', run)): + raise self.InvalidCmd('No run \'%s\' detected' % run) + + try: + self.resuls.def_current(run) + self.update_status(' Cleaning %s' % run, level=None) + except Exception: + misc.sprint('fail to update results or html status') + pass # Just ensure that html never makes crash this function + + + # Found the file to delete + + to_delete = misc.glob('*', pjoin(self.me_dir, 'Events', run)) + to_delete += misc.glob('*', pjoin(self.me_dir, 'HTML', run)) + # forbid the banner to be removed + to_delete = [os.path.basename(f) for f in to_delete if 'banner' not in f] + if tag: + to_delete = [f for f in to_delete if tag in f] + if 'parton' in mode or 'all' in mode: + try: + if self.results[run][0]['tag'] != tag: + raise Exception('dummy') + except Exception: + pass + else: + nb_rm = len(to_delete) + if os.path.exists(pjoin(self.me_dir, 'Events', run, 'events.lhe.gz')): + to_delete.append('events.lhe.gz') + if os.path.exists(pjoin(self.me_dir, 'Events', run, 'unweighted_events.lhe.gz')): + to_delete.append('unweighted_events.lhe.gz') + if os.path.exists(pjoin(self.me_dir, 'HTML', run,'plots_parton.html')): + to_delete.append(pjoin(self.me_dir, 'HTML', run,'plots_parton.html')) + if nb_rm != len(to_delete): + logger.warning('Be carefull that partonic information are on the point to be removed.') + if 'all' in mode: + pass # delete everything + else: + if 'pythia' not in mode: + to_delete = [f for f in to_delete if 'pythia' not in f] + if 'pgs' not in mode: + to_delete = [f for f in to_delete if 'pgs' not in f] + if 'delphes' not in mode: + to_delete = [f for f in to_delete if 'delphes' not in f] + if 'parton' not in mode: + to_delete = [f for f in to_delete if 'delphes' in f + or 'pgs' in f + or 'pythia' in f] + if not self.force and len(to_delete): + question = 'Do you want to delete the following files?\n %s' % \ + '\n '.join(to_delete) + ans = self.ask(question, 'y', choices=['y','n']) + else: + ans = 'y' + + if ans == 'y': + for file2rm in to_delete: + if os.path.exists(pjoin(self.me_dir, 'Events', run, file2rm)): + try: + os.remove(pjoin(self.me_dir, 'Events', run, file2rm)) + except Exception: + shutil.rmtree(pjoin(self.me_dir, 'Events', run, file2rm)) + else: + try: + os.remove(pjoin(self.me_dir, 'HTML', run, file2rm)) + except Exception: + shutil.rmtree(pjoin(self.me_dir, 'HTML', run, file2rm)) + + + + # Remove file in SubProcess directory + if 'all' in mode or 'channel' in mode: + try: + if tag and self.results[run][0]['tag'] != tag: + raise Exception('dummy') + except Exception: + pass + else: + to_delete = misc.glob('%s*' % run, pjoin(self.me_dir, 'SubProcesses')) + to_delete += misc.glob(pjoin('*','%s*' % run), pjoin(self.me_dir, 'SubProcesses')) + to_delete += misc.glob(pjoin('*','*','%s*' % run), pjoin(self.me_dir, 'SubProcesses')) + + if self.force or len(to_delete) == 0: + ans = 'y' + else: + question = 'Do you want to delete the following files?\n %s' % \ + '\n '.join(to_delete) + ans = self.ask(question, 'y', choices=['y','n']) + + if ans == 'y': + for file2rm in to_delete: + os.remove(file2rm) + + if 'banner' in mode: + to_delete = misc.glob('*', pjoin(self.me_dir, 'Events', run)) + if tag: + # remove banner + try: + os.remove(pjoin(self.me_dir, 'Events',run,'%s_%s_banner.txt' % (run,tag))) + except Exception: + logger.warning('fail to remove the banner') + # remove the run from the html output + if run in self.results: + self.results.delete_run(run, tag) + return + elif any(['banner' not in os.path.basename(p) for p in to_delete]): + if to_delete: + raise MadGraph5Error('''Some output still exists for this run. + Please remove those output first. Do for example: + remove %s all banner + ''' % run) + else: + shutil.rmtree(pjoin(self.me_dir, 'Events',run)) + if run in self.results: + self.results.delete_run(run) + return + else: + logger.info('''The banner is not removed. In order to remove it run: + remove %s all banner %s''' % (run, tag and '--tag=%s ' % tag or '')) + + # update database. + self.results.clean(mode, run, tag) + self.update_status('', level='all') + + + + ############################################################################ + def do_plot(self, line): + """Create the plot for a given run""" + + # Since in principle, all plot are already done automaticaly + self.store_result() + args = self.split_arg(line) + # Check argument's validity + self.check_plot(args) + logger.info('plot for run %s' % self.run_name) + if not self.force: + self.ask_edit_cards(['plot_card.dat'], args, plot=True) + + if any([arg in ['all','parton'] for arg in args]): + filename = pjoin(self.me_dir, 'Events', self.run_name, 'unweighted_events.lhe') + if os.path.exists(filename+'.gz'): + misc.gunzip('%s.gz' % filename, keep=True) + if os.path.exists(filename): + files.ln(filename, pjoin(self.me_dir, 'Events')) + self.create_plot('parton') + if not os.path.exists(filename+'.gz'): + misc.gzip(pjoin(self.me_dir, 'Events', 'unweighted_events.lhe'), + stdout= "%s.gz" % filename) + else: + try: + os.remove(pjoin(self.me_dir, 'Events', 'unweighted_events.lhe')) + os.remove(filename) + except Exception: + pass + else: + logger.info('No valid files for partonic plot') + + if any([arg in ['all','pythia'] for arg in args]): + filename = pjoin(self.me_dir, 'Events' ,self.run_name, + '%s_pythia_events.lhe' % self.run_tag) + if os.path.exists(filename+'.gz'): + misc.gunzip("%s.gz" % filename) + if os.path.exists(filename): + shutil.move(filename, pjoin(self.me_dir, 'Events','pythia_events.lhe')) + self.create_plot('Pythia') + misc.gzip(pjoin(self.me_dir, 'Events','pythia_events.lhe'), + stdout= "%s.gz" % filename) + else: + logger.info('No valid files for pythia plot') + + + if any([arg in ['all','pgs'] for arg in args]): + filename = pjoin(self.me_dir, 'Events', self.run_name, + '%s_pgs_events.lhco' % self.run_tag) + if os.path.exists(filename+'.gz'): + misc.gunzip("%s.gz" % filename) + if os.path.exists(filename): + self.create_plot('PGS') + misc.gzip(filename) + else: + logger.info('No valid files for pgs plot') + + if any([arg in ['all','delphes'] for arg in args]): + filename = pjoin(self.me_dir, 'Events', self.run_name, + '%s_delphes_events.lhco' % self.run_tag) + if os.path.exists(filename+'.gz'): + misc.gunzip("%s.gz" % filename) + if os.path.exists(filename): + self.create_plot('Delphes') + misc.gzip(filename) + else: + logger.info('No valid files for delphes plot') + + ############################################################################ + def do_syscalc(self, line): + """Evaluate systematics variation weights for a given run""" + + # Since in principle, all systematics run are already done automaticaly + self.store_result() + args = self.split_arg(line) + # Check argument's validity + self.check_syscalc(args) + if self.ninitial == 1: + logger.error('SysCalc can\'t be run for decay processes') + return + + logger.info('Calculating systematics for run %s' % self.run_name) + + self.ask_edit_cards(['run_card.dat'], args, plot=False) + self.run_card = banner_mod.RunCard(pjoin(self.me_dir, 'Cards', 'run_card.dat')) + if any([arg in ['all','parton'] for arg in args]): + filename = pjoin(self.me_dir, 'Events', self.run_name, 'unweighted_events.lhe') + if os.path.exists(filename+'.gz'): + misc.gunzip("%s.gz" % filename) + if os.path.exists(filename): + shutil.move(filename, pjoin(self.me_dir, 'Events', 'unweighted_events.lhe')) + self.run_syscalc('parton') + misc.gzip(pjoin(self.me_dir, 'Events', 'unweighted_events.lhe'), + stdout="%s.gz" % filename) + else: + logger.info('No valid files for parton level systematics run.') + + if any([arg in ['all','pythia'] for arg in args]): + filename = pjoin(self.me_dir, 'Events' ,self.run_name, + '%s_pythia_syst.dat' % self.run_tag) + if os.path.exists(filename+'.gz'): + misc.gunzip("%s.gz" % filename) + if os.path.exists(filename): + shutil.move(filename, pjoin(self.me_dir, 'Events','syst.dat')) + try: + self.run_syscalc('Pythia') + except SysCalcError as error: + logger.warning(str(error)) + return + misc.gzip(pjoin(self.me_dir, 'Events','syst.dat'), "%s.gz" % filename) + filename = pjoin(self.me_dir, 'Events' ,self.run_name, + '%s_syscalc.dat' % self.run_tag) + misc.gzip(pjoin(self.me_dir, 'Events','syscalc.dat'), + stdout=filename) + else: + logger.info('No valid files for pythia level') + + + def store_result(self): + """ tar the pythia results. This is done when we are quite sure that + the pythia output will not be use anymore """ + + if not self.run_name: + return + + if not self.to_store: + return + + tag = self.run_card['run_tag'] + self.update_status('storing files of previous run', level=None,\ + error=True) + if 'event' in self.to_store: + if not os.path.exists(pjoin(self.me_dir, 'Events',self.run_name, 'unweighted_events.lhe.gz')) and\ + os.path.exists(pjoin(self.me_dir, 'Events',self.run_name, 'unweighted_events.lhe')): + logger.info("gzipping output file: unweighted_events.lhe") + misc.gzip(pjoin(self.me_dir,'Events',self.run_name,"unweighted_events.lhe")) + if os.path.exists(pjoin(self.me_dir,'Events','reweight.lhe')): + os.remove(pjoin(self.me_dir,'Events', 'reweight.lhe')) + + if 'pythia' in self.to_store: + self.update_status('Storing Pythia files of previous run', level='pythia', error=True) + p = pjoin(self.me_dir,'Events') + n = self.run_name + t = tag + self.to_store.remove('pythia') + misc.gzip(pjoin(p,'pythia_events.hep'), + stdout=pjoin(p, str(n),'%s_pythia_events.hep' % t),forceexternal=True) + + if 'pythia8' in self.to_store: + p = pjoin(self.me_dir,'Events') + n = self.run_name + t = tag + file_path = pjoin(p, n ,'%s_pythia8_events.hepmc'%t) + self.to_store.remove('pythia8') + if os.path.isfile(file_path): + if 'removeHEPMC' in self.to_store: + os.remove(file_path) + else: + self.update_status('Storing Pythia8 files of previous run', level='pythia', error=True) + if 'compressHEPMC' in self.to_store: + misc.gzip(file_path,stdout=file_path) + hepmc_fileformat = ".gz" + + moveHEPMC_in_to_store = None + for to_store in self.to_store: + if "moveHEPMC" in to_store: + moveHEPMC_in_to_store = to_store + + if not moveHEPMC_in_to_store == None: + move_hepmc_path = moveHEPMC_in_to_store.split("@")[1] + os.system("mv " + file_path + hepmc_fileformat + " " + move_hepmc_path) + + self.update_status('Done', level='pythia',makehtml=False,error=True) + self.results.save() + + self.to_store = [] + + def launch_job(self,exe, cwd=None, stdout=None, argument = [], remaining=0, + run_type='', mode=None, **opt): + """ """ + argument = [str(arg) for arg in argument] + if mode is None: + mode = self.cluster_mode + + # ensure that exe is executable + if os.path.exists(exe) and not os.access(exe, os.X_OK): + os.system('chmod +x %s ' % exe) + elif (cwd and os.path.exists(pjoin(cwd, exe))) and not \ + os.access(pjoin(cwd, exe), os.X_OK): + os.system('chmod +x %s ' % pjoin(cwd, exe)) + + if mode == 0: + self.update_status((remaining, 1, + self.total_jobs - remaining -1, run_type), level=None, force=False) + start = time.time() + #os.system('cd %s; ./%s' % (cwd,exe)) + status = misc.call([exe] + argument, cwd=cwd, stdout=stdout, **opt) + logger.info('%s run in %f s' % (exe, time.time() -start)) + if status: + raise MadGraph5Error('%s didn\'t stop properly. Stop all computation' % exe) + + + elif mode in [1,2]: + exename = os.path.basename(exe) + # For condor cluster, create the input/output files + if 'ajob' in exename: + input_files = ['madevent','input_app.txt','symfact.dat','iproc.dat','dname.mg', + pjoin(self.me_dir, 'SubProcesses','randinit')] + if os.path.exists(pjoin(self.me_dir,'SubProcesses', + 'MadLoop5_resources.tar.gz')) and cluster.need_transfer(self.options): + input_files.append(pjoin(self.me_dir,'SubProcesses', 'MadLoop5_resources.tar.gz')) + + output_files = [] + required_output = [] + + + #Find the correct PDF input file + input_files.append(self.get_pdf_input_filename()) + + #Find the correct ajob + Gre = re.compile("\s*j=(G[\d\.\w]+)") + origre = re.compile("grid_directory=(G[\d\.\w]+)") + try : + fsock = open(exe) + except Exception: + fsock = open(pjoin(cwd,exe)) + text = fsock.read() + output_files = Gre.findall(text) + if not output_files: + Ire = re.compile("for i in ([\d\.\s]*) ; do") + data = Ire.findall(text) + data = ' '.join(data).split() + for nb in data: + output_files.append('G%s' % nb) + required_output.append('G%s/results.dat' % nb) + else: + for G in output_files: + if os.path.isdir(pjoin(cwd,G)): + input_files.append(G) + required_output.append('%s/results.dat' % G) + + if origre.search(text): + G_grid = origre.search(text).groups()[0] + input_files.append(pjoin(G_grid, 'ftn26')) + + #submitting + self.cluster.submit2(exe, stdout=stdout, cwd=cwd, + input_files=input_files, output_files=output_files, + required_output=required_output) + elif 'survey' in exename: + input_files = ['madevent','input_app.txt','symfact.dat','iproc.dat', 'dname.mg', + pjoin(self.me_dir, 'SubProcesses','randinit')] + if os.path.exists(pjoin(self.me_dir,'SubProcesses', + 'MadLoop5_resources.tar.gz')) and cluster.need_transfer(self.options): + input_files.append(pjoin(self.me_dir,'SubProcesses', + 'MadLoop5_resources.tar.gz')) + + #Find the correct PDF input file + input_files.append(self.get_pdf_input_filename()) + + + output_files = [] + required_output = [] + + #Find the correct ajob + suffix = "_%s" % int(float(argument[0])) + if suffix == '_0': + suffix = '' + output_files = ['G%s%s' % (i, suffix) for i in argument[1:]] + for G in output_files: + required_output.append('%s/results.dat' % G) + + # add the grid information if needed + for G in output_files: + if '.' in argument[0]: + offset = int(str(argument[0]).split('.')[1]) + else: + offset = 0 + + if offset ==0 or offset == int(float(argument[0])): + if os.path.exists(pjoin(cwd, G, 'input_app.txt')): + os.remove(pjoin(cwd, G, 'input_app.txt')) + + if os.path.exists(os.path.realpath(pjoin(cwd, G, 'ftn25'))): + if offset == 0 or offset == int(float(argument[0])): + os.remove(pjoin(cwd, G, 'ftn25')) + continue + else: + input_files.append(pjoin(cwd, G, 'ftn25')) + input_files.remove('input_app.txt') + input_files.append(pjoin(cwd, G, 'input_app.txt')) + elif os.path.lexists(pjoin(cwd, G, 'ftn25')): + try: + os.remove(pjoin(cwd,G,'ftn25')) + except: + pass + + #submitting + self.cluster.cluster_submit(exe, stdout=stdout, cwd=cwd, argument=argument, + input_files=input_files, output_files=output_files, + required_output=required_output, **opt) + elif "refine_splitted.sh" in exename: + input_files = ['madevent','symfact.dat','iproc.dat', 'dname.mg', + pjoin(self.me_dir, 'SubProcesses','randinit')] + + if os.path.exists(pjoin(self.me_dir,'SubProcesses', + 'MadLoop5_resources.tar.gz')) and cluster.need_transfer(self.options): + input_files.append(pjoin(self.me_dir,'SubProcesses', + 'MadLoop5_resources.tar.gz')) + + #Find the correct PDF input file + input_files.append(self.get_pdf_input_filename()) + + + output_files = [argument[0]] + required_output = [] + for G in output_files: + required_output.append('%s/results.dat' % G) + input_files.append(pjoin(argument[1], "input_app.txt")) + input_files.append(pjoin(argument[1], "ftn26")) + + #submitting + self.cluster.cluster_submit(exe, stdout=stdout, cwd=cwd, argument=argument, + input_files=input_files, output_files=output_files, + required_output=required_output, **opt) + + + + else: + self.cluster.submit(exe, argument=argument, stdout=stdout, cwd=cwd, **opt) + + + ############################################################################ + def find_madevent_mode(self): + """Find if Madevent is in Group mode or not""" + + # The strategy is too look in the files Source/run_configs.inc + # if we found: ChanPerJob=3 then it's a group mode. + file_path = pjoin(self.me_dir, 'Source', 'run_config.inc') + text = open(file_path).read() + if re.search(r'''s*parameter\s+\(ChanPerJob=2\)''', text, re.I+re.M): + return 'group' + else: + return 'v4' + + ############################################################################ + def monitor(self, run_type='monitor', mode=None, html=False): + """ monitor the progress of running job """ + + + starttime = time.time() + if mode is None: + mode = self.cluster_mode + if mode > 0: + if html: + update_status = lambda idle, run, finish: \ + self.update_status((idle, run, finish, run_type), level=None, + force=False, starttime=starttime) + update_first = lambda idle, run, finish: \ + self.update_status((idle, run, finish, run_type), level=None, + force=True, starttime=starttime) + else: + update_status = lambda idle, run, finish: None + update_first = None + try: + self.cluster.wait(self.me_dir, update_status, update_first=update_first) + except Exception as error: + logger.info(error) + if not self.force: + ans = self.ask('Cluster Error detected. Do you want to clean the queue? ("c"=continue the run anyway)', + default = 'y', choices=['y','n', 'c']) + else: + ans = 'y' + if ans == 'y': + self.cluster.remove() + elif ans == 'c': + return self.monitor(run_type=run_type, mode=mode, html=html) + raise + except KeyboardInterrupt as error: + self.cluster.remove() + raise + + + + ############################################################################ + def configure_directory(self, html_opening=True): + """ All action require before any type of run """ + + # Basic check + assert os.path.exists(pjoin(self.me_dir,'SubProcesses')) + + # environmental variables to be included in make_opts + self.make_opts_var = {} + + #see when the last file was modified + time_mod = max([os.path.getmtime(pjoin(self.me_dir,'Cards','run_card.dat')), + os.path.getmtime(pjoin(self.me_dir,'Cards','param_card.dat'))]) + + if self.configured >= time_mod and hasattr(self, 'random') and hasattr(self, 'run_card'): + #just ensure that cluster specific are correctly handled + if self.cluster: + self.cluster.modify_interface(self) + return + else: + self.configured = time_mod + self.update_status('compile directory', level=None, update_results=True) + if self.options['automatic_html_opening'] and html_opening: + misc.open_file(os.path.join(self.me_dir, 'crossx.html')) + self.options['automatic_html_opening'] = False + #open only once the web page + # Change current working directory + self.launching_dir = os.getcwd() + + # Check if we need the MSSM special treatment + model = self.find_model_name() + if model == 'mssm' or model.startswith('mssm-'): + param_card = pjoin(self.me_dir, 'Cards','param_card.dat') + mg5_param = pjoin(self.me_dir, 'Source', 'MODEL', 'MG5_param.dat') + check_param_card.convert_to_mg5card(param_card, mg5_param) + check_param_card.check_valid_param_card(mg5_param) + + # limit the number of event to 100k + self.check_nb_events() + + # this is in order to avoid conflicts between runs with and without + # lhapdf. not needed anymore the makefile handles it automaticallu + #misc.compile(['clean4pdf'], cwd = pjoin(self.me_dir, 'Source')) + + self.make_opts_var['pdlabel1'] = '' + self.make_opts_var['pdlabel2'] = '' + if self.run_card['pdlabel1'] in ['eva', 'iww']: + self.make_opts_var['pdlabel1'] = 'eva' + if self.run_card['pdlabel2'] in ['eva', 'iww']: + self.make_opts_var['pdlabel2'] = 'eva' + if self.run_card['pdlabel1'] in ['edff','chff']: + self.make_opts_var['pdlabel1'] = self.run_card['pdlabel1'] + if self.run_card['pdlabel2'] in ['edff','chff']: + self.make_opts_var['pdlabel2'] = self.run_card['pdlabel2'] + + + # set lhapdf. + if self.run_card['pdlabel'] == "lhapdf": + self.make_opts_var['lhapdf'] = 'True' + self.link_lhapdf(pjoin(self.me_dir,'lib')) + pdfsetsdir = self.get_lhapdf_pdfsetsdir() + lhaid_list = [int(self.run_card['lhaid'])] + self.copy_lhapdf_set(lhaid_list, pdfsetsdir) + if self.run_card['pdlabel'] != "lhapdf": + self.pdffile = None + self.make_opts_var['lhapdf'] = "" + + # this is the case of collision with dressed leptons + if abs(self.run_card['lpp1']) == abs(self.run_card['lpp2']) in [3,4]: + # copy the files for the chosen density + if self.run_card['pdlabel'] in sum(self.run_card.allowed_lep_densities.values(),[]): + self.copy_lep_densities(self.run_card['pdlabel'], pjoin(self.me_dir, 'Source')) + self.make_opts_var['pdlabel1'] = 'ee' + self.make_opts_var['pdlabel2'] = 'ee' + + # set random number + if self.run_card['iseed'] != 0: + self.random = int(self.run_card['iseed']) + self.run_card['iseed'] = 0 + # Reset seed in run_card to 0, to ensure that following runs + # will be statistically independent + self.run_card.write(pjoin(self.me_dir, 'Cards','run_card.dat'), template=pjoin(self.me_dir, 'Cards','run_card.dat')) + time_mod = max([os.path.getmtime(pjoin(self.me_dir,'Cards','run_card.dat')), + os.path.getmtime(pjoin(self.me_dir,'Cards','param_card.dat'))]) + self.configured = time_mod + elif os.path.exists(pjoin(self.me_dir,'SubProcesses','randinit')): + for line in open(pjoin(self.me_dir,'SubProcesses','randinit')): + data = line.split('=') + assert len(data) ==2 + self.random = int(data[1]) + break + else: + self.random = random.randint(1, 30107) + + #set random seed for python part of the code + if self.run_card['python_seed'] == -2: #-2 means same as run_card + import random + if not hasattr(random, 'mg_seedset'): + random.seed(self.run_card['python_seed']) + random.mg_seedset = self.run_card['python_seed'] + elif self.run_card['python_seed'] >= 0: + import random + if not hasattr(random, 'mg_seedset'): + random.seed(self.run_card['python_seed']) + random.mg_seedset = self.run_card['python_seed'] + if self.run_card['ickkw'] == 2: + logger.info('Running with CKKW matching') + self.treat_ckkw_matching() + + # add the make_opts_var to make_opts + self.update_make_opts(self.run_card) + # reset list of Gdirectory + self.Gdirs = None + + # create param_card.inc and run_card.inc + self.do_treatcards('') + + logger.info("compile Source Directory") + + # Compile + for name in [ 'all']:#, '../bin/internal/combine_events']: + self.compile(arg=[name], cwd=os.path.join(self.me_dir, 'Source')) + + force_subproc_clean = False + + bias_name = os.path.basename(self.run_card['bias_module']) + if bias_name.lower()=='none': + bias_name = 'dummy' + + # Always refresh the bias dependencies file + if os.path.exists(pjoin(self.me_dir, 'SubProcesses','bias_dependencies')): + os.remove(pjoin(self.me_dir, 'SubProcesses','bias_dependencies')) + if os.path.exists(pjoin(self.me_dir, 'Source','BIAS',bias_name,'bias_dependencies')): + files.ln(pjoin(self.me_dir, 'Source','BIAS',bias_name,'bias_dependencies'), + pjoin(self.me_dir, 'SubProcesses')) + + if self.proc_characteristics['bias_module']!=bias_name and \ + os.path.isfile(pjoin(self.me_dir, 'lib','libbias.a')): + os.remove(pjoin(self.me_dir, 'lib','libbias.a')) + force_subproc_clean = True + + + # Finally compile the bias module as well + if self.run_card['bias_module'] not in ['dummy',None]: + logger.debug("Compiling the bias module '%s'"%bias_name) + # Verify the compatibility of the specified module + bias_module_valid = misc.Popen(['make','requirements'], + cwd=os.path.join(self.me_dir, 'Source','BIAS',bias_name), + stdout=subprocess.PIPE, stderr=subprocess.PIPE).communicate()[0].decode(errors='ignore') + if 'VALID' not in str(bias_module_valid).upper() or \ + 'INVALID' in str(bias_module_valid).upper(): + raise InvalidCmd("The bias module '%s' cannot be used because of:\n%s"% + (bias_name,bias_module_valid)) + + self.compile(arg=[], cwd=os.path.join(self.me_dir, 'Source','BIAS',bias_name)) + self.proc_characteristics['bias_module']=bias_name + # Update the proc_characterstics file + self.proc_characteristics.write( + pjoin(self.me_dir,'SubProcesses','proc_characteristics')) + + if force_subproc_clean: + # Make sure that madevent will be recompiled + subproc = [l.strip() for l in open(pjoin(self.me_dir,'SubProcesses', + 'subproc.mg'))] + for nb_proc,subdir in enumerate(subproc): + Pdir = pjoin(self.me_dir, 'SubProcesses',subdir.strip()) + self.compile(['clean'], cwd=Pdir) + + #see when the last file was modified + time_mod = max([os.path.getmtime(pjoin(self.me_dir,'Cards','run_card.dat')), + os.path.getmtime(pjoin(self.me_dir,'Cards','param_card.dat'))]) + + self.configured = time_mod + + ############################################################################ + ## HELPING ROUTINE + ############################################################################ + @staticmethod + def check_dir(path, default=''): + """check if the directory exists. if so return the path otherwise the + default""" + + if os.path.isdir(path): + return path + else: + return default + + + + ############################################################################ + def get_Gdir(self, Pdir=None, symfact=None): + """get the list of Gdirectory if not yet saved.""" + + if hasattr(self, "Gdirs") and self.Gdirs: + if self.me_dir in self.Gdirs[0]: + if Pdir is None: + if not symfact: + return list(itertools.chain(*list(self.Gdirs[0].values()))) + else: + return list(itertools.chain(*list(self.Gdirs[0].values()))), self.Gdirs[1] + else: + if not symfact: + return self.Gdirs[0][Pdir] + else: + return self.Gdirs[0][Pdir], self.Gdirs[1] + + + Pdirs = self.get_Pdir() + Gdirs = {self.me_dir:[]} + mfactors = {} + for P in Pdirs: + Gdirs[P] = [] + #for the next line do not use P, since in readonly mode it might not have symfact + for line in open(pjoin(self.me_dir, 'SubProcesses',os.path.basename(P), "symfact.dat")): + tag, mfactor = line.split() + if int(mfactor) > 0: + Gdirs[P].append( pjoin(P, "G%s" % tag) ) + mfactors[pjoin(P, "G%s" % tag)] = mfactor + self.Gdirs = (Gdirs, mfactors) + return self.get_Gdir(Pdir, symfact=symfact) + + ############################################################################ + def set_run_name(self, name, tag=None, level='parton', reload_card=False, + allow_new_tag=True): + """define the run name, the run_tag, the banner and the results.""" + + def get_last_tag(self, level): + # Return the tag of the previous run having the required data for this + # tag/run to working wel. + if level == 'parton': + return + elif level in ['pythia','pythia8','madanalysis5_parton','madanalysis5_hadron']: + return self.results[self.run_name][0]['tag'] + else: + for i in range(-1,-len(self.results[self.run_name])-1,-1): + tagRun = self.results[self.run_name][i] + if tagRun.pythia or tagRun.shower or tagRun.pythia8 : + return tagRun['tag'] + + + # when are we force to change the tag new_run:previous run requiring changes + upgrade_tag = {'parton': ['parton','pythia','pgs','delphes','madanalysis5_hadron','madanalysis5_parton', 'rivet'], + 'pythia': ['pythia','pgs','delphes','madanalysis5_hadron'], + 'pythia8': ['pythia8','pgs','delphes','madanalysis5_hadron', 'rivet'], + 'pgs': ['pgs'], + 'delphes':['delphes'], + 'madanalysis5_hadron':['madanalysis5_hadron'], + 'madanalysis5_parton':['madanalysis5_parton'], + 'plot':[], + 'syscalc':[], + 'rivet':['rivet']} + + if name == self.run_name: + if reload_card: + run_card = pjoin(self.me_dir, 'Cards','run_card.dat') + self.run_card = banner_mod.RunCard(run_card) + + #check if we need to change the tag + if tag: + self.run_card['run_tag'] = tag + self.run_tag = tag + self.results.add_run(self.run_name, self.run_card) + else: + for tag in upgrade_tag[level]: + if getattr(self.results[self.run_name][-1], tag): + tag = self.get_available_tag() + self.run_card['run_tag'] = tag + self.run_tag = tag + self.results.add_run(self.run_name, self.run_card) + break + return get_last_tag(self, level) + + + # save/clean previous run + if self.run_name: + self.store_result() + # store new name + self.run_name = name + + new_tag = False + # First call for this run -> set the banner + self.banner = banner_mod.recover_banner(self.results, level, name) + if 'mgruncard' in self.banner: + self.run_card = self.banner.charge_card('run_card') + else: + # Read run_card + run_card = pjoin(self.me_dir, 'Cards','run_card.dat') + self.run_card = banner_mod.RunCard(run_card) + + if tag: + self.run_card['run_tag'] = tag + new_tag = True + elif not self.run_name in self.results and level =='parton': + pass # No results yet, so current tag is fine + elif not self.run_name in self.results: + #This is only for case when you want to trick the interface + logger.warning('Trying to run data on unknown run.') + self.results.add_run(name, self.run_card) + self.results.update('add run %s' % name, 'all', makehtml=False) + else: + for tag in upgrade_tag[level]: + + if getattr(self.results[self.run_name][-1], tag): + # LEVEL is already define in the last tag -> need to switch tag + tag = self.get_available_tag() + self.run_card['run_tag'] = tag + new_tag = True + break + if not new_tag: + # We can add the results to the current run + tag = self.results[self.run_name][-1]['tag'] + self.run_card['run_tag'] = tag # ensure that run_tag is correct + + if allow_new_tag and (name in self.results and not new_tag): + self.results.def_current(self.run_name) + else: + self.results.add_run(self.run_name, self.run_card) + + self.run_tag = self.run_card['run_tag'] + + return get_last_tag(self, level) + + + + ############################################################################ + def check_nb_events(self): + """Find the number of event in the run_card, and check that this is not + too large""" + + + nb_event = int(self.run_card['nevents']) + if nb_event > 1000000: + logger.warning("Attempting to generate more than 1M events") + logger.warning("Limiting number to 1M. Use multi_run for larger statistics.") + path = pjoin(self.me_dir, 'Cards', 'run_card.dat') + os.system(r"""perl -p -i.bak -e "s/\d+\s*=\s*nevents/1000000 = nevents/" %s""" \ + % path) + self.run_card['nevents'] = 1000000 + + return + + + ############################################################################ + def update_random(self): + """ change random number""" + + self.random += 3 + if self.random > 30081*30081: # can't use too big random number + raise MadGraph5Error('Random seed too large ' + str(self.random) + ' > 30081*30081') + if self.run_card['python_seed'] == -2: + import random + if not hasattr(random, 'mg_seedset'): + random.seed(self.random) + random.mg_seedset = self.random + + ############################################################################ + def save_random(self): + """save random number in appropirate file""" + + fsock = open(pjoin(self.me_dir, 'SubProcesses','randinit'),'w') + fsock.writelines('r=%s\n' % self.random) + + def do_quit(self, *args, **opts): + + return common_run.CommonRunCmd.do_quit(self, *args, **opts) + #return CmdExtended.do_quit(self, *args, **opts) + + ############################################################################ + def treat_CKKW_matching(self): + """check for ckkw""" + + lpp1 = self.run_card['lpp1'] + lpp2 = self.run_card['lpp2'] + e1 = self.run_card['ebeam1'] + e2 = self.run_card['ebeam2'] + pd = self.run_card['pdlabel'] + lha = self.run_card['lhaid'] + xq = self.run_card['xqcut'] + translation = {'e1': e1, 'e2':e2, 'pd':pd, + 'lha':lha, 'xq':xq} + + if lpp1 or lpp2: + # Remove ':s from pd + if pd.startswith("'"): + pd = pd[1:] + if pd.endswith("'"): + pd = pd[:-1] + + if xq >2 or xq ==2: + xq = 2 + + # find data file + if pd == "lhapdf": + issudfile = 'lib/issudgrid-%(e1)s-%(e2)s-%(pd)s-%(lha)s-%(xq)s.dat.gz' + else: + issudfile = 'lib/issudgrid-%(e1)s-%(e2)s-%(pd)s-%(xq)s.dat.gz' + if self.web: + issudfile = pjoin(self.webbin, issudfile % translation) + else: + issudfile = pjoin(self.me_dir, issudfile % translation) + + logger.info('Sudakov grid file: %s' % issudfile) + + # check that filepath exists + if os.path.exists(issudfile): + path = pjoin(self.me_dir, 'lib', 'issudgrid.dat') + misc.gunzip(issudfile, keep=True, stdout=path) + else: + msg = 'No sudakov grid file for parameter choice. Start to generate it. This might take a while' + logger.info(msg) + self.update_status('GENERATE SUDAKOV GRID', level='parton') + + for i in range(-2,6): + self.cluster.submit('%s/gensudgrid ' % self.dirbin, + argument = ['%d'%i], + cwd=self.me_dir, + stdout=open(pjoin(self.me_dir, 'gensudgrid%s.log' % i),'w')) + self.monitor() + for i in range(-2,6): + path = pjoin(self.me_dir, 'lib', 'issudgrid.dat') + os.system('cat %s/gensudgrid%s.log >> %s' % (self.me_dir, path)) + misc.gzip(path, stdout=issudfile) + + ############################################################################ + def create_root_file(self, input='unweighted_events.lhe', + output='unweighted_events.root' ): + """create the LHE root file """ + self.update_status('Creating root files', level='parton') + + eradir = self.options['exrootanalysis_path'] + totar = False + torm = False + if input.endswith('.gz'): + if not os.path.exists(input) and os.path.exists(input[:-3]): + totar = True + input = input[:-3] + else: + misc.gunzip(input, keep=True) + totar = False + torm = True + input = input[:-3] + + try: + misc.call(['%s/ExRootLHEFConverter' % eradir, + input, output], + cwd=pjoin(self.me_dir, 'Events')) + except Exception: + logger.warning('fail to produce Root output [problem with ExRootAnalysis]') + + if totar: + if os.path.exists('%s.gz' % input): + try: + os.remove('%s.gz' % input) + except: + pass + else: + misc.gzip(input) + if torm: + os.remove(input) + + def run_syscalc(self, mode='parton', event_path=None, output=None): + """create the syscalc output""" + + if self.run_card['use_syst'] not in self.true: + return + + scdir = self.options['syscalc_path'] + if not scdir or not os.path.exists(scdir): + return + + if self.run_card['event_norm'] != 'sum': + logger.critical('SysCalc works only when event_norm is on \'sum\'.') + return + logger.info('running SysCalc on mode %s' % mode) + + # Restore the old default for SysCalc+PY6 + if self.run_card['sys_matchscale']=='auto': + self.run_card['sys_matchscale'] = "30 50" + + # Check that all pdfset are correctly installed + lhaid = [self.run_card.get_lhapdf_id()] + if '&&' in self.run_card['sys_pdf']: + line = ' '.join(self.run_card['sys_pdf']) + sys_pdf = line.split('&&') + lhaid += [l.split()[0] for l in sys_pdf] + else: + lhaid += [l for l in self.run_card['sys_pdf'].split() if not l.isdigit() or int(l) > 500] + try: + pdfsets_dir = self.get_lhapdf_pdfsetsdir() + except Exception as error: + logger.debug(str(error)) + logger.warning('Systematic computation requires lhapdf to run. Bypass SysCalc') + return + + # Copy all the relevant PDF sets + [self.copy_lhapdf_set([onelha], pdfsets_dir) for onelha in lhaid] + + to_syscalc={'sys_scalefact': self.run_card['sys_scalefact'], + 'sys_alpsfact': self.run_card['sys_alpsfact'], + 'sys_matchscale': self.run_card['sys_matchscale'], + 'sys_scalecorrelation': self.run_card['sys_scalecorrelation'], + 'sys_pdf': self.run_card['sys_pdf']} + + tag = self.run_card['run_tag'] + card = pjoin(self.me_dir, 'bin','internal', 'syscalc_card.dat') + template = open(pjoin(self.me_dir, 'bin','internal', 'syscalc_template.dat')).read() + + if '&&' in to_syscalc['sys_pdf']: + to_syscalc['sys_pdf'] = to_syscalc['sys_pdf'].split('#',1)[0].replace('&&',' \n ') + else: + data = to_syscalc['sys_pdf'].split() + new = [] + for d in data: + if not d.isdigit(): + new.append(d) + elif int(d) > 500: + new.append(d) + else: + new[-1] += ' %s' % d + to_syscalc['sys_pdf'] = '\n'.join(new) + + if to_syscalc['sys_pdf'].lower() in ['', 'f', 'false', 'none', '.false.']: + to_syscalc['sys_pdf'] = '' + if to_syscalc['sys_alpsfact'].lower() in ['', 'f', 'false', 'none','.false.']: + to_syscalc['sys_alpsfact'] = '' + + + + + # check if the scalecorrelation parameter is define: + if not 'sys_scalecorrelation' in self.run_card: + self.run_card['sys_scalecorrelation'] = -1 + open(card,'w').write(template % self.run_card) + + if not os.path.exists(card): + return False + + + + event_dir = pjoin(self.me_dir, 'Events') + + if not event_path: + if mode == 'parton': + event_path = pjoin(event_dir,self.run_name, 'unweighted_events.lhe') + if not (os.path.exists(event_path) or os.path.exists(event_path+".gz")): + event_path = pjoin(event_dir, 'unweighted_events.lhe') + output = pjoin(event_dir, 'syscalc.lhe') + stdout = open(pjoin(event_dir, self.run_name, '%s_systematics.log' % (mode)),'w') + elif mode == 'Pythia': + stdout = open(pjoin(event_dir, self.run_name, '%s_%s_syscalc.log' % (tag,mode)),'w') + if 'mgpythiacard' in self.banner: + pat = re.compile('''^\s*qcut\s*=\s*([\+\-\d.e]*)''', re.M+re.I) + data = pat.search(self.banner['mgpythiacard']) + if data: + qcut = float(data.group(1)) + xqcut = abs(self.run_card['xqcut']) + for value in self.run_card['sys_matchscale'].split(): + if float(value) < qcut: + raise SysCalcError('qcut value for sys_matchscale lower than qcut in pythia_card. Bypass syscalc') + if float(value) < xqcut: + raise SysCalcError('qcut value for sys_matchscale lower than xqcut in run_card. Bypass syscalc') + + + event_path = pjoin(event_dir,'syst.dat') + output = pjoin(event_dir, 'syscalc.dat') + else: + raise self.InvalidCmd('Invalid mode %s' % mode) + + if not os.path.exists(event_path): + if os.path.exists(event_path+'.gz'): + misc.gunzip(event_path+'.gz') + else: + raise SysCalcError('Events file %s does not exits' % event_path) + + self.update_status('Calculating systematics for %s level' % mode, level = mode.lower()) + try: + proc = misc.call([os.path.join(scdir, 'sys_calc'), + event_path, card, output], + stdout = stdout, + stderr = subprocess.STDOUT, + cwd=event_dir) + # Wait 5 s to make sure file is finished writing + time.sleep(5) + except OSError as error: + logger.error('fail to run syscalc: %s. Please check that SysCalc is correctly installed.' % error) + else: + if not os.path.exists(output): + logger.warning('SysCalc Failed. Please read the associate log to see the reason. Did you install the associate PDF set?') + elif mode == 'parton': + files.mv(output, event_path) + + self.update_status('End syscalc for %s level' % mode, level = mode.lower(), + makehtml=False) + + return True + + + action_switcher = AskRun + ############################################################################ + def ask_run_configuration(self, mode=None, args=[]): + """Ask the question when launching generate_events/multi_run""" + + passing_cmd = [] + if '-R' in args or '--reweight' in args: + passing_cmd.append('reweight=ON') + if '-M' in args or '--madspin' in args: + passing_cmd.append('madspin=ON') + + switch, cmd_switch = self.ask('', '0', [], ask_class = self.action_switcher, + mode=mode, line_args=args, force=self.force, + first_cmd=passing_cmd, return_instance=True) + # + self.switch = switch # store the value of the switch for plugin purpose + if 'dynamical' in switch: + mode = 'auto' + + # Now that we know in which mode we are check that all the card + #exists (copy default if needed) + + cards = ['param_card.dat', 'run_card.dat'] + if switch['shower'] == 'Pythia6': + cards.append('pythia_card.dat') + if switch['shower'] == 'Pythia8': + cards.append('pythia8_card.dat') + if switch['detector'] in ['PGS','DELPHES+PGS']: + cards.append('pgs_card.dat') + if switch['detector'] in ['Delphes', 'DELPHES+PGS']: + cards.append('delphes_card.dat') + delphes3 = True + if os.path.exists(pjoin(self.options['delphes_path'], 'data')): + delphes3 = False + cards.append('delphes_trigger.dat') + if switch['madspin'] != 'OFF': + cards.append('madspin_card.dat') + if switch['reweight'] != 'OFF': + cards.append('reweight_card.dat') + if switch['analysis'].upper() in ['MADANALYSIS5']: + cards.append('madanalysis5_parton_card.dat') + if switch['analysis'].upper() in ['MADANALYSIS5'] and not switch['shower']=='OFF': + cards.append('madanalysis5_hadron_card.dat') + elif switch['analysis'].upper() in ['MADANALYSIS4']: + cards.append('plot_card.dat') + elif switch['analysis'].upper() in ['RIVET']: + cards.append('rivet_card.dat') + + self.keep_cards(cards) + + first_cmd = cmd_switch.get_cardcmd() + + if os.path.isfile(pjoin(self.me_dir,'Cards','MadLoopParams.dat')): + cards.append('MadLoopParams.dat') + + if self.force: + self.check_param_card(pjoin(self.me_dir,'Cards','param_card.dat' )) + return switch + + + if 'dynamical' in switch and switch['dynamical']: + self.ask_edit_cards(cards, plot=False, mode='auto', first_cmd=first_cmd) + else: + self.ask_edit_cards(cards, plot=False, first_cmd=first_cmd) + return switch + + ############################################################################ + def ask_pythia_run_configuration(self, mode=None, pythia_version=6, banner=None): + """Ask the question when launching pythia""" + + pythia_suffix = '' if pythia_version==6 else '%d'%pythia_version + + available_mode = ['0', '1'] + if pythia_version==6: + available_mode.append('2') + if self.options['delphes_path']: + available_mode.append('3') + name = {'0': 'auto', '2':'pgs', '3':'delphes'} + name['1'] = 'pythia%s'%pythia_suffix + options = available_mode + [name[val] for val in available_mode] + question = """Which programs do you want to run? + 0 / auto : running existing cards\n""" + if pythia_version==6: + question += """ 1. pythia : Pythia\n""" + question += """ 2. pgs : Pythia + PGS\n""" + else: + question += """ 1. pythia8 : Pythia8\n""" + + if '3' in available_mode: + question += """ 3. delphes : Pythia%s + Delphes.\n"""%pythia_suffix + + if not self.force: + if not mode: + mode = self.ask(question, '0', options) + elif not mode: + mode = 'auto' + + if mode.isdigit(): + mode = name[mode] + + auto = False + if mode == 'auto': + auto = True + if pythia_version==6 and os.path.exists(pjoin(self.me_dir, + 'Cards', 'pgs_card.dat')): + mode = 'pgs' + elif os.path.exists(pjoin(self.me_dir, 'Cards', 'delphes_card.dat')): + mode = 'delphes' + else: + mode = 'pythia%s'%pythia_suffix + logger.info('Will run in mode %s' % mode) + # Now that we know in which mode we are check that all the card + #exists (copy default if needed) remove pointless one + cards = ['pythia%s_card.dat'%pythia_suffix] + if mode == 'pgs' and pythia_version==6: + cards.append('pgs_card.dat') + if mode == 'delphes': + cards.append('delphes_card.dat') + delphes3 = True + if os.path.exists(pjoin(self.options['delphes_path'], 'data')): + delphes3 = False + cards.append('delphes_trigger.dat') + self.keep_cards(cards, ignore=['madanalysis5_parton_card.dat','madanalysis5_hadron_card.dat', + 'plot_card.dat']) + + if self.force: + return mode + + if not banner: + banner = self.banner + + if auto: + self.ask_edit_cards(cards, from_banner=['param', 'run'], + mode='auto', plot=(pythia_version==6), banner=banner + ) + else: + self.ask_edit_cards(cards, from_banner=['param', 'run'], + plot=(pythia_version==6), banner=banner) + + return mode + +#=============================================================================== +# MadEventCmd +#=============================================================================== +class MadEventCmdShell(MadEventCmd, cmd.CmdShell): + """The command line processor of MadGraph""" + + + +#=============================================================================== +# HELPING FUNCTION For Subprocesses +#=============================================================================== +class SubProcesses(object): + + name_to_pdg = {} + + @classmethod + def clean(cls): + cls.name_to_pdg = {} + + @staticmethod + def get_subP(me_dir): + """return the list of Subprocesses""" + + out = [] + for line in open(pjoin(me_dir,'SubProcesses', 'subproc.mg')): + if not line: + continue + name = line.strip() + if os.path.exists(pjoin(me_dir, 'SubProcesses', name)): + out.append(pjoin(me_dir, 'SubProcesses', name)) + + return out + + + + @staticmethod + def get_subP_info(path): + """ return the list of processes with their name""" + + nb_sub = 0 + names = {} + old_main = '' + + if not os.path.exists(os.path.join(path,'processes.dat')): + return SubProcesses.get_subP_info_v4(path) + + for line in open(os.path.join(path,'processes.dat')): + main = line[:8].strip() + if main == 'mirror': + main = old_main + if line[8:].strip() == 'none': + continue + else: + main = int(main) + old_main = main + + sub_proccess = line[8:] + nb_sub += sub_proccess.count(',') + 1 + if main in names: + names[main] += [sub_proccess.split(',')] + else: + names[main]= [sub_proccess.split(',')] + + return names + + @staticmethod + def get_subP_info_v4(path): + """ return the list of processes with their name in case without grouping """ + + nb_sub = 0 + names = {'':[[]]} + path = os.path.join(path, 'auto_dsig.f') + found = 0 + for line in open(path): + if line.startswith('C Process:'): + found += 1 + names[''][0].append(line[15:]) + elif found >1: + break + return names + + + @staticmethod + def get_subP_ids(path): + """return the pdg codes of the particles present in the Subprocesses""" + + all_ids = [] + for line in open(pjoin(path, 'leshouche.inc')): + if not 'IDUP' in line: + continue + particles = re.search("/([\d,-]+)/", line) + all_ids.append([int(p) for p in particles.group(1).split(',')]) + return all_ids + + +#=============================================================================== +class GridPackCmd(MadEventCmd): + """The command for the gridpack --Those are not suppose to be use interactively--""" + + def __init__(self, me_dir = None, nb_event=0, seed=0, gran=-1, *completekey, **stdin): + """Initialize the command and directly run""" + + # Initialize properly + self.readonly = False + MadEventCmd.__init__(self, me_dir, *completekey, **stdin) + self.run_mode = 0 + self.random = seed + self.random_orig = self.random + self.granularity = gran + + self.options['automatic_html_opening'] = False + #write the grid_card.dat on disk + self.nb_event = int(nb_event) + self.write_gridcard(nb_event, seed, gran) # set readonly on True if needed + self.prepare_local_dir() # move to gridpack dir or create local structure + # Now it's time to run! + print("__CUDACPP_DEBUG: GridPackCmd.__init__ will call self.launch") + if me_dir and nb_event and seed: + self.launch(nb_event, seed) + else: + raise MadGraph5Error('Gridpack run failed: ' + str(me_dir) + str(nb_event) + \ + str(seed)) + print("__CUDACPP_DEBUG: GridPackCmd.__init__ back from self.launch") + + def update_status(self, *args, **opts): + return + + def load_results_db(self): + """load the current results status""" + model = self.find_model_name() + process = self.process # define in find_model_name + self.results = gen_crossxhtml.AllResults(model, process, self.me_dir) + self.last_mode='' + + def save_random(self): + """save random number in appropirate file""" + + if not self.readonly: + fsock = open(pjoin(self.me_dir, 'SubProcesses','randinit'),'w') + else: + fsock = open('randinit','w') + fsock.writelines('r=%s\n' % self.random) + + def write_RunWeb(self, me_dir): + try: + super(GridPackCmd, self).write_RunWeb(me_dir) + except IOError: + self.readonly =True + + def write_gridcard(self, nb_event, seed, gran): + """write the grid_card.dat file at appropriate location""" + + # first try to write grid_card within the gridpack. + print("WRITE GRIDCARD", self.me_dir) + if self.readonly: + if not os.path.exists('Cards'): + os.mkdir('Cards') + fsock = open('grid_card.dat','w') + else: + fsock = open(pjoin(self.me_dir, 'Cards', 'grid_card.dat'),'w') + + gridpackcard = banner_mod.GridpackCard() + gridpackcard['GridRun'] = True + gridpackcard['gevents'] = nb_event + gridpackcard['gseed'] = seed + gridpackcard['ngran'] = gran + + gridpackcard.write(fsock) + + ############################################################################ + def get_Pdir(self): + """get the list of Pdirectory if not yet saved.""" + + if hasattr(self, "Pdirs"): + if self.me_dir in self.Pdirs[0]: + return self.Pdirs + + if not self.readonly: + self.Pdirs = [pjoin(self.me_dir, 'SubProcesses', l.strip()) + for l in open(pjoin(self.me_dir,'SubProcesses', 'subproc.mg'))] + else: + self.Pdirs = [l.strip() + for l in open(pjoin(self.me_dir,'SubProcesses', 'subproc.mg'))] + + return self.Pdirs + + def prepare_local_dir(self): + """create the P directory structure in the local directory""" + + if not self.readonly: + os.chdir(self.me_dir) + else: + for line in open(pjoin(self.me_dir,'SubProcesses','subproc.mg')): + p = line.strip() + os.mkdir(p) + files.cp(pjoin(self.me_dir,'SubProcesses',p,'symfact.dat'), + pjoin(p, 'symfact.dat')) + + + def launch(self, nb_event, seed): + """ launch the generation for the grid """ + print("__CUDACPP_DEBUG: GridPackCmd.launch starting") + cudacpp_start = time.perf_counter() + # 1) Restore the default data + print("__CUDACPP_DEBUG: GridPackCmd.launch (1) restore_data") + logger.info('generate %s events' % nb_event) + self.set_run_name('GridRun_%s' % seed) + if not self.readonly: + self.update_status('restoring default data', level=None) + misc.call([pjoin(self.me_dir,'bin','internal','restore_data'), + 'default'], cwd=self.me_dir) + + if self.run_card['python_seed'] == -2: + import random + if not hasattr(random, 'mg_seedset'): + random.seed(seed) + random.mg_seedset = seed + elif self.run_card['python_seed'] > 0: + import random + if not hasattr(random, 'mg_seedset'): + random.seed(self.run_card['python_seed']) + random.mg_seedset = self.run_card['python_seed'] + # 2) Run the refine for the grid + print("__CUDACPP_DEBUG: GridPackCmd.launch (2) refine4grid") + self.update_status('Generating Events', level=None) + #misc.call([pjoin(self.me_dir,'bin','refine4grid'), + # str(nb_event), '0', 'Madevent','1','GridRun_%s' % seed], + # cwd=self.me_dir) + self.refine4grid(nb_event) + # 3) Combine the events/pythia/... + print("__CUDACPP_DEBUG: GridPackCmd.launch (3a) combine_events") + self.exec_cmd('combine_events') + print("__CUDACPP_DEBUG: GridPackCmd.launch (3b) store_events") + if not self.readonly: + self.exec_cmd('store_events') + self.print_results_in_shell(self.results.current) + if self.run_card['systematics_program'] == 'systematics' and self.run_card['use_syst']: + self.exec_cmd('systematics %s --from_card' % self.run_name, + postcmd=False,printcmd=False) + self.exec_cmd('decay_events -from_cards', postcmd=False) + elif self.run_card['use_syst'] and self.run_card['systematics_program'] == 'systematics': + self.options['nb_core'] = 1 + self.exec_cmd('systematics %s --from_card' % + pjoin('Events', self.run_name, 'unweighted_events.lhe.gz'), + postcmd=False,printcmd=False) + cudacpp_end = time.perf_counter() + cudacpp_length = cudacpp_end - cudacpp_start + print("__CUDACPP_DEBUG: GridPackCmd.launch finished in %.4f seconds"%cudacpp_length) + + def refine4grid(self, nb_event): + """Special refine for gridpack run.""" + print("__CUDACPP_DEBUG: GridPackCmd.refine4grid starting") + cudacpp_start = time.perf_counter() + self.nb_refine += 1 + + precision = nb_event + + self.opts = dict([(key,value[1]) for (key,value) in \ + self._survey_options.items()]) + + # initialize / remove lhapdf mode + # self.configure_directory() # All this has been done before + self.cluster_mode = 0 # force single machine + + # Store seed in randinit file, to be read by ranmar.f + self.save_random() + + self.update_status('Refine results to %s' % precision, level=None) + logger.info("Using random number seed offset = %s" % self.random) + + refine_opt = {'err_goal': nb_event, 'split_channels': False, + 'ngran':self.granularity, 'readonly': self.readonly} + x_improve = gen_ximprove.gen_ximprove_gridpack(self, refine_opt) + x_improve.launch() # create the ajob for the refinment and run those! + self.gscalefact = x_improve.gscalefact #store jacobian associate to the gridpack + + + #bindir = pjoin(os.path.relpath(self.dirbin, pjoin(self.me_dir,'SubProcesses'))) + #print 'run combine!!!' + #combine_runs.CombineRuns(self.me_dir) + + cudacpp_end = time.perf_counter() + cudacpp_length = cudacpp_end - cudacpp_start + print("__CUDACPP_DEBUG: GridPackCmd.refine4grid finished in %.4f seconds"%cudacpp_length) + return # ?!? what is the code after return ?!? + #update html output + Presults = sum_html.collect_result(self) + cross, error = Presults.xsec, Presults.xerru + self.results.add_detail('cross', cross) + self.results.add_detail('error', error) + + + #self.update_status('finish refine', 'parton', makehtml=False) + #devnull.close() + + + + return + self.total_jobs = 0 + subproc = [P for P in os.listdir(pjoin(self.me_dir,'SubProcesses')) if + P.startswith('P') and os.path.isdir(pjoin(self.me_dir,'SubProcesses', P))] + devnull = open(os.devnull, 'w') + for nb_proc,subdir in enumerate(subproc): + subdir = subdir.strip() + Pdir = pjoin(self.me_dir, 'SubProcesses',subdir) + bindir = pjoin(os.path.relpath(self.dirbin, Pdir)) + + logger.info(' %s ' % subdir) + # clean previous run + for match in misc.glob('*ajob*', Pdir): + if os.path.basename(match)[:4] in ['ajob', 'wait', 'run.', 'done']: + os.remove(pjoin(Pdir, match)) + + + logfile = pjoin(Pdir, 'gen_ximprove.log') + misc.call([pjoin(bindir, 'gen_ximprove')], + stdin=subprocess.PIPE, + stdout=open(logfile,'w'), + cwd=Pdir) + + if os.path.exists(pjoin(Pdir, 'ajob1')): + alljobs = misc.glob('ajob*', Pdir) + nb_tot = len(alljobs) + self.total_jobs += nb_tot + for i, job in enumerate(alljobs): + job = os.path.basename(job) + self.launch_job('%s' % job, cwd=Pdir, remaining=(nb_tot-i-1), + run_type='Refine number %s on %s (%s/%s)' % + (self.nb_refine, subdir, nb_proc+1, len(subproc))) + if os.path.exists(pjoin(self.me_dir,'error')): + self.monitor(html=True) + raise MadEventError('Error detected in dir %s: %s' % \ + (Pdir, open(pjoin(self.me_dir,'error')).read())) + self.monitor(run_type='All job submitted for refine number %s' % + self.nb_refine) + + self.update_status("Combining runs", level='parton') + try: + os.remove(pjoin(Pdir, 'combine_runs.log')) + except Exception: + pass + + bindir = pjoin(os.path.relpath(self.dirbin, pjoin(self.me_dir,'SubProcesses'))) + combine_runs.CombineRuns(self.me_dir) + + #update html output + cross, error = self.make_make_all_html_results() + self.results.add_detail('cross', cross) + self.results.add_detail('error', error) + + + self.update_status('finish refine', 'parton', makehtml=False) + devnull.close() + + def do_combine_events(self, line): + """Advanced commands: Launch combine events""" + + if self.readonly: + outdir = 'Events' + if not os.path.exists(outdir): + os.mkdir(outdir) + else: + outdir = pjoin(self.me_dir, 'Events') + args = self.split_arg(line) + # Check argument's validity + self.check_combine_events(args) + gscalefact = self.gscalefact # {(C.get('name')): jac} + # Define The Banner + tag = self.run_card['run_tag'] + # Update the banner with the pythia card + if not self.banner: + self.banner = banner_mod.recover_banner(self.results, 'parton') + self.banner.load_basic(self.me_dir) + # Add cross-section/event information + self.banner.add_generation_info(self.results.current['cross'], self.run_card['nevents']) + if not hasattr(self, 'random_orig'): self.random_orig = 0 + self.banner.change_seed(self.random_orig) + + + if not os.path.exists(pjoin(outdir, self.run_name)): + os.mkdir(pjoin(outdir, self.run_name)) + self.banner.write(pjoin(outdir, self.run_name, + '%s_%s_banner.txt' % (self.run_name, tag))) + + get_wgt = lambda event: event.wgt + AllEvent = lhe_parser.MultiEventFile() + AllEvent.banner = self.banner + + partials = 0 # if too many file make some partial unweighting + sum_xsec, sum_xerru, sum_axsec = 0,[],0 + Gdirs = self.get_Gdir() + Gdirs.sort() + for Gdir in Gdirs: + #mfactor already taken into accoun in auto_dsig.f + if os.path.exists(pjoin(Gdir, 'events.lhe')): + result = sum_html.OneResult('') + result.read_results(pjoin(Gdir, 'results.dat')) + AllEvent.add(pjoin(Gdir, 'events.lhe'), + result.get('xsec')*gscalefact[Gdir], + result.get('xerru')*gscalefact[Gdir], + result.get('axsec')*gscalefact[Gdir] + ) + + sum_xsec += result.get('xsec')*gscalefact[Gdir] + sum_xerru.append(result.get('xerru')*gscalefact[Gdir]) + sum_axsec += result.get('axsec')*gscalefact[Gdir] + + if len(AllEvent) >= 80: #perform a partial unweighting + AllEvent.unweight(pjoin(outdir, self.run_name, "partials%s.lhe.gz" % partials), + get_wgt, log_level=5, trunc_error=1e-2, event_target=self.nb_event) + AllEvent = lhe_parser.MultiEventFile() + AllEvent.banner = self.banner + AllEvent.add(pjoin(outdir, self.run_name, "partials%s.lhe.gz" % partials), + sum_xsec, + math.sqrt(sum(x**2 for x in sum_xerru)), + sum_axsec) + partials +=1 + + if not hasattr(self,'proc_characteristic'): + self.proc_characteristic = self.get_characteristics() + + self.banner.add_generation_info(sum_xsec, self.nb_event) + nb_event = AllEvent.unweight(pjoin(outdir, self.run_name, "unweighted_events.lhe.gz"), + get_wgt, trunc_error=1e-2, event_target=self.nb_event, + log_level=logging.DEBUG, normalization=self.run_card['event_norm'], + proc_charac=self.proc_characteristic) + + + if partials: + for i in range(partials): + try: + os.remove(pjoin(outdir, self.run_name, "partials%s.lhe.gz" % i)) + except Exception: + os.remove(pjoin(outdir, self.run_name, "partials%s.lhe" % i)) + + self.results.add_detail('nb_event', nb_event) + self.banner.add_generation_info(sum_xsec, nb_event) + if self.run_card['bias_module'].lower() not in ['dummy', 'none']: + self.correct_bias() + + +class MadLoopInitializer(object): + """ A container class for the various methods for initializing MadLoop. It is + placed in MadEventInterface because it is used by Madevent for loop-induced + simulations. """ + + @staticmethod + def make_and_run(dir_name,checkRam=False): + """ Compile the check program in the directory dir_name. + Return the compilation and running time. """ + + # Make sure to recreate the executable and modified source + # (The time stamps are sometimes not actualized if it is too fast) + if os.path.isfile(pjoin(dir_name,'check')): + os.remove(pjoin(dir_name,'check')) + os.remove(pjoin(dir_name,'check_sa.o')) + os.remove(pjoin(dir_name,'loop_matrix.o')) + # Now run make + devnull = open(os.devnull, 'w') + start=time.time() + retcode = misc.compile(arg=['-j1','check'], cwd=dir_name, nb_core=1) + compilation_time = time.time()-start + if retcode != 0: + logging.info("Error while executing make in %s" % dir_name) + return None, None, None + + if not checkRam: + start=time.time() + retcode = subprocess.call('./check', + cwd=dir_name, stdout=devnull, stderr=devnull) + + run_time = time.time()-start + ram_usage = None + else: + ptimer = misc.ProcessTimer(['./check'], cwd=dir_name, shell=False, \ + stdout=devnull, stderr=devnull, close_fds=True) + try: + ptimer.execute() + #poll as often as possible; otherwise the subprocess might + # "sneak" in some extra memory usage while you aren't looking + # Accuracy of .2 seconds is enough for the timing. + while ptimer.poll(): + time.sleep(.2) + finally: + #make sure that we don't leave the process dangling. + ptimer.close() + # Notice that ptimer.max_vms_memory is also available if needed. + ram_usage = ptimer.max_rss_memory + # Unfortunately the running time is less precise than with the + # above version + run_time = (ptimer.t1 - ptimer.t0) + retcode = ptimer.p.returncode + + devnull.close() + + if retcode != 0: + logging.warning("Error while executing ./check in %s" % dir_name) + return None, None, None + + return compilation_time, run_time, ram_usage + + @staticmethod + def fix_PSPoint_in_check(dir_path, read_ps = True, npoints = 1, + hel_config = -1, mu_r=0.0, split_orders=-1): + """Set check_sa.f to be reading PS.input assuming a working dir dir_name. + if hel_config is different than -1 then check_sa.f is configured so to + evaluate only the specified helicity. + If mu_r > 0.0, then the renormalization constant value will be hardcoded + directly in check_sa.f, if is is 0 it will be set to Sqrt(s) and if it + is < 0.0 the value in the param_card.dat is used. + If the split_orders target (i.e. the target squared coupling orders for + the computation) is != -1, it will be changed in check_sa.f via the + subroutine CALL SET_COUPLINGORDERS_TARGET(split_orders).""" + + file_path = dir_path + if not os.path.isfile(dir_path) or \ + not os.path.basename(dir_path)=='check_sa.f': + file_path = pjoin(dir_path,'check_sa.f') + if not os.path.isfile(file_path): + directories = [d for d in misc.glob('P*_*', dir_path) \ + if (re.search(r'.*P\d+_\w*$', d) and os.path.isdir(d))] + if len(directories)>0: + file_path = pjoin(directories[0],'check_sa.f') + if not os.path.isfile(file_path): + raise MadGraph5Error('Could not find the location of check_sa.f'+\ + ' from the specified path %s.'%str(file_path)) + + file = open(file_path, 'r') + check_sa = file.read() + file.close() + + file = open(file_path, 'w') + check_sa = re.sub(r"READPS = \S+\)","READPS = %s)"%('.TRUE.' if read_ps \ + else '.FALSE.'), check_sa) + check_sa = re.sub(r"NPSPOINTS = \d+","NPSPOINTS = %d"%npoints, check_sa) + if hel_config != -1: + check_sa = re.sub(r"SLOOPMATRIX\S+\(\S+,MATELEM,", + "SLOOPMATRIXHEL_THRES(P,%d,MATELEM,"%hel_config, check_sa) + else: + check_sa = re.sub(r"SLOOPMATRIX\S+\(\S+,MATELEM,", + "SLOOPMATRIX_THRES(P,MATELEM,",check_sa) + if mu_r > 0.0: + check_sa = re.sub(r"MU_R=SQRTS","MU_R=%s"%\ + (("%.17e"%mu_r).replace('e','d')),check_sa) + elif mu_r < 0.0: + check_sa = re.sub(r"MU_R=SQRTS","",check_sa) + + if split_orders > 0: + check_sa = re.sub(r"SET_COUPLINGORDERS_TARGET\(-?\d+\)", + "SET_COUPLINGORDERS_TARGET(%d)"%split_orders,check_sa) + + file.write(check_sa) + file.close() + + @staticmethod + def run_initialization(run_dir=None, SubProc_dir=None, infos=None,\ + req_files = ['HelFilter.dat','LoopFilter.dat'], + attempts = [4,15]): + """ Run the initialization of the process in 'run_dir' with success + characterized by the creation of the files req_files in this directory. + The directory containing the driving source code 'check_sa.f'. + The list attempt gives the successive number of PS points the + initialization should be tried with before calling it failed. + Returns the number of PS points which were necessary for the init. + Notice at least run_dir or SubProc_dir must be provided. + A negative attempt number given in input means that quadprec will be + forced for initialization.""" + + # If the user does not want detailed info, then set the dictionary + # to a dummy one. + if infos is None: + infos={} + + if SubProc_dir is None and run_dir is None: + raise MadGraph5Error('At least one of [SubProc_dir,run_dir] must'+\ + ' be provided in run_initialization.') + + # If the user does not specify where is check_sa.f, then it is assumed + # to be one levels above run_dir + if SubProc_dir is None: + SubProc_dir = os.path.abspath(pjoin(run_dir,os.pardir)) + + if run_dir is None: + directories =[ dir for dir in misc.glob('P[0-9]*', SubProc_dir) + if os.path.isdir(dir) ] + if directories: + run_dir = directories[0] + else: + raise MadGraph5Error('Could not find a valid running directory'+\ + ' in %s.'%str(SubProc_dir)) + + # Use the presence of the file born_matrix.f to decide if it is a + # loop-induced process or not. It's not crucial, but just that because + # of the dynamic adjustment of the ref scale used for deciding what are + # the zero contributions, more points are neeeded for loop-induced. + if not os.path.isfile(pjoin(run_dir,'born_matrix.f')): + if len(attempts)>=1 and attempts[0]<8: + attempts[0]=8 + if len(attempts)>=2 and attempts[1]<25: + attempts[1]=25 + + to_attempt = list(attempts) + to_attempt.reverse() + my_req_files = list(req_files) + + MLCardPath = pjoin(SubProc_dir,'MadLoopParams.dat') + if not os.path.isfile(MLCardPath): + raise MadGraph5Error('Could not find MadLoopParams.dat at %s.'\ + %MLCardPath) + else: + MLCard = banner_mod.MadLoopParam(MLCardPath) + MLCard_orig = banner_mod.MadLoopParam(MLCard) + + # Make sure that LoopFilter really is needed. + if not MLCard['UseLoopFilter']: + try: + my_req_files.remove('LoopFilter.dat') + except ValueError: + pass + + if MLCard['HelicityFilterLevel']==0: + try: + my_req_files.remove('HelFilter.dat') + except ValueError: + pass + + def need_init(): + """ True if init not done yet.""" + proc_prefix_file = open(pjoin(run_dir,'proc_prefix.txt'),'r') + proc_prefix = proc_prefix_file.read() + proc_prefix_file.close() + return any([not os.path.exists(pjoin(run_dir,'MadLoop5_resources', + proc_prefix+fname)) for fname in my_req_files]) or \ + not os.path.isfile(pjoin(run_dir,'check')) or \ + not os.access(pjoin(run_dir,'check'), os.X_OK) + + # Check if this is a process without born by checking the presence of the + # file born_matrix.f + is_loop_induced = os.path.exists(pjoin(run_dir,'born_matrix.f')) + + # For loop induced processes, always attempt quadruple precision if + # double precision attempts fail and the user didn't specify himself + # quadruple precision initializations attempts + if not any(attempt<0 for attempt in to_attempt): + to_attempt = [-attempt for attempt in to_attempt] + to_attempt + use_quad_prec = 1 + curr_attempt = 1 + + MLCard.set('WriteOutFilters',True) + + while to_attempt!=[] and need_init(): + curr_attempt = to_attempt.pop() + # if the attempt is a negative number it means we must force + # quadruple precision at initialization time + if curr_attempt < 0: + use_quad_prec = -1 + # In quadruple precision we can lower the ZeroThres threshold + MLCard.set('CTModeInit',4) + MLCard.set('ZeroThres',1e-11) + else: + # Restore the default double precision intialization params + MLCard.set('CTModeInit',1) + MLCard.set('ZeroThres',1e-9) + # Plus one because the filter are written on the next PS point after + curr_attempt = abs(curr_attempt+1) + MLCard.set('MaxAttempts',curr_attempt) + MLCard.write(pjoin(SubProc_dir,'MadLoopParams.dat')) + + # initialization is performed. + MadLoopInitializer.fix_PSPoint_in_check(run_dir, read_ps = False, + npoints = curr_attempt) + compile_time, run_time, ram_usage = \ + MadLoopInitializer.make_and_run(run_dir) + if compile_time==None: + logging.error("Failed at running the process in %s."%run_dir) + attempts = None + return None + # Only set process_compilation time for the first compilation. + if 'Process_compilation' not in list(infos.keys()) or \ + infos['Process_compilation']==None: + infos['Process_compilation'] = compile_time + infos['Initialization'] = run_time + + MLCard_orig.write(pjoin(SubProc_dir,'MadLoopParams.dat')) + if need_init(): + return None + else: + return use_quad_prec*(curr_attempt-1) + + @staticmethod + def need_MadLoopInit(proc_dir, subproc_prefix='PV'): + """Checks whether the necessary filters are present or not.""" + + def need_init(ML_resources_path, proc_prefix, r_files): + """ Returns true if not all required files are present. """ + return any([not os.path.exists(pjoin(ML_resources_path, + proc_prefix+fname)) for fname in r_files]) + + MLCardPath = pjoin(proc_dir,'SubProcesses','MadLoopParams.dat') + if not os.path.isfile(MLCardPath): + raise MadGraph5Error('Could not find MadLoopParams.dat at %s.'\ + %MLCardPath) + MLCard = banner_mod.MadLoopParam(MLCardPath) + + req_files = ['HelFilter.dat','LoopFilter.dat'] + # Make sure that LoopFilter really is needed. + if not MLCard['UseLoopFilter']: + try: + req_files.remove('LoopFilter.dat') + except ValueError: + pass + if MLCard['HelicityFilterLevel']==0: + try: + req_files.remove('HelFilter.dat') + except ValueError: + pass + + for v_folder in glob.iglob(pjoin(proc_dir,'SubProcesses', + '%s*'%subproc_prefix)): + # Make sure it is a valid MadLoop directory + if not os.path.isdir(v_folder) or not os.path.isfile(\ + pjoin(v_folder,'loop_matrix.f')): + continue + proc_prefix_file = open(pjoin(v_folder,'proc_prefix.txt'),'r') + proc_prefix = proc_prefix_file.read() + proc_prefix_file.close() + if need_init(pjoin(proc_dir,'SubProcesses','MadLoop5_resources'), + proc_prefix, req_files): + return True + + return False + + @staticmethod + def init_MadLoop(proc_dir, n_PS=None, subproc_prefix='PV', MG_options=None, + interface = None): + """Advanced commands: Compiles and run MadLoop on RAMBO random PS points to initilize the + filters.""" + + logger.debug('Compiling Source materials necessary for MadLoop '+ + 'initialization.') + # Initialize all the virtuals directory, so as to generate the necessary + # filters (essentially Helcity filter). + # Make sure that the MadLoopCard has the loop induced settings + if interface is None: + misc.compile(arg=['treatCardsLoopNoInit'], cwd=pjoin(proc_dir,'Source')) + else: + interface.do_treatcards('all --no_MadLoopInit') + + # First make sure that IREGI and CUTTOOLS are compiled if needed + if os.path.exists(pjoin(proc_dir,'Source','CutTools')): + misc.compile(arg=['libcuttools'],cwd=pjoin(proc_dir,'Source')) + if os.path.exists(pjoin(proc_dir,'Source','IREGI')): + misc.compile(arg=['libiregi'],cwd=pjoin(proc_dir,'Source')) + # Then make sure DHELAS and MODEL are compiled + misc.compile(arg=['libmodel'],cwd=pjoin(proc_dir,'Source')) + misc.compile(arg=['libdhelas'],cwd=pjoin(proc_dir,'Source')) + + # Now initialize the MadLoop outputs + logger.info('Initializing MadLoop loop-induced matrix elements '+\ + '(this can take some time)...') + + # Setup parallelization + if MG_options: + if interface and hasattr(interface, 'cluster') and isinstance(interface.cluster, cluster.MultiCore): + mcore = interface.cluster + else: + mcore = cluster.MultiCore(**MG_options) + else: + mcore = cluster.onecore + + def run_initialization_wrapper(run_dir, infos, attempts): + if attempts is None: + n_PS = MadLoopInitializer.run_initialization( + run_dir=run_dir, infos=infos) + else: + n_PS = MadLoopInitializer.run_initialization( + run_dir=run_dir, infos=infos, attempts=attempts) + infos['nPS'] = n_PS + return 0 + + def wait_monitoring(Idle, Running, Done): + if Idle+Running+Done == 0: + return + logger.debug('MadLoop initialization jobs: %d Idle, %d Running, %d Done'\ + %(Idle, Running, Done)) + + init_info = {} + # List all virtual folders while making sure they are valid MadLoop folders + VirtualFolders = [f for f in glob.iglob(pjoin(proc_dir,'SubProcesses', + '%s*'%subproc_prefix)) if (os.path.isdir(f) or + os.path.isfile(pjoin(f,'loop_matrix.f')))] + logger.debug("Now Initializing MadLoop matrix element in %d folder%s:"%\ + (len(VirtualFolders),'s' if len(VirtualFolders)>1 else '')) + logger.debug(', '.join("'%s'"%os.path.basename(v_folder) for v_folder in + VirtualFolders)) + for v_folder in VirtualFolders: + init_info[v_folder] = {} + + # We try all multiples of n_PS from 1 to max_mult, first in DP and then + # in QP before giving up, or use default values if n_PS is None. + max_mult = 3 + if n_PS is None: + # Then use the default list of number of PS points to try + mcore.submit(run_initialization_wrapper, + [pjoin(v_folder), init_info[v_folder], None]) + else: + # Use specific set of PS points + mcore.submit(run_initialization_wrapper, [pjoin(v_folder), + init_info[v_folder], + [n_PS*multiplier for multiplier in range(1,max_mult+1)]]) + + # Wait for all jobs to finish. + mcore.wait('',wait_monitoring,update_first=wait_monitoring) + for v_folder in VirtualFolders: + init = init_info[v_folder] + if init['nPS'] is None: + raise MadGraph5Error('Failed the initialization of'+\ + " loop-induced matrix element '%s'%s."%\ + (os.path.basename(v_folder),' (using default n_PS points)' if\ + n_PS is None else ' (trying with a maximum of %d PS points)'\ + %(max_mult*n_PS))) + if init['nPS']==0: + logger.debug("Nothing to be done in '%s', all filters already "%\ + os.path.basename(v_folder)+\ + "present (use the '-r' option to force their recomputation)") + else: + logger.debug("'%s' finished using "%os.path.basename(v_folder)+ + '%d PS points (%s), in %.3g(compil.) + %.3g(init.) secs.'%( + abs(init['nPS']),'DP' if init['nPS']>0 else 'QP', + init['Process_compilation'],init['Initialization'])) + + logger.info('MadLoop initialization finished.') + +AskforEditCard = common_run.AskforEditCard + + +if '__main__' == __name__: + # Launch the interface without any check if one code is already running. + # This can ONLY run a single command !! + import sys + if sys.version_info < (3, 7): + sys.exit('MadGraph/MadEvent 5 works only with python 3.7 or later).\n'+\ + 'Please upgrate your version of python.') + + import os + import optparse + # Get the directory of the script real path (bin) + # and add it to the current PYTHONPATH + #root_path = os.path.dirname(os.path.dirname(os.path.dirname(os.path.realpath( __file__ )))) + sys.path.insert(0, root_path) + + class MyOptParser(optparse.OptionParser): + class InvalidOption(Exception): pass + def error(self, msg=''): + raise MyOptParser.InvalidOption(msg) + # Write out nice usage message if called with -h or --help + usage = "usage: %prog [options] [FILE] " + parser = MyOptParser(usage=usage) + parser.add_option("-l", "--logging", default='INFO', + help="logging level (DEBUG|INFO|WARNING|ERROR|CRITICAL) [%default]") + parser.add_option("","--web", action="store_true", default=False, dest='web', \ + help='force toce to be in secure mode') + parser.add_option("","--debug", action="store_true", default=False, dest='debug', \ + help='force to launch debug mode') + parser_error = '' + done = False + + for i in range(len(sys.argv)-1): + try: + (options, args) = parser.parse_args(sys.argv[1:len(sys.argv)-i]) + done = True + except MyOptParser.InvalidOption as error: + pass + else: + args += sys.argv[len(sys.argv)-i:] + if not done: + # raise correct error: + try: + (options, args) = parser.parse_args() + except MyOptParser.InvalidOption as error: + print(error) + sys.exit(2) + + if len(args) == 0: + args = '' + + import subprocess + import logging + import logging.config + # Set logging level according to the logging level given by options + #logging.basicConfig(level=vars(logging)[options.logging]) + import internal + import internal.coloring_logging + # internal.file = XXX/bin/internal/__init__.py + # => need three dirname to get XXX + # we use internal to have any issue with pythonpath finding the wrong file + me_dir = os.path.dirname(os.path.dirname(os.path.dirname(internal.__file__))) + print("me_dir is", me_dir) + try: + if __debug__ and options.logging == 'INFO': + options.logging = 'DEBUG' + if options.logging.isdigit(): + level = int(options.logging) + else: + level = eval('logging.' + options.logging) + log_path = os.path.join(me_dir, 'bin', 'internal', 'me5_logging.conf') + logging.config.fileConfig(log_path) + logging.root.setLevel(level) + logging.getLogger('madgraph').setLevel(level) + except: + raise + pass + + # Call the cmd interface main loop + try: + if args: + # a single command is provided + if '--web' in args: + i = args.index('--web') + args.pop(i) + cmd_line = MadEventCmd(me_dir, force_run=True) + else: + cmd_line = MadEventCmdShell(me_dir, force_run=True) + if not hasattr(cmd_line, 'do_%s' % args[0]): + if parser_error: + print(parser_error) + print('and %s can not be interpreted as a valid command.' % args[0]) + else: + print('ERROR: %s not a valid command. Please retry' % args[0]) + else: + cmd_line.use_rawinput = False + cmd_line.run_cmd(' '.join(args)) + cmd_line.run_cmd('quit') + + except KeyboardInterrupt: + print('quit on KeyboardInterrupt') + pass + + + + + + + + + + + + diff --git a/epochX/cudacpp/tlau/gridpacks/gg_tt.mad/output.txt b/epochX/cudacpp/tlau/gridpacks/gg_tt.mad/output.txt new file mode 100644 index 0000000000..f9e2591868 --- /dev/null +++ b/epochX/cudacpp/tlau/gridpacks/gg_tt.mad/output.txt @@ -0,0 +1,70 @@ +START: Wed Aug 7 04:24:41 PM CEST 2024 +On itgold91.cern.ch [CPU: Intel(R) Xeon(R) Gold 6326 CPU (nproc=32)] [GPU: none]: +CUDACPP_RUNTIME_DISABLEFPE is not set +************************************************************ +* * +* W E L C O M E to * +* M A D G R A P H 5 _ a M C @ N L O * +* M A D E V E N T * +* * +* * * * +* * * * * * +* * * * * 5 * * * * * +* * * * * * +* * * * +* * +* VERSION 3.5.3_lo_vect * +* * +* The MadGraph5_aMC@NLO Development Team - Find us at * +* https://server06.fynu.ucl.ac.be/projects/madgraph * +* * +* Type 'help' for in-line help. * +* * +************************************************************ +INFO: load configuration from /data/avalassi/GPU2024/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/Cards/me5_configuration.txt +INFO: load configuration from /data/avalassi/GPU2024/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/Cards/me5_configuration.txt +Using default text editor "vi". Set another one in ./input/mg5_configuration.txt +No valid eps viewer found. Please set in ./input/mg5_configuration.txt +Using default web browser "firefox". Set another one in ./input/mg5_configuration.txt +generate_events -f +Generating gridpack with run name run_01 +survey run_01 --accuracy=0.01 --points=2000 --iterations=8 --gridpack=.true. +INFO: compile directory +compile Source Directory +Using random number seed offset = 21 +INFO: Running Survey +Creating Jobs +Working on SubProcesses +INFO: P1_gg_ttx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'fortran' matrix elements +INFO: Idle: 1, Running: 0, Completed: 0 [ current time: 16h24 ] +INFO: Idle: 0, Running: 0, Completed: 1 [ 1.8s ] +INFO: Idle: 0, Running: 0, Completed: 1 [ 1.8s ] +sum of cpu time of last step: 2 seconds + === Results Summary for run: run_01 tag: tag_1 === + + Cross-section : 442.3 +- 4.036 pb + Nb of events : 0 + +INFO: End survey +combine_events +INFO: Combining Events +combination of events done in 0.0027005672454833984 s +store_events +INFO: Storing parton level results +INFO: End Parton +decay_events -from_cards +create_gridpack +INFO: Creating gridpack +P1_gg_ttx +Cleaning SubProcesses. +INFO: gridpack created +quit +INFO: storing files of previous run +INFO: Done +INFO: +17.02user 3.01system 0:07.92elapsed 252%CPU (0avgtext+0avgdata 91800maxresident)k +288inputs+285808outputs (283major+992446minor)pagefaults 0swaps +END: Wed Aug 7 04:24:49 PM CEST 2024 +ELAPSED: 8 seconds diff --git a/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/output.txt b/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/output.txt new file mode 100644 index 0000000000..c6e90944e5 --- /dev/null +++ b/epochX/cudacpp/tlau/gridpacks/gq_ttq.mad/output.txt @@ -0,0 +1,74 @@ +START: Tue Sep 10 06:46:17 PM CEST 2024 +On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU (nproc=4)] [GPU: 1x Tesla V100S-PCIE-32GB]: +CUDACPP_RUNTIME_DISABLEFPE is not set +************************************************************ +* * +* W E L C O M E to * +* M A D G R A P H 5 _ a M C @ N L O * +* M A D E V E N T * +* * +* * * * +* * * * * * +* * * * * 5 * * * * * +* * * * * * +* * * * +* * +* VERSION 3.5.3_lo_vect * +* * +* The MadGraph5_aMC@NLO Development Team - Find us at * +* https://server06.fynu.ucl.ac.be/projects/madgraph * +* * +* Type 'help' for in-line help. * +* * +************************************************************ +INFO: load configuration from /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gq_ttq.mad/Cards/me5_configuration.txt +INFO: load configuration from /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gq_ttq.mad/Cards/me5_configuration.txt +Using default text editor "vi". Set another one in ./input/mg5_configuration.txt +Using default eps viewer "evince". Set another one in ./input/mg5_configuration.txt +Using default web browser "firefox". Set another one in ./input/mg5_configuration.txt +generate_events -f +Generating gridpack with run name run_01 +survey run_01 --accuracy=0.01 --points=2000 --iterations=8 --gridpack=.true. +INFO: compile directory +compile Source Directory +Using random number seed offset = 21 +INFO: Running Survey +Creating Jobs +Working on SubProcesses +INFO: P1_gu_ttxu +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P1_gux_ttxux +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: Idle: 1, Running: 1, Completed: 2 [ current time: 18h47 ] +INFO: Idle: 0, Running: 2, Completed: 2 [ current time: 18h47 ] +INFO: Idle: 0, Running: 0, Completed: 4 [ 3s ] +sum of cpu time of last step: 11 seconds + === Results Summary for run: run_01 tag: tag_1 === + + Cross-section : 133.9 +- 1.921 pb + Nb of events : 0 + +INFO: End survey +combine_events +INFO: Combining Events +combination of events done in 0.006100177764892578 s +store_events +INFO: Storing parton level results +INFO: End Parton +decay_events -from_cards +create_gridpack +INFO: Creating gridpack +P1_gu_ttxu +P1_gux_ttxux +Cleaning SubProcesses.. +INFO: gridpack created +quit +INFO: storing files of previous run +INFO: Done +INFO: +167.23user 28.43system 1:33.76elapsed 208%CPU (0avgtext+0avgdata 398048maxresident)k +160inputs+1337160outputs (88major+7248399minor)pagefaults 0swaps +END: Tue Sep 10 06:47:52 PM CEST 2024 +ELAPSED: 95 seconds diff --git a/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/output.txt b/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/output.txt new file mode 100644 index 0000000000..5f2a287718 --- /dev/null +++ b/epochX/cudacpp/tlau/gridpacks/pp_dy3j.mad/output.txt @@ -0,0 +1,1081 @@ +START: Tue Sep 10 10:36:17 PM CEST 2024 +On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU (nproc=4)] [GPU: 1x Tesla V100S-PCIE-32GB]: +CUDACPP_RUNTIME_DISABLEFPE is set +************************************************************ +* * +* W E L C O M E to * +* M A D G R A P H 5 _ a M C @ N L O * +* M A D E V E N T * +* * +* * * * +* * * * * * +* * * * * 5 * * * * * +* * * * * * +* * * * +* * +* VERSION 3.5.3_lo_vect * +* * +* The MadGraph5_aMC@NLO Development Team - Find us at * +* https://server06.fynu.ucl.ac.be/projects/madgraph * +* * +* Type 'help' for in-line help. * +* * +************************************************************ +INFO: load configuration from /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/pp_dy3j.mad/Cards/me5_configuration.txt +INFO: load configuration from /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/pp_dy3j.mad/Cards/me5_configuration.txt +Using default text editor "vi". Set another one in ./input/mg5_configuration.txt +Using default eps viewer "evince". Set another one in ./input/mg5_configuration.txt +Using default web browser "firefox". Set another one in ./input/mg5_configuration.txt +generate_events -f +Generating gridpack with run name run_01 +survey run_01 --accuracy=0.01 --points=2000 --iterations=8 --gridpack=.true. +INFO: compile directory +compile Source Directory +Using random number seed offset = 21 +INFO: Running Survey +Creating Jobs +Working on SubProcesses +INFO: P0_gg_epemguux +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_gg_epemgddx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_gg_taptamguux +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_gg_taptamgddx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_gu_epemggu +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_gd_epemggd +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_gux_epemggux +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_gdx_epemggdx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_gu_taptamggu +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_gd_taptamggd +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_gux_taptamggux +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_gdx_taptamggdx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_uux_epemggg +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_ddx_epemggg +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_uux_taptamggg +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_ddx_taptamggg +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_gu_epemuuux +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_gd_epemdddx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_gux_epemuuxux +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_gdx_epemddxdx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_gu_epemuddx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_gu_epemuccx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_gd_epemudux +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_gd_epemdssx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_gux_epemduxdx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_gux_epemcuxcx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_gdx_epemuuxdx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_gdx_epemsdxsx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_gu_taptamuuux +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_gd_taptamdddx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_gux_taptamuuxux +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_gdx_taptamddxdx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_gu_taptamuddx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_gu_taptamuccx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_gd_taptamudux +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_gd_taptamdssx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_gux_taptamduxdx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_gux_taptamcuxcx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_gdx_taptamuuxdx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_gdx_taptamsdxsx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_uu_epemguu +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_uux_epemguux +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_dd_epemgdd +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_ddx_epemgddx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_uxux_epemguxux +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_dxdx_epemgdxdx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_ud_epemgud +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_uc_epemguc +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_uux_epemgddx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_uux_epemgccx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_udx_epemgudx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_ucx_epemgucx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_dc_epemgdc +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_ds_epemgds +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_dux_epemgdux +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_ddx_epemguux +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_ddx_epemgssx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_dsx_epemgdsx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_uxdx_epemguxdx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_uxcx_epemguxcx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_dxcx_epemgdxcx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_dxsx_epemgdxsx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_uu_taptamguu +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_uux_taptamguux +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_dd_taptamgdd +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_ddx_taptamgddx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_uxux_taptamguxux +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_dxdx_taptamgdxdx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_ud_taptamgud +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_uc_taptamguc +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_uux_taptamgddx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_uux_taptamgccx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_udx_taptamgudx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_ucx_taptamgucx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_dc_taptamgdc +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_ds_taptamgds +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_dux_taptamgdux +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_ddx_taptamguux +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_ddx_taptamgssx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_dsx_taptamgdsx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_uxdx_taptamguxdx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_uxcx_taptamguxcx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_dxcx_taptamgdxcx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_dxsx_taptamgdxsx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: Idle: 1636, Running: 4, Completed: 184 [ current time: 22h46 ] +INFO: Idle: 1635, Running: 4, Completed: 185 [ 5.8s ] +INFO: Idle: 1631, Running: 4, Completed: 189 [ 15.2s ] +INFO: Idle: 1627, Running: 4, Completed: 193 [ 22.3s ] +INFO: Idle: 1626, Running: 4, Completed: 194 [ 27.1s ] +INFO: Idle: 1624, Running: 4, Completed: 196 [ 32.1s ] +INFO: Idle: 1622, Running: 4, Completed: 198 [ 35.8s ] +INFO: Idle: 1621, Running: 4, Completed: 199 [ 39.8s ] +INFO: Idle: 1619, Running: 4, Completed: 201 [ 43.4s ] +INFO: Idle: 1617, Running: 4, Completed: 203 [ 50.8s ] +INFO: Idle: 1615, Running: 4, Completed: 205 [ 54.8s ] +INFO: Idle: 1613, Running: 4, Completed: 207 [ 59.3s ] +INFO: Idle: 1611, Running: 4, Completed: 209 [ 1m 4s ] +INFO: Idle: 1608, Running: 4, Completed: 212 [ 1m 8s ] +INFO: Idle: 1607, Running: 4, Completed: 213 [ 1m 15s ] +INFO: Idle: 1605, Running: 4, Completed: 215 [ 1m 19s ] +INFO: Idle: 1603, Running: 4, Completed: 217 [ 1m 26s ] +INFO: Idle: 1601, Running: 4, Completed: 219 [ 1m 31s ] +INFO: Idle: 1598, Running: 4, Completed: 222 [ 1m 38s ] +INFO: Idle: 1596, Running: 4, Completed: 224 [ 1m 41s ] +INFO: Idle: 1594, Running: 4, Completed: 226 [ 1m 44s ] +INFO: Idle: 1593, Running: 4, Completed: 227 [ 1m 48s ] +INFO: Idle: 1591, Running: 4, Completed: 229 [ 1m 52s ] +INFO: Idle: 1589, Running: 4, Completed: 231 [ 1m 57s ] +INFO: Idle: 1587, Running: 4, Completed: 233 [ 2m 3s ] +INFO: Idle: 1585, Running: 4, Completed: 235 [ 2m 7s ] +INFO: Idle: 1583, Running: 4, Completed: 237 [ 2m 13s ] +INFO: Idle: 1579, Running: 4, Completed: 241 [ 2m 23s ] +INFO: Idle: 1577, Running: 4, Completed: 243 [ 2m 26s ] +INFO: Idle: 1575, Running: 4, Completed: 245 [ 2m 35s ] +INFO: Idle: 1573, Running: 4, Completed: 247 [ 2m 39s ] +INFO: Idle: 1571, Running: 4, Completed: 249 [ 2m 43s ] +INFO: Idle: 1569, Running: 4, Completed: 251 [ 2m 47s ] +INFO: Idle: 1567, Running: 4, Completed: 253 [ 2m 51s ] +INFO: Idle: 1565, Running: 4, Completed: 255 [ 2m 57s ] +INFO: Idle: 1562, Running: 4, Completed: 258 [ 3m 1s ] +INFO: Idle: 1561, Running: 4, Completed: 259 [ 3m 6s ] +INFO: Idle: 1559, Running: 4, Completed: 261 [ 3m 11s ] +INFO: Idle: 1557, Running: 4, Completed: 263 [ 3m 16s ] +INFO: Idle: 1555, Running: 4, Completed: 265 [ 3m 21s ] +INFO: Idle: 1552, Running: 4, Completed: 268 [ 3m 25s ] +INFO: Idle: 1551, Running: 4, Completed: 269 [ 3m 32s ] +INFO: Idle: 1548, Running: 4, Completed: 272 [ 3m 36s ] +INFO: Idle: 1547, Running: 4, Completed: 273 [ 3m 41s ] +INFO: Idle: 1543, Running: 4, Completed: 277 [ 3m 49s ] +INFO: Idle: 1542, Running: 4, Completed: 278 [ 3m 54s ] +INFO: Idle: 1540, Running: 4, Completed: 280 [ 3m 57s ] +INFO: Idle: 1538, Running: 4, Completed: 282 [ 4m 2s ] +INFO: Idle: 1536, Running: 4, Completed: 284 [ 4m 8s ] +INFO: Idle: 1533, Running: 4, Completed: 287 [ 4m 15s ] +INFO: Idle: 1531, Running: 4, Completed: 289 [ 4m 20s ] +INFO: Idle: 1529, Running: 4, Completed: 291 [ 4m 25s ] +INFO: Idle: 1527, Running: 4, Completed: 293 [ 4m 30s ] +INFO: Idle: 1523, Running: 4, Completed: 297 [ 4m 41s ] +INFO: Idle: 1519, Running: 4, Completed: 301 [ 4m 50s ] +INFO: Idle: 1515, Running: 4, Completed: 305 [ 4m 57s ] +INFO: Idle: 1514, Running: 4, Completed: 306 [ 5m 2s ] +INFO: Idle: 1512, Running: 4, Completed: 308 [ 5m 6s ] +INFO: Idle: 1511, Running: 4, Completed: 309 [ 5m 9s ] +INFO: Idle: 1509, Running: 4, Completed: 311 [ 5m 13s ] +INFO: Idle: 1508, Running: 4, Completed: 312 [ 5m 17s ] +INFO: Idle: 1505, Running: 4, Completed: 315 [ 5m 23s ] +INFO: Idle: 1503, Running: 4, Completed: 317 [ 5m 29s ] +INFO: Idle: 1501, Running: 4, Completed: 319 [ 5m 33s ] +INFO: Idle: 1499, Running: 4, Completed: 321 [ 5m 39s ] +INFO: Idle: 1495, Running: 4, Completed: 325 [ 5m 50s ] +INFO: Idle: 1492, Running: 4, Completed: 328 [ 5m 54s ] +INFO: Idle: 1491, Running: 4, Completed: 329 [ 6m 3s ] +INFO: Idle: 1488, Running: 4, Completed: 332 [ 6m 7s ] +INFO: Idle: 1487, Running: 4, Completed: 333 [ 6m 15s ] +INFO: Idle: 1483, Running: 4, Completed: 337 [ 6m 22s ] +INFO: Idle: 1481, Running: 4, Completed: 339 [ 6m 26s ] +INFO: Idle: 1479, Running: 4, Completed: 341 [ 6m 36s ] +INFO: Idle: 1476, Running: 4, Completed: 344 [ 6m 40s ] +INFO: Idle: 1475, Running: 4, Completed: 345 [ 6m 49s ] +INFO: Idle: 1471, Running: 4, Completed: 349 [ 6m 59s ] +INFO: Idle: 1468, Running: 4, Completed: 352 [ 7m 3s ] +INFO: Idle: 1467, Running: 4, Completed: 353 [ 7m 8s ] +INFO: Idle: 1464, Running: 4, Completed: 356 [ 7m 11s ] +INFO: Idle: 1463, Running: 4, Completed: 357 [ 7m 19s ] +INFO: Idle: 1460, Running: 4, Completed: 360 [ 7m 23s ] +INFO: Idle: 1459, Running: 4, Completed: 361 [ 7m 31s ] +INFO: Idle: 1456, Running: 4, Completed: 364 [ 7m 35s ] +INFO: Idle: 1455, Running: 4, Completed: 365 [ 7m 43s ] +INFO: Idle: 1451, Running: 4, Completed: 369 [ 7m 51s ] +INFO: Idle: 1448, Running: 4, Completed: 372 [ 7m 54s ] +INFO: Idle: 1447, Running: 4, Completed: 373 [ 8m 4s ] +INFO: Idle: 1444, Running: 4, Completed: 376 [ 8m 7s ] +INFO: Idle: 1443, Running: 4, Completed: 377 [ 8m 16s ] +INFO: Idle: 1439, Running: 4, Completed: 381 [ 8m 26s ] +INFO: Idle: 1436, Running: 4, Completed: 384 [ 8m 30s ] +INFO: Idle: 1435, Running: 4, Completed: 385 [ 8m 33s ] +INFO: Idle: 1433, Running: 4, Completed: 387 [ 8m 38s ] +INFO: Idle: 1432, Running: 4, Completed: 388 [ 8m 44s ] +INFO: Idle: 1429, Running: 4, Completed: 391 [ 8m 48s ] +INFO: Idle: 1428, Running: 4, Completed: 392 [ 8m 52s ] +INFO: Idle: 1427, Running: 4, Completed: 393 [ 8m 56s ] +INFO: Idle: 1424, Running: 4, Completed: 396 [ 9m 0s ] +INFO: Idle: 1423, Running: 4, Completed: 397 [ 9m 6s ] +INFO: Idle: 1420, Running: 4, Completed: 400 [ 9m 9s ] +INFO: Idle: 1419, Running: 4, Completed: 401 [ 9m 13s ] +INFO: Idle: 1417, Running: 4, Completed: 403 [ 9m 17s ] +INFO: Idle: 1416, Running: 4, Completed: 404 [ 9m 23s ] +INFO: Idle: 1413, Running: 4, Completed: 407 [ 9m 27s ] +INFO: Idle: 1412, Running: 4, Completed: 408 [ 9m 31s ] +INFO: Idle: 1411, Running: 4, Completed: 409 [ 9m 36s ] +INFO: Idle: 1408, Running: 4, Completed: 412 [ 9m 39s ] +INFO: Idle: 1407, Running: 4, Completed: 413 [ 9m 46s ] +INFO: Idle: 1404, Running: 4, Completed: 416 [ 9m 49s ] +INFO: Idle: 1403, Running: 4, Completed: 417 [ 9m 55s ] +INFO: Idle: 1400, Running: 4, Completed: 420 [ 10m 3s ] +INFO: Idle: 1399, Running: 4, Completed: 421 [ 10m 6s ] +INFO: Idle: 1397, Running: 4, Completed: 423 [ 10m 9s ] +INFO: Idle: 1396, Running: 4, Completed: 424 [ 10m 13s ] +INFO: Idle: 1393, Running: 4, Completed: 427 [ 10m 20s ] +INFO: Idle: 1391, Running: 4, Completed: 429 [ 10m 23s ] +INFO: Idle: 1389, Running: 4, Completed: 431 [ 10m 28s ] +INFO: Idle: 1387, Running: 4, Completed: 433 [ 10m 34s ] +INFO: Idle: 1384, Running: 4, Completed: 436 [ 10m 43s ] +INFO: Idle: 1382, Running: 4, Completed: 438 [ 10m 48s ] +INFO: Idle: 1380, Running: 4, Completed: 440 [ 10m 52s ] +INFO: Idle: 1377, Running: 4, Completed: 443 [ 11m 0s ] +INFO: Idle: 1374, Running: 4, Completed: 446 [ 11m 6s ] +INFO: Idle: 1371, Running: 4, Completed: 449 [ 11m 9s ] +INFO: Idle: 1370, Running: 4, Completed: 450 [ 11m 16s ] +INFO: Idle: 1367, Running: 4, Completed: 453 [ 11m 23s ] +INFO: Idle: 1366, Running: 4, Completed: 454 [ 11m 27s ] +INFO: Idle: 1364, Running: 4, Completed: 456 [ 11m 30s ] +INFO: Idle: 1363, Running: 4, Completed: 457 [ 11m 33s ] +INFO: Idle: 1361, Running: 4, Completed: 459 [ 11m 38s ] +INFO: Idle: 1359, Running: 4, Completed: 461 [ 11m 43s ] +INFO: Idle: 1358, Running: 4, Completed: 462 [ 11m 46s ] +INFO: Idle: 1355, Running: 4, Completed: 465 [ 11m 50s ] +INFO: Idle: 1354, Running: 4, Completed: 466 [ 11m 53s ] +INFO: Idle: 1351, Running: 4, Completed: 469 [ 12m 3s ] +INFO: Idle: 1348, Running: 4, Completed: 472 [ 12m 9s ] +INFO: Idle: 1346, Running: 4, Completed: 474 [ 12m 13s ] +INFO: Idle: 1345, Running: 4, Completed: 475 [ 12m 16s ] +INFO: Idle: 1343, Running: 4, Completed: 477 [ 12m 20s ] +INFO: Idle: 1342, Running: 4, Completed: 478 [ 12m 24s ] +INFO: Idle: 1339, Running: 4, Completed: 481 [ 12m 30s ] +INFO: Idle: 1337, Running: 4, Completed: 483 [ 12m 35s ] +INFO: Idle: 1335, Running: 4, Completed: 485 [ 12m 42s ] +INFO: Idle: 1332, Running: 4, Completed: 488 [ 12m 47s ] +INFO: Idle: 1330, Running: 4, Completed: 490 [ 12m 51s ] +INFO: Idle: 1328, Running: 4, Completed: 492 [ 12m 58s ] +INFO: Idle: 1325, Running: 4, Completed: 495 [ 13m 4s ] +INFO: Idle: 1322, Running: 4, Completed: 498 [ 13m 8s ] +INFO: Idle: 1321, Running: 4, Completed: 499 [ 13m 11s ] +INFO: Idle: 1319, Running: 4, Completed: 501 [ 13m 19s ] +INFO: Idle: 1317, Running: 4, Completed: 503 [ 13m 25s ] +INFO: Idle: 1314, Running: 4, Completed: 506 [ 13m 29s ] +INFO: Idle: 1313, Running: 4, Completed: 507 [ 13m 33s ] +INFO: Idle: 1311, Running: 4, Completed: 509 [ 13m 37s ] +INFO: Idle: 1310, Running: 4, Completed: 510 [ 13m 40s ] +INFO: Idle: 1307, Running: 4, Completed: 513 [ 13m 46s ] +INFO: Idle: 1306, Running: 4, Completed: 514 [ 13m 50s ] +INFO: Idle: 1304, Running: 4, Completed: 516 [ 13m 53s ] +INFO: Idle: 1302, Running: 4, Completed: 518 [ 13m 57s ] +INFO: Idle: 1301, Running: 4, Completed: 519 [ 14m 4s ] +INFO: Idle: 1297, Running: 4, Completed: 523 [ 14m 12s ] +INFO: Idle: 1294, Running: 4, Completed: 526 [ 14m 17s ] +INFO: Idle: 1293, Running: 4, Completed: 527 [ 14m 20s ] +INFO: Idle: 1291, Running: 4, Completed: 529 [ 14m 24s ] +INFO: Idle: 1290, Running: 4, Completed: 530 [ 14m 29s ] +INFO: Idle: 1287, Running: 4, Completed: 533 [ 14m 34s ] +INFO: Idle: 1285, Running: 4, Completed: 535 [ 14m 42s ] +INFO: Idle: 1281, Running: 4, Completed: 539 [ 14m 52s ] +INFO: Idle: 1278, Running: 4, Completed: 542 [ 14m 55s ] +INFO: Idle: 1277, Running: 4, Completed: 543 [ 15m 0s ] +INFO: Idle: 1275, Running: 4, Completed: 545 [ 15m 3s ] +INFO: Idle: 1273, Running: 4, Completed: 547 [ 15m 10s ] +INFO: Idle: 1270, Running: 4, Completed: 550 [ 15m 13s ] +INFO: Idle: 1269, Running: 4, Completed: 551 [ 15m 17s ] +INFO: Idle: 1268, Running: 4, Completed: 552 [ 15m 22s ] +INFO: Idle: 1265, Running: 4, Completed: 555 [ 15m 26s ] +INFO: Idle: 1264, Running: 4, Completed: 556 [ 15m 32s ] +INFO: Idle: 1261, Running: 4, Completed: 559 [ 15m 37s ] +INFO: Idle: 1260, Running: 4, Completed: 560 [ 15m 40s ] +INFO: Idle: 1258, Running: 4, Completed: 562 [ 15m 47s ] +INFO: Idle: 1255, Running: 4, Completed: 565 [ 15m 52s ] +INFO: Idle: 1253, Running: 4, Completed: 567 [ 15m 58s ] +INFO: Idle: 1251, Running: 4, Completed: 569 [ 16m 2s ] +INFO: Idle: 1249, Running: 4, Completed: 571 [ 16m 8s ] +INFO: Idle: 1247, Running: 4, Completed: 573 [ 16m 12s ] +INFO: Idle: 1245, Running: 4, Completed: 575 [ 16m 17s ] +INFO: Idle: 1242, Running: 4, Completed: 578 [ 16m 23s ] +INFO: Idle: 1240, Running: 4, Completed: 580 [ 16m 29s ] +INFO: Idle: 1238, Running: 4, Completed: 582 [ 16m 33s ] +INFO: Idle: 1236, Running: 4, Completed: 584 [ 16m 37s ] +INFO: Idle: 1235, Running: 4, Completed: 585 [ 16m 43s ] +INFO: Idle: 1231, Running: 4, Completed: 589 [ 16m 52s ] +INFO: Idle: 1227, Running: 4, Completed: 593 [ 16m 59s ] +INFO: Idle: 1224, Running: 4, Completed: 596 [ 17m 7s ] +INFO: Idle: 1221, Running: 4, Completed: 599 [ 17m 11s ] +INFO: Idle: 1220, Running: 4, Completed: 600 [ 17m 15s ] +INFO: Idle: 1219, Running: 4, Completed: 601 [ 17m 21s ] +INFO: Idle: 1215, Running: 4, Completed: 605 [ 17m 30s ] +INFO: Idle: 1211, Running: 4, Completed: 609 [ 17m 39s ] +INFO: Idle: 1208, Running: 4, Completed: 612 [ 17m 46s ] +INFO: Idle: 1207, Running: 4, Completed: 613 [ 17m 50s ] +INFO: Idle: 1204, Running: 4, Completed: 616 [ 17m 55s ] +INFO: Idle: 1202, Running: 4, Completed: 618 [ 17m 59s ] +INFO: Idle: 1201, Running: 4, Completed: 619 [ 18m 3s ] +INFO: Idle: 1199, Running: 4, Completed: 621 [ 18m 7s ] +INFO: Idle: 1197, Running: 4, Completed: 623 [ 18m 11s ] +INFO: Idle: 1195, Running: 4, Completed: 625 [ 18m 17s ] +INFO: Idle: 1192, Running: 4, Completed: 628 [ 18m 25s ] +INFO: Idle: 1190, Running: 4, Completed: 630 [ 18m 30s ] +INFO: Idle: 1188, Running: 4, Completed: 632 [ 18m 35s ] +INFO: Idle: 1185, Running: 4, Completed: 635 [ 18m 43s ] +INFO: Idle: 1182, Running: 4, Completed: 638 [ 18m 48s ] +INFO: Idle: 1178, Running: 4, Completed: 642 [ 18m 58s ] +INFO: Idle: 1175, Running: 4, Completed: 645 [ 19m 4s ] +INFO: Idle: 1174, Running: 4, Completed: 646 [ 19m 8s ] +INFO: Idle: 1172, Running: 4, Completed: 648 [ 19m 11s ] +INFO: Idle: 1170, Running: 4, Completed: 650 [ 19m 16s ] +INFO: Idle: 1169, Running: 4, Completed: 651 [ 19m 19s ] +INFO: Idle: 1167, Running: 4, Completed: 653 [ 19m 23s ] +INFO: Idle: 1166, Running: 4, Completed: 654 [ 19m 26s ] +INFO: Idle: 1163, Running: 4, Completed: 657 [ 19m 29s ] +INFO: Idle: 1162, Running: 4, Completed: 658 [ 19m 34s ] +INFO: Idle: 1159, Running: 4, Completed: 661 [ 19m 42s ] +INFO: Idle: 1157, Running: 4, Completed: 663 [ 19m 47s ] +INFO: Idle: 1155, Running: 4, Completed: 665 [ 19m 52s ] +INFO: Idle: 1153, Running: 4, Completed: 667 [ 19m 57s ] +INFO: Idle: 1151, Running: 4, Completed: 669 [ 20m 1s ] +INFO: Idle: 1149, Running: 4, Completed: 671 [ 20m 4s ] +INFO: Idle: 1147, Running: 4, Completed: 673 [ 20m 10s ] +INFO: Idle: 1145, Running: 4, Completed: 675 [ 20m 15s ] +INFO: Idle: 1143, Running: 4, Completed: 677 [ 20m 21s ] +INFO: Idle: 1140, Running: 4, Completed: 680 [ 20m 27s ] +INFO: Idle: 1137, Running: 4, Completed: 683 [ 20m 33s ] +INFO: Idle: 1136, Running: 4, Completed: 684 [ 20m 37s ] +INFO: Idle: 1133, Running: 4, Completed: 687 [ 20m 43s ] +INFO: Idle: 1130, Running: 4, Completed: 690 [ 20m 47s ] +INFO: Idle: 1128, Running: 4, Completed: 692 [ 20m 53s ] +INFO: Idle: 1127, Running: 4, Completed: 693 [ 20m 58s ] +INFO: Idle: 1125, Running: 4, Completed: 695 [ 21m 3s ] +INFO: Idle: 1122, Running: 4, Completed: 698 [ 21m 8s ] +INFO: Idle: 1121, Running: 4, Completed: 699 [ 21m 12s ] +INFO: Idle: 1119, Running: 4, Completed: 701 [ 21m 15s ] +INFO: Idle: 1118, Running: 4, Completed: 702 [ 21m 18s ] +INFO: Idle: 1116, Running: 4, Completed: 704 [ 21m 21s ] +INFO: Idle: 1114, Running: 4, Completed: 706 [ 21m 28s ] +INFO: Idle: 1111, Running: 4, Completed: 709 [ 21m 33s ] +INFO: Idle: 1109, Running: 4, Completed: 711 [ 21m 41s ] +INFO: Idle: 1105, Running: 4, Completed: 715 [ 21m 50s ] +INFO: Idle: 1102, Running: 4, Completed: 718 [ 21m 54s ] +INFO: Idle: 1101, Running: 4, Completed: 719 [ 21m 58s ] +INFO: Idle: 1099, Running: 4, Completed: 721 [ 22m 1s ] +INFO: Idle: 1098, Running: 4, Completed: 722 [ 22m 5s ] +INFO: Idle: 1096, Running: 4, Completed: 724 [ 22m 9s ] +INFO: Idle: 1094, Running: 4, Completed: 726 [ 22m 13s ] +INFO: Idle: 1093, Running: 4, Completed: 727 [ 22m 19s ] +INFO: Idle: 1089, Running: 4, Completed: 731 [ 22m 27s ] +INFO: Idle: 1086, Running: 4, Completed: 734 [ 22m 31s ] +INFO: Idle: 1085, Running: 4, Completed: 735 [ 22m 35s ] +INFO: Idle: 1083, Running: 4, Completed: 737 [ 22m 38s ] +INFO: Idle: 1081, Running: 4, Completed: 739 [ 22m 47s ] +INFO: Idle: 1077, Running: 4, Completed: 743 [ 22m 53s ] +INFO: Idle: 1076, Running: 4, Completed: 744 [ 22m 57s ] +INFO: Idle: 1073, Running: 4, Completed: 747 [ 23m 2s ] +INFO: Idle: 1072, Running: 4, Completed: 748 [ 23m 6s ] +INFO: Idle: 1069, Running: 4, Completed: 751 [ 23m 12s ] +INFO: Idle: 1067, Running: 4, Completed: 753 [ 23m 17s ] +INFO: Idle: 1066, Running: 4, Completed: 754 [ 23m 21s ] +INFO: Idle: 1063, Running: 4, Completed: 757 [ 23m 26s ] +INFO: Idle: 1061, Running: 4, Completed: 759 [ 23m 34s ] +INFO: Idle: 1057, Running: 4, Completed: 763 [ 23m 42s ] +INFO: Idle: 1054, Running: 4, Completed: 766 [ 23m 46s ] +INFO: Idle: 1053, Running: 4, Completed: 767 [ 23m 50s ] +INFO: Idle: 1051, Running: 4, Completed: 769 [ 23m 54s ] +INFO: Idle: 1049, Running: 4, Completed: 771 [ 24m 2s ] +INFO: Idle: 1046, Running: 4, Completed: 774 [ 24m 6s ] +INFO: Idle: 1045, Running: 4, Completed: 775 [ 24m 12s ] +INFO: Idle: 1041, Running: 4, Completed: 779 [ 24m 21s ] +INFO: Idle: 1039, Running: 4, Completed: 781 [ 24m 24s ] +INFO: Idle: 1038, Running: 4, Completed: 782 [ 24m 28s ] +INFO: Idle: 1037, Running: 4, Completed: 783 [ 24m 32s ] +INFO: Idle: 1034, Running: 4, Completed: 786 [ 24m 37s ] +INFO: Idle: 1031, Running: 4, Completed: 789 [ 24m 46s ] +INFO: Idle: 1027, Running: 4, Completed: 793 [ 24m 55s ] +INFO: Idle: 1024, Running: 4, Completed: 796 [ 25m 1s ] +INFO: Idle: 1023, Running: 4, Completed: 797 [ 25m 5s ] +INFO: Idle: 1021, Running: 4, Completed: 799 [ 25m 9s ] +INFO: Idle: 1020, Running: 4, Completed: 800 [ 25m 15s ] +INFO: Idle: 1018, Running: 4, Completed: 802 [ 25m 18s ] +INFO: Idle: 1016, Running: 4, Completed: 804 [ 25m 23s ] +INFO: Idle: 1015, Running: 4, Completed: 805 [ 25m 28s ] +INFO: Idle: 1011, Running: 4, Completed: 809 [ 25m 37s ] +INFO: Idle: 1007, Running: 4, Completed: 813 [ 25m 47s ] +INFO: Idle: 1004, Running: 4, Completed: 816 [ 25m 50s ] +INFO: Idle: 1003, Running: 4, Completed: 817 [ 25m 57s ] +INFO: Idle: 999, Running: 4, Completed: 821 [ 26m 5s ] +INFO: Idle: 995, Running: 4, Completed: 825 [ 26m 14s ] +INFO: Idle: 991, Running: 4, Completed: 829 [ 26m 22s ] +INFO: Idle: 987, Running: 4, Completed: 833 [ 26m 32s ] +INFO: Idle: 985, Running: 4, Completed: 835 [ 26m 36s ] +INFO: Idle: 983, Running: 4, Completed: 837 [ 26m 43s ] +INFO: Idle: 980, Running: 4, Completed: 840 [ 26m 48s ] +INFO: Idle: 978, Running: 4, Completed: 842 [ 26m 52s ] +INFO: Idle: 977, Running: 4, Completed: 843 [ 26m 57s ] +INFO: Idle: 975, Running: 4, Completed: 845 [ 27m 2s ] +INFO: Idle: 973, Running: 4, Completed: 847 [ 27m 6s ] +INFO: Idle: 971, Running: 4, Completed: 849 [ 27m 13s ] +INFO: Idle: 967, Running: 4, Completed: 853 [ 27m 22s ] +INFO: Idle: 963, Running: 4, Completed: 857 [ 27m 31s ] +INFO: Idle: 960, Running: 4, Completed: 860 [ 27m 38s ] +INFO: Idle: 959, Running: 4, Completed: 861 [ 27m 42s ] +INFO: Idle: 957, Running: 4, Completed: 863 [ 27m 46s ] +INFO: Idle: 956, Running: 4, Completed: 864 [ 27m 51s ] +INFO: Idle: 953, Running: 4, Completed: 867 [ 27m 56s ] +INFO: Idle: 952, Running: 4, Completed: 868 [ 27m 59s ] +INFO: Idle: 951, Running: 4, Completed: 869 [ 28m 4s ] +INFO: Idle: 948, Running: 4, Completed: 872 [ 28m 7s ] +INFO: Idle: 947, Running: 4, Completed: 873 [ 28m 13s ] +INFO: Idle: 943, Running: 4, Completed: 877 [ 28m 23s ] +INFO: Idle: 939, Running: 4, Completed: 881 [ 28m 33s ] +INFO: Idle: 934, Running: 4, Completed: 886 [ 28m 42s ] +INFO: Idle: 931, Running: 4, Completed: 889 [ 28m 50s ] +INFO: Idle: 927, Running: 4, Completed: 893 [ 28m 59s ] +INFO: Idle: 923, Running: 4, Completed: 897 [ 29m 9s ] +INFO: Idle: 919, Running: 4, Completed: 901 [ 29m 19s ] +INFO: Idle: 916, Running: 4, Completed: 904 [ 29m 22s ] +INFO: Idle: 915, Running: 4, Completed: 905 [ 29m 30s ] +INFO: Idle: 911, Running: 4, Completed: 909 [ 29m 41s ] +INFO: Idle: 907, Running: 4, Completed: 913 [ 29m 50s ] +INFO: Idle: 903, Running: 4, Completed: 917 [ 29m 59s ] +INFO: Idle: 899, Running: 4, Completed: 921 [ 30m 9s ] +INFO: Idle: 895, Running: 4, Completed: 925 [ 30m 19s ] +INFO: Idle: 891, Running: 4, Completed: 929 [ 30m 29s ] +INFO: Idle: 888, Running: 4, Completed: 932 [ 30m 33s ] +INFO: Idle: 887, Running: 4, Completed: 933 [ 30m 42s ] +INFO: Idle: 883, Running: 4, Completed: 937 [ 30m 51s ] +INFO: Idle: 879, Running: 4, Completed: 941 [ 31m 0s ] +INFO: Idle: 875, Running: 4, Completed: 945 [ 31m 7s ] +INFO: Idle: 873, Running: 4, Completed: 947 [ 31m 11s ] +INFO: Idle: 871, Running: 4, Completed: 949 [ 31m 17s ] +INFO: Idle: 869, Running: 4, Completed: 951 [ 31m 23s ] +INFO: Idle: 867, Running: 4, Completed: 953 [ 31m 28s ] +INFO: Idle: 865, Running: 4, Completed: 955 [ 31m 32s ] +INFO: Idle: 863, Running: 4, Completed: 957 [ 31m 39s ] +INFO: Idle: 860, Running: 4, Completed: 960 [ 31m 42s ] +INFO: Idle: 859, Running: 4, Completed: 961 [ 31m 49s ] +INFO: Idle: 855, Running: 4, Completed: 965 [ 31m 57s ] +INFO: Idle: 852, Running: 4, Completed: 968 [ 32m 0s ] +INFO: Idle: 851, Running: 4, Completed: 969 [ 32m 5s ] +INFO: Idle: 849, Running: 4, Completed: 971 [ 32m 11s ] +INFO: Idle: 847, Running: 4, Completed: 973 [ 32m 14s ] +INFO: Idle: 845, Running: 4, Completed: 975 [ 32m 22s ] +INFO: Idle: 842, Running: 4, Completed: 978 [ 32m 26s ] +INFO: Idle: 841, Running: 4, Completed: 979 [ 32m 30s ] +INFO: Idle: 839, Running: 4, Completed: 981 [ 32m 36s ] +INFO: Idle: 837, Running: 4, Completed: 983 [ 32m 40s ] +INFO: Idle: 835, Running: 4, Completed: 985 [ 32m 47s ] +INFO: Idle: 831, Running: 4, Completed: 989 [ 32m 56s ] +INFO: Idle: 827, Running: 4, Completed: 993 [ 33m 5s ] +INFO: Idle: 825, Running: 4, Completed: 995 [ 33m 9s ] +INFO: Idle: 823, Running: 4, Completed: 997 [ 33m 17s ] +INFO: Idle: 819, Running: 4, Completed: 1001 [ 33m 24s ] +INFO: Idle: 816, Running: 4, Completed: 1004 [ 33m 28s ] +INFO: Idle: 815, Running: 4, Completed: 1005 [ 33m 32s ] +INFO: Idle: 812, Running: 4, Completed: 1008 [ 33m 35s ] +INFO: Idle: 811, Running: 4, Completed: 1009 [ 33m 41s ] +INFO: Idle: 809, Running: 4, Completed: 1011 [ 33m 45s ] +INFO: Idle: 808, Running: 4, Completed: 1012 [ 33m 49s ] +INFO: Idle: 807, Running: 4, Completed: 1013 [ 33m 54s ] +INFO: Idle: 803, Running: 4, Completed: 1017 [ 34m 2s ] +INFO: Idle: 800, Running: 4, Completed: 1020 [ 34m 5s ] +INFO: Idle: 799, Running: 4, Completed: 1021 [ 34m 10s ] +INFO: Idle: 795, Running: 4, Completed: 1025 [ 34m 19s ] +INFO: Idle: 793, Running: 4, Completed: 1027 [ 34m 23s ] +INFO: Idle: 791, Running: 4, Completed: 1029 [ 34m 28s ] +INFO: Idle: 789, Running: 4, Completed: 1031 [ 34m 35s ] +INFO: Idle: 787, Running: 4, Completed: 1033 [ 34m 39s ] +INFO: Idle: 785, Running: 4, Completed: 1035 [ 34m 43s ] +INFO: Idle: 783, Running: 4, Completed: 1037 [ 34m 51s ] +INFO: Idle: 779, Running: 4, Completed: 1041 [ 35m 1s ] +INFO: Idle: 775, Running: 4, Completed: 1045 [ 35m 8s ] +INFO: Idle: 771, Running: 4, Completed: 1049 [ 35m 17s ] +INFO: Idle: 769, Running: 4, Completed: 1051 [ 35m 21s ] +INFO: Idle: 768, Running: 4, Completed: 1052 [ 35m 24s ] +INFO: Idle: 765, Running: 4, Completed: 1055 [ 35m 33s ] +INFO: Idle: 762, Running: 4, Completed: 1058 [ 35m 40s ] +INFO: Idle: 760, Running: 4, Completed: 1060 [ 35m 44s ] +INFO: Idle: 759, Running: 4, Completed: 1061 [ 35m 48s ] +INFO: Idle: 757, Running: 4, Completed: 1063 [ 35m 52s ] +INFO: Idle: 755, Running: 4, Completed: 1065 [ 35m 58s ] +INFO: Idle: 751, Running: 4, Completed: 1069 [ 36m 7s ] +INFO: Idle: 747, Running: 4, Completed: 1073 [ 36m 14s ] +INFO: Idle: 745, Running: 4, Completed: 1075 [ 36m 19s ] +INFO: Idle: 744, Running: 4, Completed: 1076 [ 36m 22s ] +INFO: Idle: 741, Running: 4, Completed: 1079 [ 36m 31s ] +INFO: Idle: 739, Running: 4, Completed: 1081 [ 36m 35s ] +INFO: Idle: 737, Running: 4, Completed: 1083 [ 36m 40s ] +INFO: Idle: 735, Running: 4, Completed: 1085 [ 36m 46s ] +INFO: Idle: 733, Running: 4, Completed: 1087 [ 36m 49s ] +INFO: Idle: 731, Running: 4, Completed: 1089 [ 36m 56s ] +INFO: Idle: 727, Running: 4, Completed: 1093 [ 37m 5s ] +INFO: Idle: 723, Running: 4, Completed: 1097 [ 37m 13s ] +INFO: Idle: 721, Running: 4, Completed: 1099 [ 37m 17s ] +INFO: Idle: 720, Running: 4, Completed: 1100 [ 37m 20s ] +INFO: Idle: 717, Running: 4, Completed: 1103 [ 37m 29s ] +INFO: Idle: 715, Running: 4, Completed: 1105 [ 37m 33s ] +INFO: Idle: 713, Running: 4, Completed: 1107 [ 37m 37s ] +INFO: Idle: 711, Running: 4, Completed: 1109 [ 37m 44s ] +INFO: Idle: 708, Running: 4, Completed: 1112 [ 37m 48s ] +INFO: Idle: 707, Running: 4, Completed: 1113 [ 37m 54s ] +INFO: Idle: 703, Running: 4, Completed: 1117 [ 38m 3s ] +INFO: Idle: 699, Running: 4, Completed: 1121 [ 38m 12s ] +INFO: Idle: 697, Running: 4, Completed: 1123 [ 38m 15s ] +INFO: Idle: 695, Running: 4, Completed: 1125 [ 38m 21s ] +INFO: Idle: 693, Running: 4, Completed: 1127 [ 38m 27s ] +INFO: Idle: 691, Running: 4, Completed: 1129 [ 38m 32s ] +INFO: Idle: 689, Running: 4, Completed: 1131 [ 38m 36s ] +INFO: Idle: 687, Running: 4, Completed: 1133 [ 38m 43s ] +INFO: Idle: 684, Running: 4, Completed: 1136 [ 38m 47s ] +INFO: Idle: 683, Running: 4, Completed: 1137 [ 38m 53s ] +INFO: Idle: 679, Running: 4, Completed: 1141 [ 39m 2s ] +INFO: Idle: 675, Running: 4, Completed: 1145 [ 39m 9s ] +INFO: Idle: 673, Running: 4, Completed: 1147 [ 39m 12s ] +INFO: Idle: 671, Running: 4, Completed: 1149 [ 39m 22s ] +INFO: Idle: 668, Running: 4, Completed: 1152 [ 39m 26s ] +INFO: Idle: 667, Running: 4, Completed: 1153 [ 39m 31s ] +INFO: Idle: 664, Running: 4, Completed: 1156 [ 39m 36s ] +INFO: Idle: 662, Running: 4, Completed: 1158 [ 39m 40s ] +INFO: Idle: 659, Running: 4, Completed: 1161 [ 39m 46s ] +INFO: Idle: 658, Running: 4, Completed: 1162 [ 39m 51s ] +INFO: Idle: 655, Running: 4, Completed: 1165 [ 40m 0s ] +INFO: Idle: 651, Running: 4, Completed: 1169 [ 40m 7s ] +INFO: Idle: 648, Running: 4, Completed: 1172 [ 40m 11s ] +INFO: Idle: 647, Running: 4, Completed: 1173 [ 40m 16s ] +INFO: Idle: 643, Running: 4, Completed: 1177 [ 40m 25s ] +INFO: Idle: 641, Running: 4, Completed: 1179 [ 40m 29s ] +INFO: Idle: 639, Running: 4, Completed: 1181 [ 40m 34s ] +INFO: Idle: 637, Running: 4, Completed: 1183 [ 40m 41s ] +INFO: Idle: 635, Running: 4, Completed: 1185 [ 40m 45s ] +INFO: Idle: 633, Running: 4, Completed: 1187 [ 40m 49s ] +INFO: Idle: 631, Running: 4, Completed: 1189 [ 40m 56s ] +INFO: Idle: 629, Running: 4, Completed: 1191 [ 41m 0s ] +INFO: Idle: 627, Running: 4, Completed: 1193 [ 41m 6s ] +INFO: Idle: 623, Running: 4, Completed: 1197 [ 41m 14s ] +INFO: Idle: 619, Running: 4, Completed: 1201 [ 41m 22s ] +INFO: Idle: 617, Running: 4, Completed: 1203 [ 41m 27s ] +INFO: Idle: 615, Running: 4, Completed: 1205 [ 41m 32s ] +INFO: Idle: 613, Running: 4, Completed: 1207 [ 41m 39s ] +INFO: Idle: 611, Running: 4, Completed: 1209 [ 41m 43s ] +INFO: Idle: 609, Running: 4, Completed: 1211 [ 41m 48s ] +INFO: Idle: 607, Running: 4, Completed: 1213 [ 41m 55s ] +INFO: Idle: 604, Running: 4, Completed: 1216 [ 41m 59s ] +INFO: Idle: 603, Running: 4, Completed: 1217 [ 42m 4s ] +INFO: Idle: 599, Running: 4, Completed: 1221 [ 42m 13s ] +INFO: Idle: 595, Running: 4, Completed: 1225 [ 42m 21s ] +INFO: Idle: 593, Running: 4, Completed: 1227 [ 42m 24s ] +INFO: Idle: 591, Running: 4, Completed: 1229 [ 42m 29s ] +INFO: Idle: 589, Running: 4, Completed: 1231 [ 42m 36s ] +INFO: Idle: 587, Running: 4, Completed: 1233 [ 42m 42s ] +INFO: Idle: 585, Running: 4, Completed: 1235 [ 42m 46s ] +INFO: Idle: 583, Running: 4, Completed: 1237 [ 42m 53s ] +INFO: Idle: 580, Running: 4, Completed: 1240 [ 42m 56s ] +INFO: Idle: 579, Running: 4, Completed: 1241 [ 43m 2s ] +INFO: Idle: 575, Running: 4, Completed: 1245 [ 43m 11s ] +INFO: Idle: 571, Running: 4, Completed: 1249 [ 43m 19s ] +INFO: Idle: 569, Running: 4, Completed: 1251 [ 43m 23s ] +INFO: Idle: 567, Running: 4, Completed: 1253 [ 43m 28s ] +INFO: Idle: 565, Running: 4, Completed: 1255 [ 43m 35s ] +INFO: Idle: 563, Running: 4, Completed: 1257 [ 43m 39s ] +INFO: Idle: 561, Running: 4, Completed: 1259 [ 43m 44s ] +INFO: Idle: 559, Running: 4, Completed: 1261 [ 43m 51s ] +INFO: Idle: 555, Running: 4, Completed: 1265 [ 44m 1s ] +INFO: Idle: 551, Running: 4, Completed: 1269 [ 44m 9s ] +INFO: Idle: 547, Running: 4, Completed: 1273 [ 44m 17s ] +INFO: Idle: 545, Running: 4, Completed: 1275 [ 44m 21s ] +INFO: Idle: 543, Running: 4, Completed: 1277 [ 44m 26s ] +INFO: Idle: 541, Running: 4, Completed: 1279 [ 44m 33s ] +INFO: Idle: 539, Running: 4, Completed: 1281 [ 44m 37s ] +INFO: Idle: 537, Running: 4, Completed: 1283 [ 44m 42s ] +INFO: Idle: 535, Running: 4, Completed: 1285 [ 44m 49s ] +INFO: Idle: 533, Running: 4, Completed: 1287 [ 44m 52s ] +INFO: Idle: 531, Running: 4, Completed: 1289 [ 44m 58s ] +INFO: Idle: 527, Running: 4, Completed: 1293 [ 45m 7s ] +INFO: Idle: 523, Running: 4, Completed: 1297 [ 45m 14s ] +INFO: Idle: 521, Running: 4, Completed: 1299 [ 45m 20s ] +INFO: Idle: 519, Running: 4, Completed: 1301 [ 45m 26s ] +INFO: Idle: 516, Running: 4, Completed: 1304 [ 45m 32s ] +INFO: Idle: 513, Running: 4, Completed: 1307 [ 45m 41s ] +INFO: Idle: 511, Running: 4, Completed: 1309 [ 45m 45s ] +INFO: Idle: 509, Running: 4, Completed: 1311 [ 45m 49s ] +INFO: Idle: 507, Running: 4, Completed: 1313 [ 45m 55s ] +INFO: Idle: 504, Running: 4, Completed: 1316 [ 45m 58s ] +INFO: Idle: 503, Running: 4, Completed: 1317 [ 46m 4s ] +INFO: Idle: 499, Running: 4, Completed: 1321 [ 46m 13s ] +INFO: Idle: 496, Running: 4, Completed: 1324 [ 46m 19s ] +INFO: Idle: 495, Running: 4, Completed: 1325 [ 46m 23s ] +INFO: Idle: 493, Running: 4, Completed: 1327 [ 46m 27s ] +INFO: Idle: 492, Running: 4, Completed: 1328 [ 46m 32s ] +INFO: Idle: 489, Running: 4, Completed: 1331 [ 46m 36s ] +INFO: Idle: 488, Running: 4, Completed: 1332 [ 46m 40s ] +INFO: Idle: 487, Running: 4, Completed: 1333 [ 46m 45s ] +INFO: Idle: 483, Running: 4, Completed: 1337 [ 46m 54s ] +INFO: Idle: 479, Running: 4, Completed: 1341 [ 47m 4s ] +INFO: Idle: 475, Running: 4, Completed: 1345 [ 47m 12s ] +INFO: Idle: 471, Running: 4, Completed: 1349 [ 47m 21s ] +INFO: Idle: 467, Running: 4, Completed: 1353 [ 47m 30s ] +INFO: Idle: 463, Running: 4, Completed: 1357 [ 47m 38s ] +INFO: Idle: 459, Running: 4, Completed: 1361 [ 47m 47s ] +INFO: Idle: 457, Running: 4, Completed: 1363 [ 47m 51s ] +INFO: Idle: 455, Running: 4, Completed: 1365 [ 47m 58s ] +INFO: Idle: 452, Running: 4, Completed: 1368 [ 48m 3s ] +INFO: Idle: 450, Running: 4, Completed: 1370 [ 48m 6s ] +INFO: Idle: 449, Running: 4, Completed: 1371 [ 48m 12s ] +INFO: Idle: 447, Running: 4, Completed: 1373 [ 48m 17s ] +INFO: Idle: 445, Running: 4, Completed: 1375 [ 48m 20s ] +INFO: Idle: 443, Running: 4, Completed: 1377 [ 48m 26s ] +INFO: Idle: 439, Running: 4, Completed: 1381 [ 48m 37s ] +INFO: Idle: 435, Running: 4, Completed: 1385 [ 48m 45s ] +INFO: Idle: 432, Running: 4, Completed: 1388 [ 48m 51s ] +INFO: Idle: 431, Running: 4, Completed: 1389 [ 48m 55s ] +INFO: Idle: 429, Running: 4, Completed: 1391 [ 48m 59s ] +INFO: Idle: 428, Running: 4, Completed: 1392 [ 49m 4s ] +INFO: Idle: 425, Running: 4, Completed: 1395 [ 49m 9s ] +INFO: Idle: 424, Running: 4, Completed: 1396 [ 49m 13s ] +INFO: Idle: 423, Running: 4, Completed: 1397 [ 49m 17s ] +INFO: Idle: 419, Running: 4, Completed: 1401 [ 49m 26s ] +INFO: Idle: 415, Running: 4, Completed: 1405 [ 49m 36s ] +INFO: Idle: 411, Running: 4, Completed: 1409 [ 49m 45s ] +INFO: Idle: 407, Running: 4, Completed: 1413 [ 49m 54s ] +INFO: Idle: 403, Running: 4, Completed: 1417 [ 50m 3s ] +INFO: Idle: 399, Running: 4, Completed: 1421 [ 50m 11s ] +INFO: Idle: 395, Running: 4, Completed: 1425 [ 50m 20s ] +INFO: Idle: 391, Running: 4, Completed: 1429 [ 50m 31s ] +INFO: Idle: 387, Running: 4, Completed: 1433 [ 50m 40s ] +INFO: Idle: 384, Running: 4, Completed: 1436 [ 50m 44s ] +INFO: Idle: 383, Running: 4, Completed: 1437 [ 50m 52s ] +INFO: Idle: 379, Running: 4, Completed: 1441 [ 51m 0s ] +INFO: Idle: 375, Running: 4, Completed: 1445 [ 51m 9s ] +INFO: Idle: 371, Running: 4, Completed: 1449 [ 51m 18s ] +INFO: Idle: 367, Running: 4, Completed: 1453 [ 51m 28s ] +INFO: Idle: 363, Running: 4, Completed: 1457 [ 51m 37s ] +INFO: Idle: 360, Running: 4, Completed: 1460 [ 51m 41s ] +INFO: Idle: 359, Running: 4, Completed: 1461 [ 51m 49s ] +INFO: Idle: 355, Running: 4, Completed: 1465 [ 51m 58s ] +INFO: Idle: 351, Running: 4, Completed: 1469 [ 52m 7s ] +INFO: Idle: 347, Running: 4, Completed: 1473 [ 52m 14s ] +INFO: Idle: 345, Running: 4, Completed: 1475 [ 52m 18s ] +INFO: Idle: 343, Running: 4, Completed: 1477 [ 52m 23s ] +INFO: Idle: 341, Running: 4, Completed: 1479 [ 52m 30s ] +INFO: Idle: 339, Running: 4, Completed: 1481 [ 52m 34s ] +INFO: Idle: 337, Running: 4, Completed: 1483 [ 52m 38s ] +INFO: Idle: 335, Running: 4, Completed: 1485 [ 52m 45s ] +INFO: Idle: 332, Running: 4, Completed: 1488 [ 52m 48s ] +INFO: Idle: 331, Running: 4, Completed: 1489 [ 52m 55s ] +INFO: Idle: 327, Running: 4, Completed: 1493 [ 53m 3s ] +INFO: Idle: 323, Running: 4, Completed: 1497 [ 53m 10s ] +INFO: Idle: 321, Running: 4, Completed: 1499 [ 53m 14s ] +INFO: Idle: 319, Running: 4, Completed: 1501 [ 53m 19s ] +INFO: Idle: 317, Running: 4, Completed: 1503 [ 53m 26s ] +INFO: Idle: 315, Running: 4, Completed: 1505 [ 53m 29s ] +INFO: Idle: 313, Running: 4, Completed: 1507 [ 53m 34s ] +INFO: Idle: 311, Running: 4, Completed: 1509 [ 53m 41s ] +INFO: Idle: 308, Running: 4, Completed: 1512 [ 53m 45s ] +INFO: Idle: 307, Running: 4, Completed: 1513 [ 53m 50s ] +INFO: Idle: 303, Running: 4, Completed: 1517 [ 53m 59s ] +INFO: Idle: 299, Running: 4, Completed: 1521 [ 54m 7s ] +INFO: Idle: 297, Running: 4, Completed: 1523 [ 54m 11s ] +INFO: Idle: 295, Running: 4, Completed: 1525 [ 54m 19s ] +INFO: Idle: 291, Running: 4, Completed: 1529 [ 54m 26s ] +INFO: Idle: 288, Running: 4, Completed: 1532 [ 54m 30s ] +INFO: Idle: 287, Running: 4, Completed: 1533 [ 54m 34s ] +INFO: Idle: 283, Running: 4, Completed: 1537 [ 54m 43s ] +INFO: Idle: 281, Running: 4, Completed: 1539 [ 54m 47s ] +INFO: Idle: 279, Running: 4, Completed: 1541 [ 54m 56s ] +INFO: Idle: 275, Running: 4, Completed: 1545 [ 55m 4s ] +INFO: Idle: 271, Running: 4, Completed: 1549 [ 55m 11s ] +INFO: Idle: 267, Running: 4, Completed: 1553 [ 55m 22s ] +INFO: Idle: 265, Running: 4, Completed: 1555 [ 55m 26s ] +INFO: Idle: 263, Running: 4, Completed: 1557 [ 55m 33s ] +INFO: Idle: 261, Running: 4, Completed: 1559 [ 55m 39s ] +INFO: Idle: 259, Running: 4, Completed: 1561 [ 55m 45s ] +INFO: Idle: 256, Running: 4, Completed: 1564 [ 55m 49s ] +INFO: Idle: 255, Running: 4, Completed: 1565 [ 55m 56s ] +INFO: Idle: 251, Running: 4, Completed: 1569 [ 56m 4s ] +INFO: Idle: 247, Running: 4, Completed: 1573 [ 56m 13s ] +INFO: Idle: 243, Running: 4, Completed: 1577 [ 56m 21s ] +INFO: Idle: 241, Running: 4, Completed: 1579 [ 56m 24s ] +INFO: Idle: 239, Running: 4, Completed: 1581 [ 56m 29s ] +INFO: Idle: 237, Running: 4, Completed: 1583 [ 56m 36s ] +INFO: Idle: 235, Running: 4, Completed: 1585 [ 56m 41s ] +INFO: Idle: 232, Running: 4, Completed: 1588 [ 56m 46s ] +INFO: Idle: 231, Running: 4, Completed: 1589 [ 56m 52s ] +INFO: Idle: 227, Running: 4, Completed: 1593 [ 57m 1s ] +INFO: Idle: 223, Running: 4, Completed: 1597 [ 57m 9s ] +INFO: Idle: 219, Running: 4, Completed: 1601 [ 57m 17s ] +INFO: Idle: 217, Running: 4, Completed: 1603 [ 57m 21s ] +INFO: Idle: 215, Running: 4, Completed: 1605 [ 57m 26s ] +INFO: Idle: 213, Running: 4, Completed: 1607 [ 57m 33s ] +INFO: Idle: 211, Running: 4, Completed: 1609 [ 57m 36s ] +INFO: Idle: 209, Running: 4, Completed: 1611 [ 57m 41s ] +INFO: Idle: 207, Running: 4, Completed: 1613 [ 57m 48s ] +INFO: Idle: 204, Running: 4, Completed: 1616 [ 57m 51s ] +INFO: Idle: 203, Running: 4, Completed: 1617 [ 57m 58s ] +INFO: Idle: 199, Running: 4, Completed: 1621 [ 58m 6s ] +INFO: Idle: 195, Running: 4, Completed: 1625 [ 58m 13s ] +INFO: Idle: 193, Running: 4, Completed: 1627 [ 58m 18s ] +INFO: Idle: 191, Running: 4, Completed: 1629 [ 58m 23s ] +INFO: Idle: 189, Running: 4, Completed: 1631 [ 58m 30s ] +INFO: Idle: 186, Running: 4, Completed: 1634 [ 58m 34s ] +INFO: Idle: 185, Running: 4, Completed: 1635 [ 58m 38s ] +INFO: Idle: 183, Running: 4, Completed: 1637 [ 58m 44s ] +INFO: Idle: 181, Running: 4, Completed: 1639 [ 58m 48s ] +INFO: Idle: 179, Running: 4, Completed: 1641 [ 58m 53s ] +INFO: Idle: 175, Running: 4, Completed: 1645 [ 59m 3s ] +INFO: Idle: 171, Running: 4, Completed: 1649 [ 59m 10s ] +INFO: Idle: 169, Running: 4, Completed: 1651 [ 59m 14s ] +INFO: Idle: 167, Running: 4, Completed: 1653 [ 59m 20s ] +INFO: Idle: 165, Running: 4, Completed: 1655 [ 59m 25s ] +INFO: Idle: 163, Running: 4, Completed: 1657 [ 59m 30s ] +INFO: Idle: 161, Running: 4, Completed: 1659 [ 59m 34s ] +INFO: Idle: 159, Running: 4, Completed: 1661 [ 59m 41s ] +INFO: Idle: 155, Running: 4, Completed: 1665 [ 59m 51s ] +INFO: Idle: 151, Running: 4, Completed: 1669 [ 59m 59s ] +INFO: Idle: 147, Running: 4, Completed: 1673 [ 1h 0m ] +INFO: Idle: 143, Running: 4, Completed: 1677 [ 1h 0m ] +INFO: Idle: 140, Running: 4, Completed: 1680 [ 1h 0m ] +INFO: Idle: 139, Running: 4, Completed: 1681 [ 1h 0m ] +INFO: Idle: 136, Running: 4, Completed: 1684 [ 1h 0m ] +INFO: Idle: 134, Running: 4, Completed: 1686 [ 1h 0m ] +INFO: Idle: 131, Running: 4, Completed: 1689 [ 1h 0m ] +INFO: Idle: 130, Running: 4, Completed: 1690 [ 1h 0m ] +INFO: Idle: 128, Running: 4, Completed: 1692 [ 1h 0m ] +INFO: Idle: 127, Running: 4, Completed: 1693 [ 1h 0m ] +INFO: Idle: 123, Running: 4, Completed: 1697 [ 1h 1m ] +INFO: Idle: 119, Running: 4, Completed: 1701 [ 1h 1m ] +INFO: Idle: 115, Running: 4, Completed: 1705 [ 1h 1m ] +INFO: Idle: 113, Running: 4, Completed: 1707 [ 1h 1m ] +INFO: Idle: 111, Running: 4, Completed: 1709 [ 1h 1m ] +INFO: Idle: 109, Running: 4, Completed: 1711 [ 1h 1m ] +INFO: Idle: 107, Running: 4, Completed: 1713 [ 1h 1m ] +INFO: Idle: 105, Running: 4, Completed: 1715 [ 1h 1m ] +INFO: Idle: 103, Running: 4, Completed: 1717 [ 1h 1m ] +INFO: Idle: 99, Running: 4, Completed: 1721 [ 1h 1m ] +INFO: Idle: 95, Running: 4, Completed: 1725 [ 1h 2m ] +INFO: Idle: 91, Running: 4, Completed: 1729 [ 1h 2m ] +INFO: Idle: 89, Running: 4, Completed: 1731 [ 1h 2m ] +INFO: Idle: 87, Running: 4, Completed: 1733 [ 1h 2m ] +INFO: Idle: 85, Running: 4, Completed: 1735 [ 1h 2m ] +INFO: Idle: 83, Running: 4, Completed: 1737 [ 1h 2m ] +INFO: Idle: 81, Running: 4, Completed: 1739 [ 1h 2m ] +INFO: Idle: 79, Running: 4, Completed: 1741 [ 1h 2m ] +INFO: Idle: 76, Running: 4, Completed: 1744 [ 1h 2m ] +INFO: Idle: 75, Running: 4, Completed: 1745 [ 1h 2m ] +INFO: Idle: 71, Running: 4, Completed: 1749 [ 1h 3m ] +INFO: Idle: 67, Running: 4, Completed: 1753 [ 1h 3m ] +INFO: Idle: 64, Running: 4, Completed: 1756 [ 1h 3m ] +INFO: Idle: 63, Running: 4, Completed: 1757 [ 1h 3m ] +INFO: Idle: 61, Running: 4, Completed: 1759 [ 1h 3m ] +INFO: Idle: 59, Running: 4, Completed: 1761 [ 1h 3m ] +INFO: Idle: 57, Running: 4, Completed: 1763 [ 1h 3m ] +INFO: Idle: 55, Running: 4, Completed: 1765 [ 1h 3m ] +INFO: Idle: 51, Running: 4, Completed: 1769 [ 1h 3m ] +INFO: Idle: 47, Running: 4, Completed: 1773 [ 1h 4m ] +INFO: Idle: 43, Running: 4, Completed: 1777 [ 1h 4m ] +INFO: Idle: 40, Running: 4, Completed: 1780 [ 1h 4m ] +INFO: Idle: 39, Running: 4, Completed: 1781 [ 1h 4m ] +INFO: Idle: 37, Running: 4, Completed: 1783 [ 1h 4m ] +INFO: Idle: 35, Running: 4, Completed: 1785 [ 1h 4m ] +INFO: Idle: 33, Running: 4, Completed: 1787 [ 1h 4m ] +INFO: Idle: 31, Running: 4, Completed: 1789 [ 1h 4m ] +INFO: Idle: 28, Running: 4, Completed: 1792 [ 1h 4m ] +INFO: Idle: 27, Running: 4, Completed: 1793 [ 1h 4m ] +INFO: Idle: 23, Running: 4, Completed: 1797 [ 1h 4m ] +INFO: Idle: 19, Running: 4, Completed: 1801 [ 1h 5m ] +INFO: Idle: 17, Running: 4, Completed: 1803 [ 1h 5m ] +INFO: Idle: 15, Running: 4, Completed: 1805 [ 1h 5m ] +INFO: Idle: 13, Running: 4, Completed: 1807 [ 1h 5m ] +INFO: Idle: 11, Running: 4, Completed: 1809 [ 1h 5m ] +INFO: Idle: 9, Running: 4, Completed: 1811 [ 1h 5m ] +INFO: Idle: 7, Running: 4, Completed: 1813 [ 1h 5m ] +INFO: Idle: 3, Running: 4, Completed: 1817 [ 1h 5m ] +INFO: Idle: 0, Running: 3, Completed: 1821 [ 1h 5m ] +INFO: Idle: 0, Running: 0, Completed: 1824 [ 1h 5m ] +sum of cpu time of last step: 5h02m29s + === Results Summary for run: run_01 tag: tag_1 === + + Cross-section : 1.066e+04 +- 31.07 pb + Nb of events : 0 + +INFO: End survey +combine_events +INFO: Combining Events +combination of events done in 0.6587498188018799 s +store_events +INFO: Storing parton level results +INFO: End Parton +decay_events -from_cards +create_gridpack +INFO: Creating gridpack +P0_gg_epemguux +P0_gg_epemgddx +P0_gg_taptamguux +P0_gg_taptamgddx +P0_gu_epemggu +P0_gd_epemggd +P0_gux_epemggux +P0_gdx_epemggdx +P0_gu_taptamggu +P0_gd_taptamggd +P0_gux_taptamggux +P0_gdx_taptamggdx +P0_uux_epemggg +P0_ddx_epemggg +P0_uux_taptamggg +P0_ddx_taptamggg +P0_gu_epemuuux +P0_gd_epemdddx +P0_gux_epemuuxux +P0_gdx_epemddxdx +P0_gu_epemuddx +P0_gu_epemuccx +P0_gd_epemudux +P0_gd_epemdssx +P0_gux_epemduxdx +P0_gux_epemcuxcx +P0_gdx_epemuuxdx +P0_gdx_epemsdxsx +P0_gu_taptamuuux +P0_gd_taptamdddx +P0_gux_taptamuuxux +P0_gdx_taptamddxdx +P0_gu_taptamuddx +P0_gu_taptamuccx +P0_gd_taptamudux +P0_gd_taptamdssx +P0_gux_taptamduxdx +P0_gux_taptamcuxcx +P0_gdx_taptamuuxdx +P0_gdx_taptamsdxsx +P0_uu_epemguu +P0_uux_epemguux +P0_dd_epemgdd +P0_ddx_epemgddx +P0_uxux_epemguxux +P0_dxdx_epemgdxdx +P0_ud_epemgud +P0_uc_epemguc +P0_uux_epemgddx +P0_uux_epemgccx +P0_udx_epemgudx +P0_ucx_epemgucx +P0_dc_epemgdc +P0_ds_epemgds +P0_dux_epemgdux +P0_ddx_epemguux +P0_ddx_epemgssx +P0_dsx_epemgdsx +P0_uxdx_epemguxdx +P0_uxcx_epemguxcx +P0_dxcx_epemgdxcx +P0_dxsx_epemgdxsx +P0_uu_taptamguu +P0_uux_taptamguux +P0_dd_taptamgdd +P0_ddx_taptamgddx +P0_uxux_taptamguxux +P0_dxdx_taptamgdxdx +P0_ud_taptamgud +P0_uc_taptamguc +P0_uux_taptamgddx +P0_uux_taptamgccx +P0_udx_taptamgudx +P0_ucx_taptamgucx +P0_dc_taptamgdc +P0_ds_taptamgds +P0_dux_taptamgdux +P0_ddx_taptamguux +P0_ddx_taptamgssx +P0_dsx_taptamgdsx +P0_uxdx_taptamguxdx +P0_uxcx_taptamguxcx +P0_dxcx_taptamgdxcx +P0_dxsx_taptamgdxsx +Cleaning SubProcesses.................................................................................... +INFO: gridpack created +quit +INFO: storing files of previous run +INFO: Done +INFO: +17383.51user 624.99system 1:22:24elapsed 364%CPU (0avgtext+0avgdata 260396maxresident)k +3598240inputs+32323960outputs (696major+57456442minor)pagefaults 0swaps +END: Tue Sep 10 11:58:43 PM CEST 2024 +ELAPSED: 4946 seconds diff --git a/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/output.txt b/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/output.txt new file mode 100644 index 0000000000..44a7dd34b5 --- /dev/null +++ b/epochX/cudacpp/tlau/gridpacks/pp_dy4j.mad/output.txt @@ -0,0 +1,10781 @@ +START: Fri Sep 13 01:06:44 AM CEST 2024 +On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU (nproc=4)] [GPU: 1x Tesla V100S-PCIE-32GB]: +CUDACPP_RUNTIME_DISABLEFPE is set +************************************************************ +* * +* W E L C O M E to * +* M A D G R A P H 5 _ a M C @ N L O * +* M A D E V E N T * +* * +* * * * +* * * * * * +* * * * * 5 * * * * * +* * * * * * +* * * * +* * +* VERSION 3.5.3_lo_vect * +* * +* The MadGraph5_aMC@NLO Development Team - Find us at * +* https://server06.fynu.ucl.ac.be/projects/madgraph * +* * +* Type 'help' for in-line help. * +* * +************************************************************ +INFO: load configuration from /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/pp_dy4j.mad/Cards/me5_configuration.txt +INFO: load configuration from /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/pp_dy4j.mad/Cards/me5_configuration.txt +Using default text editor "vi". Set another one in ./input/mg5_configuration.txt +Using default eps viewer "evince". Set another one in ./input/mg5_configuration.txt +Using default web browser "firefox". Set another one in ./input/mg5_configuration.txt +generate_events -f +Generating gridpack with run name run_01 +survey run_01 --accuracy=0.01 --points=2000 --iterations=8 --gridpack=.true. +INFO: compile directory +compile Source Directory +Using random number seed offset = 21 +INFO: Running Survey +Creating Jobs +Working on SubProcesses +INFO: P0_gg_epemgguux +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_gg_epemggddx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_gg_taptamgguux +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_gg_taptamggddx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_gu_epemgggu +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_gd_epemgggd +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_gux_epemgggux +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_gdx_epemgggdx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_gu_taptamgggu +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_gd_taptamgggd +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_gux_taptamgggux +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_gdx_taptamgggdx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_uux_epemgggg +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_ddx_epemgggg +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_uux_taptamgggg +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_ddx_taptamgggg +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_gg_epemuuuxux +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_gg_epemdddxdx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_gg_epemuduxdx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_gg_epemucuxcx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_gg_epemdsdxsx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_gg_taptamuuuxux +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_gg_taptamdddxdx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_gg_taptamuduxdx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_gg_taptamucuxcx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_gg_taptamdsdxsx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_gu_epemguuux +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_gd_epemgdddx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_gux_epemguuxux +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_gdx_epemgddxdx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_gu_epemguddx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_gu_epemguccx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_gd_epemgudux +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_gd_epemgdssx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_gux_epemgduxdx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_gux_epemgcuxcx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_gdx_epemguuxdx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_gdx_epemgsdxsx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_gu_taptamguuux +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_gd_taptamgdddx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_gux_taptamguuxux +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_gdx_taptamgddxdx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_gu_taptamguddx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_gu_taptamguccx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_gd_taptamgudux +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_gd_taptamgdssx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_gux_taptamgduxdx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_gux_taptamgcuxcx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_gdx_taptamguuxdx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_gdx_taptamgsdxsx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_uu_epemgguu +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_uux_epemgguux +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_dd_epemggdd +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_ddx_epemggddx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_uxux_epemgguxux +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_dxdx_epemggdxdx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_ud_epemggud +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_uc_epemgguc +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_uux_epemggddx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_uux_epemggccx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_udx_epemggudx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_ucx_epemggucx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_dc_epemggdc +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_ds_epemggds +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_dux_epemggdux +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_ddx_epemgguux +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_ddx_epemggssx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_dsx_epemggdsx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_uxdx_epemgguxdx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_uxcx_epemgguxcx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_dxcx_epemggdxcx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_dxsx_epemggdxsx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_uu_taptamgguu +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_uux_taptamgguux +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_dd_taptamggdd +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_ddx_taptamggddx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_uxux_taptamgguxux +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_dxdx_taptamggdxdx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_ud_taptamggud +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_uc_taptamgguc +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_uux_taptamggddx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_uux_taptamggccx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_udx_taptamggudx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_ucx_taptamggucx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_dc_taptamggdc +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_ds_taptamggds +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_dux_taptamggdux +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_ddx_taptamgguux +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_ddx_taptamggssx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_dsx_taptamggdsx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_uxdx_taptamgguxdx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_uxcx_taptamgguxcx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_dxcx_taptamggdxcx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_dxsx_taptamggdxsx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_uu_epemuuuux +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_uux_epemuuuxux +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_dd_epemddddx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_ddx_epemdddxdx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_uxux_epemuuxuxux +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_dxdx_epemddxdxdx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_uu_epemuuddx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_uu_epemuuccx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_ud_epemuudux +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_ud_epemudddx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_uc_epemuucux +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_uc_epemucccx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_uux_epemuduxdx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_uux_epemucuxcx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_uux_epemdddxdx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_uux_epemcccxcx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_udx_epemuuuxdx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_udx_epemuddxdx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_ucx_epemuuuxcx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_ucx_epemuccxcx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_dd_epemuddux +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_dd_epemddssx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_dc_epemddcdx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_dc_epemdcccx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_ds_epemddsdx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_ds_epemdsssx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_dux_epemuduxux +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_dux_epemdduxdx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_ddx_epemuuuxux +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_ddx_epemuduxdx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_ddx_epemdcdxcx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_ddx_epemdsdxsx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_ddx_epemsssxsx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_dsx_epemdddxsx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_dsx_epemdssxsx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_ccx_epemucuxcx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_ccx_epemdcdxcx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_ssx_epemdsdxsx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_uxux_epemduxuxdx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_uxux_epemcuxuxcx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_uxdx_epemuuxuxdx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_uxdx_epemduxdxdx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_uxcx_epemuuxuxcx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_uxcx_epemcuxcxcx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_dxdx_epemuuxdxdx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_dxdx_epemsdxdxsx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_dxcx_epemddxdxcx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_dxcx_epemcdxcxcx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_dxsx_epemddxdxsx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_dxsx_epemsdxsxsx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_ud_epemudccx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_ud_epemudssx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_uc_epemudcdx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_uux_epemdcdxcx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_uux_epemdsdxsx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_uux_epemcscxsx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_udx_epemucdxcx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_udx_epemusdxsx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_ucx_epemuddxcx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_dc_epemudcux +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_dc_epemdcssx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_ds_epemudsux +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_ds_epemdsbbx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_dux_epemdcuxcx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_dux_epemdsuxsx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_ddx_epemucuxcx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_ddx_epemusuxsx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_ddx_epemsbsxbx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_dsx_epemuduxsx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_dsx_epemdbsxbx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_ccx_epemuduxdx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_ccx_epemdsdxsx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_ssx_epemuduxdx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_ssx_epemucuxcx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_ssx_epemdcdxcx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_ssx_epemdbdxbx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_bbx_epemdsdxsx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_uxdx_epemcuxdxcx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_uxdx_epemsuxdxsx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_uxcx_epemduxdxcx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_dxcx_epemuuxdxcx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_dxcx_epemsdxcxsx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_dxsx_epemuuxdxsx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_dxsx_epembdxsxbx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_uu_taptamuuuux +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_uux_taptamuuuxux +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_dd_taptamddddx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_ddx_taptamdddxdx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_uxux_taptamuuxuxux +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_dxdx_taptamddxdxdx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_uu_taptamuuddx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_uu_taptamuuccx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_ud_taptamuudux +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_ud_taptamudddx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_uc_taptamuucux +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_uc_taptamucccx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_uux_taptamuduxdx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_uux_taptamucuxcx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_uux_taptamdddxdx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_uux_taptamcccxcx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_udx_taptamuuuxdx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_udx_taptamuddxdx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_ucx_taptamuuuxcx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_ucx_taptamuccxcx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_dd_taptamuddux +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_dd_taptamddssx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_dc_taptamddcdx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_dc_taptamdcccx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_ds_taptamddsdx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_ds_taptamdsssx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_dux_taptamuduxux +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_dux_taptamdduxdx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_ddx_taptamuuuxux +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_ddx_taptamuduxdx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_ddx_taptamdcdxcx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_ddx_taptamdsdxsx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_ddx_taptamsssxsx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_dsx_taptamdddxsx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_dsx_taptamdssxsx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_ccx_taptamucuxcx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_ccx_taptamdcdxcx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_ssx_taptamdsdxsx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_uxux_taptamduxuxdx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_uxux_taptamcuxuxcx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_uxdx_taptamuuxuxdx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_uxdx_taptamduxdxdx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_uxcx_taptamuuxuxcx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_uxcx_taptamcuxcxcx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_dxdx_taptamuuxdxdx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_dxdx_taptamsdxdxsx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_dxcx_taptamddxdxcx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_dxcx_taptamcdxcxcx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_dxsx_taptamddxdxsx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_dxsx_taptamsdxsxsx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_ud_taptamudccx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_ud_taptamudssx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_uc_taptamudcdx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_uux_taptamdcdxcx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_uux_taptamdsdxsx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_uux_taptamcscxsx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_udx_taptamucdxcx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_udx_taptamusdxsx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_ucx_taptamuddxcx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_dc_taptamudcux +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_dc_taptamdcssx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_ds_taptamudsux +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_ds_taptamdsbbx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_dux_taptamdcuxcx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_dux_taptamdsuxsx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_ddx_taptamucuxcx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_ddx_taptamusuxsx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_ddx_taptamsbsxbx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_dsx_taptamuduxsx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_dsx_taptamdbsxbx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_ccx_taptamuduxdx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_ccx_taptamdsdxsx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_ssx_taptamuduxdx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_ssx_taptamucuxcx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_ssx_taptamdcdxcx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_ssx_taptamdbdxbx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_bbx_taptamdsdxsx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_uxdx_taptamcuxdxcx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_uxdx_taptamsuxdxsx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_uxcx_taptamduxdxcx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_dxcx_taptamuuxdxcx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_dxcx_taptamsdxcxsx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_dxsx_taptamuuxdxsx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: P0_dxsx_taptambdxsxbx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: Idle: 14954, Running: 4, Completed: 1062 [ current time: 03h24 ] +INFO: Idle: 14953, Running: 4, Completed: 1063 [ 13s ] +INFO: Idle: 14951, Running: 4, Completed: 1065 [ 18s ] +INFO: Idle: 14950, Running: 4, Completed: 1066 [ 22.9s ] +INFO: Idle: 14949, Running: 4, Completed: 1067 [ 38.5s ] +INFO: Idle: 14947, Running: 4, Completed: 1069 [ 54.6s ] +INFO: Idle: 14945, Running: 4, Completed: 1071 [ 1m 10s ] +INFO: Idle: 14943, Running: 4, Completed: 1073 [ 1m 22s ] +INFO: Idle: 14942, Running: 4, Completed: 1074 [ 1m 26s ] +INFO: Idle: 14941, Running: 4, Completed: 1075 [ 1m 33s ] +INFO: Idle: 14940, Running: 4, Completed: 1076 [ 1m 36s ] +INFO: Idle: 14939, Running: 4, Completed: 1077 [ 1m 49s ] +INFO: Idle: 14937, Running: 4, Completed: 1079 [ 1m 53s ] +INFO: Idle: 14935, Running: 4, Completed: 1081 [ 2m 8s ] +INFO: Idle: 14933, Running: 4, Completed: 1083 [ 2m 14s ] +INFO: Idle: 14932, Running: 4, Completed: 1084 [ 2m 23s ] +INFO: Idle: 14931, Running: 4, Completed: 1085 [ 2m 28s ] +INFO: Idle: 14930, Running: 4, Completed: 1086 [ 2m 34s ] +INFO: Idle: 14928, Running: 4, Completed: 1088 [ 2m 43s ] +INFO: Idle: 14927, Running: 4, Completed: 1089 [ 2m 56s ] +INFO: Idle: 14926, Running: 4, Completed: 1090 [ 3m 2s ] +INFO: Idle: 14923, Running: 4, Completed: 1093 [ 3m 16s ] +INFO: Idle: 14922, Running: 4, Completed: 1094 [ 3m 24s ] +INFO: Idle: 14921, Running: 4, Completed: 1095 [ 3m 31s ] +INFO: Idle: 14920, Running: 4, Completed: 1096 [ 3m 41s ] +INFO: Idle: 14919, Running: 4, Completed: 1097 [ 3m 45s ] +INFO: Idle: 14918, Running: 4, Completed: 1098 [ 3m 51s ] +INFO: Idle: 14917, Running: 4, Completed: 1099 [ 3m 55s ] +INFO: Idle: 14916, Running: 4, Completed: 1100 [ 4m 2s ] +INFO: Idle: 14915, Running: 4, Completed: 1101 [ 4m 6s ] +INFO: Idle: 14914, Running: 4, Completed: 1102 [ 4m 16s ] +INFO: Idle: 14912, Running: 4, Completed: 1104 [ 4m 28s ] +INFO: Idle: 14910, Running: 4, Completed: 1106 [ 4m 36s ] +INFO: Idle: 14909, Running: 4, Completed: 1107 [ 4m 41s ] +INFO: Idle: 14908, Running: 4, Completed: 1108 [ 4m 47s ] +INFO: Idle: 14906, Running: 4, Completed: 1110 [ 4m 54s ] +INFO: Idle: 14905, Running: 4, Completed: 1111 [ 4m 59s ] +INFO: Idle: 14904, Running: 4, Completed: 1112 [ 5m 4s ] +INFO: Idle: 14903, Running: 4, Completed: 1113 [ 5m 17s ] +INFO: Idle: 14902, Running: 4, Completed: 1114 [ 5m 21s ] +INFO: Idle: 14901, Running: 4, Completed: 1115 [ 5m 27s ] +INFO: Idle: 14899, Running: 4, Completed: 1117 [ 5m 38s ] +INFO: Idle: 14898, Running: 4, Completed: 1118 [ 5m 45s ] +INFO: Idle: 14897, Running: 4, Completed: 1119 [ 5m 57s ] +INFO: Idle: 14895, Running: 4, Completed: 1121 [ 6m 10s ] +INFO: Idle: 14893, Running: 4, Completed: 1123 [ 6m 18s ] +INFO: Idle: 14892, Running: 4, Completed: 1124 [ 6m 22s ] +INFO: Idle: 14891, Running: 4, Completed: 1125 [ 6m 33s ] +INFO: Idle: 14890, Running: 4, Completed: 1126 [ 6m 38s ] +INFO: Idle: 14888, Running: 4, Completed: 1128 [ 6m 51s ] +INFO: Idle: 14887, Running: 4, Completed: 1129 [ 7m 0s ] +INFO: Idle: 14886, Running: 4, Completed: 1130 [ 7m 4s ] +INFO: Idle: 14885, Running: 4, Completed: 1131 [ 7m 8s ] +INFO: Idle: 14884, Running: 4, Completed: 1132 [ 7m 15s ] +INFO: Idle: 14883, Running: 4, Completed: 1133 [ 7m 24s ] +INFO: Idle: 14881, Running: 4, Completed: 1135 [ 7m 27s ] +INFO: Idle: 14880, Running: 4, Completed: 1136 [ 7m 33s ] +INFO: Idle: 14879, Running: 4, Completed: 1137 [ 7m 48s ] +INFO: Idle: 14876, Running: 4, Completed: 1140 [ 7m 51s ] +INFO: Idle: 14875, Running: 4, Completed: 1141 [ 8m 6s ] +INFO: Idle: 14873, Running: 4, Completed: 1143 [ 8m 19s ] +INFO: Idle: 14871, Running: 4, Completed: 1145 [ 8m 24s ] +INFO: Idle: 14870, Running: 4, Completed: 1146 [ 8m 27s ] +INFO: Idle: 14869, Running: 4, Completed: 1147 [ 8m 38s ] +INFO: Idle: 14868, Running: 4, Completed: 1148 [ 8m 53s ] +INFO: Idle: 14866, Running: 4, Completed: 1150 [ 9m 0s ] +INFO: Idle: 14864, Running: 4, Completed: 1152 [ 9m 16s ] +INFO: Idle: 14862, Running: 4, Completed: 1154 [ 9m 26s ] +INFO: Idle: 14861, Running: 4, Completed: 1155 [ 9m 39s ] +INFO: Idle: 14860, Running: 4, Completed: 1156 [ 9m 44s ] +INFO: Idle: 14858, Running: 4, Completed: 1158 [ 9m 54s ] +INFO: Idle: 14857, Running: 4, Completed: 1159 [ 9m 58s ] +INFO: Idle: 14856, Running: 4, Completed: 1160 [ 10m 7s ] +INFO: Idle: 14854, Running: 4, Completed: 1162 [ 10m 26s ] +INFO: Idle: 14853, Running: 4, Completed: 1163 [ 10m 29s ] +INFO: Idle: 14851, Running: 4, Completed: 1165 [ 10m 35s ] +INFO: Idle: 14850, Running: 4, Completed: 1166 [ 10m 48s ] +INFO: Idle: 14848, Running: 4, Completed: 1168 [ 10m 59s ] +INFO: Idle: 14846, Running: 4, Completed: 1170 [ 11m 7s ] +INFO: Idle: 14844, Running: 4, Completed: 1172 [ 11m 15s ] +INFO: Idle: 14843, Running: 4, Completed: 1173 [ 11m 24s ] +INFO: Idle: 14842, Running: 4, Completed: 1174 [ 11m 28s ] +INFO: Idle: 14841, Running: 4, Completed: 1175 [ 11m 36s ] +INFO: Idle: 14840, Running: 4, Completed: 1176 [ 11m 40s ] +INFO: Idle: 14839, Running: 4, Completed: 1177 [ 11m 46s ] +INFO: Idle: 14837, Running: 4, Completed: 1179 [ 11m 52s ] +INFO: Idle: 14836, Running: 4, Completed: 1180 [ 11m 55s ] +INFO: Idle: 14835, Running: 4, Completed: 1181 [ 12m 12s ] +INFO: Idle: 14833, Running: 4, Completed: 1183 [ 12m 15s ] +INFO: Idle: 14831, Running: 4, Completed: 1185 [ 12m 31s ] +INFO: Idle: 14830, Running: 4, Completed: 1186 [ 12m 47s ] +INFO: Idle: 14827, Running: 4, Completed: 1189 [ 12m 53s ] +INFO: Idle: 14826, Running: 4, Completed: 1190 [ 13m 10s ] +INFO: Idle: 14824, Running: 4, Completed: 1192 [ 13m 19s ] +INFO: Idle: 14823, Running: 4, Completed: 1193 [ 13m 28s ] +INFO: Idle: 14822, Running: 4, Completed: 1194 [ 13m 39s ] +INFO: Idle: 14820, Running: 4, Completed: 1196 [ 13m 48s ] +INFO: Idle: 14818, Running: 4, Completed: 1198 [ 14m 2s ] +INFO: Idle: 14816, Running: 4, Completed: 1200 [ 14m 21s ] +INFO: Idle: 14813, Running: 4, Completed: 1203 [ 14m 32s ] +INFO: Idle: 14812, Running: 4, Completed: 1204 [ 14m 42s ] +INFO: Idle: 14810, Running: 4, Completed: 1206 [ 14m 55s ] +INFO: Idle: 14808, Running: 4, Completed: 1208 [ 15m 4s ] +INFO: Idle: 14806, Running: 4, Completed: 1210 [ 15m 12s ] +INFO: Idle: 14805, Running: 4, Completed: 1211 [ 15m 17s ] +INFO: Idle: 14804, Running: 4, Completed: 1212 [ 15m 26s ] +INFO: Idle: 14803, Running: 4, Completed: 1213 [ 15m 29s ] +INFO: Idle: 14802, Running: 4, Completed: 1214 [ 15m 36s ] +INFO: Idle: 14801, Running: 4, Completed: 1215 [ 15m 39s ] +INFO: Idle: 14800, Running: 4, Completed: 1216 [ 15m 45s ] +INFO: Idle: 14798, Running: 4, Completed: 1218 [ 15m 50s ] +INFO: Idle: 14797, Running: 4, Completed: 1219 [ 16m 5s ] +INFO: Idle: 14794, Running: 4, Completed: 1222 [ 16m 9s ] +INFO: Idle: 14793, Running: 4, Completed: 1223 [ 16m 25s ] +INFO: Idle: 14792, Running: 4, Completed: 1224 [ 16m 37s ] +INFO: Idle: 14790, Running: 4, Completed: 1226 [ 16m 41s ] +INFO: Idle: 14789, Running: 4, Completed: 1227 [ 16m 47s ] +INFO: Idle: 14788, Running: 4, Completed: 1228 [ 17m 3s ] +INFO: Idle: 14786, Running: 4, Completed: 1230 [ 17m 17s ] +INFO: Idle: 14785, Running: 4, Completed: 1231 [ 17m 22s ] +INFO: Idle: 14784, Running: 4, Completed: 1232 [ 17m 35s ] +INFO: Idle: 14782, Running: 4, Completed: 1234 [ 17m 44s ] +INFO: Idle: 14781, Running: 4, Completed: 1235 [ 17m 48s ] +INFO: Idle: 14780, Running: 4, Completed: 1236 [ 17m 55s ] +INFO: Idle: 14778, Running: 4, Completed: 1238 [ 18m 15s ] +INFO: Idle: 14777, Running: 4, Completed: 1239 [ 18m 20s ] +INFO: Idle: 14775, Running: 4, Completed: 1241 [ 18m 23s ] +INFO: Idle: 14774, Running: 4, Completed: 1242 [ 18m 36s ] +INFO: Idle: 14773, Running: 4, Completed: 1243 [ 18m 40s ] +INFO: Idle: 14772, Running: 4, Completed: 1244 [ 18m 44s ] +INFO: Idle: 14771, Running: 4, Completed: 1245 [ 18m 52s ] +INFO: Idle: 14770, Running: 4, Completed: 1246 [ 18m 57s ] +INFO: Idle: 14768, Running: 4, Completed: 1248 [ 19m 0s ] +INFO: Idle: 14767, Running: 4, Completed: 1249 [ 19m 13s ] +INFO: Idle: 14766, Running: 4, Completed: 1250 [ 19m 17s ] +INFO: Idle: 14765, Running: 4, Completed: 1251 [ 19m 20s ] +INFO: Idle: 14764, Running: 4, Completed: 1252 [ 19m 26s ] +INFO: Idle: 14763, Running: 4, Completed: 1253 [ 19m 36s ] +INFO: Idle: 14760, Running: 4, Completed: 1256 [ 19m 41s ] +INFO: Idle: 14759, Running: 4, Completed: 1257 [ 20m 3s ] +INFO: Idle: 14756, Running: 4, Completed: 1260 [ 20m 7s ] +INFO: Idle: 14755, Running: 4, Completed: 1261 [ 20m 23s ] +INFO: Idle: 14754, Running: 4, Completed: 1262 [ 20m 35s ] +INFO: Idle: 14753, Running: 4, Completed: 1263 [ 20m 38s ] +INFO: Idle: 14751, Running: 4, Completed: 1265 [ 20m 47s ] +INFO: Idle: 14750, Running: 4, Completed: 1266 [ 20m 59s ] +INFO: Idle: 14748, Running: 4, Completed: 1268 [ 21m 16s ] +INFO: Idle: 14746, Running: 4, Completed: 1270 [ 21m 28s ] +INFO: Idle: 14744, Running: 4, Completed: 1272 [ 21m 39s ] +INFO: Idle: 14743, Running: 4, Completed: 1273 [ 21m 44s ] +INFO: Idle: 14742, Running: 4, Completed: 1274 [ 21m 51s ] +INFO: Idle: 14740, Running: 4, Completed: 1276 [ 22m 10s ] +INFO: Idle: 14739, Running: 4, Completed: 1277 [ 22m 14s ] +INFO: Idle: 14737, Running: 4, Completed: 1279 [ 22m 20s ] +INFO: Idle: 14736, Running: 4, Completed: 1280 [ 22m 32s ] +INFO: Idle: 14735, Running: 4, Completed: 1281 [ 22m 35s ] +INFO: Idle: 14734, Running: 4, Completed: 1282 [ 22m 45s ] +INFO: Idle: 14732, Running: 4, Completed: 1284 [ 22m 54s ] +INFO: Idle: 14730, Running: 4, Completed: 1286 [ 23m 0s ] +INFO: Idle: 14729, Running: 4, Completed: 1287 [ 23m 7s ] +INFO: Idle: 14728, Running: 4, Completed: 1288 [ 23m 11s ] +INFO: Idle: 14727, Running: 4, Completed: 1289 [ 23m 14s ] +INFO: Idle: 14726, Running: 4, Completed: 1290 [ 23m 23s ] +INFO: Idle: 14725, Running: 4, Completed: 1291 [ 23m 27s ] +INFO: Idle: 14723, Running: 4, Completed: 1293 [ 23m 32s ] +INFO: Idle: 14722, Running: 4, Completed: 1294 [ 23m 48s ] +INFO: Idle: 14721, Running: 4, Completed: 1295 [ 23m 53s ] +INFO: Idle: 14719, Running: 4, Completed: 1297 [ 24m 2s ] +INFO: Idle: 14718, Running: 4, Completed: 1298 [ 24m 17s ] +INFO: Idle: 14717, Running: 4, Completed: 1299 [ 24m 20s ] +INFO: Idle: 14716, Running: 4, Completed: 1300 [ 24m 29s ] +INFO: Idle: 14714, Running: 4, Completed: 1302 [ 24m 38s ] +INFO: Idle: 14712, Running: 4, Completed: 1304 [ 24m 56s ] +INFO: Idle: 14709, Running: 4, Completed: 1307 [ 25m 8s ] +INFO: Idle: 14708, Running: 4, Completed: 1308 [ 25m 15s ] +INFO: Idle: 14707, Running: 4, Completed: 1309 [ 25m 19s ] +INFO: Idle: 14706, Running: 4, Completed: 1310 [ 25m 25s ] +INFO: Idle: 14705, Running: 4, Completed: 1311 [ 25m 29s ] +INFO: Idle: 14704, Running: 4, Completed: 1312 [ 25m 32s ] +INFO: Idle: 14703, Running: 4, Completed: 1313 [ 25m 46s ] +INFO: Idle: 14702, Running: 4, Completed: 1314 [ 25m 52s ] +INFO: Idle: 14700, Running: 4, Completed: 1316 [ 25m 59s ] +INFO: Idle: 14699, Running: 4, Completed: 1317 [ 26m 7s ] +INFO: Idle: 14698, Running: 4, Completed: 1318 [ 26m 14s ] +INFO: Idle: 14697, Running: 4, Completed: 1319 [ 26m 18s ] +INFO: Idle: 14696, Running: 4, Completed: 1320 [ 26m 24s ] +INFO: Idle: 14695, Running: 4, Completed: 1321 [ 26m 31s ] +INFO: Idle: 14693, Running: 4, Completed: 1323 [ 26m 36s ] +INFO: Idle: 14692, Running: 4, Completed: 1324 [ 26m 42s ] +INFO: Idle: 14691, Running: 4, Completed: 1325 [ 26m 47s ] +INFO: Idle: 14690, Running: 4, Completed: 1326 [ 26m 55s ] +INFO: Idle: 14687, Running: 4, Completed: 1329 [ 27m 4s ] +INFO: Idle: 14686, Running: 4, Completed: 1330 [ 27m 17s ] +INFO: Idle: 14685, Running: 4, Completed: 1331 [ 27m 24s ] +INFO: Idle: 14684, Running: 4, Completed: 1332 [ 27m 28s ] +INFO: Idle: 14683, Running: 4, Completed: 1333 [ 27m 33s ] +INFO: Idle: 14682, Running: 4, Completed: 1334 [ 27m 46s ] +INFO: Idle: 14681, Running: 4, Completed: 1335 [ 27m 53s ] +INFO: Idle: 14680, Running: 4, Completed: 1336 [ 28m 0s ] +INFO: Idle: 14678, Running: 4, Completed: 1338 [ 28m 6s ] +INFO: Idle: 14677, Running: 4, Completed: 1339 [ 28m 10s ] +INFO: Idle: 14676, Running: 4, Completed: 1340 [ 28m 27s ] +INFO: Idle: 14675, Running: 4, Completed: 1341 [ 28m 30s ] +INFO: Idle: 14673, Running: 4, Completed: 1343 [ 28m 36s ] +INFO: Idle: 14672, Running: 4, Completed: 1344 [ 28m 46s ] +INFO: Idle: 14671, Running: 4, Completed: 1345 [ 28m 51s ] +INFO: Idle: 14670, Running: 4, Completed: 1346 [ 28m 57s ] +INFO: Idle: 14668, Running: 4, Completed: 1348 [ 29m 3s ] +INFO: Idle: 14667, Running: 4, Completed: 1349 [ 29m 18s ] +INFO: Idle: 14666, Running: 4, Completed: 1350 [ 29m 23s ] +INFO: Idle: 14664, Running: 4, Completed: 1352 [ 29m 29s ] +INFO: Idle: 14663, Running: 4, Completed: 1353 [ 29m 39s ] +INFO: Idle: 14662, Running: 4, Completed: 1354 [ 29m 45s ] +INFO: Idle: 14661, Running: 4, Completed: 1355 [ 29m 51s ] +INFO: Idle: 14659, Running: 4, Completed: 1357 [ 30m 3s ] +INFO: Idle: 14657, Running: 4, Completed: 1359 [ 30m 8s ] +INFO: Idle: 14655, Running: 4, Completed: 1361 [ 30m 18s ] +INFO: Idle: 14654, Running: 4, Completed: 1362 [ 30m 24s ] +INFO: Idle: 14652, Running: 4, Completed: 1364 [ 30m 28s ] +INFO: Idle: 14651, Running: 4, Completed: 1365 [ 30m 33s ] +INFO: Idle: 14650, Running: 4, Completed: 1366 [ 30m 47s ] +INFO: Idle: 14649, Running: 4, Completed: 1367 [ 30m 52s ] +INFO: Idle: 14647, Running: 4, Completed: 1369 [ 30m 59s ] +INFO: Idle: 14646, Running: 4, Completed: 1370 [ 31m 13s ] +INFO: Idle: 14645, Running: 4, Completed: 1371 [ 31m 19s ] +INFO: Idle: 14644, Running: 4, Completed: 1372 [ 31m 24s ] +INFO: Idle: 14643, Running: 4, Completed: 1373 [ 31m 28s ] +INFO: Idle: 14642, Running: 4, Completed: 1374 [ 31m 31s ] +INFO: Idle: 14641, Running: 4, Completed: 1375 [ 31m 34s ] +INFO: Idle: 14640, Running: 4, Completed: 1376 [ 31m 49s ] +INFO: Idle: 14639, Running: 4, Completed: 1377 [ 31m 53s ] +INFO: Idle: 14637, Running: 4, Completed: 1379 [ 32m 0s ] +INFO: Idle: 14636, Running: 4, Completed: 1380 [ 32m 6s ] +INFO: Idle: 14635, Running: 4, Completed: 1381 [ 32m 12s ] +INFO: Idle: 14634, Running: 4, Completed: 1382 [ 32m 18s ] +INFO: Idle: 14632, Running: 4, Completed: 1384 [ 32m 22s ] +INFO: Idle: 14631, Running: 4, Completed: 1385 [ 32m 37s ] +INFO: Idle: 14629, Running: 4, Completed: 1387 [ 32m 45s ] +INFO: Idle: 14627, Running: 4, Completed: 1389 [ 32m 57s ] +INFO: Idle: 14626, Running: 4, Completed: 1390 [ 33m 1s ] +INFO: Idle: 14625, Running: 4, Completed: 1391 [ 33m 7s ] +INFO: Idle: 14623, Running: 4, Completed: 1393 [ 33m 17s ] +INFO: Idle: 14621, Running: 4, Completed: 1395 [ 33m 23s ] +INFO: Idle: 14620, Running: 4, Completed: 1396 [ 33m 26s ] +INFO: Idle: 14619, Running: 4, Completed: 1397 [ 33m 31s ] +INFO: Idle: 14618, Running: 4, Completed: 1398 [ 33m 40s ] +INFO: Idle: 14615, Running: 4, Completed: 1401 [ 33m 46s ] +INFO: Idle: 14614, Running: 4, Completed: 1402 [ 34m 1s ] +INFO: Idle: 14613, Running: 4, Completed: 1403 [ 34m 8s ] +INFO: Idle: 14611, Running: 4, Completed: 1405 [ 34m 15s ] +INFO: Idle: 14610, Running: 4, Completed: 1406 [ 34m 27s ] +INFO: Idle: 14609, Running: 4, Completed: 1407 [ 34m 33s ] +INFO: Idle: 14608, Running: 4, Completed: 1408 [ 34m 39s ] +INFO: Idle: 14607, Running: 4, Completed: 1409 [ 34m 44s ] +INFO: Idle: 14605, Running: 4, Completed: 1411 [ 34m 48s ] +INFO: Idle: 14604, Running: 4, Completed: 1412 [ 35m 4s ] +INFO: Idle: 14602, Running: 4, Completed: 1414 [ 35m 10s ] +INFO: Idle: 14601, Running: 4, Completed: 1415 [ 35m 13s ] +INFO: Idle: 14600, Running: 4, Completed: 1416 [ 35m 22s ] +INFO: Idle: 14599, Running: 4, Completed: 1417 [ 35m 26s ] +INFO: Idle: 14598, Running: 4, Completed: 1418 [ 35m 32s ] +INFO: Idle: 14596, Running: 4, Completed: 1420 [ 35m 37s ] +INFO: Idle: 14595, Running: 4, Completed: 1421 [ 35m 52s ] +INFO: Idle: 14594, Running: 4, Completed: 1422 [ 35m 56s ] +INFO: Idle: 14593, Running: 4, Completed: 1423 [ 36m 0s ] +INFO: Idle: 14591, Running: 4, Completed: 1425 [ 36m 11s ] +INFO: Idle: 14590, Running: 4, Completed: 1426 [ 36m 17s ] +INFO: Idle: 14589, Running: 4, Completed: 1427 [ 36m 23s ] +INFO: Idle: 14587, Running: 4, Completed: 1429 [ 36m 32s ] +INFO: Idle: 14585, Running: 4, Completed: 1431 [ 36m 38s ] +INFO: Idle: 14584, Running: 4, Completed: 1432 [ 36m 41s ] +INFO: Idle: 14583, Running: 4, Completed: 1433 [ 36m 47s ] +INFO: Idle: 14582, Running: 4, Completed: 1434 [ 36m 55s ] +INFO: Idle: 14580, Running: 4, Completed: 1436 [ 36m 58s ] +INFO: Idle: 14579, Running: 4, Completed: 1437 [ 37m 2s ] +INFO: Idle: 14578, Running: 4, Completed: 1438 [ 37m 16s ] +INFO: Idle: 14577, Running: 4, Completed: 1439 [ 37m 22s ] +INFO: Idle: 14575, Running: 4, Completed: 1441 [ 37m 29s ] +INFO: Idle: 14574, Running: 4, Completed: 1442 [ 37m 43s ] +INFO: Idle: 14573, Running: 4, Completed: 1443 [ 37m 48s ] +INFO: Idle: 14572, Running: 4, Completed: 1444 [ 37m 54s ] +INFO: Idle: 14571, Running: 4, Completed: 1445 [ 37m 58s ] +INFO: Idle: 14570, Running: 4, Completed: 1446 [ 38m 1s ] +INFO: Idle: 14568, Running: 4, Completed: 1448 [ 38m 19s ] +INFO: Idle: 14566, Running: 4, Completed: 1450 [ 38m 24s ] +INFO: Idle: 14565, Running: 4, Completed: 1451 [ 38m 30s ] +INFO: Idle: 14564, Running: 4, Completed: 1452 [ 38m 36s ] +INFO: Idle: 14563, Running: 4, Completed: 1453 [ 38m 42s ] +INFO: Idle: 14562, Running: 4, Completed: 1454 [ 38m 48s ] +INFO: Idle: 14560, Running: 4, Completed: 1456 [ 38m 52s ] +INFO: Idle: 14559, Running: 4, Completed: 1457 [ 39m 7s ] +INFO: Idle: 14557, Running: 4, Completed: 1459 [ 39m 15s ] +INFO: Idle: 14555, Running: 4, Completed: 1461 [ 39m 27s ] +INFO: Idle: 14554, Running: 4, Completed: 1462 [ 39m 30s ] +INFO: Idle: 14553, Running: 4, Completed: 1463 [ 39m 38s ] +INFO: Idle: 14551, Running: 4, Completed: 1465 [ 39m 46s ] +INFO: Idle: 14549, Running: 4, Completed: 1467 [ 39m 54s ] +INFO: Idle: 14547, Running: 4, Completed: 1469 [ 40m 2s ] +INFO: Idle: 14546, Running: 4, Completed: 1470 [ 40m 11s ] +INFO: Idle: 14543, Running: 4, Completed: 1473 [ 40m 19s ] +INFO: Idle: 14542, Running: 4, Completed: 1474 [ 40m 34s ] +INFO: Idle: 14541, Running: 4, Completed: 1475 [ 40m 40s ] +INFO: Idle: 14539, Running: 4, Completed: 1477 [ 40m 47s ] +INFO: Idle: 14538, Running: 4, Completed: 1478 [ 41m 2s ] +INFO: Idle: 14537, Running: 4, Completed: 1479 [ 41m 8s ] +INFO: Idle: 14536, Running: 4, Completed: 1480 [ 41m 13s ] +INFO: Idle: 14534, Running: 4, Completed: 1482 [ 41m 22s ] +INFO: Idle: 14533, Running: 4, Completed: 1483 [ 41m 25s ] +INFO: Idle: 14532, Running: 4, Completed: 1484 [ 41m 40s ] +INFO: Idle: 14530, Running: 4, Completed: 1486 [ 41m 45s ] +INFO: Idle: 14529, Running: 4, Completed: 1487 [ 41m 51s ] +INFO: Idle: 14528, Running: 4, Completed: 1488 [ 41m 58s ] +INFO: Idle: 14527, Running: 4, Completed: 1489 [ 42m 4s ] +INFO: Idle: 14526, Running: 4, Completed: 1490 [ 42m 10s ] +INFO: Idle: 14524, Running: 4, Completed: 1492 [ 42m 16s ] +INFO: Idle: 14523, Running: 4, Completed: 1493 [ 42m 30s ] +INFO: Idle: 14522, Running: 4, Completed: 1494 [ 42m 35s ] +INFO: Idle: 14520, Running: 4, Completed: 1496 [ 42m 41s ] +INFO: Idle: 14519, Running: 4, Completed: 1497 [ 42m 51s ] +INFO: Idle: 14518, Running: 4, Completed: 1498 [ 42m 57s ] +INFO: Idle: 14517, Running: 4, Completed: 1499 [ 43m 2s ] +INFO: Idle: 14515, Running: 4, Completed: 1501 [ 43m 15s ] +INFO: Idle: 14513, Running: 4, Completed: 1503 [ 43m 20s ] +INFO: Idle: 14512, Running: 4, Completed: 1504 [ 43m 23s ] +INFO: Idle: 14511, Running: 4, Completed: 1505 [ 43m 32s ] +INFO: Idle: 14510, Running: 4, Completed: 1506 [ 43m 38s ] +INFO: Idle: 14509, Running: 4, Completed: 1507 [ 43m 41s ] +INFO: Idle: 14507, Running: 4, Completed: 1509 [ 43m 51s ] +INFO: Idle: 14506, Running: 4, Completed: 1510 [ 44m 3s ] +INFO: Idle: 14505, Running: 4, Completed: 1511 [ 44m 7s ] +INFO: Idle: 14504, Running: 4, Completed: 1512 [ 44m 16s ] +INFO: Idle: 14502, Running: 4, Completed: 1514 [ 44m 31s ] +INFO: Idle: 14501, Running: 4, Completed: 1515 [ 44m 42s ] +INFO: Idle: 14498, Running: 4, Completed: 1518 [ 44m 53s ] +INFO: Idle: 14497, Running: 4, Completed: 1519 [ 44m 59s ] +INFO: Idle: 14496, Running: 4, Completed: 1520 [ 45m 11s ] +INFO: Idle: 14494, Running: 4, Completed: 1522 [ 45m 19s ] +INFO: Idle: 14493, Running: 4, Completed: 1523 [ 45m 22s ] +INFO: Idle: 14492, Running: 4, Completed: 1524 [ 45m 30s ] +INFO: Idle: 14491, Running: 4, Completed: 1525 [ 45m 33s ] +INFO: Idle: 14490, Running: 4, Completed: 1526 [ 45m 43s ] +INFO: Idle: 14488, Running: 4, Completed: 1528 [ 45m 48s ] +INFO: Idle: 14487, Running: 4, Completed: 1529 [ 46m 1s ] +INFO: Idle: 14486, Running: 4, Completed: 1530 [ 46m 9s ] +INFO: Idle: 14485, Running: 4, Completed: 1531 [ 46m 12s ] +INFO: Idle: 14483, Running: 4, Completed: 1533 [ 46m 21s ] +INFO: Idle: 14482, Running: 4, Completed: 1534 [ 46m 31s ] +INFO: Idle: 14481, Running: 4, Completed: 1535 [ 46m 36s ] +INFO: Idle: 14479, Running: 4, Completed: 1537 [ 46m 44s ] +INFO: Idle: 14478, Running: 4, Completed: 1538 [ 46m 48s ] +INFO: Idle: 14477, Running: 4, Completed: 1539 [ 46m 54s ] +INFO: Idle: 14475, Running: 4, Completed: 1541 [ 47m 0s ] +INFO: Idle: 14474, Running: 4, Completed: 1542 [ 47m 9s ] +INFO: Idle: 14472, Running: 4, Completed: 1544 [ 47m 12s ] +INFO: Idle: 14470, Running: 4, Completed: 1546 [ 47m 31s ] +INFO: Idle: 14469, Running: 4, Completed: 1547 [ 47m 36s ] +INFO: Idle: 14467, Running: 4, Completed: 1549 [ 47m 43s ] +INFO: Idle: 14466, Running: 4, Completed: 1550 [ 47m 57s ] +INFO: Idle: 14465, Running: 4, Completed: 1551 [ 48m 1s ] +INFO: Idle: 14464, Running: 4, Completed: 1552 [ 48m 6s ] +INFO: Idle: 14463, Running: 4, Completed: 1553 [ 48m 10s ] +INFO: Idle: 14462, Running: 4, Completed: 1554 [ 48m 14s ] +INFO: Idle: 14460, Running: 4, Completed: 1556 [ 48m 30s ] +INFO: Idle: 14459, Running: 4, Completed: 1557 [ 48m 34s ] +INFO: Idle: 14457, Running: 4, Completed: 1559 [ 48m 41s ] +INFO: Idle: 14456, Running: 4, Completed: 1560 [ 48m 47s ] +INFO: Idle: 14455, Running: 4, Completed: 1561 [ 48m 53s ] +INFO: Idle: 14454, Running: 4, Completed: 1562 [ 48m 58s ] +INFO: Idle: 14452, Running: 4, Completed: 1564 [ 49m 3s ] +INFO: Idle: 14451, Running: 4, Completed: 1565 [ 49m 17s ] +INFO: Idle: 14450, Running: 4, Completed: 1566 [ 49m 21s ] +INFO: Idle: 14448, Running: 4, Completed: 1568 [ 49m 26s ] +INFO: Idle: 14447, Running: 4, Completed: 1569 [ 49m 37s ] +INFO: Idle: 14446, Running: 4, Completed: 1570 [ 49m 42s ] +INFO: Idle: 14445, Running: 4, Completed: 1571 [ 49m 46s ] +INFO: Idle: 14443, Running: 4, Completed: 1573 [ 49m 57s ] +INFO: Idle: 14441, Running: 4, Completed: 1575 [ 50m 2s ] +INFO: Idle: 14439, Running: 4, Completed: 1577 [ 50m 11s ] +INFO: Idle: 14438, Running: 4, Completed: 1578 [ 50m 19s ] +INFO: Idle: 14435, Running: 4, Completed: 1581 [ 50m 27s ] +INFO: Idle: 14434, Running: 4, Completed: 1582 [ 50m 38s ] +INFO: Idle: 14433, Running: 4, Completed: 1583 [ 50m 46s ] +INFO: Idle: 14432, Running: 4, Completed: 1584 [ 50m 49s ] +INFO: Idle: 14430, Running: 4, Completed: 1586 [ 51m 5s ] +INFO: Idle: 14429, Running: 4, Completed: 1587 [ 51m 14s ] +INFO: Idle: 14427, Running: 4, Completed: 1589 [ 51m 20s ] +INFO: Idle: 14425, Running: 4, Completed: 1591 [ 51m 29s ] +INFO: Idle: 14424, Running: 4, Completed: 1592 [ 51m 39s ] +INFO: Idle: 14423, Running: 4, Completed: 1593 [ 51m 45s ] +INFO: Idle: 14421, Running: 4, Completed: 1595 [ 51m 50s ] +INFO: Idle: 14420, Running: 4, Completed: 1596 [ 51m 56s ] +INFO: Idle: 14419, Running: 4, Completed: 1597 [ 52m 5s ] +INFO: Idle: 14418, Running: 4, Completed: 1598 [ 52m 8s ] +INFO: Idle: 14416, Running: 4, Completed: 1600 [ 52m 12s ] +INFO: Idle: 14415, Running: 4, Completed: 1601 [ 52m 29s ] +INFO: Idle: 14413, Running: 4, Completed: 1603 [ 52m 35s ] +INFO: Idle: 14411, Running: 4, Completed: 1605 [ 52m 49s ] +INFO: Idle: 14409, Running: 4, Completed: 1607 [ 52m 56s ] +INFO: Idle: 14408, Running: 4, Completed: 1608 [ 53m 0s ] +INFO: Idle: 14407, Running: 4, Completed: 1609 [ 53m 6s ] +INFO: Idle: 14406, Running: 4, Completed: 1610 [ 53m 10s ] +INFO: Idle: 14404, Running: 4, Completed: 1612 [ 53m 16s ] +INFO: Idle: 14403, Running: 4, Completed: 1613 [ 53m 21s ] +INFO: Idle: 14402, Running: 4, Completed: 1614 [ 53m 30s ] +INFO: Idle: 14399, Running: 4, Completed: 1617 [ 53m 36s ] +INFO: Idle: 14398, Running: 4, Completed: 1618 [ 53m 49s ] +INFO: Idle: 14397, Running: 4, Completed: 1619 [ 53m 57s ] +INFO: Idle: 14395, Running: 4, Completed: 1621 [ 54m 3s ] +INFO: Idle: 14394, Running: 4, Completed: 1622 [ 54m 16s ] +INFO: Idle: 14393, Running: 4, Completed: 1623 [ 54m 22s ] +INFO: Idle: 14392, Running: 4, Completed: 1624 [ 54m 27s ] +INFO: Idle: 14391, Running: 4, Completed: 1625 [ 54m 33s ] +INFO: Idle: 14389, Running: 4, Completed: 1627 [ 54m 37s ] +INFO: Idle: 14388, Running: 4, Completed: 1628 [ 54m 51s ] +INFO: Idle: 14387, Running: 4, Completed: 1629 [ 54m 56s ] +INFO: Idle: 14385, Running: 4, Completed: 1631 [ 55m 1s ] +INFO: Idle: 14384, Running: 4, Completed: 1632 [ 55m 9s ] +INFO: Idle: 14383, Running: 4, Completed: 1633 [ 55m 15s ] +INFO: Idle: 14382, Running: 4, Completed: 1634 [ 55m 20s ] +INFO: Idle: 14380, Running: 4, Completed: 1636 [ 55m 25s ] +INFO: Idle: 14379, Running: 4, Completed: 1637 [ 55m 39s ] +INFO: Idle: 14378, Running: 4, Completed: 1638 [ 55m 43s ] +INFO: Idle: 14376, Running: 4, Completed: 1640 [ 55m 48s ] +INFO: Idle: 14375, Running: 4, Completed: 1641 [ 55m 59s ] +INFO: Idle: 14374, Running: 4, Completed: 1642 [ 56m 4s ] +INFO: Idle: 14373, Running: 4, Completed: 1643 [ 56m 8s ] +INFO: Idle: 14371, Running: 4, Completed: 1645 [ 56m 19s ] +INFO: Idle: 14369, Running: 4, Completed: 1647 [ 56m 24s ] +INFO: Idle: 14367, Running: 4, Completed: 1649 [ 56m 41s ] +INFO: Idle: 14365, Running: 4, Completed: 1651 [ 56m 44s ] +INFO: Idle: 14363, Running: 4, Completed: 1653 [ 57m 7s ] +INFO: Idle: 14361, Running: 4, Completed: 1655 [ 57m 15s ] +INFO: Idle: 14359, Running: 4, Completed: 1657 [ 57m 36s ] +INFO: Idle: 14357, Running: 4, Completed: 1659 [ 57m 42s ] +INFO: Idle: 14356, Running: 4, Completed: 1660 [ 57m 47s ] +INFO: Idle: 14355, Running: 4, Completed: 1661 [ 57m 54s ] +INFO: Idle: 14353, Running: 4, Completed: 1663 [ 58m 1s ] +INFO: Idle: 14352, Running: 4, Completed: 1664 [ 58m 8s ] +INFO: Idle: 14351, Running: 4, Completed: 1665 [ 58m 21s ] +INFO: Idle: 14349, Running: 4, Completed: 1667 [ 58m 29s ] +INFO: Idle: 14348, Running: 4, Completed: 1668 [ 58m 39s ] +INFO: Idle: 14345, Running: 4, Completed: 1671 [ 58m 50s ] +INFO: Idle: 14344, Running: 4, Completed: 1672 [ 59m 1s ] +INFO: Idle: 14343, Running: 4, Completed: 1673 [ 59m 11s ] +INFO: Idle: 14342, Running: 4, Completed: 1674 [ 59m 15s ] +INFO: Idle: 14340, Running: 4, Completed: 1676 [ 59m 20s ] +INFO: Idle: 14339, Running: 4, Completed: 1677 [ 59m 35s ] +INFO: Idle: 14338, Running: 4, Completed: 1678 [ 59m 43s ] +INFO: Idle: 14337, Running: 4, Completed: 1679 [ 59m 49s ] +INFO: Idle: 14336, Running: 4, Completed: 1680 [ 59m 54s ] +INFO: Idle: 14334, Running: 4, Completed: 1682 [ 1h 0m ] +INFO: Idle: 14333, Running: 4, Completed: 1683 [ 1h 0m ] +INFO: Idle: 14330, Running: 4, Completed: 1686 [ 1h 0m ] +INFO: Idle: 14328, Running: 4, Completed: 1688 [ 1h 0m ] +INFO: Idle: 14327, Running: 4, Completed: 1689 [ 1h 0m ] +INFO: Idle: 14326, Running: 4, Completed: 1690 [ 1h 0m ] +INFO: Idle: 14323, Running: 4, Completed: 1693 [ 1h 1m ] +INFO: Idle: 14320, Running: 4, Completed: 1696 [ 1h 1m ] +INFO: Idle: 14319, Running: 4, Completed: 1697 [ 1h 1m ] +INFO: Idle: 14318, Running: 4, Completed: 1698 [ 1h 1m ] +INFO: Idle: 14317, Running: 4, Completed: 1699 [ 1h 1m ] +INFO: Idle: 14316, Running: 4, Completed: 1700 [ 1h 1m ] +INFO: Idle: 14314, Running: 4, Completed: 1702 [ 1h 2m ] +INFO: Idle: 14312, Running: 4, Completed: 1704 [ 1h 2m ] +INFO: Idle: 14311, Running: 4, Completed: 1705 [ 1h 2m ] +INFO: Idle: 14308, Running: 4, Completed: 1708 [ 1h 2m ] +INFO: Idle: 14307, Running: 4, Completed: 1709 [ 1h 2m ] +INFO: Idle: 14306, Running: 4, Completed: 1710 [ 1h 2m ] +INFO: Idle: 14305, Running: 4, Completed: 1711 [ 1h 2m ] +INFO: Idle: 14302, Running: 4, Completed: 1714 [ 1h 2m ] +INFO: Idle: 14301, Running: 4, Completed: 1715 [ 1h 3m ] +INFO: Idle: 14299, Running: 4, Completed: 1717 [ 1h 3m ] +INFO: Idle: 14297, Running: 4, Completed: 1719 [ 1h 3m ] +INFO: Idle: 14295, Running: 4, Completed: 1721 [ 1h 3m ] +INFO: Idle: 14293, Running: 4, Completed: 1723 [ 1h 3m ] +INFO: Idle: 14291, Running: 4, Completed: 1725 [ 1h 3m ] +INFO: Idle: 14290, Running: 4, Completed: 1726 [ 1h 4m ] +INFO: Idle: 14289, Running: 4, Completed: 1727 [ 1h 4m ] +INFO: Idle: 14288, Running: 4, Completed: 1728 [ 1h 4m ] +INFO: Idle: 14286, Running: 4, Completed: 1730 [ 1h 4m ] +INFO: Idle: 14283, Running: 4, Completed: 1733 [ 1h 4m ] +INFO: Idle: 14282, Running: 4, Completed: 1734 [ 1h 4m ] +INFO: Idle: 14280, Running: 4, Completed: 1736 [ 1h 5m ] +INFO: Idle: 14279, Running: 4, Completed: 1737 [ 1h 5m ] +INFO: Idle: 14277, Running: 4, Completed: 1739 [ 1h 5m ] +INFO: Idle: 14275, Running: 4, Completed: 1741 [ 1h 5m ] +INFO: Idle: 14274, Running: 4, Completed: 1742 [ 1h 5m ] +INFO: Idle: 14271, Running: 4, Completed: 1745 [ 1h 5m ] +INFO: Idle: 14270, Running: 4, Completed: 1746 [ 1h 5m ] +INFO: Idle: 14269, Running: 4, Completed: 1747 [ 1h 6m ] +INFO: Idle: 14268, Running: 4, Completed: 1748 [ 1h 6m ] +INFO: Idle: 14267, Running: 4, Completed: 1749 [ 1h 6m ] +INFO: Idle: 14266, Running: 4, Completed: 1750 [ 1h 6m ] +INFO: Idle: 14265, Running: 4, Completed: 1751 [ 1h 6m ] +INFO: Idle: 14264, Running: 4, Completed: 1752 [ 1h 6m ] +INFO: Idle: 14262, Running: 4, Completed: 1754 [ 1h 6m ] +INFO: Idle: 14261, Running: 4, Completed: 1755 [ 1h 6m ] +INFO: Idle: 14260, Running: 4, Completed: 1756 [ 1h 6m ] +INFO: Idle: 14259, Running: 4, Completed: 1757 [ 1h 7m ] +INFO: Idle: 14257, Running: 4, Completed: 1759 [ 1h 7m ] +INFO: Idle: 14255, Running: 4, Completed: 1761 [ 1h 7m ] +INFO: Idle: 14253, Running: 4, Completed: 1763 [ 1h 7m ] +INFO: Idle: 14251, Running: 4, Completed: 1765 [ 1h 7m ] +INFO: Idle: 14248, Running: 4, Completed: 1768 [ 1h 7m ] +INFO: Idle: 14247, Running: 4, Completed: 1769 [ 1h 8m ] +INFO: Idle: 14245, Running: 4, Completed: 1771 [ 1h 8m ] +INFO: Idle: 14243, Running: 4, Completed: 1773 [ 1h 8m ] +INFO: Idle: 14241, Running: 4, Completed: 1775 [ 1h 8m ] +INFO: Idle: 14239, Running: 4, Completed: 1777 [ 1h 9m ] +INFO: Idle: 14237, Running: 4, Completed: 1779 [ 1h 9m ] +INFO: Idle: 14236, Running: 4, Completed: 1780 [ 1h 9m ] +INFO: Idle: 14235, Running: 4, Completed: 1781 [ 1h 9m ] +INFO: Idle: 14233, Running: 4, Completed: 1783 [ 1h 9m ] +INFO: Idle: 14232, Running: 4, Completed: 1784 [ 1h 9m ] +INFO: Idle: 14231, Running: 4, Completed: 1785 [ 1h 9m ] +INFO: Idle: 14229, Running: 4, Completed: 1787 [ 1h 9m ] +INFO: Idle: 14228, Running: 4, Completed: 1788 [ 1h 10m ] +INFO: Idle: 14227, Running: 4, Completed: 1789 [ 1h 10m ] +INFO: Idle: 14225, Running: 4, Completed: 1791 [ 1h 10m ] +INFO: Idle: 14224, Running: 4, Completed: 1792 [ 1h 10m ] +INFO: Idle: 14223, Running: 4, Completed: 1793 [ 1h 10m ] +INFO: Idle: 14222, Running: 4, Completed: 1794 [ 1h 10m ] +INFO: Idle: 14220, Running: 4, Completed: 1796 [ 1h 10m ] +INFO: Idle: 14219, Running: 4, Completed: 1797 [ 1h 11m ] +INFO: Idle: 14218, Running: 4, Completed: 1798 [ 1h 11m ] +INFO: Idle: 14217, Running: 4, Completed: 1799 [ 1h 11m ] +INFO: Idle: 14216, Running: 4, Completed: 1800 [ 1h 11m ] +INFO: Idle: 14214, Running: 4, Completed: 1802 [ 1h 11m ] +INFO: Idle: 14213, Running: 4, Completed: 1803 [ 1h 11m ] +INFO: Idle: 14210, Running: 4, Completed: 1806 [ 1h 12m ] +INFO: Idle: 14208, Running: 4, Completed: 1808 [ 1h 12m ] +INFO: Idle: 14207, Running: 4, Completed: 1809 [ 1h 12m ] +INFO: Idle: 14206, Running: 4, Completed: 1810 [ 1h 12m ] +INFO: Idle: 14203, Running: 4, Completed: 1813 [ 1h 12m ] +INFO: Idle: 14200, Running: 4, Completed: 1816 [ 1h 12m ] +INFO: Idle: 14199, Running: 4, Completed: 1817 [ 1h 13m ] +INFO: Idle: 14198, Running: 4, Completed: 1818 [ 1h 13m ] +INFO: Idle: 14197, Running: 4, Completed: 1819 [ 1h 13m ] +INFO: Idle: 14196, Running: 4, Completed: 1820 [ 1h 13m ] +INFO: Idle: 14195, Running: 4, Completed: 1821 [ 1h 13m ] +INFO: Idle: 14194, Running: 4, Completed: 1822 [ 1h 13m ] +INFO: Idle: 14192, Running: 4, Completed: 1824 [ 1h 13m ] +INFO: Idle: 14191, Running: 4, Completed: 1825 [ 1h 13m ] +INFO: Idle: 14188, Running: 4, Completed: 1828 [ 1h 13m ] +INFO: Idle: 14187, Running: 4, Completed: 1829 [ 1h 14m ] +INFO: Idle: 14186, Running: 4, Completed: 1830 [ 1h 14m ] +INFO: Idle: 14185, Running: 4, Completed: 1831 [ 1h 14m ] +INFO: Idle: 14183, Running: 4, Completed: 1833 [ 1h 14m ] +INFO: Idle: 14181, Running: 4, Completed: 1835 [ 1h 14m ] +INFO: Idle: 14179, Running: 4, Completed: 1837 [ 1h 14m ] +INFO: Idle: 14178, Running: 4, Completed: 1838 [ 1h 14m ] +INFO: Idle: 14177, Running: 4, Completed: 1839 [ 1h 14m ] +INFO: Idle: 14175, Running: 4, Completed: 1841 [ 1h 15m ] +INFO: Idle: 14174, Running: 4, Completed: 1842 [ 1h 15m ] +INFO: Idle: 14173, Running: 4, Completed: 1843 [ 1h 15m ] +INFO: Idle: 14172, Running: 4, Completed: 1844 [ 1h 15m ] +INFO: Idle: 14171, Running: 4, Completed: 1845 [ 1h 15m ] +INFO: Idle: 14170, Running: 4, Completed: 1846 [ 1h 15m ] +INFO: Idle: 14169, Running: 4, Completed: 1847 [ 1h 15m ] +INFO: Idle: 14168, Running: 4, Completed: 1848 [ 1h 15m ] +INFO: Idle: 14167, Running: 4, Completed: 1849 [ 1h 15m ] +INFO: Idle: 14166, Running: 4, Completed: 1850 [ 1h 15m ] +INFO: Idle: 14163, Running: 4, Completed: 1853 [ 1h 16m ] +INFO: Idle: 14162, Running: 4, Completed: 1854 [ 1h 16m ] +INFO: Idle: 14161, Running: 4, Completed: 1855 [ 1h 16m ] +INFO: Idle: 14160, Running: 4, Completed: 1856 [ 1h 16m ] +INFO: Idle: 14159, Running: 4, Completed: 1857 [ 1h 16m ] +INFO: Idle: 14158, Running: 4, Completed: 1858 [ 1h 16m ] +INFO: Idle: 14157, Running: 4, Completed: 1859 [ 1h 16m ] +INFO: Idle: 14155, Running: 4, Completed: 1861 [ 1h 16m ] +INFO: Idle: 14154, Running: 4, Completed: 1862 [ 1h 17m ] +INFO: Idle: 14151, Running: 4, Completed: 1865 [ 1h 17m ] +INFO: Idle: 14150, Running: 4, Completed: 1866 [ 1h 17m ] +INFO: Idle: 14149, Running: 4, Completed: 1867 [ 1h 17m ] +INFO: Idle: 14148, Running: 4, Completed: 1868 [ 1h 17m ] +INFO: Idle: 14147, Running: 4, Completed: 1869 [ 1h 17m ] +INFO: Idle: 14146, Running: 4, Completed: 1870 [ 1h 17m ] +INFO: Idle: 14145, Running: 4, Completed: 1871 [ 1h 17m ] +INFO: Idle: 14144, Running: 4, Completed: 1872 [ 1h 18m ] +INFO: Idle: 14142, Running: 4, Completed: 1874 [ 1h 18m ] +INFO: Idle: 14141, Running: 4, Completed: 1875 [ 1h 18m ] +INFO: Idle: 14140, Running: 4, Completed: 1876 [ 1h 18m ] +INFO: Idle: 14139, Running: 4, Completed: 1877 [ 1h 18m ] +INFO: Idle: 14137, Running: 4, Completed: 1879 [ 1h 18m ] +INFO: Idle: 14136, Running: 4, Completed: 1880 [ 1h 18m ] +INFO: Idle: 14135, Running: 4, Completed: 1881 [ 1h 18m ] +INFO: Idle: 14133, Running: 4, Completed: 1883 [ 1h 18m ] +INFO: Idle: 14131, Running: 4, Completed: 1885 [ 1h 19m ] +INFO: Idle: 14129, Running: 4, Completed: 1887 [ 1h 19m ] +INFO: Idle: 14127, Running: 4, Completed: 1889 [ 1h 19m ] +INFO: Idle: 14124, Running: 4, Completed: 1892 [ 1h 19m ] +INFO: Idle: 14123, Running: 4, Completed: 1893 [ 1h 19m ] +INFO: Idle: 14121, Running: 4, Completed: 1895 [ 1h 20m ] +INFO: Idle: 14119, Running: 4, Completed: 1897 [ 1h 20m ] +INFO: Idle: 14117, Running: 4, Completed: 1899 [ 1h 20m ] +INFO: Idle: 14116, Running: 4, Completed: 1900 [ 1h 20m ] +INFO: Idle: 14115, Running: 4, Completed: 1901 [ 1h 20m ] +INFO: Idle: 14113, Running: 4, Completed: 1903 [ 1h 20m ] +INFO: Idle: 14112, Running: 4, Completed: 1904 [ 1h 20m ] +INFO: Idle: 14111, Running: 4, Completed: 1905 [ 1h 21m ] +INFO: Idle: 14109, Running: 4, Completed: 1907 [ 1h 21m ] +INFO: Idle: 14108, Running: 4, Completed: 1908 [ 1h 21m ] +INFO: Idle: 14107, Running: 4, Completed: 1909 [ 1h 21m ] +INFO: Idle: 14105, Running: 4, Completed: 1911 [ 1h 21m ] +INFO: Idle: 14103, Running: 4, Completed: 1913 [ 1h 21m ] +INFO: Idle: 14101, Running: 4, Completed: 1915 [ 1h 21m ] +INFO: Idle: 14100, Running: 4, Completed: 1916 [ 1h 22m ] +INFO: Idle: 14098, Running: 4, Completed: 1918 [ 1h 22m ] +INFO: Idle: 14096, Running: 4, Completed: 1920 [ 1h 22m ] +INFO: Idle: 14095, Running: 4, Completed: 1921 [ 1h 22m ] +INFO: Idle: 14093, Running: 4, Completed: 1923 [ 1h 22m ] +INFO: Idle: 14091, Running: 4, Completed: 1925 [ 1h 23m ] +INFO: Idle: 14088, Running: 4, Completed: 1928 [ 1h 23m ] +INFO: Idle: 14087, Running: 4, Completed: 1929 [ 1h 23m ] +INFO: Idle: 14085, Running: 4, Completed: 1931 [ 1h 23m ] +INFO: Idle: 14084, Running: 4, Completed: 1932 [ 1h 23m ] +INFO: Idle: 14082, Running: 4, Completed: 1934 [ 1h 23m ] +INFO: Idle: 14081, Running: 4, Completed: 1935 [ 1h 23m ] +INFO: Idle: 14079, Running: 4, Completed: 1937 [ 1h 24m ] +INFO: Idle: 14077, Running: 4, Completed: 1939 [ 1h 24m ] +INFO: Idle: 14075, Running: 4, Completed: 1941 [ 1h 24m ] +INFO: Idle: 14074, Running: 4, Completed: 1942 [ 1h 24m ] +INFO: Idle: 14073, Running: 4, Completed: 1943 [ 1h 24m ] +INFO: Idle: 14072, Running: 4, Completed: 1944 [ 1h 24m ] +INFO: Idle: 14071, Running: 4, Completed: 1945 [ 1h 24m ] +INFO: Idle: 14070, Running: 4, Completed: 1946 [ 1h 25m ] +INFO: Idle: 14069, Running: 4, Completed: 1947 [ 1h 25m ] +INFO: Idle: 14067, Running: 4, Completed: 1949 [ 1h 25m ] +INFO: Idle: 14065, Running: 4, Completed: 1951 [ 1h 25m ] +INFO: Idle: 14064, Running: 4, Completed: 1952 [ 1h 25m ] +INFO: Idle: 14062, Running: 4, Completed: 1954 [ 1h 25m ] +INFO: Idle: 14061, Running: 4, Completed: 1955 [ 1h 26m ] +INFO: Idle: 14060, Running: 4, Completed: 1956 [ 1h 26m ] +INFO: Idle: 14058, Running: 4, Completed: 1958 [ 1h 26m ] +INFO: Idle: 14057, Running: 4, Completed: 1959 [ 1h 26m ] +INFO: Idle: 14056, Running: 4, Completed: 1960 [ 1h 26m ] +INFO: Idle: 14055, Running: 4, Completed: 1961 [ 1h 26m ] +INFO: Idle: 14053, Running: 4, Completed: 1963 [ 1h 26m ] +INFO: Idle: 14052, Running: 4, Completed: 1964 [ 1h 26m ] +INFO: Idle: 14050, Running: 4, Completed: 1966 [ 1h 27m ] +INFO: Idle: 14048, Running: 4, Completed: 1968 [ 1h 27m ] +INFO: Idle: 14046, Running: 4, Completed: 1970 [ 1h 27m ] +INFO: Idle: 14045, Running: 4, Completed: 1971 [ 1h 27m ] +INFO: Idle: 14044, Running: 4, Completed: 1972 [ 1h 27m ] +INFO: Idle: 14043, Running: 4, Completed: 1973 [ 1h 27m ] +INFO: Idle: 14041, Running: 4, Completed: 1975 [ 1h 27m ] +INFO: Idle: 14040, Running: 4, Completed: 1976 [ 1h 28m ] +INFO: Idle: 14039, Running: 4, Completed: 1977 [ 1h 28m ] +INFO: Idle: 14037, Running: 4, Completed: 1979 [ 1h 28m ] +INFO: Idle: 14036, Running: 4, Completed: 1980 [ 1h 28m ] +INFO: Idle: 14035, Running: 4, Completed: 1981 [ 1h 28m ] +INFO: Idle: 14034, Running: 4, Completed: 1982 [ 1h 28m ] +INFO: Idle: 14032, Running: 4, Completed: 1984 [ 1h 28m ] +INFO: Idle: 14031, Running: 4, Completed: 1985 [ 1h 28m ] +INFO: Idle: 14030, Running: 4, Completed: 1986 [ 1h 28m ] +INFO: Idle: 14028, Running: 4, Completed: 1988 [ 1h 29m ] +INFO: Idle: 14027, Running: 4, Completed: 1989 [ 1h 29m ] +INFO: Idle: 14026, Running: 4, Completed: 1990 [ 1h 29m ] +INFO: Idle: 14025, Running: 4, Completed: 1991 [ 1h 29m ] +INFO: Idle: 14024, Running: 4, Completed: 1992 [ 1h 29m ] +INFO: Idle: 14023, Running: 4, Completed: 1993 [ 1h 29m ] +INFO: Idle: 14022, Running: 4, Completed: 1994 [ 1h 29m ] +INFO: Idle: 14021, Running: 4, Completed: 1995 [ 1h 29m ] +INFO: Idle: 14019, Running: 4, Completed: 1997 [ 1h 29m ] +INFO: Idle: 14018, Running: 4, Completed: 1998 [ 1h 30m ] +INFO: Idle: 14017, Running: 4, Completed: 1999 [ 1h 30m ] +INFO: Idle: 14016, Running: 4, Completed: 2000 [ 1h 30m ] +INFO: Idle: 14015, Running: 4, Completed: 2001 [ 1h 30m ] +INFO: Idle: 14014, Running: 4, Completed: 2002 [ 1h 30m ] +INFO: Idle: 14013, Running: 4, Completed: 2003 [ 1h 30m ] +INFO: Idle: 14011, Running: 4, Completed: 2005 [ 1h 30m ] +INFO: Idle: 14010, Running: 4, Completed: 2006 [ 1h 30m ] +INFO: Idle: 14009, Running: 4, Completed: 2007 [ 1h 30m ] +INFO: Idle: 14007, Running: 4, Completed: 2009 [ 1h 30m ] +INFO: Idle: 14005, Running: 4, Completed: 2011 [ 1h 31m ] +INFO: Idle: 14003, Running: 4, Completed: 2013 [ 1h 31m ] +INFO: Idle: 14001, Running: 4, Completed: 2015 [ 1h 31m ] +INFO: Idle: 13999, Running: 4, Completed: 2017 [ 1h 31m ] +INFO: Idle: 13997, Running: 4, Completed: 2019 [ 1h 31m ] +INFO: Idle: 13996, Running: 4, Completed: 2020 [ 1h 32m ] +INFO: Idle: 13995, Running: 4, Completed: 2021 [ 1h 32m ] +INFO: Idle: 13993, Running: 4, Completed: 2023 [ 1h 32m ] +INFO: Idle: 13992, Running: 4, Completed: 2024 [ 1h 32m ] +INFO: Idle: 13991, Running: 4, Completed: 2025 [ 1h 32m ] +INFO: Idle: 13990, Running: 4, Completed: 2026 [ 1h 32m ] +INFO: Idle: 13988, Running: 4, Completed: 2028 [ 1h 32m ] +INFO: Idle: 13987, Running: 4, Completed: 2029 [ 1h 32m ] +INFO: Idle: 13985, Running: 4, Completed: 2031 [ 1h 33m ] +INFO: Idle: 13983, Running: 4, Completed: 2033 [ 1h 33m ] +INFO: Idle: 13981, Running: 4, Completed: 2035 [ 1h 33m ] +INFO: Idle: 13980, Running: 4, Completed: 2036 [ 1h 33m ] +INFO: Idle: 13978, Running: 4, Completed: 2038 [ 1h 33m ] +INFO: Idle: 13976, Running: 4, Completed: 2040 [ 1h 33m ] +INFO: Idle: 13975, Running: 4, Completed: 2041 [ 1h 33m ] +INFO: Idle: 13973, Running: 4, Completed: 2043 [ 1h 34m ] +INFO: Idle: 13972, Running: 4, Completed: 2044 [ 1h 34m ] +INFO: Idle: 13971, Running: 4, Completed: 2045 [ 1h 34m ] +INFO: Idle: 13969, Running: 4, Completed: 2047 [ 1h 34m ] +INFO: Idle: 13968, Running: 4, Completed: 2048 [ 1h 34m ] +INFO: Idle: 13967, Running: 4, Completed: 2049 [ 1h 34m ] +INFO: Idle: 13965, Running: 4, Completed: 2051 [ 1h 35m ] +INFO: Idle: 13964, Running: 4, Completed: 2052 [ 1h 35m ] +INFO: Idle: 13961, Running: 4, Completed: 2055 [ 1h 35m ] +INFO: Idle: 13958, Running: 4, Completed: 2058 [ 1h 35m ] +INFO: Idle: 13957, Running: 4, Completed: 2059 [ 1h 35m ] +INFO: Idle: 13955, Running: 4, Completed: 2061 [ 1h 36m ] +INFO: Idle: 13953, Running: 4, Completed: 2063 [ 1h 36m ] +INFO: Idle: 13952, Running: 4, Completed: 2064 [ 1h 36m ] +INFO: Idle: 13951, Running: 4, Completed: 2065 [ 1h 36m ] +INFO: Idle: 13950, Running: 4, Completed: 2066 [ 1h 36m ] +INFO: Idle: 13949, Running: 4, Completed: 2067 [ 1h 36m ] +INFO: Idle: 13947, Running: 4, Completed: 2069 [ 1h 36m ] +INFO: Idle: 13945, Running: 4, Completed: 2071 [ 1h 37m ] +INFO: Idle: 13944, Running: 4, Completed: 2072 [ 1h 37m ] +INFO: Idle: 13943, Running: 4, Completed: 2073 [ 1h 37m ] +INFO: Idle: 13941, Running: 4, Completed: 2075 [ 1h 37m ] +INFO: Idle: 13940, Running: 4, Completed: 2076 [ 1h 37m ] +INFO: Idle: 13938, Running: 4, Completed: 2078 [ 1h 37m ] +INFO: Idle: 13937, Running: 4, Completed: 2079 [ 1h 37m ] +INFO: Idle: 13936, Running: 4, Completed: 2080 [ 1h 37m ] +INFO: Idle: 13935, Running: 4, Completed: 2081 [ 1h 38m ] +INFO: Idle: 13932, Running: 4, Completed: 2084 [ 1h 38m ] +INFO: Idle: 13930, Running: 4, Completed: 2086 [ 1h 38m ] +INFO: Idle: 13928, Running: 4, Completed: 2088 [ 1h 38m ] +INFO: Idle: 13926, Running: 4, Completed: 2090 [ 1h 38m ] +INFO: Idle: 13924, Running: 4, Completed: 2092 [ 1h 39m ] +INFO: Idle: 13923, Running: 4, Completed: 2093 [ 1h 39m ] +INFO: Idle: 13922, Running: 4, Completed: 2094 [ 1h 39m ] +INFO: Idle: 13921, Running: 4, Completed: 2095 [ 1h 39m ] +INFO: Idle: 13920, Running: 4, Completed: 2096 [ 1h 39m ] +INFO: Idle: 13919, Running: 4, Completed: 2097 [ 1h 39m ] +INFO: Idle: 13917, Running: 4, Completed: 2099 [ 1h 39m ] +INFO: Idle: 13916, Running: 4, Completed: 2100 [ 1h 39m ] +INFO: Idle: 13915, Running: 4, Completed: 2101 [ 1h 39m ] +INFO: Idle: 13914, Running: 4, Completed: 2102 [ 1h 40m ] +INFO: Idle: 13912, Running: 4, Completed: 2104 [ 1h 40m ] +INFO: Idle: 13911, Running: 4, Completed: 2105 [ 1h 40m ] +INFO: Idle: 13910, Running: 4, Completed: 2106 [ 1h 40m ] +INFO: Idle: 13909, Running: 4, Completed: 2107 [ 1h 40m ] +INFO: Idle: 13908, Running: 4, Completed: 2108 [ 1h 40m ] +INFO: Idle: 13907, Running: 4, Completed: 2109 [ 1h 40m ] +INFO: Idle: 13906, Running: 4, Completed: 2110 [ 1h 40m ] +INFO: Idle: 13905, Running: 4, Completed: 2111 [ 1h 40m ] +INFO: Idle: 13904, Running: 4, Completed: 2112 [ 1h 40m ] +INFO: Idle: 13903, Running: 4, Completed: 2113 [ 1h 41m ] +INFO: Idle: 13902, Running: 4, Completed: 2114 [ 1h 41m ] +INFO: Idle: 13901, Running: 4, Completed: 2115 [ 1h 41m ] +INFO: Idle: 13900, Running: 4, Completed: 2116 [ 1h 41m ] +INFO: Idle: 13899, Running: 4, Completed: 2117 [ 1h 41m ] +INFO: Idle: 13898, Running: 4, Completed: 2118 [ 1h 41m ] +INFO: Idle: 13897, Running: 4, Completed: 2119 [ 1h 41m ] +INFO: Idle: 13896, Running: 4, Completed: 2120 [ 1h 41m ] +INFO: Idle: 13895, Running: 4, Completed: 2121 [ 1h 41m ] +INFO: Idle: 13894, Running: 4, Completed: 2122 [ 1h 41m ] +INFO: Idle: 13893, Running: 4, Completed: 2123 [ 1h 41m ] +INFO: Idle: 13891, Running: 4, Completed: 2125 [ 1h 42m ] +INFO: Idle: 13890, Running: 4, Completed: 2126 [ 1h 42m ] +INFO: Idle: 13889, Running: 4, Completed: 2127 [ 1h 42m ] +INFO: Idle: 13887, Running: 4, Completed: 2129 [ 1h 42m ] +INFO: Idle: 13885, Running: 4, Completed: 2131 [ 1h 42m ] +INFO: Idle: 13883, Running: 4, Completed: 2133 [ 1h 42m ] +INFO: Idle: 13881, Running: 4, Completed: 2135 [ 1h 42m ] +INFO: Idle: 13879, Running: 4, Completed: 2137 [ 1h 43m ] +INFO: Idle: 13877, Running: 4, Completed: 2139 [ 1h 43m ] +INFO: Idle: 13876, Running: 4, Completed: 2140 [ 1h 43m ] +INFO: Idle: 13875, Running: 4, Completed: 2141 [ 1h 43m ] +INFO: Idle: 13873, Running: 4, Completed: 2143 [ 1h 43m ] +INFO: Idle: 13872, Running: 4, Completed: 2144 [ 1h 43m ] +INFO: Idle: 13871, Running: 4, Completed: 2145 [ 1h 43m ] +INFO: Idle: 13869, Running: 4, Completed: 2147 [ 1h 43m ] +INFO: Idle: 13868, Running: 4, Completed: 2148 [ 1h 44m ] +INFO: Idle: 13867, Running: 4, Completed: 2149 [ 1h 44m ] +INFO: Idle: 13865, Running: 4, Completed: 2151 [ 1h 44m ] +INFO: Idle: 13864, Running: 4, Completed: 2152 [ 1h 44m ] +INFO: Idle: 13863, Running: 4, Completed: 2153 [ 1h 44m ] +INFO: Idle: 13862, Running: 4, Completed: 2154 [ 1h 44m ] +INFO: Idle: 13861, Running: 4, Completed: 2155 [ 1h 44m ] +INFO: Idle: 13860, Running: 4, Completed: 2156 [ 1h 44m ] +INFO: Idle: 13858, Running: 4, Completed: 2158 [ 1h 45m ] +INFO: Idle: 13857, Running: 4, Completed: 2159 [ 1h 45m ] +INFO: Idle: 13856, Running: 4, Completed: 2160 [ 1h 45m ] +INFO: Idle: 13855, Running: 4, Completed: 2161 [ 1h 45m ] +INFO: Idle: 13854, Running: 4, Completed: 2162 [ 1h 45m ] +INFO: Idle: 13852, Running: 4, Completed: 2164 [ 1h 45m ] +INFO: Idle: 13851, Running: 4, Completed: 2165 [ 1h 45m ] +INFO: Idle: 13849, Running: 4, Completed: 2167 [ 1h 46m ] +INFO: Idle: 13848, Running: 4, Completed: 2168 [ 1h 46m ] +INFO: Idle: 13847, Running: 4, Completed: 2169 [ 1h 46m ] +INFO: Idle: 13846, Running: 4, Completed: 2170 [ 1h 46m ] +INFO: Idle: 13845, Running: 4, Completed: 2171 [ 1h 46m ] +INFO: Idle: 13844, Running: 4, Completed: 2172 [ 1h 46m ] +INFO: Idle: 13842, Running: 4, Completed: 2174 [ 1h 46m ] +INFO: Idle: 13841, Running: 4, Completed: 2175 [ 1h 46m ] +INFO: Idle: 13840, Running: 4, Completed: 2176 [ 1h 46m ] +INFO: Idle: 13838, Running: 4, Completed: 2178 [ 1h 46m ] +INFO: Idle: 13837, Running: 4, Completed: 2179 [ 1h 47m ] +INFO: Idle: 13835, Running: 4, Completed: 2181 [ 1h 47m ] +INFO: Idle: 13834, Running: 4, Completed: 2182 [ 1h 47m ] +INFO: Idle: 13833, Running: 4, Completed: 2183 [ 1h 47m ] +INFO: Idle: 13832, Running: 4, Completed: 2184 [ 1h 47m ] +INFO: Idle: 13830, Running: 4, Completed: 2186 [ 1h 47m ] +INFO: Idle: 13828, Running: 4, Completed: 2188 [ 1h 47m ] +INFO: Idle: 13826, Running: 4, Completed: 2190 [ 1h 48m ] +INFO: Idle: 13824, Running: 4, Completed: 2192 [ 1h 48m ] +INFO: Idle: 13821, Running: 4, Completed: 2195 [ 1h 48m ] +INFO: Idle: 13820, Running: 4, Completed: 2196 [ 1h 48m ] +INFO: Idle: 13819, Running: 4, Completed: 2197 [ 1h 48m ] +INFO: Idle: 13816, Running: 4, Completed: 2200 [ 1h 48m ] +INFO: Idle: 13815, Running: 4, Completed: 2201 [ 1h 48m ] +INFO: Idle: 13814, Running: 4, Completed: 2202 [ 1h 48m ] +INFO: Idle: 13813, Running: 4, Completed: 2203 [ 1h 49m ] +INFO: Idle: 13812, Running: 4, Completed: 2204 [ 1h 49m ] +INFO: Idle: 13811, Running: 4, Completed: 2205 [ 1h 49m ] +INFO: Idle: 13809, Running: 4, Completed: 2207 [ 1h 49m ] +INFO: Idle: 13807, Running: 4, Completed: 2209 [ 1h 49m ] +INFO: Idle: 13805, Running: 4, Completed: 2211 [ 1h 49m ] +INFO: Idle: 13802, Running: 4, Completed: 2214 [ 1h 49m ] +INFO: Idle: 13801, Running: 4, Completed: 2215 [ 1h 50m ] +INFO: Idle: 13800, Running: 4, Completed: 2216 [ 1h 50m ] +INFO: Idle: 13798, Running: 4, Completed: 2218 [ 1h 50m ] +INFO: Idle: 13796, Running: 4, Completed: 2220 [ 1h 50m ] +INFO: Idle: 13795, Running: 4, Completed: 2221 [ 1h 50m ] +INFO: Idle: 13794, Running: 4, Completed: 2222 [ 1h 50m ] +INFO: Idle: 13793, Running: 4, Completed: 2223 [ 1h 50m ] +INFO: Idle: 13791, Running: 4, Completed: 2225 [ 1h 51m ] +INFO: Idle: 13790, Running: 4, Completed: 2226 [ 1h 51m ] +INFO: Idle: 13788, Running: 4, Completed: 2228 [ 1h 51m ] +INFO: Idle: 13787, Running: 4, Completed: 2229 [ 1h 51m ] +INFO: Idle: 13786, Running: 4, Completed: 2230 [ 1h 51m ] +INFO: Idle: 13785, Running: 4, Completed: 2231 [ 1h 51m ] +INFO: Idle: 13784, Running: 4, Completed: 2232 [ 1h 51m ] +INFO: Idle: 13783, Running: 4, Completed: 2233 [ 1h 51m ] +INFO: Idle: 13781, Running: 4, Completed: 2235 [ 1h 51m ] +INFO: Idle: 13780, Running: 4, Completed: 2236 [ 1h 51m ] +INFO: Idle: 13779, Running: 4, Completed: 2237 [ 1h 52m ] +INFO: Idle: 13777, Running: 4, Completed: 2239 [ 1h 52m ] +INFO: Idle: 13776, Running: 4, Completed: 2240 [ 1h 52m ] +INFO: Idle: 13775, Running: 4, Completed: 2241 [ 1h 52m ] +INFO: Idle: 13772, Running: 4, Completed: 2244 [ 1h 52m ] +INFO: Idle: 13771, Running: 4, Completed: 2245 [ 1h 52m ] +INFO: Idle: 13767, Running: 4, Completed: 2249 [ 1h 52m ] +INFO: Idle: 13765, Running: 4, Completed: 2251 [ 1h 53m ] +INFO: Idle: 13763, Running: 4, Completed: 2253 [ 1h 53m ] +INFO: Idle: 13761, Running: 4, Completed: 2255 [ 1h 53m ] +INFO: Idle: 13759, Running: 4, Completed: 2257 [ 1h 53m ] +INFO: Idle: 13757, Running: 4, Completed: 2259 [ 1h 53m ] +INFO: Idle: 13756, Running: 4, Completed: 2260 [ 1h 54m ] +INFO: Idle: 13755, Running: 4, Completed: 2261 [ 1h 54m ] +INFO: Idle: 13753, Running: 4, Completed: 2263 [ 1h 54m ] +INFO: Idle: 13752, Running: 4, Completed: 2264 [ 1h 54m ] +INFO: Idle: 13751, Running: 4, Completed: 2265 [ 1h 54m ] +INFO: Idle: 13750, Running: 4, Completed: 2266 [ 1h 54m ] +INFO: Idle: 13748, Running: 4, Completed: 2268 [ 1h 54m ] +INFO: Idle: 13747, Running: 4, Completed: 2269 [ 1h 54m ] +INFO: Idle: 13746, Running: 4, Completed: 2270 [ 1h 55m ] +INFO: Idle: 13744, Running: 4, Completed: 2272 [ 1h 55m ] +INFO: Idle: 13743, Running: 4, Completed: 2273 [ 1h 55m ] +INFO: Idle: 13741, Running: 4, Completed: 2275 [ 1h 55m ] +INFO: Idle: 13740, Running: 4, Completed: 2276 [ 1h 55m ] +INFO: Idle: 13738, Running: 4, Completed: 2278 [ 1h 55m ] +INFO: Idle: 13736, Running: 4, Completed: 2280 [ 1h 55m ] +INFO: Idle: 13735, Running: 4, Completed: 2281 [ 1h 56m ] +INFO: Idle: 13733, Running: 4, Completed: 2283 [ 1h 56m ] +INFO: Idle: 13731, Running: 4, Completed: 2285 [ 1h 56m ] +INFO: Idle: 13729, Running: 4, Completed: 2287 [ 1h 56m ] +INFO: Idle: 13728, Running: 4, Completed: 2288 [ 1h 56m ] +INFO: Idle: 13727, Running: 4, Completed: 2289 [ 1h 56m ] +INFO: Idle: 13726, Running: 4, Completed: 2290 [ 1h 56m ] +INFO: Idle: 13725, Running: 4, Completed: 2291 [ 1h 57m ] +INFO: Idle: 13724, Running: 4, Completed: 2292 [ 1h 57m ] +INFO: Idle: 13722, Running: 4, Completed: 2294 [ 1h 57m ] +INFO: Idle: 13721, Running: 4, Completed: 2295 [ 1h 57m ] +INFO: Idle: 13718, Running: 4, Completed: 2298 [ 1h 57m ] +INFO: Idle: 13717, Running: 4, Completed: 2299 [ 1h 57m ] +INFO: Idle: 13716, Running: 4, Completed: 2300 [ 1h 57m ] +INFO: Idle: 13714, Running: 4, Completed: 2302 [ 1h 57m ] +INFO: Idle: 13713, Running: 4, Completed: 2303 [ 1h 58m ] +INFO: Idle: 13712, Running: 4, Completed: 2304 [ 1h 58m ] +INFO: Idle: 13710, Running: 4, Completed: 2306 [ 1h 58m ] +INFO: Idle: 13708, Running: 4, Completed: 2308 [ 1h 58m ] +INFO: Idle: 13704, Running: 4, Completed: 2312 [ 1h 58m ] +INFO: Idle: 13702, Running: 4, Completed: 2314 [ 1h 58m ] +INFO: Idle: 13701, Running: 4, Completed: 2315 [ 1h 58m ] +INFO: Idle: 13700, Running: 4, Completed: 2316 [ 1h 59m ] +INFO: Idle: 13699, Running: 4, Completed: 2317 [ 1h 59m ] +INFO: Idle: 13696, Running: 4, Completed: 2320 [ 1h 59m ] +INFO: Idle: 13695, Running: 4, Completed: 2321 [ 1h 59m ] +INFO: Idle: 13693, Running: 4, Completed: 2323 [ 1h 59m ] +INFO: Idle: 13692, Running: 4, Completed: 2324 [ 1h 59m ] +INFO: Idle: 13691, Running: 4, Completed: 2325 [ 1h 59m ] +INFO: Idle: 13689, Running: 4, Completed: 2327 [ 1h 59m ] +INFO: Idle: 13688, Running: 4, Completed: 2328 [ 2h 0m ] +INFO: Idle: 13687, Running: 4, Completed: 2329 [ 2h 0m ] +INFO: Idle: 13685, Running: 4, Completed: 2331 [ 2h 0m ] +INFO: Idle: 13684, Running: 4, Completed: 2332 [ 2h 0m ] +INFO: Idle: 13682, Running: 4, Completed: 2334 [ 2h 0m ] +INFO: Idle: 13681, Running: 4, Completed: 2335 [ 2h 0m ] +INFO: Idle: 13680, Running: 4, Completed: 2336 [ 2h 0m ] +INFO: Idle: 13677, Running: 4, Completed: 2339 [ 2h 0m ] +INFO: Idle: 13676, Running: 4, Completed: 2340 [ 2h 1m ] +INFO: Idle: 13675, Running: 4, Completed: 2341 [ 2h 1m ] +INFO: Idle: 13674, Running: 4, Completed: 2342 [ 2h 1m ] +INFO: Idle: 13672, Running: 4, Completed: 2344 [ 2h 1m ] +INFO: Idle: 13671, Running: 4, Completed: 2345 [ 2h 1m ] +INFO: Idle: 13670, Running: 4, Completed: 2346 [ 2h 1m ] +INFO: Idle: 13668, Running: 4, Completed: 2348 [ 2h 1m ] +INFO: Idle: 13667, Running: 4, Completed: 2349 [ 2h 1m ] +INFO: Idle: 13666, Running: 4, Completed: 2350 [ 2h 2m ] +INFO: Idle: 13665, Running: 4, Completed: 2351 [ 2h 2m ] +INFO: Idle: 13664, Running: 4, Completed: 2352 [ 2h 2m ] +INFO: Idle: 13663, Running: 4, Completed: 2353 [ 2h 2m ] +INFO: Idle: 13661, Running: 4, Completed: 2355 [ 2h 2m ] +INFO: Idle: 13660, Running: 4, Completed: 2356 [ 2h 2m ] +INFO: Idle: 13659, Running: 4, Completed: 2357 [ 2h 2m ] +INFO: Idle: 13657, Running: 4, Completed: 2359 [ 2h 2m ] +INFO: Idle: 13656, Running: 4, Completed: 2360 [ 2h 2m ] +INFO: Idle: 13655, Running: 4, Completed: 2361 [ 2h 2m ] +INFO: Idle: 13651, Running: 4, Completed: 2365 [ 2h 3m ] +INFO: Idle: 13647, Running: 4, Completed: 2369 [ 2h 3m ] +INFO: Idle: 13645, Running: 4, Completed: 2371 [ 2h 3m ] +INFO: Idle: 13644, Running: 4, Completed: 2372 [ 2h 3m ] +INFO: Idle: 13643, Running: 4, Completed: 2373 [ 2h 3m ] +INFO: Idle: 13641, Running: 4, Completed: 2375 [ 2h 3m ] +INFO: Idle: 13640, Running: 4, Completed: 2376 [ 2h 4m ] +INFO: Idle: 13639, Running: 4, Completed: 2377 [ 2h 4m ] +INFO: Idle: 13637, Running: 4, Completed: 2379 [ 2h 4m ] +INFO: Idle: 13636, Running: 4, Completed: 2380 [ 2h 4m ] +INFO: Idle: 13635, Running: 4, Completed: 2381 [ 2h 4m ] +INFO: Idle: 13634, Running: 4, Completed: 2382 [ 2h 4m ] +INFO: Idle: 13632, Running: 4, Completed: 2384 [ 2h 4m ] +INFO: Idle: 13631, Running: 4, Completed: 2385 [ 2h 5m ] +INFO: Idle: 13628, Running: 4, Completed: 2388 [ 2h 5m ] +INFO: Idle: 13627, Running: 4, Completed: 2389 [ 2h 5m ] +INFO: Idle: 13626, Running: 4, Completed: 2390 [ 2h 5m ] +INFO: Idle: 13625, Running: 4, Completed: 2391 [ 2h 5m ] +INFO: Idle: 13624, Running: 4, Completed: 2392 [ 2h 5m ] +INFO: Idle: 13623, Running: 4, Completed: 2393 [ 2h 5m ] +INFO: Idle: 13622, Running: 4, Completed: 2394 [ 2h 5m ] +INFO: Idle: 13621, Running: 4, Completed: 2395 [ 2h 6m ] +INFO: Idle: 13619, Running: 4, Completed: 2397 [ 2h 6m ] +INFO: Idle: 13618, Running: 4, Completed: 2398 [ 2h 6m ] +INFO: Idle: 13616, Running: 4, Completed: 2400 [ 2h 6m ] +INFO: Idle: 13614, Running: 4, Completed: 2402 [ 2h 6m ] +INFO: Idle: 13613, Running: 4, Completed: 2403 [ 2h 6m ] +INFO: Idle: 13610, Running: 4, Completed: 2406 [ 2h 7m ] +INFO: Idle: 13609, Running: 4, Completed: 2407 [ 2h 7m ] +INFO: Idle: 13608, Running: 4, Completed: 2408 [ 2h 7m ] +INFO: Idle: 13607, Running: 4, Completed: 2409 [ 2h 7m ] +INFO: Idle: 13606, Running: 4, Completed: 2410 [ 2h 7m ] +INFO: Idle: 13604, Running: 4, Completed: 2412 [ 2h 7m ] +INFO: Idle: 13603, Running: 4, Completed: 2413 [ 2h 7m ] +INFO: Idle: 13602, Running: 4, Completed: 2414 [ 2h 7m ] +INFO: Idle: 13601, Running: 4, Completed: 2415 [ 2h 7m ] +INFO: Idle: 13600, Running: 4, Completed: 2416 [ 2h 8m ] +INFO: Idle: 13599, Running: 4, Completed: 2417 [ 2h 8m ] +INFO: Idle: 13597, Running: 4, Completed: 2419 [ 2h 8m ] +INFO: Idle: 13595, Running: 4, Completed: 2421 [ 2h 8m ] +INFO: Idle: 13594, Running: 4, Completed: 2422 [ 2h 8m ] +INFO: Idle: 13593, Running: 4, Completed: 2423 [ 2h 8m ] +INFO: Idle: 13592, Running: 4, Completed: 2424 [ 2h 8m ] +INFO: Idle: 13590, Running: 4, Completed: 2426 [ 2h 8m ] +INFO: Idle: 13588, Running: 4, Completed: 2428 [ 2h 8m ] +INFO: Idle: 13587, Running: 4, Completed: 2429 [ 2h 9m ] +INFO: Idle: 13586, Running: 4, Completed: 2430 [ 2h 9m ] +INFO: Idle: 13585, Running: 4, Completed: 2431 [ 2h 9m ] +INFO: Idle: 13583, Running: 4, Completed: 2433 [ 2h 9m ] +INFO: Idle: 13581, Running: 4, Completed: 2435 [ 2h 9m ] +INFO: Idle: 13579, Running: 4, Completed: 2437 [ 2h 9m ] +INFO: Idle: 13578, Running: 4, Completed: 2438 [ 2h 9m ] +INFO: Idle: 13577, Running: 4, Completed: 2439 [ 2h 9m ] +INFO: Idle: 13576, Running: 4, Completed: 2440 [ 2h 9m ] +INFO: Idle: 13575, Running: 4, Completed: 2441 [ 2h 9m ] +INFO: Idle: 13574, Running: 4, Completed: 2442 [ 2h 10m ] +INFO: Idle: 13572, Running: 4, Completed: 2444 [ 2h 10m ] +INFO: Idle: 13571, Running: 4, Completed: 2445 [ 2h 10m ] +INFO: Idle: 13570, Running: 4, Completed: 2446 [ 2h 10m ] +INFO: Idle: 13569, Running: 4, Completed: 2447 [ 2h 10m ] +INFO: Idle: 13567, Running: 4, Completed: 2449 [ 2h 10m ] +INFO: Idle: 13566, Running: 4, Completed: 2450 [ 2h 10m ] +INFO: Idle: 13565, Running: 4, Completed: 2451 [ 2h 10m ] +INFO: Idle: 13563, Running: 4, Completed: 2453 [ 2h 11m ] +INFO: Idle: 13561, Running: 4, Completed: 2455 [ 2h 11m ] +INFO: Idle: 13560, Running: 4, Completed: 2456 [ 2h 11m ] +INFO: Idle: 13559, Running: 4, Completed: 2457 [ 2h 11m ] +INFO: Idle: 13558, Running: 4, Completed: 2458 [ 2h 11m ] +INFO: Idle: 13557, Running: 4, Completed: 2459 [ 2h 11m ] +INFO: Idle: 13556, Running: 4, Completed: 2460 [ 2h 11m ] +INFO: Idle: 13555, Running: 4, Completed: 2461 [ 2h 11m ] +INFO: Idle: 13554, Running: 4, Completed: 2462 [ 2h 11m ] +INFO: Idle: 13551, Running: 4, Completed: 2465 [ 2h 12m ] +INFO: Idle: 13549, Running: 4, Completed: 2467 [ 2h 12m ] +INFO: Idle: 13547, Running: 4, Completed: 2469 [ 2h 12m ] +INFO: Idle: 13545, Running: 4, Completed: 2471 [ 2h 12m ] +INFO: Idle: 13544, Running: 4, Completed: 2472 [ 2h 12m ] +INFO: Idle: 13543, Running: 4, Completed: 2473 [ 2h 12m ] +INFO: Idle: 13541, Running: 4, Completed: 2475 [ 2h 13m ] +INFO: Idle: 13538, Running: 4, Completed: 2478 [ 2h 13m ] +INFO: Idle: 13537, Running: 4, Completed: 2479 [ 2h 13m ] +INFO: Idle: 13533, Running: 4, Completed: 2483 [ 2h 13m ] +INFO: Idle: 13532, Running: 4, Completed: 2484 [ 2h 13m ] +INFO: Idle: 13530, Running: 4, Completed: 2486 [ 2h 13m ] +INFO: Idle: 13529, Running: 4, Completed: 2487 [ 2h 13m ] +INFO: Idle: 13528, Running: 4, Completed: 2488 [ 2h 13m ] +INFO: Idle: 13527, Running: 4, Completed: 2489 [ 2h 14m ] +INFO: Idle: 13524, Running: 4, Completed: 2492 [ 2h 14m ] +INFO: Idle: 13523, Running: 4, Completed: 2493 [ 2h 14m ] +INFO: Idle: 13522, Running: 4, Completed: 2494 [ 2h 14m ] +INFO: Idle: 13521, Running: 4, Completed: 2495 [ 2h 14m ] +INFO: Idle: 13519, Running: 4, Completed: 2497 [ 2h 14m ] +INFO: Idle: 13518, Running: 4, Completed: 2498 [ 2h 14m ] +INFO: Idle: 13517, Running: 4, Completed: 2499 [ 2h 15m ] +INFO: Idle: 13516, Running: 4, Completed: 2500 [ 2h 15m ] +INFO: Idle: 13514, Running: 4, Completed: 2502 [ 2h 15m ] +INFO: Idle: 13513, Running: 4, Completed: 2503 [ 2h 15m ] +INFO: Idle: 13512, Running: 4, Completed: 2504 [ 2h 15m ] +INFO: Idle: 13511, Running: 4, Completed: 2505 [ 2h 15m ] +INFO: Idle: 13509, Running: 4, Completed: 2507 [ 2h 15m ] +INFO: Idle: 13508, Running: 4, Completed: 2508 [ 2h 15m ] +INFO: Idle: 13507, Running: 4, Completed: 2509 [ 2h 16m ] +INFO: Idle: 13506, Running: 4, Completed: 2510 [ 2h 16m ] +INFO: Idle: 13505, Running: 4, Completed: 2511 [ 2h 16m ] +INFO: Idle: 13503, Running: 4, Completed: 2513 [ 2h 16m ] +INFO: Idle: 13501, Running: 4, Completed: 2515 [ 2h 16m ] +INFO: Idle: 13499, Running: 4, Completed: 2517 [ 2h 16m ] +INFO: Idle: 13498, Running: 4, Completed: 2518 [ 2h 16m ] +INFO: Idle: 13497, Running: 4, Completed: 2519 [ 2h 16m ] +INFO: Idle: 13496, Running: 4, Completed: 2520 [ 2h 17m ] +INFO: Idle: 13495, Running: 4, Completed: 2521 [ 2h 17m ] +INFO: Idle: 13494, Running: 4, Completed: 2522 [ 2h 17m ] +INFO: Idle: 13493, Running: 4, Completed: 2523 [ 2h 17m ] +INFO: Idle: 13491, Running: 4, Completed: 2525 [ 2h 17m ] +INFO: Idle: 13489, Running: 4, Completed: 2527 [ 2h 17m ] +INFO: Idle: 13488, Running: 4, Completed: 2528 [ 2h 17m ] +INFO: Idle: 13487, Running: 4, Completed: 2529 [ 2h 17m ] +INFO: Idle: 13486, Running: 4, Completed: 2530 [ 2h 18m ] +INFO: Idle: 13485, Running: 4, Completed: 2531 [ 2h 18m ] +INFO: Idle: 13484, Running: 4, Completed: 2532 [ 2h 18m ] +INFO: Idle: 13482, Running: 4, Completed: 2534 [ 2h 18m ] +INFO: Idle: 13481, Running: 4, Completed: 2535 [ 2h 18m ] +INFO: Idle: 13478, Running: 4, Completed: 2538 [ 2h 18m ] +INFO: Idle: 13477, Running: 4, Completed: 2539 [ 2h 18m ] +INFO: Idle: 13476, Running: 4, Completed: 2540 [ 2h 19m ] +INFO: Idle: 13474, Running: 4, Completed: 2542 [ 2h 19m ] +INFO: Idle: 13473, Running: 4, Completed: 2543 [ 2h 19m ] +INFO: Idle: 13472, Running: 4, Completed: 2544 [ 2h 19m ] +INFO: Idle: 13470, Running: 4, Completed: 2546 [ 2h 19m ] +INFO: Idle: 13468, Running: 4, Completed: 2548 [ 2h 19m ] +INFO: Idle: 13466, Running: 4, Completed: 2550 [ 2h 19m ] +INFO: Idle: 13464, Running: 4, Completed: 2552 [ 2h 19m ] +INFO: Idle: 13461, Running: 4, Completed: 2555 [ 2h 20m ] +INFO: Idle: 13460, Running: 4, Completed: 2556 [ 2h 20m ] +INFO: Idle: 13459, Running: 4, Completed: 2557 [ 2h 20m ] +INFO: Idle: 13456, Running: 4, Completed: 2560 [ 2h 20m ] +INFO: Idle: 13455, Running: 4, Completed: 2561 [ 2h 20m ] +INFO: Idle: 13454, Running: 4, Completed: 2562 [ 2h 20m ] +INFO: Idle: 13453, Running: 4, Completed: 2563 [ 2h 20m ] +INFO: Idle: 13452, Running: 4, Completed: 2564 [ 2h 20m ] +INFO: Idle: 13451, Running: 4, Completed: 2565 [ 2h 20m ] +INFO: Idle: 13449, Running: 4, Completed: 2567 [ 2h 21m ] +INFO: Idle: 13448, Running: 4, Completed: 2568 [ 2h 21m ] +INFO: Idle: 13447, Running: 4, Completed: 2569 [ 2h 21m ] +INFO: Idle: 13445, Running: 4, Completed: 2571 [ 2h 21m ] +INFO: Idle: 13443, Running: 4, Completed: 2573 [ 2h 21m ] +INFO: Idle: 13442, Running: 4, Completed: 2574 [ 2h 21m ] +INFO: Idle: 13441, Running: 4, Completed: 2575 [ 2h 21m ] +INFO: Idle: 13440, Running: 4, Completed: 2576 [ 2h 21m ] +INFO: Idle: 13438, Running: 4, Completed: 2578 [ 2h 22m ] +INFO: Idle: 13436, Running: 4, Completed: 2580 [ 2h 22m ] +INFO: Idle: 13435, Running: 4, Completed: 2581 [ 2h 22m ] +INFO: Idle: 13434, Running: 4, Completed: 2582 [ 2h 22m ] +INFO: Idle: 13433, Running: 4, Completed: 2583 [ 2h 22m ] +INFO: Idle: 13432, Running: 4, Completed: 2584 [ 2h 22m ] +INFO: Idle: 13431, Running: 4, Completed: 2585 [ 2h 22m ] +INFO: Idle: 13430, Running: 4, Completed: 2586 [ 2h 22m ] +INFO: Idle: 13428, Running: 4, Completed: 2588 [ 2h 22m ] +INFO: Idle: 13427, Running: 4, Completed: 2589 [ 2h 23m ] +INFO: Idle: 13426, Running: 4, Completed: 2590 [ 2h 23m ] +INFO: Idle: 13425, Running: 4, Completed: 2591 [ 2h 23m ] +INFO: Idle: 13424, Running: 4, Completed: 2592 [ 2h 23m ] +INFO: Idle: 13423, Running: 4, Completed: 2593 [ 2h 23m ] +INFO: Idle: 13422, Running: 4, Completed: 2594 [ 2h 23m ] +INFO: Idle: 13420, Running: 4, Completed: 2596 [ 2h 23m ] +INFO: Idle: 13419, Running: 4, Completed: 2597 [ 2h 23m ] +INFO: Idle: 13417, Running: 4, Completed: 2599 [ 2h 23m ] +INFO: Idle: 13416, Running: 4, Completed: 2600 [ 2h 23m ] +INFO: Idle: 13415, Running: 4, Completed: 2601 [ 2h 24m ] +INFO: Idle: 13411, Running: 4, Completed: 2605 [ 2h 24m ] +INFO: Idle: 13407, Running: 4, Completed: 2609 [ 2h 24m ] +INFO: Idle: 13405, Running: 4, Completed: 2611 [ 2h 24m ] +INFO: Idle: 13403, Running: 4, Completed: 2613 [ 2h 25m ] +INFO: Idle: 13401, Running: 4, Completed: 2615 [ 2h 25m ] +INFO: Idle: 13399, Running: 4, Completed: 2617 [ 2h 25m ] +INFO: Idle: 13397, Running: 4, Completed: 2619 [ 2h 25m ] +INFO: Idle: 13396, Running: 4, Completed: 2620 [ 2h 25m ] +INFO: Idle: 13395, Running: 4, Completed: 2621 [ 2h 25m ] +INFO: Idle: 13393, Running: 4, Completed: 2623 [ 2h 25m ] +INFO: Idle: 13392, Running: 4, Completed: 2624 [ 2h 26m ] +INFO: Idle: 13391, Running: 4, Completed: 2625 [ 2h 26m ] +INFO: Idle: 13388, Running: 4, Completed: 2628 [ 2h 26m ] +INFO: Idle: 13387, Running: 4, Completed: 2629 [ 2h 26m ] +INFO: Idle: 13385, Running: 4, Completed: 2631 [ 2h 26m ] +INFO: Idle: 13383, Running: 4, Completed: 2633 [ 2h 26m ] +INFO: Idle: 13382, Running: 4, Completed: 2634 [ 2h 26m ] +INFO: Idle: 13381, Running: 4, Completed: 2635 [ 2h 26m ] +INFO: Idle: 13379, Running: 4, Completed: 2637 [ 2h 26m ] +INFO: Idle: 13378, Running: 4, Completed: 2638 [ 2h 27m ] +INFO: Idle: 13377, Running: 4, Completed: 2639 [ 2h 27m ] +INFO: Idle: 13376, Running: 4, Completed: 2640 [ 2h 27m ] +INFO: Idle: 13375, Running: 4, Completed: 2641 [ 2h 27m ] +INFO: Idle: 13373, Running: 4, Completed: 2643 [ 2h 27m ] +INFO: Idle: 13372, Running: 4, Completed: 2644 [ 2h 27m ] +INFO: Idle: 13371, Running: 4, Completed: 2645 [ 2h 27m ] +INFO: Idle: 13368, Running: 4, Completed: 2648 [ 2h 27m ] +INFO: Idle: 13367, Running: 4, Completed: 2649 [ 2h 28m ] +INFO: Idle: 13366, Running: 4, Completed: 2650 [ 2h 28m ] +INFO: Idle: 13364, Running: 4, Completed: 2652 [ 2h 28m ] +INFO: Idle: 13362, Running: 4, Completed: 2654 [ 2h 28m ] +INFO: Idle: 13361, Running: 4, Completed: 2655 [ 2h 28m ] +INFO: Idle: 13358, Running: 4, Completed: 2658 [ 2h 28m ] +INFO: Idle: 13357, Running: 4, Completed: 2659 [ 2h 29m ] +INFO: Idle: 13356, Running: 4, Completed: 2660 [ 2h 29m ] +INFO: Idle: 13353, Running: 4, Completed: 2663 [ 2h 29m ] +INFO: Idle: 13352, Running: 4, Completed: 2664 [ 2h 29m ] +INFO: Idle: 13351, Running: 4, Completed: 2665 [ 2h 29m ] +INFO: Idle: 13350, Running: 4, Completed: 2666 [ 2h 29m ] +INFO: Idle: 13349, Running: 4, Completed: 2667 [ 2h 29m ] +INFO: Idle: 13348, Running: 4, Completed: 2668 [ 2h 30m ] +INFO: Idle: 13347, Running: 4, Completed: 2669 [ 2h 30m ] +INFO: Idle: 13345, Running: 4, Completed: 2671 [ 2h 30m ] +INFO: Idle: 13344, Running: 4, Completed: 2672 [ 2h 30m ] +INFO: Idle: 13342, Running: 4, Completed: 2674 [ 2h 30m ] +INFO: Idle: 13341, Running: 4, Completed: 2675 [ 2h 30m ] +INFO: Idle: 13340, Running: 4, Completed: 2676 [ 2h 30m ] +INFO: Idle: 13339, Running: 4, Completed: 2677 [ 2h 30m ] +INFO: Idle: 13338, Running: 4, Completed: 2678 [ 2h 30m ] +INFO: Idle: 13337, Running: 4, Completed: 2679 [ 2h 30m ] +INFO: Idle: 13336, Running: 4, Completed: 2680 [ 2h 31m ] +INFO: Idle: 13335, Running: 4, Completed: 2681 [ 2h 31m ] +INFO: Idle: 13332, Running: 4, Completed: 2684 [ 2h 31m ] +INFO: Idle: 13330, Running: 4, Completed: 2686 [ 2h 31m ] +INFO: Idle: 13328, Running: 4, Completed: 2688 [ 2h 31m ] +INFO: Idle: 13327, Running: 4, Completed: 2689 [ 2h 31m ] +INFO: Idle: 13326, Running: 4, Completed: 2690 [ 2h 31m ] +INFO: Idle: 13325, Running: 4, Completed: 2691 [ 2h 32m ] +INFO: Idle: 13323, Running: 4, Completed: 2693 [ 2h 32m ] +INFO: Idle: 13322, Running: 4, Completed: 2694 [ 2h 32m ] +INFO: Idle: 13321, Running: 4, Completed: 2695 [ 2h 32m ] +INFO: Idle: 13320, Running: 4, Completed: 2696 [ 2h 32m ] +INFO: Idle: 13318, Running: 4, Completed: 2698 [ 2h 32m ] +INFO: Idle: 13317, Running: 4, Completed: 2699 [ 2h 32m ] +INFO: Idle: 13316, Running: 4, Completed: 2700 [ 2h 32m ] +INFO: Idle: 13315, Running: 4, Completed: 2701 [ 2h 32m ] +INFO: Idle: 13314, Running: 4, Completed: 2702 [ 2h 33m ] +INFO: Idle: 13312, Running: 4, Completed: 2704 [ 2h 33m ] +INFO: Idle: 13311, Running: 4, Completed: 2705 [ 2h 33m ] +INFO: Idle: 13310, Running: 4, Completed: 2706 [ 2h 33m ] +INFO: Idle: 13309, Running: 4, Completed: 2707 [ 2h 33m ] +INFO: Idle: 13308, Running: 4, Completed: 2708 [ 2h 33m ] +INFO: Idle: 13307, Running: 4, Completed: 2709 [ 2h 33m ] +INFO: Idle: 13306, Running: 4, Completed: 2710 [ 2h 33m ] +INFO: Idle: 13305, Running: 4, Completed: 2711 [ 2h 33m ] +INFO: Idle: 13304, Running: 4, Completed: 2712 [ 2h 33m ] +INFO: Idle: 13302, Running: 4, Completed: 2714 [ 2h 34m ] +INFO: Idle: 13301, Running: 4, Completed: 2715 [ 2h 34m ] +INFO: Idle: 13300, Running: 4, Completed: 2716 [ 2h 34m ] +INFO: Idle: 13298, Running: 4, Completed: 2718 [ 2h 34m ] +INFO: Idle: 13297, Running: 4, Completed: 2719 [ 2h 34m ] +INFO: Idle: 13296, Running: 4, Completed: 2720 [ 2h 34m ] +INFO: Idle: 13294, Running: 4, Completed: 2722 [ 2h 34m ] +INFO: Idle: 13293, Running: 4, Completed: 2723 [ 2h 34m ] +INFO: Idle: 13292, Running: 4, Completed: 2724 [ 2h 34m ] +INFO: Idle: 13290, Running: 4, Completed: 2726 [ 2h 34m ] +INFO: Idle: 13289, Running: 4, Completed: 2727 [ 2h 35m ] +INFO: Idle: 13288, Running: 4, Completed: 2728 [ 2h 35m ] +INFO: Idle: 13287, Running: 4, Completed: 2729 [ 2h 35m ] +INFO: Idle: 13286, Running: 4, Completed: 2730 [ 2h 35m ] +INFO: Idle: 13284, Running: 4, Completed: 2732 [ 2h 35m ] +INFO: Idle: 13283, Running: 4, Completed: 2733 [ 2h 35m ] +INFO: Idle: 13282, Running: 4, Completed: 2734 [ 2h 35m ] +INFO: Idle: 13280, Running: 4, Completed: 2736 [ 2h 35m ] +INFO: Idle: 13279, Running: 4, Completed: 2737 [ 2h 36m ] +INFO: Idle: 13278, Running: 4, Completed: 2738 [ 2h 36m ] +INFO: Idle: 13276, Running: 4, Completed: 2740 [ 2h 36m ] +INFO: Idle: 13275, Running: 4, Completed: 2741 [ 2h 36m ] +INFO: Idle: 13274, Running: 4, Completed: 2742 [ 2h 36m ] +INFO: Idle: 13272, Running: 4, Completed: 2744 [ 2h 36m ] +INFO: Idle: 13271, Running: 4, Completed: 2745 [ 2h 36m ] +INFO: Idle: 13269, Running: 4, Completed: 2747 [ 2h 36m ] +INFO: Idle: 13268, Running: 4, Completed: 2748 [ 2h 36m ] +INFO: Idle: 13267, Running: 4, Completed: 2749 [ 2h 37m ] +INFO: Idle: 13265, Running: 4, Completed: 2751 [ 2h 37m ] +INFO: Idle: 13263, Running: 4, Completed: 2753 [ 2h 37m ] +INFO: Idle: 13261, Running: 4, Completed: 2755 [ 2h 37m ] +INFO: Idle: 13260, Running: 4, Completed: 2756 [ 2h 37m ] +INFO: Idle: 13258, Running: 4, Completed: 2758 [ 2h 37m ] +INFO: Idle: 13256, Running: 4, Completed: 2760 [ 2h 37m ] +INFO: Idle: 13255, Running: 4, Completed: 2761 [ 2h 37m ] +INFO: Idle: 13253, Running: 4, Completed: 2763 [ 2h 38m ] +INFO: Idle: 13252, Running: 4, Completed: 2764 [ 2h 38m ] +INFO: Idle: 13251, Running: 4, Completed: 2765 [ 2h 38m ] +INFO: Idle: 13248, Running: 4, Completed: 2768 [ 2h 38m ] +INFO: Idle: 13247, Running: 4, Completed: 2769 [ 2h 38m ] +INFO: Idle: 13246, Running: 4, Completed: 2770 [ 2h 38m ] +INFO: Idle: 13244, Running: 4, Completed: 2772 [ 2h 39m ] +INFO: Idle: 13242, Running: 4, Completed: 2774 [ 2h 39m ] +INFO: Idle: 13240, Running: 4, Completed: 2776 [ 2h 39m ] +INFO: Idle: 13238, Running: 4, Completed: 2778 [ 2h 39m ] +INFO: Idle: 13237, Running: 4, Completed: 2779 [ 2h 39m ] +INFO: Idle: 13236, Running: 4, Completed: 2780 [ 2h 39m ] +INFO: Idle: 13233, Running: 4, Completed: 2783 [ 2h 39m ] +INFO: Idle: 13232, Running: 4, Completed: 2784 [ 2h 40m ] +INFO: Idle: 13231, Running: 4, Completed: 2785 [ 2h 40m ] +INFO: Idle: 13230, Running: 4, Completed: 2786 [ 2h 40m ] +INFO: Idle: 13229, Running: 4, Completed: 2787 [ 2h 40m ] +INFO: Idle: 13228, Running: 4, Completed: 2788 [ 2h 40m ] +INFO: Idle: 13227, Running: 4, Completed: 2789 [ 2h 40m ] +INFO: Idle: 13225, Running: 4, Completed: 2791 [ 2h 40m ] +INFO: Idle: 13224, Running: 4, Completed: 2792 [ 2h 40m ] +INFO: Idle: 13221, Running: 4, Completed: 2795 [ 2h 41m ] +INFO: Idle: 13220, Running: 4, Completed: 2796 [ 2h 41m ] +INFO: Idle: 13219, Running: 4, Completed: 2797 [ 2h 41m ] +INFO: Idle: 13218, Running: 4, Completed: 2798 [ 2h 41m ] +INFO: Idle: 13216, Running: 4, Completed: 2800 [ 2h 41m ] +INFO: Idle: 13215, Running: 4, Completed: 2801 [ 2h 41m ] +INFO: Idle: 13212, Running: 4, Completed: 2804 [ 2h 42m ] +INFO: Idle: 13211, Running: 4, Completed: 2805 [ 2h 42m ] +INFO: Idle: 13210, Running: 4, Completed: 2806 [ 2h 42m ] +INFO: Idle: 13208, Running: 4, Completed: 2808 [ 2h 42m ] +INFO: Idle: 13207, Running: 4, Completed: 2809 [ 2h 42m ] +INFO: Idle: 13206, Running: 4, Completed: 2810 [ 2h 42m ] +INFO: Idle: 13205, Running: 4, Completed: 2811 [ 2h 42m ] +INFO: Idle: 13204, Running: 4, Completed: 2812 [ 2h 42m ] +INFO: Idle: 13203, Running: 4, Completed: 2813 [ 2h 42m ] +INFO: Idle: 13202, Running: 4, Completed: 2814 [ 2h 42m ] +INFO: Idle: 13201, Running: 4, Completed: 2815 [ 2h 42m ] +INFO: Idle: 13200, Running: 4, Completed: 2816 [ 2h 43m ] +INFO: Idle: 13197, Running: 4, Completed: 2819 [ 2h 43m ] +INFO: Idle: 13196, Running: 4, Completed: 2820 [ 2h 43m ] +INFO: Idle: 13194, Running: 4, Completed: 2822 [ 2h 43m ] +INFO: Idle: 13192, Running: 4, Completed: 2824 [ 2h 43m ] +INFO: Idle: 13191, Running: 4, Completed: 2825 [ 2h 43m ] +INFO: Idle: 13190, Running: 4, Completed: 2826 [ 2h 43m ] +INFO: Idle: 13189, Running: 4, Completed: 2827 [ 2h 44m ] +INFO: Idle: 13187, Running: 4, Completed: 2829 [ 2h 44m ] +INFO: Idle: 13186, Running: 4, Completed: 2830 [ 2h 44m ] +INFO: Idle: 13185, Running: 4, Completed: 2831 [ 2h 44m ] +INFO: Idle: 13184, Running: 4, Completed: 2832 [ 2h 44m ] +INFO: Idle: 13183, Running: 4, Completed: 2833 [ 2h 44m ] +INFO: Idle: 13181, Running: 4, Completed: 2835 [ 2h 44m ] +INFO: Idle: 13179, Running: 4, Completed: 2837 [ 2h 44m ] +INFO: Idle: 13178, Running: 4, Completed: 2838 [ 2h 44m ] +INFO: Idle: 13177, Running: 4, Completed: 2839 [ 2h 45m ] +INFO: Idle: 13176, Running: 4, Completed: 2840 [ 2h 45m ] +INFO: Idle: 13175, Running: 4, Completed: 2841 [ 2h 45m ] +INFO: Idle: 13174, Running: 4, Completed: 2842 [ 2h 45m ] +INFO: Idle: 13173, Running: 4, Completed: 2843 [ 2h 45m ] +INFO: Idle: 13171, Running: 4, Completed: 2845 [ 2h 45m ] +INFO: Idle: 13170, Running: 4, Completed: 2846 [ 2h 45m ] +INFO: Idle: 13169, Running: 4, Completed: 2847 [ 2h 45m ] +INFO: Idle: 13167, Running: 4, Completed: 2849 [ 2h 45m ] +INFO: Idle: 13165, Running: 4, Completed: 2851 [ 2h 45m ] +INFO: Idle: 13163, Running: 4, Completed: 2853 [ 2h 46m ] +INFO: Idle: 13161, Running: 4, Completed: 2855 [ 2h 46m ] +INFO: Idle: 13160, Running: 4, Completed: 2856 [ 2h 46m ] +INFO: Idle: 13159, Running: 4, Completed: 2857 [ 2h 46m ] +INFO: Idle: 13157, Running: 4, Completed: 2859 [ 2h 46m ] +INFO: Idle: 13156, Running: 4, Completed: 2860 [ 2h 46m ] +INFO: Idle: 13155, Running: 4, Completed: 2861 [ 2h 46m ] +INFO: Idle: 13154, Running: 4, Completed: 2862 [ 2h 47m ] +INFO: Idle: 13151, Running: 4, Completed: 2865 [ 2h 47m ] +INFO: Idle: 13149, Running: 4, Completed: 2867 [ 2h 47m ] +INFO: Idle: 13148, Running: 4, Completed: 2868 [ 2h 47m ] +INFO: Idle: 13147, Running: 4, Completed: 2869 [ 2h 47m ] +INFO: Idle: 13146, Running: 4, Completed: 2870 [ 2h 47m ] +INFO: Idle: 13145, Running: 4, Completed: 2871 [ 2h 48m ] +INFO: Idle: 13143, Running: 4, Completed: 2873 [ 2h 48m ] +INFO: Idle: 13142, Running: 4, Completed: 2874 [ 2h 48m ] +INFO: Idle: 13141, Running: 4, Completed: 2875 [ 2h 48m ] +INFO: Idle: 13139, Running: 4, Completed: 2877 [ 2h 48m ] +INFO: Idle: 13138, Running: 4, Completed: 2878 [ 2h 48m ] +INFO: Idle: 13137, Running: 4, Completed: 2879 [ 2h 48m ] +INFO: Idle: 13136, Running: 4, Completed: 2880 [ 2h 48m ] +INFO: Idle: 13134, Running: 4, Completed: 2882 [ 2h 48m ] +INFO: Idle: 13133, Running: 4, Completed: 2883 [ 2h 49m ] +INFO: Idle: 13131, Running: 4, Completed: 2885 [ 2h 49m ] +INFO: Idle: 13130, Running: 4, Completed: 2886 [ 2h 49m ] +INFO: Idle: 13129, Running: 4, Completed: 2887 [ 2h 49m ] +INFO: Idle: 13128, Running: 4, Completed: 2888 [ 2h 49m ] +INFO: Idle: 13126, Running: 4, Completed: 2890 [ 2h 49m ] +INFO: Idle: 13125, Running: 4, Completed: 2891 [ 2h 49m ] +INFO: Idle: 13124, Running: 4, Completed: 2892 [ 2h 49m ] +INFO: Idle: 13123, Running: 4, Completed: 2893 [ 2h 49m ] +INFO: Idle: 13122, Running: 4, Completed: 2894 [ 2h 50m ] +INFO: Idle: 13121, Running: 4, Completed: 2895 [ 2h 50m ] +INFO: Idle: 13119, Running: 4, Completed: 2897 [ 2h 50m ] +INFO: Idle: 13117, Running: 4, Completed: 2899 [ 2h 50m ] +INFO: Idle: 13115, Running: 4, Completed: 2901 [ 2h 50m ] +INFO: Idle: 13114, Running: 4, Completed: 2902 [ 2h 50m ] +INFO: Idle: 13113, Running: 4, Completed: 2903 [ 2h 50m ] +INFO: Idle: 13112, Running: 4, Completed: 2904 [ 2h 50m ] +INFO: Idle: 13111, Running: 4, Completed: 2905 [ 2h 50m ] +INFO: Idle: 13109, Running: 4, Completed: 2907 [ 2h 51m ] +INFO: Idle: 13107, Running: 4, Completed: 2909 [ 2h 51m ] +INFO: Idle: 13105, Running: 4, Completed: 2911 [ 2h 51m ] +INFO: Idle: 13104, Running: 4, Completed: 2912 [ 2h 51m ] +INFO: Idle: 13103, Running: 4, Completed: 2913 [ 2h 51m ] +INFO: Idle: 13101, Running: 4, Completed: 2915 [ 2h 51m ] +INFO: Idle: 13100, Running: 4, Completed: 2916 [ 2h 51m ] +INFO: Idle: 13098, Running: 4, Completed: 2918 [ 2h 52m ] +INFO: Idle: 13096, Running: 4, Completed: 2920 [ 2h 52m ] +INFO: Idle: 13094, Running: 4, Completed: 2922 [ 2h 52m ] +INFO: Idle: 13091, Running: 4, Completed: 2925 [ 2h 52m ] +INFO: Idle: 13090, Running: 4, Completed: 2926 [ 2h 52m ] +INFO: Idle: 13089, Running: 4, Completed: 2927 [ 2h 52m ] +INFO: Idle: 13087, Running: 4, Completed: 2929 [ 2h 53m ] +INFO: Idle: 13085, Running: 4, Completed: 2931 [ 2h 53m ] +INFO: Idle: 13084, Running: 4, Completed: 2932 [ 2h 53m ] +INFO: Idle: 13083, Running: 4, Completed: 2933 [ 2h 53m ] +INFO: Idle: 13082, Running: 4, Completed: 2934 [ 2h 53m ] +INFO: Idle: 13081, Running: 4, Completed: 2935 [ 2h 53m ] +INFO: Idle: 13080, Running: 4, Completed: 2936 [ 2h 53m ] +INFO: Idle: 13079, Running: 4, Completed: 2937 [ 2h 53m ] +INFO: Idle: 13077, Running: 4, Completed: 2939 [ 2h 53m ] +INFO: Idle: 13076, Running: 4, Completed: 2940 [ 2h 54m ] +INFO: Idle: 13075, Running: 4, Completed: 2941 [ 2h 54m ] +INFO: Idle: 13074, Running: 4, Completed: 2942 [ 2h 54m ] +INFO: Idle: 13072, Running: 4, Completed: 2944 [ 2h 54m ] +INFO: Idle: 13071, Running: 4, Completed: 2945 [ 2h 54m ] +INFO: Idle: 13069, Running: 4, Completed: 2947 [ 2h 54m ] +INFO: Idle: 13068, Running: 4, Completed: 2948 [ 2h 54m ] +INFO: Idle: 13067, Running: 4, Completed: 2949 [ 2h 54m ] +INFO: Idle: 13065, Running: 4, Completed: 2951 [ 2h 54m ] +INFO: Idle: 13064, Running: 4, Completed: 2952 [ 2h 55m ] +INFO: Idle: 13063, Running: 4, Completed: 2953 [ 2h 55m ] +INFO: Idle: 13062, Running: 4, Completed: 2954 [ 2h 55m ] +INFO: Idle: 13061, Running: 4, Completed: 2955 [ 2h 55m ] +INFO: Idle: 13060, Running: 4, Completed: 2956 [ 2h 55m ] +INFO: Idle: 13059, Running: 4, Completed: 2957 [ 2h 55m ] +INFO: Idle: 13057, Running: 4, Completed: 2959 [ 2h 55m ] +INFO: Idle: 13056, Running: 4, Completed: 2960 [ 2h 55m ] +INFO: Idle: 13055, Running: 4, Completed: 2961 [ 2h 55m ] +INFO: Idle: 13054, Running: 4, Completed: 2962 [ 2h 55m ] +INFO: Idle: 13053, Running: 4, Completed: 2963 [ 2h 55m ] +INFO: Idle: 13052, Running: 4, Completed: 2964 [ 2h 56m ] +INFO: Idle: 13050, Running: 4, Completed: 2966 [ 2h 56m ] +INFO: Idle: 13049, Running: 4, Completed: 2967 [ 2h 56m ] +INFO: Idle: 13048, Running: 4, Completed: 2968 [ 2h 56m ] +INFO: Idle: 13047, Running: 4, Completed: 2969 [ 2h 56m ] +INFO: Idle: 13045, Running: 4, Completed: 2971 [ 2h 56m ] +INFO: Idle: 13044, Running: 4, Completed: 2972 [ 2h 56m ] +INFO: Idle: 13043, Running: 4, Completed: 2973 [ 2h 56m ] +INFO: Idle: 13042, Running: 4, Completed: 2974 [ 2h 56m ] +INFO: Idle: 13041, Running: 4, Completed: 2975 [ 2h 57m ] +INFO: Idle: 13039, Running: 4, Completed: 2977 [ 2h 57m ] +INFO: Idle: 13038, Running: 4, Completed: 2978 [ 2h 57m ] +INFO: Idle: 13037, Running: 4, Completed: 2979 [ 2h 57m ] +INFO: Idle: 13036, Running: 4, Completed: 2980 [ 2h 57m ] +INFO: Idle: 13035, Running: 4, Completed: 2981 [ 2h 57m ] +INFO: Idle: 13034, Running: 4, Completed: 2982 [ 2h 57m ] +INFO: Idle: 13033, Running: 4, Completed: 2983 [ 2h 57m ] +INFO: Idle: 13032, Running: 4, Completed: 2984 [ 2h 57m ] +INFO: Idle: 13031, Running: 4, Completed: 2985 [ 2h 57m ] +INFO: Idle: 13029, Running: 4, Completed: 2987 [ 2h 57m ] +INFO: Idle: 13028, Running: 4, Completed: 2988 [ 2h 58m ] +INFO: Idle: 13027, Running: 4, Completed: 2989 [ 2h 58m ] +INFO: Idle: 13025, Running: 4, Completed: 2991 [ 2h 58m ] +INFO: Idle: 13023, Running: 4, Completed: 2993 [ 2h 58m ] +INFO: Idle: 13021, Running: 4, Completed: 2995 [ 2h 58m ] +INFO: Idle: 13019, Running: 4, Completed: 2997 [ 2h 58m ] +INFO: Idle: 13018, Running: 4, Completed: 2998 [ 2h 58m ] +INFO: Idle: 13017, Running: 4, Completed: 2999 [ 2h 58m ] +INFO: Idle: 13016, Running: 4, Completed: 3000 [ 2h 58m ] +INFO: Idle: 13015, Running: 4, Completed: 3001 [ 2h 59m ] +INFO: Idle: 13013, Running: 4, Completed: 3003 [ 2h 59m ] +INFO: Idle: 13011, Running: 4, Completed: 3005 [ 2h 59m ] +INFO: Idle: 13008, Running: 4, Completed: 3008 [ 2h 59m ] +INFO: Idle: 13007, Running: 4, Completed: 3009 [ 2h 59m ] +INFO: Idle: 13006, Running: 4, Completed: 3010 [ 3h 0m ] +INFO: Idle: 13004, Running: 4, Completed: 3012 [ 3h 0m ] +INFO: Idle: 13002, Running: 4, Completed: 3014 [ 3h 0m ] +INFO: Idle: 13000, Running: 4, Completed: 3016 [ 3h 0m ] +INFO: Idle: 12998, Running: 4, Completed: 3018 [ 3h 0m ] +INFO: Idle: 12997, Running: 4, Completed: 3019 [ 3h 0m ] +INFO: Idle: 12996, Running: 4, Completed: 3020 [ 3h 1m ] +INFO: Idle: 12994, Running: 4, Completed: 3022 [ 3h 1m ] +INFO: Idle: 12993, Running: 4, Completed: 3023 [ 3h 1m ] +INFO: Idle: 12992, Running: 4, Completed: 3024 [ 3h 1m ] +INFO: Idle: 12991, Running: 4, Completed: 3025 [ 3h 1m ] +INFO: Idle: 12990, Running: 4, Completed: 3026 [ 3h 1m ] +INFO: Idle: 12989, Running: 4, Completed: 3027 [ 3h 1m ] +INFO: Idle: 12988, Running: 4, Completed: 3028 [ 3h 1m ] +INFO: Idle: 12987, Running: 4, Completed: 3029 [ 3h 1m ] +INFO: Idle: 12986, Running: 4, Completed: 3030 [ 3h 1m ] +INFO: Idle: 12985, Running: 4, Completed: 3031 [ 3h 2m ] +INFO: Idle: 12984, Running: 4, Completed: 3032 [ 3h 2m ] +INFO: Idle: 12981, Running: 4, Completed: 3035 [ 3h 2m ] +INFO: Idle: 12980, Running: 4, Completed: 3036 [ 3h 2m ] +INFO: Idle: 12979, Running: 4, Completed: 3037 [ 3h 2m ] +INFO: Idle: 12978, Running: 4, Completed: 3038 [ 3h 2m ] +INFO: Idle: 12977, Running: 4, Completed: 3039 [ 3h 2m ] +INFO: Idle: 12976, Running: 4, Completed: 3040 [ 3h 2m ] +INFO: Idle: 12975, Running: 4, Completed: 3041 [ 3h 2m ] +INFO: Idle: 12972, Running: 4, Completed: 3044 [ 3h 3m ] +INFO: Idle: 12971, Running: 4, Completed: 3045 [ 3h 3m ] +INFO: Idle: 12970, Running: 4, Completed: 3046 [ 3h 3m ] +INFO: Idle: 12968, Running: 4, Completed: 3048 [ 3h 3m ] +INFO: Idle: 12967, Running: 4, Completed: 3049 [ 3h 3m ] +INFO: Idle: 12966, Running: 4, Completed: 3050 [ 3h 3m ] +INFO: Idle: 12965, Running: 4, Completed: 3051 [ 3h 3m ] +INFO: Idle: 12964, Running: 4, Completed: 3052 [ 3h 3m ] +INFO: Idle: 12962, Running: 4, Completed: 3054 [ 3h 4m ] +INFO: Idle: 12961, Running: 4, Completed: 3055 [ 3h 4m ] +INFO: Idle: 12960, Running: 4, Completed: 3056 [ 3h 4m ] +INFO: Idle: 12958, Running: 4, Completed: 3058 [ 3h 4m ] +INFO: Idle: 12957, Running: 4, Completed: 3059 [ 3h 4m ] +INFO: Idle: 12956, Running: 4, Completed: 3060 [ 3h 4m ] +INFO: Idle: 12954, Running: 4, Completed: 3062 [ 3h 4m ] +INFO: Idle: 12952, Running: 4, Completed: 3064 [ 3h 4m ] +INFO: Idle: 12951, Running: 4, Completed: 3065 [ 3h 5m ] +INFO: Idle: 12950, Running: 4, Completed: 3066 [ 3h 5m ] +INFO: Idle: 12949, Running: 4, Completed: 3067 [ 3h 5m ] +INFO: Idle: 12948, Running: 4, Completed: 3068 [ 3h 5m ] +INFO: Idle: 12947, Running: 4, Completed: 3069 [ 3h 5m ] +INFO: Idle: 12946, Running: 4, Completed: 3070 [ 3h 5m ] +INFO: Idle: 12945, Running: 4, Completed: 3071 [ 3h 5m ] +INFO: Idle: 12944, Running: 4, Completed: 3072 [ 3h 5m ] +INFO: Idle: 12942, Running: 4, Completed: 3074 [ 3h 5m ] +INFO: Idle: 12941, Running: 4, Completed: 3075 [ 3h 5m ] +INFO: Idle: 12940, Running: 4, Completed: 3076 [ 3h 5m ] +INFO: Idle: 12939, Running: 4, Completed: 3077 [ 3h 6m ] +INFO: Idle: 12938, Running: 4, Completed: 3078 [ 3h 6m ] +INFO: Idle: 12937, Running: 4, Completed: 3079 [ 3h 6m ] +INFO: Idle: 12936, Running: 4, Completed: 3080 [ 3h 6m ] +INFO: Idle: 12934, Running: 4, Completed: 3082 [ 3h 6m ] +INFO: Idle: 12933, Running: 4, Completed: 3083 [ 3h 6m ] +INFO: Idle: 12932, Running: 4, Completed: 3084 [ 3h 6m ] +INFO: Idle: 12930, Running: 4, Completed: 3086 [ 3h 6m ] +INFO: Idle: 12929, Running: 4, Completed: 3087 [ 3h 6m ] +INFO: Idle: 12928, Running: 4, Completed: 3088 [ 3h 6m ] +INFO: Idle: 12927, Running: 4, Completed: 3089 [ 3h 7m ] +INFO: Idle: 12925, Running: 4, Completed: 3091 [ 3h 7m ] +INFO: Idle: 12923, Running: 4, Completed: 3093 [ 3h 7m ] +INFO: Idle: 12921, Running: 4, Completed: 3095 [ 3h 7m ] +INFO: Idle: 12919, Running: 4, Completed: 3097 [ 3h 7m ] +INFO: Idle: 12916, Running: 4, Completed: 3100 [ 3h 8m ] +INFO: Idle: 12915, Running: 4, Completed: 3101 [ 3h 8m ] +INFO: Idle: 12912, Running: 4, Completed: 3104 [ 3h 8m ] +INFO: Idle: 12911, Running: 4, Completed: 3105 [ 3h 8m ] +INFO: Idle: 12908, Running: 4, Completed: 3108 [ 3h 8m ] +INFO: Idle: 12907, Running: 4, Completed: 3109 [ 3h 9m ] +INFO: Idle: 12905, Running: 4, Completed: 3111 [ 3h 9m ] +INFO: Idle: 12904, Running: 4, Completed: 3112 [ 3h 9m ] +INFO: Idle: 12903, Running: 4, Completed: 3113 [ 3h 9m ] +INFO: Idle: 12901, Running: 4, Completed: 3115 [ 3h 9m ] +INFO: Idle: 12899, Running: 4, Completed: 3117 [ 3h 9m ] +INFO: Idle: 12898, Running: 4, Completed: 3118 [ 3h 10m ] +INFO: Idle: 12897, Running: 4, Completed: 3119 [ 3h 10m ] +INFO: Idle: 12896, Running: 4, Completed: 3120 [ 3h 10m ] +INFO: Idle: 12895, Running: 4, Completed: 3121 [ 3h 10m ] +INFO: Idle: 12894, Running: 4, Completed: 3122 [ 3h 10m ] +INFO: Idle: 12893, Running: 4, Completed: 3123 [ 3h 10m ] +INFO: Idle: 12892, Running: 4, Completed: 3124 [ 3h 10m ] +INFO: Idle: 12890, Running: 4, Completed: 3126 [ 3h 10m ] +INFO: Idle: 12888, Running: 4, Completed: 3128 [ 3h 11m ] +INFO: Idle: 12886, Running: 4, Completed: 3130 [ 3h 11m ] +INFO: Idle: 12883, Running: 4, Completed: 3133 [ 3h 11m ] +INFO: Idle: 12882, Running: 4, Completed: 3134 [ 3h 11m ] +INFO: Idle: 12880, Running: 4, Completed: 3136 [ 3h 11m ] +INFO: Idle: 12879, Running: 4, Completed: 3137 [ 3h 11m ] +INFO: Idle: 12878, Running: 4, Completed: 3138 [ 3h 12m ] +INFO: Idle: 12877, Running: 4, Completed: 3139 [ 3h 12m ] +INFO: Idle: 12876, Running: 4, Completed: 3140 [ 3h 12m ] +INFO: Idle: 12875, Running: 4, Completed: 3141 [ 3h 12m ] +INFO: Idle: 12873, Running: 4, Completed: 3143 [ 3h 12m ] +INFO: Idle: 12872, Running: 4, Completed: 3144 [ 3h 12m ] +INFO: Idle: 12871, Running: 4, Completed: 3145 [ 3h 12m ] +INFO: Idle: 12869, Running: 4, Completed: 3147 [ 3h 12m ] +INFO: Idle: 12868, Running: 4, Completed: 3148 [ 3h 12m ] +INFO: Idle: 12867, Running: 4, Completed: 3149 [ 3h 12m ] +INFO: Idle: 12866, Running: 4, Completed: 3150 [ 3h 13m ] +INFO: Idle: 12865, Running: 4, Completed: 3151 [ 3h 13m ] +INFO: Idle: 12863, Running: 4, Completed: 3153 [ 3h 13m ] +INFO: Idle: 12862, Running: 4, Completed: 3154 [ 3h 13m ] +INFO: Idle: 12861, Running: 4, Completed: 3155 [ 3h 13m ] +INFO: Idle: 12860, Running: 4, Completed: 3156 [ 3h 13m ] +INFO: Idle: 12858, Running: 4, Completed: 3158 [ 3h 13m ] +INFO: Idle: 12857, Running: 4, Completed: 3159 [ 3h 13m ] +INFO: Idle: 12856, Running: 4, Completed: 3160 [ 3h 13m ] +INFO: Idle: 12855, Running: 4, Completed: 3161 [ 3h 13m ] +INFO: Idle: 12853, Running: 4, Completed: 3163 [ 3h 14m ] +INFO: Idle: 12852, Running: 4, Completed: 3164 [ 3h 14m ] +INFO: Idle: 12851, Running: 4, Completed: 3165 [ 3h 14m ] +INFO: Idle: 12850, Running: 4, Completed: 3166 [ 3h 14m ] +INFO: Idle: 12849, Running: 4, Completed: 3167 [ 3h 14m ] +INFO: Idle: 12848, Running: 4, Completed: 3168 [ 3h 14m ] +INFO: Idle: 12846, Running: 4, Completed: 3170 [ 3h 14m ] +INFO: Idle: 12844, Running: 4, Completed: 3172 [ 3h 14m ] +INFO: Idle: 12843, Running: 4, Completed: 3173 [ 3h 14m ] +INFO: Idle: 12842, Running: 4, Completed: 3174 [ 3h 15m ] +INFO: Idle: 12841, Running: 4, Completed: 3175 [ 3h 15m ] +INFO: Idle: 12840, Running: 4, Completed: 3176 [ 3h 15m ] +INFO: Idle: 12839, Running: 4, Completed: 3177 [ 3h 15m ] +INFO: Idle: 12838, Running: 4, Completed: 3178 [ 3h 15m ] +INFO: Idle: 12837, Running: 4, Completed: 3179 [ 3h 15m ] +INFO: Idle: 12835, Running: 4, Completed: 3181 [ 3h 15m ] +INFO: Idle: 12834, Running: 4, Completed: 3182 [ 3h 15m ] +INFO: Idle: 12831, Running: 4, Completed: 3185 [ 3h 16m ] +INFO: Idle: 12830, Running: 4, Completed: 3186 [ 3h 16m ] +INFO: Idle: 12829, Running: 4, Completed: 3187 [ 3h 16m ] +INFO: Idle: 12827, Running: 4, Completed: 3189 [ 3h 16m ] +INFO: Idle: 12826, Running: 4, Completed: 3190 [ 3h 16m ] +INFO: Idle: 12825, Running: 4, Completed: 3191 [ 3h 16m ] +INFO: Idle: 12823, Running: 4, Completed: 3193 [ 3h 16m ] +INFO: Idle: 12822, Running: 4, Completed: 3194 [ 3h 16m ] +INFO: Idle: 12821, Running: 4, Completed: 3195 [ 3h 17m ] +INFO: Idle: 12819, Running: 4, Completed: 3197 [ 3h 17m ] +INFO: Idle: 12817, Running: 4, Completed: 3199 [ 3h 17m ] +INFO: Idle: 12815, Running: 4, Completed: 3201 [ 3h 17m ] +INFO: Idle: 12812, Running: 4, Completed: 3204 [ 3h 17m ] +INFO: Idle: 12811, Running: 4, Completed: 3205 [ 3h 17m ] +INFO: Idle: 12809, Running: 4, Completed: 3207 [ 3h 17m ] +INFO: Idle: 12807, Running: 4, Completed: 3209 [ 3h 18m ] +INFO: Idle: 12805, Running: 4, Completed: 3211 [ 3h 18m ] +INFO: Idle: 12803, Running: 4, Completed: 3213 [ 3h 18m ] +INFO: Idle: 12800, Running: 4, Completed: 3216 [ 3h 18m ] +INFO: Idle: 12799, Running: 4, Completed: 3217 [ 3h 19m ] +INFO: Idle: 12796, Running: 4, Completed: 3220 [ 3h 19m ] +INFO: Idle: 12795, Running: 4, Completed: 3221 [ 3h 19m ] +INFO: Idle: 12792, Running: 4, Completed: 3224 [ 3h 19m ] +INFO: Idle: 12791, Running: 4, Completed: 3225 [ 3h 19m ] +INFO: Idle: 12788, Running: 4, Completed: 3228 [ 3h 20m ] +INFO: Idle: 12787, Running: 4, Completed: 3229 [ 3h 20m ] +INFO: Idle: 12785, Running: 4, Completed: 3231 [ 3h 20m ] +INFO: Idle: 12784, Running: 4, Completed: 3232 [ 3h 20m ] +INFO: Idle: 12783, Running: 4, Completed: 3233 [ 3h 20m ] +INFO: Idle: 12782, Running: 4, Completed: 3234 [ 3h 20m ] +INFO: Idle: 12780, Running: 4, Completed: 3236 [ 3h 20m ] +INFO: Idle: 12779, Running: 4, Completed: 3237 [ 3h 21m ] +INFO: Idle: 12778, Running: 4, Completed: 3238 [ 3h 21m ] +INFO: Idle: 12777, Running: 4, Completed: 3239 [ 3h 21m ] +INFO: Idle: 12776, Running: 4, Completed: 3240 [ 3h 21m ] +INFO: Idle: 12774, Running: 4, Completed: 3242 [ 3h 21m ] +INFO: Idle: 12773, Running: 4, Completed: 3243 [ 3h 21m ] +INFO: Idle: 12770, Running: 4, Completed: 3246 [ 3h 22m ] +INFO: Idle: 12768, Running: 4, Completed: 3248 [ 3h 22m ] +INFO: Idle: 12767, Running: 4, Completed: 3249 [ 3h 22m ] +INFO: Idle: 12766, Running: 4, Completed: 3250 [ 3h 22m ] +INFO: Idle: 12763, Running: 4, Completed: 3253 [ 3h 22m ] +INFO: Idle: 12760, Running: 4, Completed: 3256 [ 3h 23m ] +INFO: Idle: 12759, Running: 4, Completed: 3257 [ 3h 23m ] +INFO: Idle: 12758, Running: 4, Completed: 3258 [ 3h 23m ] +INFO: Idle: 12757, Running: 4, Completed: 3259 [ 3h 23m ] +INFO: Idle: 12756, Running: 4, Completed: 3260 [ 3h 23m ] +INFO: Idle: 12755, Running: 4, Completed: 3261 [ 3h 23m ] +INFO: Idle: 12753, Running: 4, Completed: 3263 [ 3h 23m ] +INFO: Idle: 12752, Running: 4, Completed: 3264 [ 3h 23m ] +INFO: Idle: 12751, Running: 4, Completed: 3265 [ 3h 23m ] +INFO: Idle: 12748, Running: 4, Completed: 3268 [ 3h 24m ] +INFO: Idle: 12747, Running: 4, Completed: 3269 [ 3h 24m ] +INFO: Idle: 12746, Running: 4, Completed: 3270 [ 3h 24m ] +INFO: Idle: 12745, Running: 4, Completed: 3271 [ 3h 24m ] +INFO: Idle: 12742, Running: 4, Completed: 3274 [ 3h 24m ] +INFO: Idle: 12741, Running: 4, Completed: 3275 [ 3h 24m ] +INFO: Idle: 12739, Running: 4, Completed: 3277 [ 3h 24m ] +INFO: Idle: 12737, Running: 4, Completed: 3279 [ 3h 25m ] +INFO: Idle: 12736, Running: 4, Completed: 3280 [ 3h 25m ] +INFO: Idle: 12735, Running: 4, Completed: 3281 [ 3h 25m ] +INFO: Idle: 12733, Running: 4, Completed: 3283 [ 3h 25m ] +INFO: Idle: 12732, Running: 4, Completed: 3284 [ 3h 25m ] +INFO: Idle: 12730, Running: 4, Completed: 3286 [ 3h 25m ] +INFO: Idle: 12729, Running: 4, Completed: 3287 [ 3h 25m ] +INFO: Idle: 12728, Running: 4, Completed: 3288 [ 3h 25m ] +INFO: Idle: 12726, Running: 4, Completed: 3290 [ 3h 26m ] +INFO: Idle: 12725, Running: 4, Completed: 3291 [ 3h 26m ] +INFO: Idle: 12723, Running: 4, Completed: 3293 [ 3h 26m ] +INFO: Idle: 12722, Running: 4, Completed: 3294 [ 3h 26m ] +INFO: Idle: 12721, Running: 4, Completed: 3295 [ 3h 26m ] +INFO: Idle: 12719, Running: 4, Completed: 3297 [ 3h 26m ] +INFO: Idle: 12718, Running: 4, Completed: 3298 [ 3h 26m ] +INFO: Idle: 12717, Running: 4, Completed: 3299 [ 3h 26m ] +INFO: Idle: 12715, Running: 4, Completed: 3301 [ 3h 27m ] +INFO: Idle: 12714, Running: 4, Completed: 3302 [ 3h 27m ] +INFO: Idle: 12711, Running: 4, Completed: 3305 [ 3h 27m ] +INFO: Idle: 12710, Running: 4, Completed: 3306 [ 3h 27m ] +INFO: Idle: 12709, Running: 4, Completed: 3307 [ 3h 27m ] +INFO: Idle: 12707, Running: 4, Completed: 3309 [ 3h 27m ] +INFO: Idle: 12706, Running: 4, Completed: 3310 [ 3h 27m ] +INFO: Idle: 12705, Running: 4, Completed: 3311 [ 3h 28m ] +INFO: Idle: 12704, Running: 4, Completed: 3312 [ 3h 28m ] +INFO: Idle: 12703, Running: 4, Completed: 3313 [ 3h 28m ] +INFO: Idle: 12701, Running: 4, Completed: 3315 [ 3h 28m ] +INFO: Idle: 12700, Running: 4, Completed: 3316 [ 3h 28m ] +INFO: Idle: 12699, Running: 4, Completed: 3317 [ 3h 28m ] +INFO: Idle: 12697, Running: 4, Completed: 3319 [ 3h 28m ] +INFO: Idle: 12696, Running: 4, Completed: 3320 [ 3h 28m ] +INFO: Idle: 12695, Running: 4, Completed: 3321 [ 3h 28m ] +INFO: Idle: 12692, Running: 4, Completed: 3324 [ 3h 29m ] +INFO: Idle: 12691, Running: 4, Completed: 3325 [ 3h 29m ] +INFO: Idle: 12687, Running: 4, Completed: 3329 [ 3h 29m ] +INFO: Idle: 12685, Running: 4, Completed: 3331 [ 3h 29m ] +INFO: Idle: 12683, Running: 4, Completed: 3333 [ 3h 30m ] +INFO: Idle: 12681, Running: 4, Completed: 3335 [ 3h 30m ] +INFO: Idle: 12679, Running: 4, Completed: 3337 [ 3h 30m ] +INFO: Idle: 12678, Running: 4, Completed: 3338 [ 3h 30m ] +INFO: Idle: 12677, Running: 4, Completed: 3339 [ 3h 30m ] +INFO: Idle: 12676, Running: 4, Completed: 3340 [ 3h 30m ] +INFO: Idle: 12675, Running: 4, Completed: 3341 [ 3h 30m ] +INFO: Idle: 12674, Running: 4, Completed: 3342 [ 3h 30m ] +INFO: Idle: 12672, Running: 4, Completed: 3344 [ 3h 31m ] +INFO: Idle: 12671, Running: 4, Completed: 3345 [ 3h 31m ] +INFO: Idle: 12669, Running: 4, Completed: 3347 [ 3h 31m ] +INFO: Idle: 12668, Running: 4, Completed: 3348 [ 3h 31m ] +INFO: Idle: 12667, Running: 4, Completed: 3349 [ 3h 31m ] +INFO: Idle: 12665, Running: 4, Completed: 3351 [ 3h 31m ] +INFO: Idle: 12663, Running: 4, Completed: 3353 [ 3h 31m ] +INFO: Idle: 12661, Running: 4, Completed: 3355 [ 3h 31m ] +INFO: Idle: 12659, Running: 4, Completed: 3357 [ 3h 32m ] +INFO: Idle: 12658, Running: 4, Completed: 3358 [ 3h 32m ] +INFO: Idle: 12656, Running: 4, Completed: 3360 [ 3h 32m ] +INFO: Idle: 12655, Running: 4, Completed: 3361 [ 3h 32m ] +INFO: Idle: 12653, Running: 4, Completed: 3363 [ 3h 32m ] +INFO: Idle: 12652, Running: 4, Completed: 3364 [ 3h 32m ] +INFO: Idle: 12651, Running: 4, Completed: 3365 [ 3h 33m ] +INFO: Idle: 12649, Running: 4, Completed: 3367 [ 3h 33m ] +INFO: Idle: 12648, Running: 4, Completed: 3368 [ 3h 33m ] +INFO: Idle: 12647, Running: 4, Completed: 3369 [ 3h 33m ] +INFO: Idle: 12645, Running: 4, Completed: 3371 [ 3h 33m ] +INFO: Idle: 12644, Running: 4, Completed: 3372 [ 3h 33m ] +INFO: Idle: 12643, Running: 4, Completed: 3373 [ 3h 33m ] +INFO: Idle: 12641, Running: 4, Completed: 3375 [ 3h 33m ] +INFO: Idle: 12640, Running: 4, Completed: 3376 [ 3h 34m ] +INFO: Idle: 12637, Running: 4, Completed: 3379 [ 3h 34m ] +INFO: Idle: 12635, Running: 4, Completed: 3381 [ 3h 34m ] +INFO: Idle: 12634, Running: 4, Completed: 3382 [ 3h 34m ] +INFO: Idle: 12633, Running: 4, Completed: 3383 [ 3h 34m ] +INFO: Idle: 12632, Running: 4, Completed: 3384 [ 3h 34m ] +INFO: Idle: 12631, Running: 4, Completed: 3385 [ 3h 34m ] +INFO: Idle: 12630, Running: 4, Completed: 3386 [ 3h 35m ] +INFO: Idle: 12628, Running: 4, Completed: 3388 [ 3h 35m ] +INFO: Idle: 12627, Running: 4, Completed: 3389 [ 3h 35m ] +INFO: Idle: 12626, Running: 4, Completed: 3390 [ 3h 35m ] +INFO: Idle: 12625, Running: 4, Completed: 3391 [ 3h 35m ] +INFO: Idle: 12624, Running: 4, Completed: 3392 [ 3h 35m ] +INFO: Idle: 12623, Running: 4, Completed: 3393 [ 3h 35m ] +INFO: Idle: 12621, Running: 4, Completed: 3395 [ 3h 36m ] +INFO: Idle: 12620, Running: 4, Completed: 3396 [ 3h 36m ] +INFO: Idle: 12618, Running: 4, Completed: 3398 [ 3h 36m ] +INFO: Idle: 12616, Running: 4, Completed: 3400 [ 3h 36m ] +INFO: Idle: 12615, Running: 4, Completed: 3401 [ 3h 36m ] +INFO: Idle: 12612, Running: 4, Completed: 3404 [ 3h 36m ] +INFO: Idle: 12610, Running: 4, Completed: 3406 [ 3h 37m ] +INFO: Idle: 12608, Running: 4, Completed: 3408 [ 3h 37m ] +INFO: Idle: 12606, Running: 4, Completed: 3410 [ 3h 37m ] +INFO: Idle: 12605, Running: 4, Completed: 3411 [ 3h 37m ] +INFO: Idle: 12604, Running: 4, Completed: 3412 [ 3h 37m ] +INFO: Idle: 12603, Running: 4, Completed: 3413 [ 3h 37m ] +INFO: Idle: 12601, Running: 4, Completed: 3415 [ 3h 37m ] +INFO: Idle: 12600, Running: 4, Completed: 3416 [ 3h 38m ] +INFO: Idle: 12598, Running: 4, Completed: 3418 [ 3h 38m ] +INFO: Idle: 12597, Running: 4, Completed: 3419 [ 3h 38m ] +INFO: Idle: 12596, Running: 4, Completed: 3420 [ 3h 38m ] +INFO: Idle: 12595, Running: 4, Completed: 3421 [ 3h 38m ] +INFO: Idle: 12594, Running: 4, Completed: 3422 [ 3h 38m ] +INFO: Idle: 12592, Running: 4, Completed: 3424 [ 3h 38m ] +INFO: Idle: 12591, Running: 4, Completed: 3425 [ 3h 38m ] +INFO: Idle: 12590, Running: 4, Completed: 3426 [ 3h 38m ] +INFO: Idle: 12588, Running: 4, Completed: 3428 [ 3h 39m ] +INFO: Idle: 12587, Running: 4, Completed: 3429 [ 3h 39m ] +INFO: Idle: 12586, Running: 4, Completed: 3430 [ 3h 39m ] +INFO: Idle: 12585, Running: 4, Completed: 3431 [ 3h 39m ] +INFO: Idle: 12584, Running: 4, Completed: 3432 [ 3h 39m ] +INFO: Idle: 12583, Running: 4, Completed: 3433 [ 3h 39m ] +INFO: Idle: 12581, Running: 4, Completed: 3435 [ 3h 39m ] +INFO: Idle: 12580, Running: 4, Completed: 3436 [ 3h 39m ] +INFO: Idle: 12579, Running: 4, Completed: 3437 [ 3h 39m ] +INFO: Idle: 12578, Running: 4, Completed: 3438 [ 3h 39m ] +INFO: Idle: 12576, Running: 4, Completed: 3440 [ 3h 40m ] +INFO: Idle: 12575, Running: 4, Completed: 3441 [ 3h 40m ] +INFO: Idle: 12573, Running: 4, Completed: 3443 [ 3h 40m ] +INFO: Idle: 12571, Running: 4, Completed: 3445 [ 3h 40m ] +INFO: Idle: 12569, Running: 4, Completed: 3447 [ 3h 40m ] +INFO: Idle: 12568, Running: 4, Completed: 3448 [ 3h 40m ] +INFO: Idle: 12567, Running: 4, Completed: 3449 [ 3h 40m ] +INFO: Idle: 12564, Running: 4, Completed: 3452 [ 3h 40m ] +INFO: Idle: 12563, Running: 4, Completed: 3453 [ 3h 41m ] +INFO: Idle: 12562, Running: 4, Completed: 3454 [ 3h 41m ] +INFO: Idle: 12561, Running: 4, Completed: 3455 [ 3h 41m ] +INFO: Idle: 12559, Running: 4, Completed: 3457 [ 3h 41m ] +INFO: Idle: 12558, Running: 4, Completed: 3458 [ 3h 41m ] +INFO: Idle: 12557, Running: 4, Completed: 3459 [ 3h 41m ] +INFO: Idle: 12556, Running: 4, Completed: 3460 [ 3h 42m ] +INFO: Idle: 12555, Running: 4, Completed: 3461 [ 3h 42m ] +INFO: Idle: 12554, Running: 4, Completed: 3462 [ 3h 42m ] +INFO: Idle: 12553, Running: 4, Completed: 3463 [ 3h 42m ] +INFO: Idle: 12552, Running: 4, Completed: 3464 [ 3h 42m ] +INFO: Idle: 12551, Running: 4, Completed: 3465 [ 3h 42m ] +INFO: Idle: 12550, Running: 4, Completed: 3466 [ 3h 42m ] +INFO: Idle: 12548, Running: 4, Completed: 3468 [ 3h 42m ] +INFO: Idle: 12547, Running: 4, Completed: 3469 [ 3h 42m ] +INFO: Idle: 12546, Running: 4, Completed: 3470 [ 3h 42m ] +INFO: Idle: 12545, Running: 4, Completed: 3471 [ 3h 42m ] +INFO: Idle: 12543, Running: 4, Completed: 3473 [ 3h 43m ] +INFO: Idle: 12541, Running: 4, Completed: 3475 [ 3h 43m ] +INFO: Idle: 12539, Running: 4, Completed: 3477 [ 3h 43m ] +INFO: Idle: 12538, Running: 4, Completed: 3478 [ 3h 43m ] +INFO: Idle: 12537, Running: 4, Completed: 3479 [ 3h 43m ] +INFO: Idle: 12536, Running: 4, Completed: 3480 [ 3h 43m ] +INFO: Idle: 12535, Running: 4, Completed: 3481 [ 3h 43m ] +INFO: Idle: 12534, Running: 4, Completed: 3482 [ 3h 43m ] +INFO: Idle: 12533, Running: 4, Completed: 3483 [ 3h 44m ] +INFO: Idle: 12531, Running: 4, Completed: 3485 [ 3h 44m ] +INFO: Idle: 12530, Running: 4, Completed: 3486 [ 3h 44m ] +INFO: Idle: 12528, Running: 4, Completed: 3488 [ 3h 44m ] +INFO: Idle: 12527, Running: 4, Completed: 3489 [ 3h 44m ] +INFO: Idle: 12526, Running: 4, Completed: 3490 [ 3h 44m ] +INFO: Idle: 12525, Running: 4, Completed: 3491 [ 3h 44m ] +INFO: Idle: 12524, Running: 4, Completed: 3492 [ 3h 44m ] +INFO: Idle: 12522, Running: 4, Completed: 3494 [ 3h 45m ] +INFO: Idle: 12521, Running: 4, Completed: 3495 [ 3h 45m ] +INFO: Idle: 12519, Running: 4, Completed: 3497 [ 3h 45m ] +INFO: Idle: 12517, Running: 4, Completed: 3499 [ 3h 45m ] +INFO: Idle: 12515, Running: 4, Completed: 3501 [ 3h 45m ] +INFO: Idle: 12514, Running: 4, Completed: 3502 [ 3h 45m ] +INFO: Idle: 12513, Running: 4, Completed: 3503 [ 3h 45m ] +INFO: Idle: 12512, Running: 4, Completed: 3504 [ 3h 46m ] +INFO: Idle: 12511, Running: 4, Completed: 3505 [ 3h 46m ] +INFO: Idle: 12510, Running: 4, Completed: 3506 [ 3h 46m ] +INFO: Idle: 12509, Running: 4, Completed: 3507 [ 3h 46m ] +INFO: Idle: 12508, Running: 4, Completed: 3508 [ 3h 46m ] +INFO: Idle: 12507, Running: 4, Completed: 3509 [ 3h 46m ] +INFO: Idle: 12505, Running: 4, Completed: 3511 [ 3h 46m ] +INFO: Idle: 12504, Running: 4, Completed: 3512 [ 3h 46m ] +INFO: Idle: 12502, Running: 4, Completed: 3514 [ 3h 47m ] +INFO: Idle: 12501, Running: 4, Completed: 3515 [ 3h 47m ] +INFO: Idle: 12500, Running: 4, Completed: 3516 [ 3h 47m ] +INFO: Idle: 12498, Running: 4, Completed: 3518 [ 3h 47m ] +INFO: Idle: 12497, Running: 4, Completed: 3519 [ 3h 47m ] +INFO: Idle: 12496, Running: 4, Completed: 3520 [ 3h 47m ] +INFO: Idle: 12495, Running: 4, Completed: 3521 [ 3h 47m ] +INFO: Idle: 12493, Running: 4, Completed: 3523 [ 3h 47m ] +INFO: Idle: 12492, Running: 4, Completed: 3524 [ 3h 48m ] +INFO: Idle: 12490, Running: 4, Completed: 3526 [ 3h 48m ] +INFO: Idle: 12488, Running: 4, Completed: 3528 [ 3h 48m ] +INFO: Idle: 12487, Running: 4, Completed: 3529 [ 3h 48m ] +INFO: Idle: 12485, Running: 4, Completed: 3531 [ 3h 48m ] +INFO: Idle: 12484, Running: 4, Completed: 3532 [ 3h 48m ] +INFO: Idle: 12483, Running: 4, Completed: 3533 [ 3h 48m ] +INFO: Idle: 12481, Running: 4, Completed: 3535 [ 3h 48m ] +INFO: Idle: 12480, Running: 4, Completed: 3536 [ 3h 49m ] +INFO: Idle: 12477, Running: 4, Completed: 3539 [ 3h 49m ] +INFO: Idle: 12476, Running: 4, Completed: 3540 [ 3h 49m ] +INFO: Idle: 12474, Running: 4, Completed: 3542 [ 3h 49m ] +INFO: Idle: 12472, Running: 4, Completed: 3544 [ 3h 49m ] +INFO: Idle: 12471, Running: 4, Completed: 3545 [ 3h 50m ] +INFO: Idle: 12470, Running: 4, Completed: 3546 [ 3h 50m ] +INFO: Idle: 12468, Running: 4, Completed: 3548 [ 3h 50m ] +INFO: Idle: 12467, Running: 4, Completed: 3549 [ 3h 50m ] +INFO: Idle: 12466, Running: 4, Completed: 3550 [ 3h 50m ] +INFO: Idle: 12465, Running: 4, Completed: 3551 [ 3h 50m ] +INFO: Idle: 12463, Running: 4, Completed: 3553 [ 3h 50m ] +INFO: Idle: 12461, Running: 4, Completed: 3555 [ 3h 50m ] +INFO: Idle: 12459, Running: 4, Completed: 3557 [ 3h 51m ] +INFO: Idle: 12458, Running: 4, Completed: 3558 [ 3h 51m ] +INFO: Idle: 12456, Running: 4, Completed: 3560 [ 3h 51m ] +INFO: Idle: 12455, Running: 4, Completed: 3561 [ 3h 51m ] +INFO: Idle: 12453, Running: 4, Completed: 3563 [ 3h 51m ] +INFO: Idle: 12451, Running: 4, Completed: 3565 [ 3h 51m ] +INFO: Idle: 12450, Running: 4, Completed: 3566 [ 3h 51m ] +INFO: Idle: 12449, Running: 4, Completed: 3567 [ 3h 51m ] +INFO: Idle: 12447, Running: 4, Completed: 3569 [ 3h 52m ] +INFO: Idle: 12444, Running: 4, Completed: 3572 [ 3h 52m ] +INFO: Idle: 12443, Running: 4, Completed: 3573 [ 3h 52m ] +INFO: Idle: 12441, Running: 4, Completed: 3575 [ 3h 52m ] +INFO: Idle: 12439, Running: 4, Completed: 3577 [ 3h 52m ] +INFO: Idle: 12437, Running: 4, Completed: 3579 [ 3h 53m ] +INFO: Idle: 12436, Running: 4, Completed: 3580 [ 3h 53m ] +INFO: Idle: 12435, Running: 4, Completed: 3581 [ 3h 53m ] +INFO: Idle: 12433, Running: 4, Completed: 3583 [ 3h 53m ] +INFO: Idle: 12432, Running: 4, Completed: 3584 [ 3h 53m ] +INFO: Idle: 12431, Running: 4, Completed: 3585 [ 3h 53m ] +INFO: Idle: 12428, Running: 4, Completed: 3588 [ 3h 53m ] +INFO: Idle: 12427, Running: 4, Completed: 3589 [ 3h 54m ] +INFO: Idle: 12425, Running: 4, Completed: 3591 [ 3h 54m ] +INFO: Idle: 12423, Running: 4, Completed: 3593 [ 3h 54m ] +INFO: Idle: 12421, Running: 4, Completed: 3595 [ 3h 54m ] +INFO: Idle: 12418, Running: 4, Completed: 3598 [ 3h 54m ] +INFO: Idle: 12417, Running: 4, Completed: 3599 [ 3h 54m ] +INFO: Idle: 12415, Running: 4, Completed: 3601 [ 3h 55m ] +INFO: Idle: 12414, Running: 4, Completed: 3602 [ 3h 55m ] +INFO: Idle: 12412, Running: 4, Completed: 3604 [ 3h 55m ] +INFO: Idle: 12411, Running: 4, Completed: 3605 [ 3h 55m ] +INFO: Idle: 12409, Running: 4, Completed: 3607 [ 3h 55m ] +INFO: Idle: 12408, Running: 4, Completed: 3608 [ 3h 55m ] +INFO: Idle: 12407, Running: 4, Completed: 3609 [ 3h 55m ] +INFO: Idle: 12406, Running: 4, Completed: 3610 [ 3h 55m ] +INFO: Idle: 12405, Running: 4, Completed: 3611 [ 3h 56m ] +INFO: Idle: 12402, Running: 4, Completed: 3614 [ 3h 56m ] +INFO: Idle: 12401, Running: 4, Completed: 3615 [ 3h 56m ] +INFO: Idle: 12399, Running: 4, Completed: 3617 [ 3h 56m ] +INFO: Idle: 12398, Running: 4, Completed: 3618 [ 3h 56m ] +INFO: Idle: 12397, Running: 4, Completed: 3619 [ 3h 56m ] +INFO: Idle: 12396, Running: 4, Completed: 3620 [ 3h 56m ] +INFO: Idle: 12394, Running: 4, Completed: 3622 [ 3h 57m ] +INFO: Idle: 12393, Running: 4, Completed: 3623 [ 3h 57m ] +INFO: Idle: 12392, Running: 4, Completed: 3624 [ 3h 57m ] +INFO: Idle: 12390, Running: 4, Completed: 3626 [ 3h 57m ] +INFO: Idle: 12388, Running: 4, Completed: 3628 [ 3h 57m ] +INFO: Idle: 12385, Running: 4, Completed: 3631 [ 3h 57m ] +INFO: Idle: 12384, Running: 4, Completed: 3632 [ 3h 57m ] +INFO: Idle: 12381, Running: 4, Completed: 3635 [ 3h 57m ] +INFO: Idle: 12380, Running: 4, Completed: 3636 [ 3h 58m ] +INFO: Idle: 12379, Running: 4, Completed: 3637 [ 3h 58m ] +INFO: Idle: 12376, Running: 4, Completed: 3640 [ 3h 58m ] +INFO: Idle: 12375, Running: 4, Completed: 3641 [ 3h 58m ] +INFO: Idle: 12373, Running: 4, Completed: 3643 [ 3h 58m ] +INFO: Idle: 12372, Running: 4, Completed: 3644 [ 3h 58m ] +INFO: Idle: 12371, Running: 4, Completed: 3645 [ 3h 58m ] +INFO: Idle: 12369, Running: 4, Completed: 3647 [ 3h 58m ] +INFO: Idle: 12368, Running: 4, Completed: 3648 [ 3h 59m ] +INFO: Idle: 12367, Running: 4, Completed: 3649 [ 3h 59m ] +INFO: Idle: 12365, Running: 4, Completed: 3651 [ 3h 59m ] +INFO: Idle: 12364, Running: 4, Completed: 3652 [ 3h 59m ] +INFO: Idle: 12362, Running: 4, Completed: 3654 [ 3h 59m ] +INFO: Idle: 12361, Running: 4, Completed: 3655 [ 3h 59m ] +INFO: Idle: 12360, Running: 4, Completed: 3656 [ 3h 59m ] +INFO: Idle: 12358, Running: 4, Completed: 3658 [ 3h 59m ] +INFO: Idle: 12356, Running: 4, Completed: 3660 [ 4h 0m ] +INFO: Idle: 12355, Running: 4, Completed: 3661 [ 4h 0m ] +INFO: Idle: 12354, Running: 4, Completed: 3662 [ 4h 0m ] +INFO: Idle: 12352, Running: 4, Completed: 3664 [ 4h 0m ] +INFO: Idle: 12351, Running: 4, Completed: 3665 [ 4h 0m ] +INFO: Idle: 12350, Running: 4, Completed: 3666 [ 4h 0m ] +INFO: Idle: 12347, Running: 4, Completed: 3669 [ 4h 0m ] +INFO: Idle: 12346, Running: 4, Completed: 3670 [ 4h 0m ] +INFO: Idle: 12344, Running: 4, Completed: 3672 [ 4h 1m ] +INFO: Idle: 12343, Running: 4, Completed: 3673 [ 4h 1m ] +INFO: Idle: 12340, Running: 4, Completed: 3676 [ 4h 1m ] +INFO: Idle: 12339, Running: 4, Completed: 3677 [ 4h 1m ] +INFO: Idle: 12338, Running: 4, Completed: 3678 [ 4h 1m ] +INFO: Idle: 12336, Running: 4, Completed: 3680 [ 4h 1m ] +INFO: Idle: 12335, Running: 4, Completed: 3681 [ 4h 1m ] +INFO: Idle: 12332, Running: 4, Completed: 3684 [ 4h 1m ] +INFO: Idle: 12331, Running: 4, Completed: 3685 [ 4h 2m ] +INFO: Idle: 12328, Running: 4, Completed: 3688 [ 4h 2m ] +INFO: Idle: 12327, Running: 4, Completed: 3689 [ 4h 2m ] +INFO: Idle: 12325, Running: 4, Completed: 3691 [ 4h 2m ] +INFO: Idle: 12323, Running: 4, Completed: 3693 [ 4h 2m ] +INFO: Idle: 12321, Running: 4, Completed: 3695 [ 4h 2m ] +INFO: Idle: 12319, Running: 4, Completed: 3697 [ 4h 3m ] +INFO: Idle: 12317, Running: 4, Completed: 3699 [ 4h 3m ] +INFO: Idle: 12316, Running: 4, Completed: 3700 [ 4h 3m ] +INFO: Idle: 12315, Running: 4, Completed: 3701 [ 4h 3m ] +INFO: Idle: 12313, Running: 4, Completed: 3703 [ 4h 3m ] +INFO: Idle: 12312, Running: 4, Completed: 3704 [ 4h 3m ] +INFO: Idle: 12311, Running: 4, Completed: 3705 [ 4h 3m ] +INFO: Idle: 12309, Running: 4, Completed: 3707 [ 4h 3m ] +INFO: Idle: 12308, Running: 4, Completed: 3708 [ 4h 4m ] +INFO: Idle: 12307, Running: 4, Completed: 3709 [ 4h 4m ] +INFO: Idle: 12305, Running: 4, Completed: 3711 [ 4h 4m ] +INFO: Idle: 12304, Running: 4, Completed: 3712 [ 4h 4m ] +INFO: Idle: 12303, Running: 4, Completed: 3713 [ 4h 4m ] +INFO: Idle: 12301, Running: 4, Completed: 3715 [ 4h 4m ] +INFO: Idle: 12300, Running: 4, Completed: 3716 [ 4h 4m ] +INFO: Idle: 12298, Running: 4, Completed: 3718 [ 4h 5m ] +INFO: Idle: 12297, Running: 4, Completed: 3719 [ 4h 5m ] +INFO: Idle: 12296, Running: 4, Completed: 3720 [ 4h 5m ] +INFO: Idle: 12295, Running: 4, Completed: 3721 [ 4h 5m ] +INFO: Idle: 12294, Running: 4, Completed: 3722 [ 4h 5m ] +INFO: Idle: 12293, Running: 4, Completed: 3723 [ 4h 5m ] +INFO: Idle: 12291, Running: 4, Completed: 3725 [ 4h 5m ] +INFO: Idle: 12289, Running: 4, Completed: 3727 [ 4h 5m ] +INFO: Idle: 12288, Running: 4, Completed: 3728 [ 4h 5m ] +INFO: Idle: 12287, Running: 4, Completed: 3729 [ 4h 6m ] +INFO: Idle: 12286, Running: 4, Completed: 3730 [ 4h 6m ] +INFO: Idle: 12285, Running: 4, Completed: 3731 [ 4h 6m ] +INFO: Idle: 12284, Running: 4, Completed: 3732 [ 4h 6m ] +INFO: Idle: 12282, Running: 4, Completed: 3734 [ 4h 6m ] +INFO: Idle: 12281, Running: 4, Completed: 3735 [ 4h 6m ] +INFO: Idle: 12279, Running: 4, Completed: 3737 [ 4h 6m ] +INFO: Idle: 12278, Running: 4, Completed: 3738 [ 4h 6m ] +INFO: Idle: 12277, Running: 4, Completed: 3739 [ 4h 7m ] +INFO: Idle: 12276, Running: 4, Completed: 3740 [ 4h 7m ] +INFO: Idle: 12274, Running: 4, Completed: 3742 [ 4h 7m ] +INFO: Idle: 12273, Running: 4, Completed: 3743 [ 4h 7m ] +INFO: Idle: 12272, Running: 4, Completed: 3744 [ 4h 7m ] +INFO: Idle: 12270, Running: 4, Completed: 3746 [ 4h 7m ] +INFO: Idle: 12268, Running: 4, Completed: 3748 [ 4h 7m ] +INFO: Idle: 12266, Running: 4, Completed: 3750 [ 4h 7m ] +INFO: Idle: 12264, Running: 4, Completed: 3752 [ 4h 8m ] +INFO: Idle: 12261, Running: 4, Completed: 3755 [ 4h 8m ] +INFO: Idle: 12260, Running: 4, Completed: 3756 [ 4h 8m ] +INFO: Idle: 12259, Running: 4, Completed: 3757 [ 4h 8m ] +INFO: Idle: 12256, Running: 4, Completed: 3760 [ 4h 8m ] +INFO: Idle: 12255, Running: 4, Completed: 3761 [ 4h 8m ] +INFO: Idle: 12254, Running: 4, Completed: 3762 [ 4h 8m ] +INFO: Idle: 12253, Running: 4, Completed: 3763 [ 4h 8m ] +INFO: Idle: 12252, Running: 4, Completed: 3764 [ 4h 8m ] +INFO: Idle: 12251, Running: 4, Completed: 3765 [ 4h 9m ] +INFO: Idle: 12249, Running: 4, Completed: 3767 [ 4h 9m ] +INFO: Idle: 12248, Running: 4, Completed: 3768 [ 4h 9m ] +INFO: Idle: 12247, Running: 4, Completed: 3769 [ 4h 9m ] +INFO: Idle: 12245, Running: 4, Completed: 3771 [ 4h 9m ] +INFO: Idle: 12244, Running: 4, Completed: 3772 [ 4h 9m ] +INFO: Idle: 12242, Running: 4, Completed: 3774 [ 4h 9m ] +INFO: Idle: 12241, Running: 4, Completed: 3775 [ 4h 9m ] +INFO: Idle: 12240, Running: 4, Completed: 3776 [ 4h 10m ] +INFO: Idle: 12238, Running: 4, Completed: 3778 [ 4h 10m ] +INFO: Idle: 12236, Running: 4, Completed: 3780 [ 4h 10m ] +INFO: Idle: 12235, Running: 4, Completed: 3781 [ 4h 10m ] +INFO: Idle: 12234, Running: 4, Completed: 3782 [ 4h 10m ] +INFO: Idle: 12232, Running: 4, Completed: 3784 [ 4h 10m ] +INFO: Idle: 12231, Running: 4, Completed: 3785 [ 4h 10m ] +INFO: Idle: 12230, Running: 4, Completed: 3786 [ 4h 10m ] +INFO: Idle: 12228, Running: 4, Completed: 3788 [ 4h 10m ] +INFO: Idle: 12227, Running: 4, Completed: 3789 [ 4h 11m ] +INFO: Idle: 12226, Running: 4, Completed: 3790 [ 4h 11m ] +INFO: Idle: 12225, Running: 4, Completed: 3791 [ 4h 11m ] +INFO: Idle: 12224, Running: 4, Completed: 3792 [ 4h 11m ] +INFO: Idle: 12223, Running: 4, Completed: 3793 [ 4h 11m ] +INFO: Idle: 12221, Running: 4, Completed: 3795 [ 4h 11m ] +INFO: Idle: 12220, Running: 4, Completed: 3796 [ 4h 11m ] +INFO: Idle: 12219, Running: 4, Completed: 3797 [ 4h 11m ] +INFO: Idle: 12217, Running: 4, Completed: 3799 [ 4h 11m ] +INFO: Idle: 12215, Running: 4, Completed: 3801 [ 4h 12m ] +INFO: Idle: 12211, Running: 4, Completed: 3805 [ 4h 12m ] +INFO: Idle: 12208, Running: 4, Completed: 3808 [ 4h 12m ] +INFO: Idle: 12207, Running: 4, Completed: 3809 [ 4h 12m ] +INFO: Idle: 12206, Running: 4, Completed: 3810 [ 4h 12m ] +INFO: Idle: 12204, Running: 4, Completed: 3812 [ 4h 12m ] +INFO: Idle: 12203, Running: 4, Completed: 3813 [ 4h 13m ] +INFO: Idle: 12201, Running: 4, Completed: 3815 [ 4h 13m ] +INFO: Idle: 12200, Running: 4, Completed: 3816 [ 4h 13m ] +INFO: Idle: 12199, Running: 4, Completed: 3817 [ 4h 13m ] +INFO: Idle: 12197, Running: 4, Completed: 3819 [ 4h 13m ] +INFO: Idle: 12196, Running: 4, Completed: 3820 [ 4h 13m ] +INFO: Idle: 12195, Running: 4, Completed: 3821 [ 4h 13m ] +INFO: Idle: 12194, Running: 4, Completed: 3822 [ 4h 13m ] +INFO: Idle: 12192, Running: 4, Completed: 3824 [ 4h 13m ] +INFO: Idle: 12191, Running: 4, Completed: 3825 [ 4h 14m ] +INFO: Idle: 12188, Running: 4, Completed: 3828 [ 4h 14m ] +INFO: Idle: 12187, Running: 4, Completed: 3829 [ 4h 14m ] +INFO: Idle: 12186, Running: 4, Completed: 3830 [ 4h 14m ] +INFO: Idle: 12185, Running: 4, Completed: 3831 [ 4h 14m ] +INFO: Idle: 12184, Running: 4, Completed: 3832 [ 4h 14m ] +INFO: Idle: 12183, Running: 4, Completed: 3833 [ 4h 14m ] +INFO: Idle: 12182, Running: 4, Completed: 3834 [ 4h 15m ] +INFO: Idle: 12181, Running: 4, Completed: 3835 [ 4h 15m ] +INFO: Idle: 12179, Running: 4, Completed: 3837 [ 4h 15m ] +INFO: Idle: 12178, Running: 4, Completed: 3838 [ 4h 15m ] +INFO: Idle: 12176, Running: 4, Completed: 3840 [ 4h 15m ] +INFO: Idle: 12175, Running: 4, Completed: 3841 [ 4h 15m ] +INFO: Idle: 12174, Running: 4, Completed: 3842 [ 4h 15m ] +INFO: Idle: 12173, Running: 4, Completed: 3843 [ 4h 15m ] +INFO: Idle: 12172, Running: 4, Completed: 3844 [ 4h 16m ] +INFO: Idle: 12170, Running: 4, Completed: 3846 [ 4h 16m ] +INFO: Idle: 12168, Running: 4, Completed: 3848 [ 4h 16m ] +INFO: Idle: 12167, Running: 4, Completed: 3849 [ 4h 16m ] +INFO: Idle: 12166, Running: 4, Completed: 3850 [ 4h 16m ] +INFO: Idle: 12165, Running: 4, Completed: 3851 [ 4h 16m ] +INFO: Idle: 12164, Running: 4, Completed: 3852 [ 4h 16m ] +INFO: Idle: 12163, Running: 4, Completed: 3853 [ 4h 16m ] +INFO: Idle: 12162, Running: 4, Completed: 3854 [ 4h 17m ] +INFO: Idle: 12161, Running: 4, Completed: 3855 [ 4h 17m ] +INFO: Idle: 12159, Running: 4, Completed: 3857 [ 4h 17m ] +INFO: Idle: 12158, Running: 4, Completed: 3858 [ 4h 17m ] +INFO: Idle: 12157, Running: 4, Completed: 3859 [ 4h 17m ] +INFO: Idle: 12155, Running: 4, Completed: 3861 [ 4h 17m ] +INFO: Idle: 12154, Running: 4, Completed: 3862 [ 4h 17m ] +INFO: Idle: 12153, Running: 4, Completed: 3863 [ 4h 17m ] +INFO: Idle: 12152, Running: 4, Completed: 3864 [ 4h 17m ] +INFO: Idle: 12151, Running: 4, Completed: 3865 [ 4h 17m ] +INFO: Idle: 12150, Running: 4, Completed: 3866 [ 4h 17m ] +INFO: Idle: 12149, Running: 4, Completed: 3867 [ 4h 18m ] +INFO: Idle: 12148, Running: 4, Completed: 3868 [ 4h 18m ] +INFO: Idle: 12147, Running: 4, Completed: 3869 [ 4h 18m ] +INFO: Idle: 12146, Running: 4, Completed: 3870 [ 4h 18m ] +INFO: Idle: 12145, Running: 4, Completed: 3871 [ 4h 18m ] +INFO: Idle: 12143, Running: 4, Completed: 3873 [ 4h 18m ] +INFO: Idle: 12141, Running: 4, Completed: 3875 [ 4h 18m ] +INFO: Idle: 12138, Running: 4, Completed: 3878 [ 4h 18m ] +INFO: Idle: 12137, Running: 4, Completed: 3879 [ 4h 18m ] +INFO: Idle: 12136, Running: 4, Completed: 3880 [ 4h 18m ] +INFO: Idle: 12135, Running: 4, Completed: 3881 [ 4h 19m ] +INFO: Idle: 12134, Running: 4, Completed: 3882 [ 4h 19m ] +INFO: Idle: 12131, Running: 4, Completed: 3885 [ 4h 19m ] +INFO: Idle: 12130, Running: 4, Completed: 3886 [ 4h 19m ] +INFO: Idle: 12129, Running: 4, Completed: 3887 [ 4h 19m ] +INFO: Idle: 12127, Running: 4, Completed: 3889 [ 4h 19m ] +INFO: Idle: 12125, Running: 4, Completed: 3891 [ 4h 19m ] +INFO: Idle: 12123, Running: 4, Completed: 3893 [ 4h 20m ] +INFO: Idle: 12122, Running: 4, Completed: 3894 [ 4h 20m ] +INFO: Idle: 12120, Running: 4, Completed: 3896 [ 4h 20m ] +INFO: Idle: 12119, Running: 4, Completed: 3897 [ 4h 20m ] +INFO: Idle: 12117, Running: 4, Completed: 3899 [ 4h 20m ] +INFO: Idle: 12115, Running: 4, Completed: 3901 [ 4h 20m ] +INFO: Idle: 12114, Running: 4, Completed: 3902 [ 4h 20m ] +INFO: Idle: 12113, Running: 4, Completed: 3903 [ 4h 20m ] +INFO: Idle: 12111, Running: 4, Completed: 3905 [ 4h 21m ] +INFO: Idle: 12109, Running: 4, Completed: 3907 [ 4h 21m ] +INFO: Idle: 12108, Running: 4, Completed: 3908 [ 4h 21m ] +INFO: Idle: 12107, Running: 4, Completed: 3909 [ 4h 21m ] +INFO: Idle: 12105, Running: 4, Completed: 3911 [ 4h 21m ] +INFO: Idle: 12103, Running: 4, Completed: 3913 [ 4h 21m ] +INFO: Idle: 12101, Running: 4, Completed: 3915 [ 4h 21m ] +INFO: Idle: 12099, Running: 4, Completed: 3917 [ 4h 22m ] +INFO: Idle: 12097, Running: 4, Completed: 3919 [ 4h 22m ] +INFO: Idle: 12095, Running: 4, Completed: 3921 [ 4h 22m ] +INFO: Idle: 12093, Running: 4, Completed: 3923 [ 4h 22m ] +INFO: Idle: 12090, Running: 4, Completed: 3926 [ 4h 22m ] +INFO: Idle: 12089, Running: 4, Completed: 3927 [ 4h 22m ] +INFO: Idle: 12087, Running: 4, Completed: 3929 [ 4h 23m ] +INFO: Idle: 12086, Running: 4, Completed: 3930 [ 4h 23m ] +INFO: Idle: 12085, Running: 4, Completed: 3931 [ 4h 23m ] +INFO: Idle: 12083, Running: 4, Completed: 3933 [ 4h 23m ] +INFO: Idle: 12080, Running: 4, Completed: 3936 [ 4h 23m ] +INFO: Idle: 12079, Running: 4, Completed: 3937 [ 4h 23m ] +INFO: Idle: 12076, Running: 4, Completed: 3940 [ 4h 24m ] +INFO: Idle: 12075, Running: 4, Completed: 3941 [ 4h 24m ] +INFO: Idle: 12072, Running: 4, Completed: 3944 [ 4h 24m ] +INFO: Idle: 12071, Running: 4, Completed: 3945 [ 4h 24m ] +INFO: Idle: 12070, Running: 4, Completed: 3946 [ 4h 24m ] +INFO: Idle: 12068, Running: 4, Completed: 3948 [ 4h 24m ] +INFO: Idle: 12067, Running: 4, Completed: 3949 [ 4h 24m ] +INFO: Idle: 12066, Running: 4, Completed: 3950 [ 4h 25m ] +INFO: Idle: 12064, Running: 4, Completed: 3952 [ 4h 25m ] +INFO: Idle: 12063, Running: 4, Completed: 3953 [ 4h 25m ] +INFO: Idle: 12061, Running: 4, Completed: 3955 [ 4h 25m ] +INFO: Idle: 12060, Running: 4, Completed: 3956 [ 4h 25m ] +INFO: Idle: 12058, Running: 4, Completed: 3958 [ 4h 25m ] +INFO: Idle: 12057, Running: 4, Completed: 3959 [ 4h 25m ] +INFO: Idle: 12056, Running: 4, Completed: 3960 [ 4h 25m ] +INFO: Idle: 12055, Running: 4, Completed: 3961 [ 4h 26m ] +INFO: Idle: 12054, Running: 4, Completed: 3962 [ 4h 26m ] +INFO: Idle: 12052, Running: 4, Completed: 3964 [ 4h 26m ] +INFO: Idle: 12051, Running: 4, Completed: 3965 [ 4h 26m ] +INFO: Idle: 12049, Running: 4, Completed: 3967 [ 4h 26m ] +INFO: Idle: 12047, Running: 4, Completed: 3969 [ 4h 26m ] +INFO: Idle: 12046, Running: 4, Completed: 3970 [ 4h 26m ] +INFO: Idle: 12045, Running: 4, Completed: 3971 [ 4h 27m ] +INFO: Idle: 12044, Running: 4, Completed: 3972 [ 4h 27m ] +INFO: Idle: 12042, Running: 4, Completed: 3974 [ 4h 27m ] +INFO: Idle: 12041, Running: 4, Completed: 3975 [ 4h 27m ] +INFO: Idle: 12039, Running: 4, Completed: 3977 [ 4h 27m ] +INFO: Idle: 12038, Running: 4, Completed: 3978 [ 4h 27m ] +INFO: Idle: 12037, Running: 4, Completed: 3979 [ 4h 27m ] +INFO: Idle: 12036, Running: 4, Completed: 3980 [ 4h 27m ] +INFO: Idle: 12034, Running: 4, Completed: 3982 [ 4h 27m ] +INFO: Idle: 12033, Running: 4, Completed: 3983 [ 4h 28m ] +INFO: Idle: 12032, Running: 4, Completed: 3984 [ 4h 28m ] +INFO: Idle: 12030, Running: 4, Completed: 3986 [ 4h 28m ] +INFO: Idle: 12028, Running: 4, Completed: 3988 [ 4h 28m ] +INFO: Idle: 12026, Running: 4, Completed: 3990 [ 4h 28m ] +INFO: Idle: 12024, Running: 4, Completed: 3992 [ 4h 28m ] +INFO: Idle: 12021, Running: 4, Completed: 3995 [ 4h 28m ] +INFO: Idle: 12020, Running: 4, Completed: 3996 [ 4h 29m ] +INFO: Idle: 12019, Running: 4, Completed: 3997 [ 4h 29m ] +INFO: Idle: 12017, Running: 4, Completed: 3999 [ 4h 29m ] +INFO: Idle: 12016, Running: 4, Completed: 4000 [ 4h 29m ] +INFO: Idle: 12015, Running: 4, Completed: 4001 [ 4h 29m ] +INFO: Idle: 12014, Running: 4, Completed: 4002 [ 4h 29m ] +INFO: Idle: 12013, Running: 4, Completed: 4003 [ 4h 29m ] +INFO: Idle: 12012, Running: 4, Completed: 4004 [ 4h 29m ] +INFO: Idle: 12011, Running: 4, Completed: 4005 [ 4h 29m ] +INFO: Idle: 12009, Running: 4, Completed: 4007 [ 4h 29m ] +INFO: Idle: 12008, Running: 4, Completed: 4008 [ 4h 29m ] +INFO: Idle: 12007, Running: 4, Completed: 4009 [ 4h 30m ] +INFO: Idle: 12005, Running: 4, Completed: 4011 [ 4h 30m ] +INFO: Idle: 12004, Running: 4, Completed: 4012 [ 4h 30m ] +INFO: Idle: 12002, Running: 4, Completed: 4014 [ 4h 30m ] +INFO: Idle: 12001, Running: 4, Completed: 4015 [ 4h 30m ] +INFO: Idle: 12000, Running: 4, Completed: 4016 [ 4h 30m ] +INFO: Idle: 11998, Running: 4, Completed: 4018 [ 4h 30m ] +INFO: Idle: 11996, Running: 4, Completed: 4020 [ 4h 31m ] +INFO: Idle: 11994, Running: 4, Completed: 4022 [ 4h 31m ] +INFO: Idle: 11992, Running: 4, Completed: 4024 [ 4h 31m ] +INFO: Idle: 11991, Running: 4, Completed: 4025 [ 4h 31m ] +INFO: Idle: 11990, Running: 4, Completed: 4026 [ 4h 31m ] +INFO: Idle: 11988, Running: 4, Completed: 4028 [ 4h 31m ] +INFO: Idle: 11987, Running: 4, Completed: 4029 [ 4h 31m ] +INFO: Idle: 11986, Running: 4, Completed: 4030 [ 4h 31m ] +INFO: Idle: 11984, Running: 4, Completed: 4032 [ 4h 32m ] +INFO: Idle: 11983, Running: 4, Completed: 4033 [ 4h 32m ] +INFO: Idle: 11981, Running: 4, Completed: 4035 [ 4h 32m ] +INFO: Idle: 11980, Running: 4, Completed: 4036 [ 4h 32m ] +INFO: Idle: 11979, Running: 4, Completed: 4037 [ 4h 32m ] +INFO: Idle: 11978, Running: 4, Completed: 4038 [ 4h 32m ] +INFO: Idle: 11976, Running: 4, Completed: 4040 [ 4h 32m ] +INFO: Idle: 11975, Running: 4, Completed: 4041 [ 4h 32m ] +INFO: Idle: 11971, Running: 4, Completed: 4045 [ 4h 33m ] +INFO: Idle: 11967, Running: 4, Completed: 4049 [ 4h 33m ] +INFO: Idle: 11965, Running: 4, Completed: 4051 [ 4h 33m ] +INFO: Idle: 11963, Running: 4, Completed: 4053 [ 4h 33m ] +INFO: Idle: 11961, Running: 4, Completed: 4055 [ 4h 33m ] +INFO: Idle: 11959, Running: 4, Completed: 4057 [ 4h 34m ] +INFO: Idle: 11957, Running: 4, Completed: 4059 [ 4h 34m ] +INFO: Idle: 11956, Running: 4, Completed: 4060 [ 4h 34m ] +INFO: Idle: 11955, Running: 4, Completed: 4061 [ 4h 34m ] +INFO: Idle: 11953, Running: 4, Completed: 4063 [ 4h 34m ] +INFO: Idle: 11952, Running: 4, Completed: 4064 [ 4h 34m ] +INFO: Idle: 11951, Running: 4, Completed: 4065 [ 4h 34m ] +INFO: Idle: 11948, Running: 4, Completed: 4068 [ 4h 34m ] +INFO: Idle: 11947, Running: 4, Completed: 4069 [ 4h 35m ] +INFO: Idle: 11945, Running: 4, Completed: 4071 [ 4h 35m ] +INFO: Idle: 11943, Running: 4, Completed: 4073 [ 4h 35m ] +INFO: Idle: 11941, Running: 4, Completed: 4075 [ 4h 35m ] +INFO: Idle: 11939, Running: 4, Completed: 4077 [ 4h 35m ] +INFO: Idle: 11938, Running: 4, Completed: 4078 [ 4h 35m ] +INFO: Idle: 11937, Running: 4, Completed: 4079 [ 4h 35m ] +INFO: Idle: 11936, Running: 4, Completed: 4080 [ 4h 35m ] +INFO: Idle: 11935, Running: 4, Completed: 4081 [ 4h 36m ] +INFO: Idle: 11933, Running: 4, Completed: 4083 [ 4h 36m ] +INFO: Idle: 11932, Running: 4, Completed: 4084 [ 4h 36m ] +INFO: Idle: 11931, Running: 4, Completed: 4085 [ 4h 36m ] +INFO: Idle: 11928, Running: 4, Completed: 4088 [ 4h 36m ] +INFO: Idle: 11927, Running: 4, Completed: 4089 [ 4h 36m ] +INFO: Idle: 11926, Running: 4, Completed: 4090 [ 4h 36m ] +INFO: Idle: 11924, Running: 4, Completed: 4092 [ 4h 37m ] +INFO: Idle: 11923, Running: 4, Completed: 4093 [ 4h 37m ] +INFO: Idle: 11922, Running: 4, Completed: 4094 [ 4h 37m ] +INFO: Idle: 11920, Running: 4, Completed: 4096 [ 4h 37m ] +INFO: Idle: 11919, Running: 4, Completed: 4097 [ 4h 37m ] +INFO: Idle: 11918, Running: 4, Completed: 4098 [ 4h 37m ] +INFO: Idle: 11917, Running: 4, Completed: 4099 [ 4h 37m ] +INFO: Idle: 11916, Running: 4, Completed: 4100 [ 4h 37m ] +INFO: Idle: 11914, Running: 4, Completed: 4102 [ 4h 37m ] +INFO: Idle: 11913, Running: 4, Completed: 4103 [ 4h 37m ] +INFO: Idle: 11912, Running: 4, Completed: 4104 [ 4h 38m ] +INFO: Idle: 11911, Running: 4, Completed: 4105 [ 4h 38m ] +INFO: Idle: 11910, Running: 4, Completed: 4106 [ 4h 38m ] +INFO: Idle: 11909, Running: 4, Completed: 4107 [ 4h 38m ] +INFO: Idle: 11908, Running: 4, Completed: 4108 [ 4h 38m ] +INFO: Idle: 11907, Running: 4, Completed: 4109 [ 4h 38m ] +INFO: Idle: 11905, Running: 4, Completed: 4111 [ 4h 38m ] +INFO: Idle: 11904, Running: 4, Completed: 4112 [ 4h 38m ] +INFO: Idle: 11901, Running: 4, Completed: 4115 [ 4h 39m ] +INFO: Idle: 11900, Running: 4, Completed: 4116 [ 4h 39m ] +INFO: Idle: 11899, Running: 4, Completed: 4117 [ 4h 39m ] +INFO: Idle: 11897, Running: 4, Completed: 4119 [ 4h 39m ] +INFO: Idle: 11896, Running: 4, Completed: 4120 [ 4h 39m ] +INFO: Idle: 11895, Running: 4, Completed: 4121 [ 4h 39m ] +INFO: Idle: 11893, Running: 4, Completed: 4123 [ 4h 39m ] +INFO: Idle: 11892, Running: 4, Completed: 4124 [ 4h 40m ] +INFO: Idle: 11891, Running: 4, Completed: 4125 [ 4h 40m ] +INFO: Idle: 11890, Running: 4, Completed: 4126 [ 4h 40m ] +INFO: Idle: 11888, Running: 4, Completed: 4128 [ 4h 40m ] +INFO: Idle: 11887, Running: 4, Completed: 4129 [ 4h 40m ] +INFO: Idle: 11885, Running: 4, Completed: 4131 [ 4h 40m ] +INFO: Idle: 11884, Running: 4, Completed: 4132 [ 4h 40m ] +INFO: Idle: 11883, Running: 4, Completed: 4133 [ 4h 40m ] +INFO: Idle: 11881, Running: 4, Completed: 4135 [ 4h 40m ] +INFO: Idle: 11880, Running: 4, Completed: 4136 [ 4h 41m ] +INFO: Idle: 11877, Running: 4, Completed: 4139 [ 4h 41m ] +INFO: Idle: 11876, Running: 4, Completed: 4140 [ 4h 41m ] +INFO: Idle: 11875, Running: 4, Completed: 4141 [ 4h 41m ] +INFO: Idle: 11873, Running: 4, Completed: 4143 [ 4h 41m ] +INFO: Idle: 11872, Running: 4, Completed: 4144 [ 4h 41m ] +INFO: Idle: 11871, Running: 4, Completed: 4145 [ 4h 41m ] +INFO: Idle: 11870, Running: 4, Completed: 4146 [ 4h 41m ] +INFO: Idle: 11868, Running: 4, Completed: 4148 [ 4h 41m ] +INFO: Idle: 11867, Running: 4, Completed: 4149 [ 4h 42m ] +INFO: Idle: 11866, Running: 4, Completed: 4150 [ 4h 42m ] +INFO: Idle: 11865, Running: 4, Completed: 4151 [ 4h 42m ] +INFO: Idle: 11864, Running: 4, Completed: 4152 [ 4h 42m ] +INFO: Idle: 11863, Running: 4, Completed: 4153 [ 4h 42m ] +INFO: Idle: 11862, Running: 4, Completed: 4154 [ 4h 42m ] +INFO: Idle: 11861, Running: 4, Completed: 4155 [ 4h 42m ] +INFO: Idle: 11860, Running: 4, Completed: 4156 [ 4h 42m ] +INFO: Idle: 11859, Running: 4, Completed: 4157 [ 4h 42m ] +INFO: Idle: 11858, Running: 4, Completed: 4158 [ 4h 42m ] +INFO: Idle: 11856, Running: 4, Completed: 4160 [ 4h 42m ] +INFO: Idle: 11855, Running: 4, Completed: 4161 [ 4h 43m ] +INFO: Idle: 11853, Running: 4, Completed: 4163 [ 4h 43m ] +INFO: Idle: 11851, Running: 4, Completed: 4165 [ 4h 43m ] +INFO: Idle: 11850, Running: 4, Completed: 4166 [ 4h 43m ] +INFO: Idle: 11849, Running: 4, Completed: 4167 [ 4h 43m ] +INFO: Idle: 11847, Running: 4, Completed: 4169 [ 4h 43m ] +INFO: Idle: 11846, Running: 4, Completed: 4170 [ 4h 43m ] +INFO: Idle: 11843, Running: 4, Completed: 4173 [ 4h 44m ] +INFO: Idle: 11841, Running: 4, Completed: 4175 [ 4h 44m ] +INFO: Idle: 11840, Running: 4, Completed: 4176 [ 4h 44m ] +INFO: Idle: 11839, Running: 4, Completed: 4177 [ 4h 44m ] +INFO: Idle: 11837, Running: 4, Completed: 4179 [ 4h 44m ] +INFO: Idle: 11836, Running: 4, Completed: 4180 [ 4h 44m ] +INFO: Idle: 11835, Running: 4, Completed: 4181 [ 4h 44m ] +INFO: Idle: 11833, Running: 4, Completed: 4183 [ 4h 44m ] +INFO: Idle: 11832, Running: 4, Completed: 4184 [ 4h 45m ] +INFO: Idle: 11831, Running: 4, Completed: 4185 [ 4h 45m ] +INFO: Idle: 11828, Running: 4, Completed: 4188 [ 4h 45m ] +INFO: Idle: 11827, Running: 4, Completed: 4189 [ 4h 45m ] +INFO: Idle: 11825, Running: 4, Completed: 4191 [ 4h 45m ] +INFO: Idle: 11824, Running: 4, Completed: 4192 [ 4h 45m ] +INFO: Idle: 11823, Running: 4, Completed: 4193 [ 4h 45m ] +INFO: Idle: 11821, Running: 4, Completed: 4195 [ 4h 45m ] +INFO: Idle: 11820, Running: 4, Completed: 4196 [ 4h 45m ] +INFO: Idle: 11818, Running: 4, Completed: 4198 [ 4h 45m ] +INFO: Idle: 11816, Running: 4, Completed: 4200 [ 4h 46m ] +INFO: Idle: 11815, Running: 4, Completed: 4201 [ 4h 46m ] +INFO: Idle: 11813, Running: 4, Completed: 4203 [ 4h 46m ] +INFO: Idle: 11811, Running: 4, Completed: 4205 [ 4h 46m ] +INFO: Idle: 11809, Running: 4, Completed: 4207 [ 4h 46m ] +INFO: Idle: 11807, Running: 4, Completed: 4209 [ 4h 47m ] +INFO: Idle: 11806, Running: 4, Completed: 4210 [ 4h 47m ] +INFO: Idle: 11805, Running: 4, Completed: 4211 [ 4h 47m ] +INFO: Idle: 11804, Running: 4, Completed: 4212 [ 4h 47m ] +INFO: Idle: 11802, Running: 4, Completed: 4214 [ 4h 47m ] +INFO: Idle: 11801, Running: 4, Completed: 4215 [ 4h 47m ] +INFO: Idle: 11799, Running: 4, Completed: 4217 [ 4h 47m ] +INFO: Idle: 11798, Running: 4, Completed: 4218 [ 4h 47m ] +INFO: Idle: 11797, Running: 4, Completed: 4219 [ 4h 47m ] +INFO: Idle: 11796, Running: 4, Completed: 4220 [ 4h 48m ] +INFO: Idle: 11794, Running: 4, Completed: 4222 [ 4h 48m ] +INFO: Idle: 11793, Running: 4, Completed: 4223 [ 4h 48m ] +INFO: Idle: 11792, Running: 4, Completed: 4224 [ 4h 48m ] +INFO: Idle: 11791, Running: 4, Completed: 4225 [ 4h 48m ] +INFO: Idle: 11790, Running: 4, Completed: 4226 [ 4h 48m ] +INFO: Idle: 11789, Running: 4, Completed: 4227 [ 4h 48m ] +INFO: Idle: 11788, Running: 4, Completed: 4228 [ 4h 48m ] +INFO: Idle: 11787, Running: 4, Completed: 4229 [ 4h 48m ] +INFO: Idle: 11785, Running: 4, Completed: 4231 [ 4h 49m ] +INFO: Idle: 11784, Running: 4, Completed: 4232 [ 4h 49m ] +INFO: Idle: 11782, Running: 4, Completed: 4234 [ 4h 49m ] +INFO: Idle: 11781, Running: 4, Completed: 4235 [ 4h 49m ] +INFO: Idle: 11780, Running: 4, Completed: 4236 [ 4h 49m ] +INFO: Idle: 11779, Running: 4, Completed: 4237 [ 4h 49m ] +INFO: Idle: 11778, Running: 4, Completed: 4238 [ 4h 49m ] +INFO: Idle: 11777, Running: 4, Completed: 4239 [ 4h 49m ] +INFO: Idle: 11776, Running: 4, Completed: 4240 [ 4h 49m ] +INFO: Idle: 11775, Running: 4, Completed: 4241 [ 4h 50m ] +INFO: Idle: 11772, Running: 4, Completed: 4244 [ 4h 50m ] +INFO: Idle: 11770, Running: 4, Completed: 4246 [ 4h 50m ] +INFO: Idle: 11768, Running: 4, Completed: 4248 [ 4h 50m ] +INFO: Idle: 11766, Running: 4, Completed: 4250 [ 4h 50m ] +INFO: Idle: 11765, Running: 4, Completed: 4251 [ 4h 50m ] +INFO: Idle: 11764, Running: 4, Completed: 4252 [ 4h 50m ] +INFO: Idle: 11763, Running: 4, Completed: 4253 [ 4h 50m ] +INFO: Idle: 11761, Running: 4, Completed: 4255 [ 4h 51m ] +INFO: Idle: 11760, Running: 4, Completed: 4256 [ 4h 51m ] +INFO: Idle: 11758, Running: 4, Completed: 4258 [ 4h 51m ] +INFO: Idle: 11756, Running: 4, Completed: 4260 [ 4h 51m ] +INFO: Idle: 11755, Running: 4, Completed: 4261 [ 4h 51m ] +INFO: Idle: 11753, Running: 4, Completed: 4263 [ 4h 51m ] +INFO: Idle: 11752, Running: 4, Completed: 4264 [ 4h 51m ] +INFO: Idle: 11751, Running: 4, Completed: 4265 [ 4h 52m ] +INFO: Idle: 11750, Running: 4, Completed: 4266 [ 4h 52m ] +INFO: Idle: 11748, Running: 4, Completed: 4268 [ 4h 52m ] +INFO: Idle: 11747, Running: 4, Completed: 4269 [ 4h 52m ] +INFO: Idle: 11746, Running: 4, Completed: 4270 [ 4h 52m ] +INFO: Idle: 11745, Running: 4, Completed: 4271 [ 4h 52m ] +INFO: Idle: 11744, Running: 4, Completed: 4272 [ 4h 52m ] +INFO: Idle: 11743, Running: 4, Completed: 4273 [ 4h 52m ] +INFO: Idle: 11741, Running: 4, Completed: 4275 [ 4h 52m ] +INFO: Idle: 11740, Running: 4, Completed: 4276 [ 4h 52m ] +INFO: Idle: 11739, Running: 4, Completed: 4277 [ 4h 52m ] +INFO: Idle: 11738, Running: 4, Completed: 4278 [ 4h 53m ] +INFO: Idle: 11736, Running: 4, Completed: 4280 [ 4h 53m ] +INFO: Idle: 11735, Running: 4, Completed: 4281 [ 4h 53m ] +INFO: Idle: 11733, Running: 4, Completed: 4283 [ 4h 53m ] +INFO: Idle: 11731, Running: 4, Completed: 4285 [ 4h 53m ] +INFO: Idle: 11729, Running: 4, Completed: 4287 [ 4h 53m ] +INFO: Idle: 11727, Running: 4, Completed: 4289 [ 4h 53m ] +INFO: Idle: 11726, Running: 4, Completed: 4290 [ 4h 53m ] +INFO: Idle: 11724, Running: 4, Completed: 4292 [ 4h 54m ] +INFO: Idle: 11723, Running: 4, Completed: 4293 [ 4h 54m ] +INFO: Idle: 11721, Running: 4, Completed: 4295 [ 4h 54m ] +INFO: Idle: 11720, Running: 4, Completed: 4296 [ 4h 54m ] +INFO: Idle: 11719, Running: 4, Completed: 4297 [ 4h 54m ] +INFO: Idle: 11717, Running: 4, Completed: 4299 [ 4h 54m ] +INFO: Idle: 11716, Running: 4, Completed: 4300 [ 4h 54m ] +INFO: Idle: 11715, Running: 4, Completed: 4301 [ 4h 55m ] +INFO: Idle: 11714, Running: 4, Completed: 4302 [ 4h 55m ] +INFO: Idle: 11711, Running: 4, Completed: 4305 [ 4h 55m ] +INFO: Idle: 11709, Running: 4, Completed: 4307 [ 4h 55m ] +INFO: Idle: 11708, Running: 4, Completed: 4308 [ 4h 55m ] +INFO: Idle: 11707, Running: 4, Completed: 4309 [ 4h 55m ] +INFO: Idle: 11706, Running: 4, Completed: 4310 [ 4h 55m ] +INFO: Idle: 11705, Running: 4, Completed: 4311 [ 4h 56m ] +INFO: Idle: 11704, Running: 4, Completed: 4312 [ 4h 56m ] +INFO: Idle: 11703, Running: 4, Completed: 4313 [ 4h 56m ] +INFO: Idle: 11702, Running: 4, Completed: 4314 [ 4h 56m ] +INFO: Idle: 11700, Running: 4, Completed: 4316 [ 4h 56m ] +INFO: Idle: 11699, Running: 4, Completed: 4317 [ 4h 56m ] +INFO: Idle: 11698, Running: 4, Completed: 4318 [ 4h 56m ] +INFO: Idle: 11696, Running: 4, Completed: 4320 [ 4h 56m ] +INFO: Idle: 11695, Running: 4, Completed: 4321 [ 4h 56m ] +INFO: Idle: 11694, Running: 4, Completed: 4322 [ 4h 56m ] +INFO: Idle: 11693, Running: 4, Completed: 4323 [ 4h 57m ] +INFO: Idle: 11692, Running: 4, Completed: 4324 [ 4h 57m ] +INFO: Idle: 11690, Running: 4, Completed: 4326 [ 4h 57m ] +INFO: Idle: 11689, Running: 4, Completed: 4327 [ 4h 57m ] +INFO: Idle: 11688, Running: 4, Completed: 4328 [ 4h 57m ] +INFO: Idle: 11686, Running: 4, Completed: 4330 [ 4h 57m ] +INFO: Idle: 11684, Running: 4, Completed: 4332 [ 4h 57m ] +INFO: Idle: 11683, Running: 4, Completed: 4333 [ 4h 57m ] +INFO: Idle: 11681, Running: 4, Completed: 4335 [ 4h 58m ] +INFO: Idle: 11679, Running: 4, Completed: 4337 [ 4h 58m ] +INFO: Idle: 11677, Running: 4, Completed: 4339 [ 4h 58m ] +INFO: Idle: 11675, Running: 4, Completed: 4341 [ 4h 58m ] +INFO: Idle: 11674, Running: 4, Completed: 4342 [ 4h 58m ] +INFO: Idle: 11673, Running: 4, Completed: 4343 [ 4h 58m ] +INFO: Idle: 11672, Running: 4, Completed: 4344 [ 4h 58m ] +INFO: Idle: 11671, Running: 4, Completed: 4345 [ 4h 58m ] +INFO: Idle: 11669, Running: 4, Completed: 4347 [ 4h 58m ] +INFO: Idle: 11668, Running: 4, Completed: 4348 [ 4h 58m ] +INFO: Idle: 11667, Running: 4, Completed: 4349 [ 4h 59m ] +INFO: Idle: 11666, Running: 4, Completed: 4350 [ 4h 59m ] +INFO: Idle: 11664, Running: 4, Completed: 4352 [ 4h 59m ] +INFO: Idle: 11663, Running: 4, Completed: 4353 [ 4h 59m ] +INFO: Idle: 11661, Running: 4, Completed: 4355 [ 4h 59m ] +INFO: Idle: 11660, Running: 4, Completed: 4356 [ 4h 59m ] +INFO: Idle: 11658, Running: 4, Completed: 4358 [ 4h 59m ] +INFO: Idle: 11656, Running: 4, Completed: 4360 [ 5h 0m ] +INFO: Idle: 11654, Running: 4, Completed: 4362 [ 5h 0m ] +INFO: Idle: 11653, Running: 4, Completed: 4363 [ 5h 0m ] +INFO: Idle: 11650, Running: 4, Completed: 4366 [ 5h 0m ] +INFO: Idle: 11649, Running: 4, Completed: 4367 [ 5h 0m ] +INFO: Idle: 11647, Running: 4, Completed: 4369 [ 5h 0m ] +INFO: Idle: 11645, Running: 4, Completed: 4371 [ 5h 1m ] +INFO: Idle: 11644, Running: 4, Completed: 4372 [ 5h 1m ] +INFO: Idle: 11643, Running: 4, Completed: 4373 [ 5h 1m ] +INFO: Idle: 11641, Running: 4, Completed: 4375 [ 5h 1m ] +INFO: Idle: 11640, Running: 4, Completed: 4376 [ 5h 1m ] +INFO: Idle: 11639, Running: 4, Completed: 4377 [ 5h 1m ] +INFO: Idle: 11636, Running: 4, Completed: 4380 [ 5h 1m ] +INFO: Idle: 11635, Running: 4, Completed: 4381 [ 5h 1m ] +INFO: Idle: 11633, Running: 4, Completed: 4383 [ 5h 2m ] +INFO: Idle: 11632, Running: 4, Completed: 4384 [ 5h 2m ] +INFO: Idle: 11631, Running: 4, Completed: 4385 [ 5h 2m ] +INFO: Idle: 11629, Running: 4, Completed: 4387 [ 5h 2m ] +INFO: Idle: 11628, Running: 4, Completed: 4388 [ 5h 2m ] +INFO: Idle: 11627, Running: 4, Completed: 4389 [ 5h 2m ] +INFO: Idle: 11626, Running: 4, Completed: 4390 [ 5h 2m ] +INFO: Idle: 11624, Running: 4, Completed: 4392 [ 5h 2m ] +INFO: Idle: 11623, Running: 4, Completed: 4393 [ 5h 2m ] +INFO: Idle: 11622, Running: 4, Completed: 4394 [ 5h 3m ] +INFO: Idle: 11620, Running: 4, Completed: 4396 [ 5h 3m ] +INFO: Idle: 11619, Running: 4, Completed: 4397 [ 5h 3m ] +INFO: Idle: 11618, Running: 4, Completed: 4398 [ 5h 3m ] +INFO: Idle: 11616, Running: 4, Completed: 4400 [ 5h 3m ] +INFO: Idle: 11615, Running: 4, Completed: 4401 [ 5h 3m ] +INFO: Idle: 11614, Running: 4, Completed: 4402 [ 5h 3m ] +INFO: Idle: 11613, Running: 4, Completed: 4403 [ 5h 3m ] +INFO: Idle: 11612, Running: 4, Completed: 4404 [ 5h 3m ] +INFO: Idle: 11609, Running: 4, Completed: 4407 [ 5h 3m ] +INFO: Idle: 11608, Running: 4, Completed: 4408 [ 5h 4m ] +INFO: Idle: 11607, Running: 4, Completed: 4409 [ 5h 4m ] +INFO: Idle: 11605, Running: 4, Completed: 4411 [ 5h 4m ] +INFO: Idle: 11604, Running: 4, Completed: 4412 [ 5h 4m ] +INFO: Idle: 11603, Running: 4, Completed: 4413 [ 5h 4m ] +INFO: Idle: 11602, Running: 4, Completed: 4414 [ 5h 4m ] +INFO: Idle: 11599, Running: 4, Completed: 4417 [ 5h 4m ] +INFO: Idle: 11598, Running: 4, Completed: 4418 [ 5h 5m ] +INFO: Idle: 11596, Running: 4, Completed: 4420 [ 5h 5m ] +INFO: Idle: 11595, Running: 4, Completed: 4421 [ 5h 5m ] +INFO: Idle: 11594, Running: 4, Completed: 4422 [ 5h 5m ] +INFO: Idle: 11592, Running: 4, Completed: 4424 [ 5h 5m ] +INFO: Idle: 11591, Running: 4, Completed: 4425 [ 5h 5m ] +INFO: Idle: 11589, Running: 4, Completed: 4427 [ 5h 5m ] +INFO: Idle: 11587, Running: 4, Completed: 4429 [ 5h 5m ] +INFO: Idle: 11586, Running: 4, Completed: 4430 [ 5h 6m ] +INFO: Idle: 11583, Running: 4, Completed: 4433 [ 5h 6m ] +INFO: Idle: 11581, Running: 4, Completed: 4435 [ 5h 6m ] +INFO: Idle: 11579, Running: 4, Completed: 4437 [ 5h 6m ] +INFO: Idle: 11578, Running: 4, Completed: 4438 [ 5h 6m ] +INFO: Idle: 11576, Running: 4, Completed: 4440 [ 5h 6m ] +INFO: Idle: 11575, Running: 4, Completed: 4441 [ 5h 6m ] +INFO: Idle: 11573, Running: 4, Completed: 4443 [ 5h 7m ] +INFO: Idle: 11571, Running: 4, Completed: 4445 [ 5h 7m ] +INFO: Idle: 11568, Running: 4, Completed: 4448 [ 5h 7m ] +INFO: Idle: 11567, Running: 4, Completed: 4449 [ 5h 7m ] +INFO: Idle: 11566, Running: 4, Completed: 4450 [ 5h 7m ] +INFO: Idle: 11565, Running: 4, Completed: 4451 [ 5h 7m ] +INFO: Idle: 11564, Running: 4, Completed: 4452 [ 5h 7m ] +INFO: Idle: 11563, Running: 4, Completed: 4453 [ 5h 7m ] +INFO: Idle: 11562, Running: 4, Completed: 4454 [ 5h 8m ] +INFO: Idle: 11561, Running: 4, Completed: 4455 [ 5h 8m ] +INFO: Idle: 11559, Running: 4, Completed: 4457 [ 5h 8m ] +INFO: Idle: 11558, Running: 4, Completed: 4458 [ 5h 8m ] +INFO: Idle: 11557, Running: 4, Completed: 4459 [ 5h 8m ] +INFO: Idle: 11556, Running: 4, Completed: 4460 [ 5h 8m ] +INFO: Idle: 11554, Running: 4, Completed: 4462 [ 5h 8m ] +INFO: Idle: 11553, Running: 4, Completed: 4463 [ 5h 8m ] +INFO: Idle: 11552, Running: 4, Completed: 4464 [ 5h 8m ] +INFO: Idle: 11551, Running: 4, Completed: 4465 [ 5h 9m ] +INFO: Idle: 11550, Running: 4, Completed: 4466 [ 5h 9m ] +INFO: Idle: 11549, Running: 4, Completed: 4467 [ 5h 9m ] +INFO: Idle: 11548, Running: 4, Completed: 4468 [ 5h 9m ] +INFO: Idle: 11547, Running: 4, Completed: 4469 [ 5h 9m ] +INFO: Idle: 11545, Running: 4, Completed: 4471 [ 5h 9m ] +INFO: Idle: 11544, Running: 4, Completed: 4472 [ 5h 9m ] +INFO: Idle: 11542, Running: 4, Completed: 4474 [ 5h 9m ] +INFO: Idle: 11541, Running: 4, Completed: 4475 [ 5h 9m ] +INFO: Idle: 11540, Running: 4, Completed: 4476 [ 5h 10m ] +INFO: Idle: 11539, Running: 4, Completed: 4477 [ 5h 10m ] +INFO: Idle: 11538, Running: 4, Completed: 4478 [ 5h 10m ] +INFO: Idle: 11537, Running: 4, Completed: 4479 [ 5h 10m ] +INFO: Idle: 11536, Running: 4, Completed: 4480 [ 5h 10m ] +INFO: Idle: 11535, Running: 4, Completed: 4481 [ 5h 10m ] +INFO: Idle: 11532, Running: 4, Completed: 4484 [ 5h 10m ] +INFO: Idle: 11530, Running: 4, Completed: 4486 [ 5h 10m ] +INFO: Idle: 11528, Running: 4, Completed: 4488 [ 5h 11m ] +INFO: Idle: 11526, Running: 4, Completed: 4490 [ 5h 11m ] +INFO: Idle: 11524, Running: 4, Completed: 4492 [ 5h 11m ] +INFO: Idle: 11523, Running: 4, Completed: 4493 [ 5h 11m ] +INFO: Idle: 11521, Running: 4, Completed: 4495 [ 5h 11m ] +INFO: Idle: 11520, Running: 4, Completed: 4496 [ 5h 11m ] +INFO: Idle: 11518, Running: 4, Completed: 4498 [ 5h 11m ] +INFO: Idle: 11517, Running: 4, Completed: 4499 [ 5h 12m ] +INFO: Idle: 11516, Running: 4, Completed: 4500 [ 5h 12m ] +INFO: Idle: 11515, Running: 4, Completed: 4501 [ 5h 12m ] +INFO: Idle: 11514, Running: 4, Completed: 4502 [ 5h 12m ] +INFO: Idle: 11512, Running: 4, Completed: 4504 [ 5h 12m ] +INFO: Idle: 11511, Running: 4, Completed: 4505 [ 5h 12m ] +INFO: Idle: 11510, Running: 4, Completed: 4506 [ 5h 12m ] +INFO: Idle: 11508, Running: 4, Completed: 4508 [ 5h 12m ] +INFO: Idle: 11507, Running: 4, Completed: 4509 [ 5h 12m ] +INFO: Idle: 11506, Running: 4, Completed: 4510 [ 5h 12m ] +INFO: Idle: 11505, Running: 4, Completed: 4511 [ 5h 13m ] +INFO: Idle: 11504, Running: 4, Completed: 4512 [ 5h 13m ] +INFO: Idle: 11503, Running: 4, Completed: 4513 [ 5h 13m ] +INFO: Idle: 11501, Running: 4, Completed: 4515 [ 5h 13m ] +INFO: Idle: 11500, Running: 4, Completed: 4516 [ 5h 13m ] +INFO: Idle: 11499, Running: 4, Completed: 4517 [ 5h 13m ] +INFO: Idle: 11498, Running: 4, Completed: 4518 [ 5h 13m ] +INFO: Idle: 11497, Running: 4, Completed: 4519 [ 5h 13m ] +INFO: Idle: 11496, Running: 4, Completed: 4520 [ 5h 13m ] +INFO: Idle: 11495, Running: 4, Completed: 4521 [ 5h 13m ] +INFO: Idle: 11493, Running: 4, Completed: 4523 [ 5h 14m ] +INFO: Idle: 11491, Running: 4, Completed: 4525 [ 5h 14m ] +INFO: Idle: 11490, Running: 4, Completed: 4526 [ 5h 14m ] +INFO: Idle: 11489, Running: 4, Completed: 4527 [ 5h 14m ] +INFO: Idle: 11488, Running: 4, Completed: 4528 [ 5h 14m ] +INFO: Idle: 11487, Running: 4, Completed: 4529 [ 5h 14m ] +INFO: Idle: 11485, Running: 4, Completed: 4531 [ 5h 14m ] +INFO: Idle: 11484, Running: 4, Completed: 4532 [ 5h 14m ] +INFO: Idle: 11483, Running: 4, Completed: 4533 [ 5h 14m ] +INFO: Idle: 11481, Running: 4, Completed: 4535 [ 5h 14m ] +INFO: Idle: 11480, Running: 4, Completed: 4536 [ 5h 14m ] +INFO: Idle: 11479, Running: 4, Completed: 4537 [ 5h 15m ] +INFO: Idle: 11477, Running: 4, Completed: 4539 [ 5h 15m ] +INFO: Idle: 11475, Running: 4, Completed: 4541 [ 5h 15m ] +INFO: Idle: 11474, Running: 4, Completed: 4542 [ 5h 15m ] +INFO: Idle: 11471, Running: 4, Completed: 4545 [ 5h 16m ] +INFO: Idle: 11469, Running: 4, Completed: 4547 [ 5h 16m ] +INFO: Idle: 11467, Running: 4, Completed: 4549 [ 5h 16m ] +INFO: Idle: 11466, Running: 4, Completed: 4550 [ 5h 16m ] +INFO: Idle: 11464, Running: 4, Completed: 4552 [ 5h 16m ] +INFO: Idle: 11463, Running: 4, Completed: 4553 [ 5h 16m ] +INFO: Idle: 11462, Running: 4, Completed: 4554 [ 5h 16m ] +INFO: Idle: 11459, Running: 4, Completed: 4557 [ 5h 17m ] +INFO: Idle: 11457, Running: 4, Completed: 4559 [ 5h 17m ] +INFO: Idle: 11455, Running: 4, Completed: 4561 [ 5h 17m ] +INFO: Idle: 11453, Running: 4, Completed: 4563 [ 5h 17m ] +INFO: Idle: 11452, Running: 4, Completed: 4564 [ 5h 17m ] +INFO: Idle: 11451, Running: 4, Completed: 4565 [ 5h 18m ] +INFO: Idle: 11449, Running: 4, Completed: 4567 [ 5h 18m ] +INFO: Idle: 11447, Running: 4, Completed: 4569 [ 5h 18m ] +INFO: Idle: 11445, Running: 4, Completed: 4571 [ 5h 18m ] +INFO: Idle: 11444, Running: 4, Completed: 4572 [ 5h 18m ] +INFO: Idle: 11442, Running: 4, Completed: 4574 [ 5h 18m ] +INFO: Idle: 11441, Running: 4, Completed: 4575 [ 5h 19m ] +INFO: Idle: 11439, Running: 4, Completed: 4577 [ 5h 19m ] +INFO: Idle: 11437, Running: 4, Completed: 4579 [ 5h 19m ] +INFO: Idle: 11435, Running: 4, Completed: 4581 [ 5h 19m ] +INFO: Idle: 11433, Running: 4, Completed: 4583 [ 5h 19m ] +INFO: Idle: 11431, Running: 4, Completed: 4585 [ 5h 19m ] +INFO: Idle: 11429, Running: 4, Completed: 4587 [ 5h 20m ] +INFO: Idle: 11428, Running: 4, Completed: 4588 [ 5h 20m ] +INFO: Idle: 11427, Running: 4, Completed: 4589 [ 5h 20m ] +INFO: Idle: 11425, Running: 4, Completed: 4591 [ 5h 20m ] +INFO: Idle: 11424, Running: 4, Completed: 4592 [ 5h 20m ] +INFO: Idle: 11423, Running: 4, Completed: 4593 [ 5h 20m ] +INFO: Idle: 11422, Running: 4, Completed: 4594 [ 5h 20m ] +INFO: Idle: 11420, Running: 4, Completed: 4596 [ 5h 20m ] +INFO: Idle: 11419, Running: 4, Completed: 4597 [ 5h 21m ] +INFO: Idle: 11418, Running: 4, Completed: 4598 [ 5h 21m ] +INFO: Idle: 11416, Running: 4, Completed: 4600 [ 5h 21m ] +INFO: Idle: 11415, Running: 4, Completed: 4601 [ 5h 21m ] +INFO: Idle: 11414, Running: 4, Completed: 4602 [ 5h 21m ] +INFO: Idle: 11412, Running: 4, Completed: 4604 [ 5h 21m ] +INFO: Idle: 11410, Running: 4, Completed: 4606 [ 5h 21m ] +INFO: Idle: 11408, Running: 4, Completed: 4608 [ 5h 22m ] +INFO: Idle: 11407, Running: 4, Completed: 4609 [ 5h 22m ] +INFO: Idle: 11406, Running: 4, Completed: 4610 [ 5h 22m ] +INFO: Idle: 11404, Running: 4, Completed: 4612 [ 5h 22m ] +INFO: Idle: 11403, Running: 4, Completed: 4613 [ 5h 22m ] +INFO: Idle: 11402, Running: 4, Completed: 4614 [ 5h 22m ] +INFO: Idle: 11401, Running: 4, Completed: 4615 [ 5h 23m ] +INFO: Idle: 11400, Running: 4, Completed: 4616 [ 5h 23m ] +INFO: Idle: 11398, Running: 4, Completed: 4618 [ 5h 23m ] +INFO: Idle: 11396, Running: 4, Completed: 4620 [ 5h 23m ] +INFO: Idle: 11394, Running: 4, Completed: 4622 [ 5h 23m ] +INFO: Idle: 11393, Running: 4, Completed: 4623 [ 5h 23m ] +INFO: Idle: 11389, Running: 4, Completed: 4627 [ 5h 24m ] +INFO: Idle: 11386, Running: 4, Completed: 4630 [ 5h 24m ] +INFO: Idle: 11385, Running: 4, Completed: 4631 [ 5h 24m ] +INFO: Idle: 11383, Running: 4, Completed: 4633 [ 5h 24m ] +INFO: Idle: 11382, Running: 4, Completed: 4634 [ 5h 24m ] +INFO: Idle: 11381, Running: 4, Completed: 4635 [ 5h 24m ] +INFO: Idle: 11380, Running: 4, Completed: 4636 [ 5h 25m ] +INFO: Idle: 11379, Running: 4, Completed: 4637 [ 5h 25m ] +INFO: Idle: 11378, Running: 4, Completed: 4638 [ 5h 25m ] +INFO: Idle: 11377, Running: 4, Completed: 4639 [ 5h 25m ] +INFO: Idle: 11376, Running: 4, Completed: 4640 [ 5h 25m ] +INFO: Idle: 11375, Running: 4, Completed: 4641 [ 5h 25m ] +INFO: Idle: 11373, Running: 4, Completed: 4643 [ 5h 25m ] +INFO: Idle: 11372, Running: 4, Completed: 4644 [ 5h 25m ] +INFO: Idle: 11370, Running: 4, Completed: 4646 [ 5h 25m ] +INFO: Idle: 11369, Running: 4, Completed: 4647 [ 5h 25m ] +INFO: Idle: 11368, Running: 4, Completed: 4648 [ 5h 26m ] +INFO: Idle: 11367, Running: 4, Completed: 4649 [ 5h 26m ] +INFO: Idle: 11365, Running: 4, Completed: 4651 [ 5h 26m ] +INFO: Idle: 11362, Running: 4, Completed: 4654 [ 5h 26m ] +INFO: Idle: 11361, Running: 4, Completed: 4655 [ 5h 26m ] +INFO: Idle: 11359, Running: 4, Completed: 4657 [ 5h 26m ] +INFO: Idle: 11357, Running: 4, Completed: 4659 [ 5h 27m ] +INFO: Idle: 11355, Running: 4, Completed: 4661 [ 5h 27m ] +INFO: Idle: 11353, Running: 4, Completed: 4663 [ 5h 27m ] +INFO: Idle: 11351, Running: 4, Completed: 4665 [ 5h 27m ] +INFO: Idle: 11349, Running: 4, Completed: 4667 [ 5h 27m ] +INFO: Idle: 11347, Running: 4, Completed: 4669 [ 5h 27m ] +INFO: Idle: 11346, Running: 4, Completed: 4670 [ 5h 27m ] +INFO: Idle: 11343, Running: 4, Completed: 4673 [ 5h 28m ] +INFO: Idle: 11341, Running: 4, Completed: 4675 [ 5h 28m ] +INFO: Idle: 11339, Running: 4, Completed: 4677 [ 5h 28m ] +INFO: Idle: 11338, Running: 4, Completed: 4678 [ 5h 28m ] +INFO: Idle: 11335, Running: 4, Completed: 4681 [ 5h 29m ] +INFO: Idle: 11333, Running: 4, Completed: 4683 [ 5h 29m ] +INFO: Idle: 11331, Running: 4, Completed: 4685 [ 5h 29m ] +INFO: Idle: 11329, Running: 4, Completed: 4687 [ 5h 29m ] +INFO: Idle: 11327, Running: 4, Completed: 4689 [ 5h 29m ] +INFO: Idle: 11326, Running: 4, Completed: 4690 [ 5h 29m ] +INFO: Idle: 11323, Running: 4, Completed: 4693 [ 5h 30m ] +INFO: Idle: 11322, Running: 4, Completed: 4694 [ 5h 30m ] +INFO: Idle: 11319, Running: 4, Completed: 4697 [ 5h 30m ] +INFO: Idle: 11318, Running: 4, Completed: 4698 [ 5h 30m ] +INFO: Idle: 11317, Running: 4, Completed: 4699 [ 5h 30m ] +INFO: Idle: 11315, Running: 4, Completed: 4701 [ 5h 31m ] +INFO: Idle: 11313, Running: 4, Completed: 4703 [ 5h 31m ] +INFO: Idle: 11311, Running: 4, Completed: 4705 [ 5h 31m ] +INFO: Idle: 11310, Running: 4, Completed: 4706 [ 5h 31m ] +INFO: Idle: 11308, Running: 4, Completed: 4708 [ 5h 31m ] +INFO: Idle: 11307, Running: 4, Completed: 4709 [ 5h 31m ] +INFO: Idle: 11306, Running: 4, Completed: 4710 [ 5h 31m ] +INFO: Idle: 11305, Running: 4, Completed: 4711 [ 5h 32m ] +INFO: Idle: 11303, Running: 4, Completed: 4713 [ 5h 32m ] +INFO: Idle: 11302, Running: 4, Completed: 4714 [ 5h 32m ] +INFO: Idle: 11300, Running: 4, Completed: 4716 [ 5h 32m ] +INFO: Idle: 11298, Running: 4, Completed: 4718 [ 5h 32m ] +INFO: Idle: 11297, Running: 4, Completed: 4719 [ 5h 32m ] +INFO: Idle: 11295, Running: 4, Completed: 4721 [ 5h 33m ] +INFO: Idle: 11294, Running: 4, Completed: 4722 [ 5h 33m ] +INFO: Idle: 11293, Running: 4, Completed: 4723 [ 5h 33m ] +INFO: Idle: 11291, Running: 4, Completed: 4725 [ 5h 33m ] +INFO: Idle: 11289, Running: 4, Completed: 4727 [ 5h 33m ] +INFO: Idle: 11288, Running: 4, Completed: 4728 [ 5h 33m ] +INFO: Idle: 11287, Running: 4, Completed: 4729 [ 5h 33m ] +INFO: Idle: 11285, Running: 4, Completed: 4731 [ 5h 34m ] +INFO: Idle: 11284, Running: 4, Completed: 4732 [ 5h 34m ] +INFO: Idle: 11283, Running: 4, Completed: 4733 [ 5h 34m ] +INFO: Idle: 11282, Running: 4, Completed: 4734 [ 5h 34m ] +INFO: Idle: 11281, Running: 4, Completed: 4735 [ 5h 34m ] +INFO: Idle: 11280, Running: 4, Completed: 4736 [ 5h 34m ] +INFO: Idle: 11279, Running: 4, Completed: 4737 [ 5h 34m ] +INFO: Idle: 11278, Running: 4, Completed: 4738 [ 5h 34m ] +INFO: Idle: 11276, Running: 4, Completed: 4740 [ 5h 35m ] +INFO: Idle: 11275, Running: 4, Completed: 4741 [ 5h 35m ] +INFO: Idle: 11274, Running: 4, Completed: 4742 [ 5h 35m ] +INFO: Idle: 11273, Running: 4, Completed: 4743 [ 5h 35m ] +INFO: Idle: 11272, Running: 4, Completed: 4744 [ 5h 35m ] +INFO: Idle: 11270, Running: 4, Completed: 4746 [ 5h 35m ] +INFO: Idle: 11269, Running: 4, Completed: 4747 [ 5h 35m ] +INFO: Idle: 11267, Running: 4, Completed: 4749 [ 5h 35m ] +INFO: Idle: 11266, Running: 4, Completed: 4750 [ 5h 36m ] +INFO: Idle: 11265, Running: 4, Completed: 4751 [ 5h 36m ] +INFO: Idle: 11264, Running: 4, Completed: 4752 [ 5h 36m ] +INFO: Idle: 11262, Running: 4, Completed: 4754 [ 5h 36m ] +INFO: Idle: 11261, Running: 4, Completed: 4755 [ 5h 36m ] +INFO: Idle: 11260, Running: 4, Completed: 4756 [ 5h 36m ] +INFO: Idle: 11258, Running: 4, Completed: 4758 [ 5h 36m ] +INFO: Idle: 11256, Running: 4, Completed: 4760 [ 5h 36m ] +INFO: Idle: 11254, Running: 4, Completed: 4762 [ 5h 37m ] +INFO: Idle: 11252, Running: 4, Completed: 4764 [ 5h 37m ] +INFO: Idle: 11251, Running: 4, Completed: 4765 [ 5h 37m ] +INFO: Idle: 11250, Running: 4, Completed: 4766 [ 5h 37m ] +INFO: Idle: 11249, Running: 4, Completed: 4767 [ 5h 37m ] +INFO: Idle: 11247, Running: 4, Completed: 4769 [ 5h 37m ] +INFO: Idle: 11246, Running: 4, Completed: 4770 [ 5h 37m ] +INFO: Idle: 11245, Running: 4, Completed: 4771 [ 5h 37m ] +INFO: Idle: 11242, Running: 4, Completed: 4774 [ 5h 38m ] +INFO: Idle: 11241, Running: 4, Completed: 4775 [ 5h 38m ] +INFO: Idle: 11239, Running: 4, Completed: 4777 [ 5h 38m ] +INFO: Idle: 11238, Running: 4, Completed: 4778 [ 5h 38m ] +INFO: Idle: 11237, Running: 4, Completed: 4779 [ 5h 38m ] +INFO: Idle: 11236, Running: 4, Completed: 4780 [ 5h 38m ] +INFO: Idle: 11235, Running: 4, Completed: 4781 [ 5h 39m ] +INFO: Idle: 11234, Running: 4, Completed: 4782 [ 5h 39m ] +INFO: Idle: 11232, Running: 4, Completed: 4784 [ 5h 39m ] +INFO: Idle: 11231, Running: 4, Completed: 4785 [ 5h 39m ] +INFO: Idle: 11230, Running: 4, Completed: 4786 [ 5h 39m ] +INFO: Idle: 11228, Running: 4, Completed: 4788 [ 5h 39m ] +INFO: Idle: 11227, Running: 4, Completed: 4789 [ 5h 39m ] +INFO: Idle: 11224, Running: 4, Completed: 4792 [ 5h 39m ] +INFO: Idle: 11223, Running: 4, Completed: 4793 [ 5h 40m ] +INFO: Idle: 11222, Running: 4, Completed: 4794 [ 5h 40m ] +INFO: Idle: 11221, Running: 4, Completed: 4795 [ 5h 40m ] +INFO: Idle: 11220, Running: 4, Completed: 4796 [ 5h 40m ] +INFO: Idle: 11218, Running: 4, Completed: 4798 [ 5h 40m ] +INFO: Idle: 11217, Running: 4, Completed: 4799 [ 5h 40m ] +INFO: Idle: 11216, Running: 4, Completed: 4800 [ 5h 40m ] +INFO: Idle: 11215, Running: 4, Completed: 4801 [ 5h 40m ] +INFO: Idle: 11214, Running: 4, Completed: 4802 [ 5h 41m ] +INFO: Idle: 11212, Running: 4, Completed: 4804 [ 5h 41m ] +INFO: Idle: 11211, Running: 4, Completed: 4805 [ 5h 41m ] +INFO: Idle: 11210, Running: 4, Completed: 4806 [ 5h 41m ] +INFO: Idle: 11209, Running: 4, Completed: 4807 [ 5h 41m ] +INFO: Idle: 11207, Running: 4, Completed: 4809 [ 5h 41m ] +INFO: Idle: 11206, Running: 4, Completed: 4810 [ 5h 41m ] +INFO: Idle: 11204, Running: 4, Completed: 4812 [ 5h 42m ] +INFO: Idle: 11202, Running: 4, Completed: 4814 [ 5h 42m ] +INFO: Idle: 11198, Running: 4, Completed: 4818 [ 5h 42m ] +INFO: Idle: 11196, Running: 4, Completed: 4820 [ 5h 42m ] +INFO: Idle: 11195, Running: 4, Completed: 4821 [ 5h 42m ] +INFO: Idle: 11194, Running: 4, Completed: 4822 [ 5h 43m ] +INFO: Idle: 11193, Running: 4, Completed: 4823 [ 5h 43m ] +INFO: Idle: 11190, Running: 4, Completed: 4826 [ 5h 43m ] +INFO: Idle: 11189, Running: 4, Completed: 4827 [ 5h 43m ] +INFO: Idle: 11188, Running: 4, Completed: 4828 [ 5h 43m ] +INFO: Idle: 11186, Running: 4, Completed: 4830 [ 5h 43m ] +INFO: Idle: 11184, Running: 4, Completed: 4832 [ 5h 43m ] +INFO: Idle: 11182, Running: 4, Completed: 4834 [ 5h 43m ] +INFO: Idle: 11180, Running: 4, Completed: 4836 [ 5h 44m ] +INFO: Idle: 11179, Running: 4, Completed: 4837 [ 5h 44m ] +INFO: Idle: 11177, Running: 4, Completed: 4839 [ 5h 44m ] +INFO: Idle: 11176, Running: 4, Completed: 4840 [ 5h 44m ] +INFO: Idle: 11175, Running: 4, Completed: 4841 [ 5h 44m ] +INFO: Idle: 11173, Running: 4, Completed: 4843 [ 5h 44m ] +INFO: Idle: 11172, Running: 4, Completed: 4844 [ 5h 44m ] +INFO: Idle: 11171, Running: 4, Completed: 4845 [ 5h 44m ] +INFO: Idle: 11169, Running: 4, Completed: 4847 [ 5h 45m ] +INFO: Idle: 11168, Running: 4, Completed: 4848 [ 5h 45m ] +INFO: Idle: 11167, Running: 4, Completed: 4849 [ 5h 45m ] +INFO: Idle: 11165, Running: 4, Completed: 4851 [ 5h 45m ] +INFO: Idle: 11164, Running: 4, Completed: 4852 [ 5h 45m ] +INFO: Idle: 11163, Running: 4, Completed: 4853 [ 5h 45m ] +INFO: Idle: 11161, Running: 4, Completed: 4855 [ 5h 45m ] +INFO: Idle: 11160, Running: 4, Completed: 4856 [ 5h 45m ] +INFO: Idle: 11159, Running: 4, Completed: 4857 [ 5h 46m ] +INFO: Idle: 11157, Running: 4, Completed: 4859 [ 5h 46m ] +INFO: Idle: 11156, Running: 4, Completed: 4860 [ 5h 46m ] +INFO: Idle: 11155, Running: 4, Completed: 4861 [ 5h 46m ] +INFO: Idle: 11153, Running: 4, Completed: 4863 [ 5h 46m ] +INFO: Idle: 11151, Running: 4, Completed: 4865 [ 5h 46m ] +INFO: Idle: 11149, Running: 4, Completed: 4867 [ 5h 46m ] +INFO: Idle: 11147, Running: 4, Completed: 4869 [ 5h 46m ] +INFO: Idle: 11146, Running: 4, Completed: 4870 [ 5h 46m ] +INFO: Idle: 11145, Running: 4, Completed: 4871 [ 5h 47m ] +INFO: Idle: 11143, Running: 4, Completed: 4873 [ 5h 47m ] +INFO: Idle: 11142, Running: 4, Completed: 4874 [ 5h 47m ] +INFO: Idle: 11141, Running: 4, Completed: 4875 [ 5h 47m ] +INFO: Idle: 11139, Running: 4, Completed: 4877 [ 5h 47m ] +INFO: Idle: 11137, Running: 4, Completed: 4879 [ 5h 47m ] +INFO: Idle: 11136, Running: 4, Completed: 4880 [ 5h 47m ] +INFO: Idle: 11134, Running: 4, Completed: 4882 [ 5h 47m ] +INFO: Idle: 11133, Running: 4, Completed: 4883 [ 5h 48m ] +INFO: Idle: 11131, Running: 4, Completed: 4885 [ 5h 48m ] +INFO: Idle: 11129, Running: 4, Completed: 4887 [ 5h 48m ] +INFO: Idle: 11128, Running: 4, Completed: 4888 [ 5h 48m ] +INFO: Idle: 11126, Running: 4, Completed: 4890 [ 5h 48m ] +INFO: Idle: 11125, Running: 4, Completed: 4891 [ 5h 49m ] +INFO: Idle: 11124, Running: 4, Completed: 4892 [ 5h 49m ] +INFO: Idle: 11121, Running: 4, Completed: 4895 [ 5h 49m ] +INFO: Idle: 11119, Running: 4, Completed: 4897 [ 5h 49m ] +INFO: Idle: 11117, Running: 4, Completed: 4899 [ 5h 49m ] +INFO: Idle: 11115, Running: 4, Completed: 4901 [ 5h 50m ] +INFO: Idle: 11114, Running: 4, Completed: 4902 [ 5h 50m ] +INFO: Idle: 11113, Running: 4, Completed: 4903 [ 5h 50m ] +INFO: Idle: 11111, Running: 4, Completed: 4905 [ 5h 50m ] +INFO: Idle: 11109, Running: 4, Completed: 4907 [ 5h 50m ] +INFO: Idle: 11107, Running: 4, Completed: 4909 [ 5h 50m ] +INFO: Idle: 11105, Running: 4, Completed: 4911 [ 5h 51m ] +INFO: Idle: 11104, Running: 4, Completed: 4912 [ 5h 51m ] +INFO: Idle: 11103, Running: 4, Completed: 4913 [ 5h 51m ] +INFO: Idle: 11101, Running: 4, Completed: 4915 [ 5h 51m ] +INFO: Idle: 11100, Running: 4, Completed: 4916 [ 5h 51m ] +INFO: Idle: 11099, Running: 4, Completed: 4917 [ 5h 51m ] +INFO: Idle: 11097, Running: 4, Completed: 4919 [ 5h 51m ] +INFO: Idle: 11095, Running: 4, Completed: 4921 [ 5h 51m ] +INFO: Idle: 11093, Running: 4, Completed: 4923 [ 5h 52m ] +INFO: Idle: 11091, Running: 4, Completed: 4925 [ 5h 52m ] +INFO: Idle: 11090, Running: 4, Completed: 4926 [ 5h 52m ] +INFO: Idle: 11089, Running: 4, Completed: 4927 [ 5h 52m ] +INFO: Idle: 11087, Running: 4, Completed: 4929 [ 5h 52m ] +INFO: Idle: 11085, Running: 4, Completed: 4931 [ 5h 52m ] +INFO: Idle: 11084, Running: 4, Completed: 4932 [ 5h 53m ] +INFO: Idle: 11082, Running: 4, Completed: 4934 [ 5h 53m ] +INFO: Idle: 11081, Running: 4, Completed: 4935 [ 5h 53m ] +INFO: Idle: 11079, Running: 4, Completed: 4937 [ 5h 53m ] +INFO: Idle: 11078, Running: 4, Completed: 4938 [ 5h 53m ] +INFO: Idle: 11077, Running: 4, Completed: 4939 [ 5h 53m ] +INFO: Idle: 11076, Running: 4, Completed: 4940 [ 5h 53m ] +INFO: Idle: 11074, Running: 4, Completed: 4942 [ 5h 54m ] +INFO: Idle: 11072, Running: 4, Completed: 4944 [ 5h 54m ] +INFO: Idle: 11071, Running: 4, Completed: 4945 [ 5h 54m ] +INFO: Idle: 11070, Running: 4, Completed: 4946 [ 5h 54m ] +INFO: Idle: 11069, Running: 4, Completed: 4947 [ 5h 54m ] +INFO: Idle: 11068, Running: 4, Completed: 4948 [ 5h 54m ] +INFO: Idle: 11066, Running: 4, Completed: 4950 [ 5h 54m ] +INFO: Idle: 11065, Running: 4, Completed: 4951 [ 5h 54m ] +INFO: Idle: 11064, Running: 4, Completed: 4952 [ 5h 55m ] +INFO: Idle: 11063, Running: 4, Completed: 4953 [ 5h 55m ] +INFO: Idle: 11062, Running: 4, Completed: 4954 [ 5h 55m ] +INFO: Idle: 11060, Running: 4, Completed: 4956 [ 5h 55m ] +INFO: Idle: 11058, Running: 4, Completed: 4958 [ 5h 55m ] +INFO: Idle: 11056, Running: 4, Completed: 4960 [ 5h 55m ] +INFO: Idle: 11055, Running: 4, Completed: 4961 [ 5h 56m ] +INFO: Idle: 11051, Running: 4, Completed: 4965 [ 5h 56m ] +INFO: Idle: 11048, Running: 4, Completed: 4968 [ 5h 56m ] +INFO: Idle: 11047, Running: 4, Completed: 4969 [ 5h 56m ] +INFO: Idle: 11045, Running: 4, Completed: 4971 [ 5h 56m ] +INFO: Idle: 11044, Running: 4, Completed: 4972 [ 5h 57m ] +INFO: Idle: 11043, Running: 4, Completed: 4973 [ 5h 57m ] +INFO: Idle: 11042, Running: 4, Completed: 4974 [ 5h 57m ] +INFO: Idle: 11041, Running: 4, Completed: 4975 [ 5h 57m ] +INFO: Idle: 11040, Running: 4, Completed: 4976 [ 5h 57m ] +INFO: Idle: 11039, Running: 4, Completed: 4977 [ 5h 57m ] +INFO: Idle: 11038, Running: 4, Completed: 4978 [ 5h 57m ] +INFO: Idle: 11035, Running: 4, Completed: 4981 [ 5h 57m ] +INFO: Idle: 11034, Running: 4, Completed: 4982 [ 5h 57m ] +INFO: Idle: 11031, Running: 4, Completed: 4985 [ 5h 58m ] +INFO: Idle: 11030, Running: 4, Completed: 4986 [ 5h 58m ] +INFO: Idle: 11029, Running: 4, Completed: 4987 [ 5h 58m ] +INFO: Idle: 11027, Running: 4, Completed: 4989 [ 5h 58m ] +INFO: Idle: 11024, Running: 4, Completed: 4992 [ 5h 58m ] +INFO: Idle: 11023, Running: 4, Completed: 4993 [ 5h 58m ] +INFO: Idle: 11021, Running: 4, Completed: 4995 [ 5h 59m ] +INFO: Idle: 11019, Running: 4, Completed: 4997 [ 5h 59m ] +INFO: Idle: 11017, Running: 4, Completed: 4999 [ 5h 59m ] +INFO: Idle: 11015, Running: 4, Completed: 5001 [ 5h 59m ] +INFO: Idle: 11013, Running: 4, Completed: 5003 [ 5h 59m ] +INFO: Idle: 11011, Running: 4, Completed: 5005 [ 5h 59m ] +INFO: Idle: 11009, Running: 4, Completed: 5007 [ 6h 0m ] +INFO: Idle: 11006, Running: 4, Completed: 5010 [ 6h 0m ] +INFO: Idle: 11005, Running: 4, Completed: 5011 [ 6h 0m ] +INFO: Idle: 11002, Running: 4, Completed: 5014 [ 6h 0m ] +INFO: Idle: 11001, Running: 4, Completed: 5015 [ 6h 0m ] +INFO: Idle: 11000, Running: 4, Completed: 5016 [ 6h 1m ] +INFO: Idle: 10997, Running: 4, Completed: 5019 [ 6h 1m ] +INFO: Idle: 10995, Running: 4, Completed: 5021 [ 6h 1m ] +INFO: Idle: 10993, Running: 4, Completed: 5023 [ 6h 1m ] +INFO: Idle: 10990, Running: 4, Completed: 5026 [ 6h 1m ] +INFO: Idle: 10989, Running: 4, Completed: 5027 [ 6h 2m ] +INFO: Idle: 10987, Running: 4, Completed: 5029 [ 6h 2m ] +INFO: Idle: 10985, Running: 4, Completed: 5031 [ 6h 2m ] +INFO: Idle: 10983, Running: 4, Completed: 5033 [ 6h 2m ] +INFO: Idle: 10981, Running: 4, Completed: 5035 [ 6h 2m ] +INFO: Idle: 10979, Running: 4, Completed: 5037 [ 6h 3m ] +INFO: Idle: 10977, Running: 4, Completed: 5039 [ 6h 3m ] +INFO: Idle: 10975, Running: 4, Completed: 5041 [ 6h 3m ] +INFO: Idle: 10973, Running: 4, Completed: 5043 [ 6h 3m ] +INFO: Idle: 10972, Running: 4, Completed: 5044 [ 6h 3m ] +INFO: Idle: 10970, Running: 4, Completed: 5046 [ 6h 3m ] +INFO: Idle: 10969, Running: 4, Completed: 5047 [ 6h 4m ] +INFO: Idle: 10968, Running: 4, Completed: 5048 [ 6h 4m ] +INFO: Idle: 10967, Running: 4, Completed: 5049 [ 6h 4m ] +INFO: Idle: 10966, Running: 4, Completed: 5050 [ 6h 4m ] +INFO: Idle: 10964, Running: 4, Completed: 5052 [ 6h 4m ] +INFO: Idle: 10962, Running: 4, Completed: 5054 [ 6h 4m ] +INFO: Idle: 10960, Running: 4, Completed: 5056 [ 6h 5m ] +INFO: Idle: 10959, Running: 4, Completed: 5057 [ 6h 5m ] +INFO: Idle: 10957, Running: 4, Completed: 5059 [ 6h 5m ] +INFO: Idle: 10956, Running: 4, Completed: 5060 [ 6h 5m ] +INFO: Idle: 10955, Running: 4, Completed: 5061 [ 6h 5m ] +INFO: Idle: 10953, Running: 4, Completed: 5063 [ 6h 5m ] +INFO: Idle: 10951, Running: 4, Completed: 5065 [ 6h 6m ] +INFO: Idle: 10950, Running: 4, Completed: 5066 [ 6h 6m ] +INFO: Idle: 10949, Running: 4, Completed: 5067 [ 6h 6m ] +INFO: Idle: 10947, Running: 4, Completed: 5069 [ 6h 6m ] +INFO: Idle: 10946, Running: 4, Completed: 5070 [ 6h 6m ] +INFO: Idle: 10944, Running: 4, Completed: 5072 [ 6h 6m ] +INFO: Idle: 10943, Running: 4, Completed: 5073 [ 6h 6m ] +INFO: Idle: 10942, Running: 4, Completed: 5074 [ 6h 6m ] +INFO: Idle: 10941, Running: 4, Completed: 5075 [ 6h 7m ] +INFO: Idle: 10940, Running: 4, Completed: 5076 [ 6h 7m ] +INFO: Idle: 10939, Running: 4, Completed: 5077 [ 6h 7m ] +INFO: Idle: 10938, Running: 4, Completed: 5078 [ 6h 7m ] +INFO: Idle: 10936, Running: 4, Completed: 5080 [ 6h 7m ] +INFO: Idle: 10935, Running: 4, Completed: 5081 [ 6h 7m ] +INFO: Idle: 10934, Running: 4, Completed: 5082 [ 6h 7m ] +INFO: Idle: 10932, Running: 4, Completed: 5084 [ 6h 8m ] +INFO: Idle: 10931, Running: 4, Completed: 5085 [ 6h 8m ] +INFO: Idle: 10929, Running: 4, Completed: 5087 [ 6h 8m ] +INFO: Idle: 10928, Running: 4, Completed: 5088 [ 6h 8m ] +INFO: Idle: 10927, Running: 4, Completed: 5089 [ 6h 8m ] +INFO: Idle: 10926, Running: 4, Completed: 5090 [ 6h 8m ] +INFO: Idle: 10924, Running: 4, Completed: 5092 [ 6h 8m ] +INFO: Idle: 10923, Running: 4, Completed: 5093 [ 6h 8m ] +INFO: Idle: 10922, Running: 4, Completed: 5094 [ 6h 9m ] +INFO: Idle: 10920, Running: 4, Completed: 5096 [ 6h 9m ] +INFO: Idle: 10918, Running: 4, Completed: 5098 [ 6h 9m ] +INFO: Idle: 10916, Running: 4, Completed: 5100 [ 6h 9m ] +INFO: Idle: 10914, Running: 4, Completed: 5102 [ 6h 9m ] +INFO: Idle: 10913, Running: 4, Completed: 5103 [ 6h 9m ] +INFO: Idle: 10912, Running: 4, Completed: 5104 [ 6h 9m ] +INFO: Idle: 10909, Running: 4, Completed: 5107 [ 6h 9m ] +INFO: Idle: 10908, Running: 4, Completed: 5108 [ 6h 10m ] +INFO: Idle: 10907, Running: 4, Completed: 5109 [ 6h 10m ] +INFO: Idle: 10905, Running: 4, Completed: 5111 [ 6h 10m ] +INFO: Idle: 10903, Running: 4, Completed: 5113 [ 6h 10m ] +INFO: Idle: 10901, Running: 4, Completed: 5115 [ 6h 10m ] +INFO: Idle: 10900, Running: 4, Completed: 5116 [ 6h 11m ] +INFO: Idle: 10898, Running: 4, Completed: 5118 [ 6h 11m ] +INFO: Idle: 10897, Running: 4, Completed: 5119 [ 6h 11m ] +INFO: Idle: 10896, Running: 4, Completed: 5120 [ 6h 11m ] +INFO: Idle: 10895, Running: 4, Completed: 5121 [ 6h 11m ] +INFO: Idle: 10894, Running: 4, Completed: 5122 [ 6h 11m ] +INFO: Idle: 10893, Running: 4, Completed: 5123 [ 6h 11m ] +INFO: Idle: 10891, Running: 4, Completed: 5125 [ 6h 11m ] +INFO: Idle: 10890, Running: 4, Completed: 5126 [ 6h 12m ] +INFO: Idle: 10889, Running: 4, Completed: 5127 [ 6h 12m ] +INFO: Idle: 10887, Running: 4, Completed: 5129 [ 6h 12m ] +INFO: Idle: 10885, Running: 4, Completed: 5131 [ 6h 12m ] +INFO: Idle: 10884, Running: 4, Completed: 5132 [ 6h 12m ] +INFO: Idle: 10882, Running: 4, Completed: 5134 [ 6h 12m ] +INFO: Idle: 10881, Running: 4, Completed: 5135 [ 6h 12m ] +INFO: Idle: 10880, Running: 4, Completed: 5136 [ 6h 13m ] +INFO: Idle: 10879, Running: 4, Completed: 5137 [ 6h 13m ] +INFO: Idle: 10878, Running: 4, Completed: 5138 [ 6h 13m ] +INFO: Idle: 10877, Running: 4, Completed: 5139 [ 6h 13m ] +INFO: Idle: 10876, Running: 4, Completed: 5140 [ 6h 13m ] +INFO: Idle: 10875, Running: 4, Completed: 5141 [ 6h 13m ] +INFO: Idle: 10874, Running: 4, Completed: 5142 [ 6h 13m ] +INFO: Idle: 10873, Running: 4, Completed: 5143 [ 6h 13m ] +INFO: Idle: 10872, Running: 4, Completed: 5144 [ 6h 13m ] +INFO: Idle: 10870, Running: 4, Completed: 5146 [ 6h 14m ] +INFO: Idle: 10868, Running: 4, Completed: 5148 [ 6h 14m ] +INFO: Idle: 10867, Running: 4, Completed: 5149 [ 6h 14m ] +INFO: Idle: 10864, Running: 4, Completed: 5152 [ 6h 14m ] +INFO: Idle: 10862, Running: 4, Completed: 5154 [ 6h 14m ] +INFO: Idle: 10860, Running: 4, Completed: 5156 [ 6h 15m ] +INFO: Idle: 10858, Running: 4, Completed: 5158 [ 6h 15m ] +INFO: Idle: 10857, Running: 4, Completed: 5159 [ 6h 15m ] +INFO: Idle: 10856, Running: 4, Completed: 5160 [ 6h 15m ] +INFO: Idle: 10855, Running: 4, Completed: 5161 [ 6h 15m ] +INFO: Idle: 10854, Running: 4, Completed: 5162 [ 6h 15m ] +INFO: Idle: 10852, Running: 4, Completed: 5164 [ 6h 15m ] +INFO: Idle: 10851, Running: 4, Completed: 5165 [ 6h 15m ] +INFO: Idle: 10849, Running: 4, Completed: 5167 [ 6h 15m ] +INFO: Idle: 10848, Running: 4, Completed: 5168 [ 6h 15m ] +INFO: Idle: 10847, Running: 4, Completed: 5169 [ 6h 16m ] +INFO: Idle: 10846, Running: 4, Completed: 5170 [ 6h 16m ] +INFO: Idle: 10844, Running: 4, Completed: 5172 [ 6h 16m ] +INFO: Idle: 10843, Running: 4, Completed: 5173 [ 6h 16m ] +INFO: Idle: 10842, Running: 4, Completed: 5174 [ 6h 16m ] +INFO: Idle: 10841, Running: 4, Completed: 5175 [ 6h 16m ] +INFO: Idle: 10839, Running: 4, Completed: 5177 [ 6h 16m ] +INFO: Idle: 10838, Running: 4, Completed: 5178 [ 6h 16m ] +INFO: Idle: 10837, Running: 4, Completed: 5179 [ 6h 16m ] +INFO: Idle: 10835, Running: 4, Completed: 5181 [ 6h 16m ] +INFO: Idle: 10834, Running: 4, Completed: 5182 [ 6h 17m ] +INFO: Idle: 10833, Running: 4, Completed: 5183 [ 6h 17m ] +INFO: Idle: 10831, Running: 4, Completed: 5185 [ 6h 17m ] +INFO: Idle: 10830, Running: 4, Completed: 5186 [ 6h 17m ] +INFO: Idle: 10829, Running: 4, Completed: 5187 [ 6h 17m ] +INFO: Idle: 10827, Running: 4, Completed: 5189 [ 6h 17m ] +INFO: Idle: 10826, Running: 4, Completed: 5190 [ 6h 17m ] +INFO: Idle: 10825, Running: 4, Completed: 5191 [ 6h 17m ] +INFO: Idle: 10823, Running: 4, Completed: 5193 [ 6h 17m ] +INFO: Idle: 10822, Running: 4, Completed: 5194 [ 6h 18m ] +INFO: Idle: 10821, Running: 4, Completed: 5195 [ 6h 18m ] +INFO: Idle: 10819, Running: 4, Completed: 5197 [ 6h 18m ] +INFO: Idle: 10818, Running: 4, Completed: 5198 [ 6h 18m ] +INFO: Idle: 10817, Running: 4, Completed: 5199 [ 6h 18m ] +INFO: Idle: 10816, Running: 4, Completed: 5200 [ 6h 18m ] +INFO: Idle: 10814, Running: 4, Completed: 5202 [ 6h 18m ] +INFO: Idle: 10813, Running: 4, Completed: 5203 [ 6h 18m ] +INFO: Idle: 10811, Running: 4, Completed: 5205 [ 6h 18m ] +INFO: Idle: 10810, Running: 4, Completed: 5206 [ 6h 19m ] +INFO: Idle: 10809, Running: 4, Completed: 5207 [ 6h 19m ] +INFO: Idle: 10808, Running: 4, Completed: 5208 [ 6h 19m ] +INFO: Idle: 10806, Running: 4, Completed: 5210 [ 6h 19m ] +INFO: Idle: 10805, Running: 4, Completed: 5211 [ 6h 19m ] +INFO: Idle: 10804, Running: 4, Completed: 5212 [ 6h 19m ] +INFO: Idle: 10803, Running: 4, Completed: 5213 [ 6h 19m ] +INFO: Idle: 10802, Running: 4, Completed: 5214 [ 6h 19m ] +INFO: Idle: 10801, Running: 4, Completed: 5215 [ 6h 19m ] +INFO: Idle: 10800, Running: 4, Completed: 5216 [ 6h 19m ] +INFO: Idle: 10799, Running: 4, Completed: 5217 [ 6h 19m ] +INFO: Idle: 10798, Running: 4, Completed: 5218 [ 6h 20m ] +INFO: Idle: 10797, Running: 4, Completed: 5219 [ 6h 20m ] +INFO: Idle: 10796, Running: 4, Completed: 5220 [ 6h 20m ] +INFO: Idle: 10795, Running: 4, Completed: 5221 [ 6h 20m ] +INFO: Idle: 10794, Running: 4, Completed: 5222 [ 6h 20m ] +INFO: Idle: 10793, Running: 4, Completed: 5223 [ 6h 20m ] +INFO: Idle: 10792, Running: 4, Completed: 5224 [ 6h 20m ] +INFO: Idle: 10791, Running: 4, Completed: 5225 [ 6h 20m ] +INFO: Idle: 10790, Running: 4, Completed: 5226 [ 6h 20m ] +INFO: Idle: 10789, Running: 4, Completed: 5227 [ 6h 21m ] +INFO: Idle: 10787, Running: 4, Completed: 5229 [ 6h 21m ] +INFO: Idle: 10785, Running: 4, Completed: 5231 [ 6h 21m ] +INFO: Idle: 10784, Running: 4, Completed: 5232 [ 6h 21m ] +INFO: Idle: 10783, Running: 4, Completed: 5233 [ 6h 21m ] +INFO: Idle: 10781, Running: 4, Completed: 5235 [ 6h 21m ] +INFO: Idle: 10780, Running: 4, Completed: 5236 [ 6h 21m ] +INFO: Idle: 10779, Running: 4, Completed: 5237 [ 6h 22m ] +INFO: Idle: 10777, Running: 4, Completed: 5239 [ 6h 22m ] +INFO: Idle: 10776, Running: 4, Completed: 5240 [ 6h 22m ] +INFO: Idle: 10774, Running: 4, Completed: 5242 [ 6h 22m ] +INFO: Idle: 10773, Running: 4, Completed: 5243 [ 6h 22m ] +INFO: Idle: 10771, Running: 4, Completed: 5245 [ 6h 22m ] +INFO: Idle: 10769, Running: 4, Completed: 5247 [ 6h 23m ] +INFO: Idle: 10767, Running: 4, Completed: 5249 [ 6h 23m ] +INFO: Idle: 10766, Running: 4, Completed: 5250 [ 6h 23m ] +INFO: Idle: 10765, Running: 4, Completed: 5251 [ 6h 23m ] +INFO: Idle: 10764, Running: 4, Completed: 5252 [ 6h 23m ] +INFO: Idle: 10763, Running: 4, Completed: 5253 [ 6h 23m ] +INFO: Idle: 10762, Running: 4, Completed: 5254 [ 6h 23m ] +INFO: Idle: 10761, Running: 4, Completed: 5255 [ 6h 23m ] +INFO: Idle: 10760, Running: 4, Completed: 5256 [ 6h 23m ] +INFO: Idle: 10759, Running: 4, Completed: 5257 [ 6h 24m ] +INFO: Idle: 10757, Running: 4, Completed: 5259 [ 6h 24m ] +INFO: Idle: 10756, Running: 4, Completed: 5260 [ 6h 24m ] +INFO: Idle: 10754, Running: 4, Completed: 5262 [ 6h 24m ] +INFO: Idle: 10753, Running: 4, Completed: 5263 [ 6h 24m ] +INFO: Idle: 10751, Running: 4, Completed: 5265 [ 6h 24m ] +INFO: Idle: 10749, Running: 4, Completed: 5267 [ 6h 25m ] +INFO: Idle: 10748, Running: 4, Completed: 5268 [ 6h 25m ] +INFO: Idle: 10746, Running: 4, Completed: 5270 [ 6h 25m ] +INFO: Idle: 10745, Running: 4, Completed: 5271 [ 6h 25m ] +INFO: Idle: 10744, Running: 4, Completed: 5272 [ 6h 25m ] +INFO: Idle: 10743, Running: 4, Completed: 5273 [ 6h 25m ] +INFO: Idle: 10742, Running: 4, Completed: 5274 [ 6h 25m ] +INFO: Idle: 10741, Running: 4, Completed: 5275 [ 6h 26m ] +INFO: Idle: 10739, Running: 4, Completed: 5277 [ 6h 26m ] +INFO: Idle: 10737, Running: 4, Completed: 5279 [ 6h 26m ] +INFO: Idle: 10736, Running: 4, Completed: 5280 [ 6h 26m ] +INFO: Idle: 10735, Running: 4, Completed: 5281 [ 6h 26m ] +INFO: Idle: 10734, Running: 4, Completed: 5282 [ 6h 26m ] +INFO: Idle: 10732, Running: 4, Completed: 5284 [ 6h 26m ] +INFO: Idle: 10731, Running: 4, Completed: 5285 [ 6h 27m ] +INFO: Idle: 10729, Running: 4, Completed: 5287 [ 6h 27m ] +INFO: Idle: 10728, Running: 4, Completed: 5288 [ 6h 27m ] +INFO: Idle: 10726, Running: 4, Completed: 5290 [ 6h 27m ] +INFO: Idle: 10725, Running: 4, Completed: 5291 [ 6h 27m ] +INFO: Idle: 10724, Running: 4, Completed: 5292 [ 6h 27m ] +INFO: Idle: 10723, Running: 4, Completed: 5293 [ 6h 27m ] +INFO: Idle: 10722, Running: 4, Completed: 5294 [ 6h 27m ] +INFO: Idle: 10720, Running: 4, Completed: 5296 [ 6h 27m ] +INFO: Idle: 10719, Running: 4, Completed: 5297 [ 6h 28m ] +INFO: Idle: 10718, Running: 4, Completed: 5298 [ 6h 28m ] +INFO: Idle: 10717, Running: 4, Completed: 5299 [ 6h 28m ] +INFO: Idle: 10715, Running: 4, Completed: 5301 [ 6h 28m ] +INFO: Idle: 10714, Running: 4, Completed: 5302 [ 6h 28m ] +INFO: Idle: 10713, Running: 4, Completed: 5303 [ 6h 28m ] +INFO: Idle: 10711, Running: 4, Completed: 5305 [ 6h 28m ] +INFO: Idle: 10710, Running: 4, Completed: 5306 [ 6h 28m ] +INFO: Idle: 10708, Running: 4, Completed: 5308 [ 6h 29m ] +INFO: Idle: 10707, Running: 4, Completed: 5309 [ 6h 29m ] +INFO: Idle: 10705, Running: 4, Completed: 5311 [ 6h 29m ] +INFO: Idle: 10704, Running: 4, Completed: 5312 [ 6h 29m ] +INFO: Idle: 10703, Running: 4, Completed: 5313 [ 6h 29m ] +INFO: Idle: 10701, Running: 4, Completed: 5315 [ 6h 29m ] +INFO: Idle: 10699, Running: 4, Completed: 5317 [ 6h 29m ] +INFO: Idle: 10698, Running: 4, Completed: 5318 [ 6h 30m ] +INFO: Idle: 10695, Running: 4, Completed: 5321 [ 6h 30m ] +INFO: Idle: 10694, Running: 4, Completed: 5322 [ 6h 30m ] +INFO: Idle: 10693, Running: 4, Completed: 5323 [ 6h 30m ] +INFO: Idle: 10692, Running: 4, Completed: 5324 [ 6h 30m ] +INFO: Idle: 10691, Running: 4, Completed: 5325 [ 6h 30m ] +INFO: Idle: 10690, Running: 4, Completed: 5326 [ 6h 30m ] +INFO: Idle: 10689, Running: 4, Completed: 5327 [ 6h 30m ] +INFO: Idle: 10688, Running: 4, Completed: 5328 [ 6h 30m ] +INFO: Idle: 10687, Running: 4, Completed: 5329 [ 6h 30m ] +INFO: Idle: 10686, Running: 4, Completed: 5330 [ 6h 31m ] +INFO: Idle: 10685, Running: 4, Completed: 5331 [ 6h 31m ] +INFO: Idle: 10684, Running: 4, Completed: 5332 [ 6h 31m ] +INFO: Idle: 10683, Running: 4, Completed: 5333 [ 6h 31m ] +INFO: Idle: 10682, Running: 4, Completed: 5334 [ 6h 31m ] +INFO: Idle: 10680, Running: 4, Completed: 5336 [ 6h 31m ] +INFO: Idle: 10678, Running: 4, Completed: 5338 [ 6h 31m ] +INFO: Idle: 10677, Running: 4, Completed: 5339 [ 6h 31m ] +INFO: Idle: 10676, Running: 4, Completed: 5340 [ 6h 31m ] +INFO: Idle: 10675, Running: 4, Completed: 5341 [ 6h 32m ] +INFO: Idle: 10674, Running: 4, Completed: 5342 [ 6h 32m ] +INFO: Idle: 10673, Running: 4, Completed: 5343 [ 6h 32m ] +INFO: Idle: 10672, Running: 4, Completed: 5344 [ 6h 32m ] +INFO: Idle: 10671, Running: 4, Completed: 5345 [ 6h 32m ] +INFO: Idle: 10670, Running: 4, Completed: 5346 [ 6h 32m ] +INFO: Idle: 10669, Running: 4, Completed: 5347 [ 6h 32m ] +INFO: Idle: 10668, Running: 4, Completed: 5348 [ 6h 32m ] +INFO: Idle: 10667, Running: 4, Completed: 5349 [ 6h 32m ] +INFO: Idle: 10666, Running: 4, Completed: 5350 [ 6h 32m ] +INFO: Idle: 10665, Running: 4, Completed: 5351 [ 6h 32m ] +INFO: Idle: 10664, Running: 4, Completed: 5352 [ 6h 33m ] +INFO: Idle: 10663, Running: 4, Completed: 5353 [ 6h 33m ] +INFO: Idle: 10662, Running: 4, Completed: 5354 [ 6h 33m ] +INFO: Idle: 10661, Running: 4, Completed: 5355 [ 6h 33m ] +INFO: Idle: 10660, Running: 4, Completed: 5356 [ 6h 33m ] +INFO: Idle: 10659, Running: 4, Completed: 5357 [ 6h 33m ] +INFO: Idle: 10658, Running: 4, Completed: 5358 [ 6h 33m ] +INFO: Idle: 10657, Running: 4, Completed: 5359 [ 6h 33m ] +INFO: Idle: 10655, Running: 4, Completed: 5361 [ 6h 34m ] +INFO: Idle: 10654, Running: 4, Completed: 5362 [ 6h 34m ] +INFO: Idle: 10652, Running: 4, Completed: 5364 [ 6h 34m ] +INFO: Idle: 10651, Running: 4, Completed: 5365 [ 6h 34m ] +INFO: Idle: 10650, Running: 4, Completed: 5366 [ 6h 34m ] +INFO: Idle: 10649, Running: 4, Completed: 5367 [ 6h 34m ] +INFO: Idle: 10648, Running: 4, Completed: 5368 [ 6h 34m ] +INFO: Idle: 10646, Running: 4, Completed: 5370 [ 6h 34m ] +INFO: Idle: 10645, Running: 4, Completed: 5371 [ 6h 35m ] +INFO: Idle: 10643, Running: 4, Completed: 5373 [ 6h 35m ] +INFO: Idle: 10641, Running: 4, Completed: 5375 [ 6h 35m ] +INFO: Idle: 10639, Running: 4, Completed: 5377 [ 6h 35m ] +INFO: Idle: 10638, Running: 4, Completed: 5378 [ 6h 35m ] +INFO: Idle: 10637, Running: 4, Completed: 5379 [ 6h 35m ] +INFO: Idle: 10636, Running: 4, Completed: 5380 [ 6h 35m ] +INFO: Idle: 10635, Running: 4, Completed: 5381 [ 6h 36m ] +INFO: Idle: 10634, Running: 4, Completed: 5382 [ 6h 36m ] +INFO: Idle: 10632, Running: 4, Completed: 5384 [ 6h 36m ] +INFO: Idle: 10631, Running: 4, Completed: 5385 [ 6h 36m ] +INFO: Idle: 10630, Running: 4, Completed: 5386 [ 6h 36m ] +INFO: Idle: 10629, Running: 4, Completed: 5387 [ 6h 36m ] +INFO: Idle: 10628, Running: 4, Completed: 5388 [ 6h 36m ] +INFO: Idle: 10626, Running: 4, Completed: 5390 [ 6h 36m ] +INFO: Idle: 10625, Running: 4, Completed: 5391 [ 6h 37m ] +INFO: Idle: 10622, Running: 4, Completed: 5394 [ 6h 37m ] +INFO: Idle: 10621, Running: 4, Completed: 5395 [ 6h 37m ] +INFO: Idle: 10620, Running: 4, Completed: 5396 [ 6h 37m ] +INFO: Idle: 10618, Running: 4, Completed: 5398 [ 6h 37m ] +INFO: Idle: 10617, Running: 4, Completed: 5399 [ 6h 38m ] +INFO: Idle: 10616, Running: 4, Completed: 5400 [ 6h 38m ] +INFO: Idle: 10615, Running: 4, Completed: 5401 [ 6h 38m ] +INFO: Idle: 10614, Running: 4, Completed: 5402 [ 6h 38m ] +INFO: Idle: 10613, Running: 4, Completed: 5403 [ 6h 38m ] +INFO: Idle: 10612, Running: 4, Completed: 5404 [ 6h 38m ] +INFO: Idle: 10611, Running: 4, Completed: 5405 [ 6h 38m ] +INFO: Idle: 10610, Running: 4, Completed: 5406 [ 6h 38m ] +INFO: Idle: 10609, Running: 4, Completed: 5407 [ 6h 38m ] +INFO: Idle: 10607, Running: 4, Completed: 5409 [ 6h 39m ] +INFO: Idle: 10606, Running: 4, Completed: 5410 [ 6h 39m ] +INFO: Idle: 10605, Running: 4, Completed: 5411 [ 6h 39m ] +INFO: Idle: 10604, Running: 4, Completed: 5412 [ 6h 39m ] +INFO: Idle: 10603, Running: 4, Completed: 5413 [ 6h 39m ] +INFO: Idle: 10602, Running: 4, Completed: 5414 [ 6h 39m ] +INFO: Idle: 10601, Running: 4, Completed: 5415 [ 6h 39m ] +INFO: Idle: 10598, Running: 4, Completed: 5418 [ 6h 39m ] +INFO: Idle: 10596, Running: 4, Completed: 5420 [ 6h 39m ] +INFO: Idle: 10595, Running: 4, Completed: 5421 [ 6h 40m ] +INFO: Idle: 10594, Running: 4, Completed: 5422 [ 6h 40m ] +INFO: Idle: 10591, Running: 4, Completed: 5425 [ 6h 40m ] +INFO: Idle: 10590, Running: 4, Completed: 5426 [ 6h 40m ] +INFO: Idle: 10589, Running: 4, Completed: 5427 [ 6h 40m ] +INFO: Idle: 10587, Running: 4, Completed: 5429 [ 6h 40m ] +INFO: Idle: 10585, Running: 4, Completed: 5431 [ 6h 41m ] +INFO: Idle: 10583, Running: 4, Completed: 5433 [ 6h 41m ] +INFO: Idle: 10582, Running: 4, Completed: 5434 [ 6h 41m ] +INFO: Idle: 10580, Running: 4, Completed: 5436 [ 6h 41m ] +INFO: Idle: 10579, Running: 4, Completed: 5437 [ 6h 41m ] +INFO: Idle: 10577, Running: 4, Completed: 5439 [ 6h 41m ] +INFO: Idle: 10576, Running: 4, Completed: 5440 [ 6h 41m ] +INFO: Idle: 10575, Running: 4, Completed: 5441 [ 6h 42m ] +INFO: Idle: 10574, Running: 4, Completed: 5442 [ 6h 42m ] +INFO: Idle: 10573, Running: 4, Completed: 5443 [ 6h 42m ] +INFO: Idle: 10571, Running: 4, Completed: 5445 [ 6h 42m ] +INFO: Idle: 10570, Running: 4, Completed: 5446 [ 6h 42m ] +INFO: Idle: 10568, Running: 4, Completed: 5448 [ 6h 42m ] +INFO: Idle: 10567, Running: 4, Completed: 5449 [ 6h 42m ] +INFO: Idle: 10566, Running: 4, Completed: 5450 [ 6h 42m ] +INFO: Idle: 10565, Running: 4, Completed: 5451 [ 6h 42m ] +INFO: Idle: 10564, Running: 4, Completed: 5452 [ 6h 42m ] +INFO: Idle: 10563, Running: 4, Completed: 5453 [ 6h 43m ] +INFO: Idle: 10562, Running: 4, Completed: 5454 [ 6h 43m ] +INFO: Idle: 10561, Running: 4, Completed: 5455 [ 6h 43m ] +INFO: Idle: 10560, Running: 4, Completed: 5456 [ 6h 43m ] +INFO: Idle: 10558, Running: 4, Completed: 5458 [ 6h 43m ] +INFO: Idle: 10557, Running: 4, Completed: 5459 [ 6h 43m ] +INFO: Idle: 10556, Running: 4, Completed: 5460 [ 6h 43m ] +INFO: Idle: 10555, Running: 4, Completed: 5461 [ 6h 43m ] +INFO: Idle: 10554, Running: 4, Completed: 5462 [ 6h 43m ] +INFO: Idle: 10552, Running: 4, Completed: 5464 [ 6h 43m ] +INFO: Idle: 10550, Running: 4, Completed: 5466 [ 6h 43m ] +INFO: Idle: 10548, Running: 4, Completed: 5468 [ 6h 44m ] +INFO: Idle: 10547, Running: 4, Completed: 5469 [ 6h 44m ] +INFO: Idle: 10546, Running: 4, Completed: 5470 [ 6h 44m ] +INFO: Idle: 10544, Running: 4, Completed: 5472 [ 6h 44m ] +INFO: Idle: 10543, Running: 4, Completed: 5473 [ 6h 44m ] +INFO: Idle: 10542, Running: 4, Completed: 5474 [ 6h 44m ] +INFO: Idle: 10540, Running: 4, Completed: 5476 [ 6h 44m ] +INFO: Idle: 10539, Running: 4, Completed: 5477 [ 6h 45m ] +INFO: Idle: 10538, Running: 4, Completed: 5478 [ 6h 45m ] +INFO: Idle: 10535, Running: 4, Completed: 5481 [ 6h 45m ] +INFO: Idle: 10534, Running: 4, Completed: 5482 [ 6h 45m ] +INFO: Idle: 10531, Running: 4, Completed: 5485 [ 6h 45m ] +INFO: Idle: 10530, Running: 4, Completed: 5486 [ 6h 45m ] +INFO: Idle: 10527, Running: 4, Completed: 5489 [ 6h 46m ] +INFO: Idle: 10525, Running: 4, Completed: 5491 [ 6h 46m ] +INFO: Idle: 10524, Running: 4, Completed: 5492 [ 6h 46m ] +INFO: Idle: 10523, Running: 4, Completed: 5493 [ 6h 46m ] +INFO: Idle: 10521, Running: 4, Completed: 5495 [ 6h 46m ] +INFO: Idle: 10520, Running: 4, Completed: 5496 [ 6h 46m ] +INFO: Idle: 10519, Running: 4, Completed: 5497 [ 6h 46m ] +INFO: Idle: 10517, Running: 4, Completed: 5499 [ 6h 47m ] +INFO: Idle: 10515, Running: 4, Completed: 5501 [ 6h 47m ] +INFO: Idle: 10513, Running: 4, Completed: 5503 [ 6h 47m ] +INFO: Idle: 10511, Running: 4, Completed: 5505 [ 6h 47m ] +INFO: Idle: 10510, Running: 4, Completed: 5506 [ 6h 47m ] +INFO: Idle: 10509, Running: 4, Completed: 5507 [ 6h 47m ] +INFO: Idle: 10507, Running: 4, Completed: 5509 [ 6h 48m ] +INFO: Idle: 10506, Running: 4, Completed: 5510 [ 6h 48m ] +INFO: Idle: 10504, Running: 4, Completed: 5512 [ 6h 48m ] +INFO: Idle: 10503, Running: 4, Completed: 5513 [ 6h 48m ] +INFO: Idle: 10502, Running: 4, Completed: 5514 [ 6h 48m ] +INFO: Idle: 10501, Running: 4, Completed: 5515 [ 6h 48m ] +INFO: Idle: 10500, Running: 4, Completed: 5516 [ 6h 48m ] +INFO: Idle: 10499, Running: 4, Completed: 5517 [ 6h 48m ] +INFO: Idle: 10498, Running: 4, Completed: 5518 [ 6h 48m ] +INFO: Idle: 10496, Running: 4, Completed: 5520 [ 6h 48m ] +INFO: Idle: 10495, Running: 4, Completed: 5521 [ 6h 49m ] +INFO: Idle: 10493, Running: 4, Completed: 5523 [ 6h 49m ] +INFO: Idle: 10492, Running: 4, Completed: 5524 [ 6h 49m ] +INFO: Idle: 10490, Running: 4, Completed: 5526 [ 6h 49m ] +INFO: Idle: 10489, Running: 4, Completed: 5527 [ 6h 49m ] +INFO: Idle: 10487, Running: 4, Completed: 5529 [ 6h 49m ] +INFO: Idle: 10486, Running: 4, Completed: 5530 [ 6h 50m ] +INFO: Idle: 10485, Running: 4, Completed: 5531 [ 6h 50m ] +INFO: Idle: 10484, Running: 4, Completed: 5532 [ 6h 50m ] +INFO: Idle: 10483, Running: 4, Completed: 5533 [ 6h 50m ] +INFO: Idle: 10482, Running: 4, Completed: 5534 [ 6h 50m ] +INFO: Idle: 10481, Running: 4, Completed: 5535 [ 6h 50m ] +INFO: Idle: 10480, Running: 4, Completed: 5536 [ 6h 50m ] +INFO: Idle: 10477, Running: 4, Completed: 5539 [ 6h 50m ] +INFO: Idle: 10475, Running: 4, Completed: 5541 [ 6h 51m ] +INFO: Idle: 10474, Running: 4, Completed: 5542 [ 6h 51m ] +INFO: Idle: 10472, Running: 4, Completed: 5544 [ 6h 51m ] +INFO: Idle: 10470, Running: 4, Completed: 5546 [ 6h 51m ] +INFO: Idle: 10469, Running: 4, Completed: 5547 [ 6h 51m ] +INFO: Idle: 10466, Running: 4, Completed: 5550 [ 6h 51m ] +INFO: Idle: 10464, Running: 4, Completed: 5552 [ 6h 51m ] +INFO: Idle: 10463, Running: 4, Completed: 5553 [ 6h 52m ] +INFO: Idle: 10462, Running: 4, Completed: 5554 [ 6h 52m ] +INFO: Idle: 10461, Running: 4, Completed: 5555 [ 6h 52m ] +INFO: Idle: 10458, Running: 4, Completed: 5558 [ 6h 52m ] +INFO: Idle: 10457, Running: 4, Completed: 5559 [ 6h 52m ] +INFO: Idle: 10453, Running: 4, Completed: 5563 [ 6h 53m ] +INFO: Idle: 10451, Running: 4, Completed: 5565 [ 6h 53m ] +INFO: Idle: 10450, Running: 4, Completed: 5566 [ 6h 53m ] +INFO: Idle: 10449, Running: 4, Completed: 5567 [ 6h 53m ] +INFO: Idle: 10448, Running: 4, Completed: 5568 [ 6h 53m ] +INFO: Idle: 10447, Running: 4, Completed: 5569 [ 6h 53m ] +INFO: Idle: 10445, Running: 4, Completed: 5571 [ 6h 53m ] +INFO: Idle: 10444, Running: 4, Completed: 5572 [ 6h 53m ] +INFO: Idle: 10442, Running: 4, Completed: 5574 [ 6h 53m ] +INFO: Idle: 10441, Running: 4, Completed: 5575 [ 6h 53m ] +INFO: Idle: 10440, Running: 4, Completed: 5576 [ 6h 54m ] +INFO: Idle: 10438, Running: 4, Completed: 5578 [ 6h 54m ] +INFO: Idle: 10437, Running: 4, Completed: 5579 [ 6h 54m ] +INFO: Idle: 10436, Running: 4, Completed: 5580 [ 6h 54m ] +INFO: Idle: 10434, Running: 4, Completed: 5582 [ 6h 54m ] +INFO: Idle: 10433, Running: 4, Completed: 5583 [ 6h 54m ] +INFO: Idle: 10430, Running: 4, Completed: 5586 [ 6h 54m ] +INFO: Idle: 10429, Running: 4, Completed: 5587 [ 6h 54m ] +INFO: Idle: 10427, Running: 4, Completed: 5589 [ 6h 55m ] +INFO: Idle: 10426, Running: 4, Completed: 5590 [ 6h 55m ] +INFO: Idle: 10425, Running: 4, Completed: 5591 [ 6h 55m ] +INFO: Idle: 10423, Running: 4, Completed: 5593 [ 6h 55m ] +INFO: Idle: 10421, Running: 4, Completed: 5595 [ 6h 55m ] +INFO: Idle: 10419, Running: 4, Completed: 5597 [ 6h 55m ] +INFO: Idle: 10418, Running: 4, Completed: 5598 [ 6h 55m ] +INFO: Idle: 10417, Running: 4, Completed: 5599 [ 6h 55m ] +INFO: Idle: 10415, Running: 4, Completed: 5601 [ 6h 56m ] +INFO: Idle: 10414, Running: 4, Completed: 5602 [ 6h 56m ] +INFO: Idle: 10413, Running: 4, Completed: 5603 [ 6h 56m ] +INFO: Idle: 10411, Running: 4, Completed: 5605 [ 6h 56m ] +INFO: Idle: 10410, Running: 4, Completed: 5606 [ 6h 56m ] +INFO: Idle: 10409, Running: 4, Completed: 5607 [ 6h 56m ] +INFO: Idle: 10407, Running: 4, Completed: 5609 [ 6h 56m ] +INFO: Idle: 10406, Running: 4, Completed: 5610 [ 6h 57m ] +INFO: Idle: 10405, Running: 4, Completed: 5611 [ 6h 57m ] +INFO: Idle: 10403, Running: 4, Completed: 5613 [ 6h 57m ] +INFO: Idle: 10400, Running: 4, Completed: 5616 [ 6h 57m ] +INFO: Idle: 10399, Running: 4, Completed: 5617 [ 6h 57m ] +INFO: Idle: 10398, Running: 4, Completed: 5618 [ 6h 57m ] +INFO: Idle: 10396, Running: 4, Completed: 5620 [ 6h 57m ] +INFO: Idle: 10395, Running: 4, Completed: 5621 [ 6h 58m ] +INFO: Idle: 10393, Running: 4, Completed: 5623 [ 6h 58m ] +INFO: Idle: 10391, Running: 4, Completed: 5625 [ 6h 58m ] +INFO: Idle: 10389, Running: 4, Completed: 5627 [ 6h 58m ] +INFO: Idle: 10387, Running: 4, Completed: 5629 [ 6h 58m ] +INFO: Idle: 10385, Running: 4, Completed: 5631 [ 6h 58m ] +INFO: Idle: 10383, Running: 4, Completed: 5633 [ 6h 59m ] +INFO: Idle: 10382, Running: 4, Completed: 5634 [ 6h 59m ] +INFO: Idle: 10381, Running: 4, Completed: 5635 [ 6h 59m ] +INFO: Idle: 10379, Running: 4, Completed: 5637 [ 6h 59m ] +INFO: Idle: 10378, Running: 4, Completed: 5638 [ 6h 59m ] +INFO: Idle: 10375, Running: 4, Completed: 5641 [ 6h 59m ] +INFO: Idle: 10373, Running: 4, Completed: 5643 [ 7h 0m ] +INFO: Idle: 10372, Running: 4, Completed: 5644 [ 7h 0m ] +INFO: Idle: 10371, Running: 4, Completed: 5645 [ 7h 0m ] +INFO: Idle: 10370, Running: 4, Completed: 5646 [ 7h 0m ] +INFO: Idle: 10368, Running: 4, Completed: 5648 [ 7h 0m ] +INFO: Idle: 10367, Running: 4, Completed: 5649 [ 7h 0m ] +INFO: Idle: 10365, Running: 4, Completed: 5651 [ 7h 0m ] +INFO: Idle: 10364, Running: 4, Completed: 5652 [ 7h 0m ] +INFO: Idle: 10362, Running: 4, Completed: 5654 [ 7h 0m ] +INFO: Idle: 10361, Running: 4, Completed: 5655 [ 7h 1m ] +INFO: Idle: 10359, Running: 4, Completed: 5657 [ 7h 1m ] +INFO: Idle: 10358, Running: 4, Completed: 5658 [ 7h 1m ] +INFO: Idle: 10357, Running: 4, Completed: 5659 [ 7h 1m ] +INFO: Idle: 10355, Running: 4, Completed: 5661 [ 7h 1m ] +INFO: Idle: 10354, Running: 4, Completed: 5662 [ 7h 1m ] +INFO: Idle: 10353, Running: 4, Completed: 5663 [ 7h 1m ] +INFO: Idle: 10352, Running: 4, Completed: 5664 [ 7h 2m ] +INFO: Idle: 10350, Running: 4, Completed: 5666 [ 7h 2m ] +INFO: Idle: 10349, Running: 4, Completed: 5667 [ 7h 2m ] +INFO: Idle: 10348, Running: 4, Completed: 5668 [ 7h 2m ] +INFO: Idle: 10347, Running: 4, Completed: 5669 [ 7h 2m ] +INFO: Idle: 10346, Running: 4, Completed: 5670 [ 7h 2m ] +INFO: Idle: 10344, Running: 4, Completed: 5672 [ 7h 2m ] +INFO: Idle: 10342, Running: 4, Completed: 5674 [ 7h 2m ] +INFO: Idle: 10341, Running: 4, Completed: 5675 [ 7h 3m ] +INFO: Idle: 10339, Running: 4, Completed: 5677 [ 7h 3m ] +INFO: Idle: 10338, Running: 4, Completed: 5678 [ 7h 3m ] +INFO: Idle: 10336, Running: 4, Completed: 5680 [ 7h 3m ] +INFO: Idle: 10334, Running: 4, Completed: 5682 [ 7h 3m ] +INFO: Idle: 10333, Running: 4, Completed: 5683 [ 7h 3m ] +INFO: Idle: 10331, Running: 4, Completed: 5685 [ 7h 3m ] +INFO: Idle: 10330, Running: 4, Completed: 5686 [ 7h 4m ] +INFO: Idle: 10329, Running: 4, Completed: 5687 [ 7h 4m ] +INFO: Idle: 10328, Running: 4, Completed: 5688 [ 7h 4m ] +INFO: Idle: 10325, Running: 4, Completed: 5691 [ 7h 4m ] +INFO: Idle: 10323, Running: 4, Completed: 5693 [ 7h 4m ] +INFO: Idle: 10322, Running: 4, Completed: 5694 [ 7h 4m ] +INFO: Idle: 10321, Running: 4, Completed: 5695 [ 7h 4m ] +INFO: Idle: 10320, Running: 4, Completed: 5696 [ 7h 4m ] +INFO: Idle: 10319, Running: 4, Completed: 5697 [ 7h 5m ] +INFO: Idle: 10317, Running: 4, Completed: 5699 [ 7h 5m ] +INFO: Idle: 10316, Running: 4, Completed: 5700 [ 7h 5m ] +INFO: Idle: 10314, Running: 4, Completed: 5702 [ 7h 5m ] +INFO: Idle: 10313, Running: 4, Completed: 5703 [ 7h 5m ] +INFO: Idle: 10312, Running: 4, Completed: 5704 [ 7h 5m ] +INFO: Idle: 10309, Running: 4, Completed: 5707 [ 7h 5m ] +INFO: Idle: 10308, Running: 4, Completed: 5708 [ 7h 5m ] +INFO: Idle: 10307, Running: 4, Completed: 5709 [ 7h 5m ] +INFO: Idle: 10305, Running: 4, Completed: 5711 [ 7h 6m ] +INFO: Idle: 10303, Running: 4, Completed: 5713 [ 7h 6m ] +INFO: Idle: 10301, Running: 4, Completed: 5715 [ 7h 6m ] +INFO: Idle: 10299, Running: 4, Completed: 5717 [ 7h 6m ] +INFO: Idle: 10297, Running: 4, Completed: 5719 [ 7h 6m ] +INFO: Idle: 10295, Running: 4, Completed: 5721 [ 7h 6m ] +INFO: Idle: 10292, Running: 4, Completed: 5724 [ 7h 7m ] +INFO: Idle: 10291, Running: 4, Completed: 5725 [ 7h 7m ] +INFO: Idle: 10288, Running: 4, Completed: 5728 [ 7h 7m ] +INFO: Idle: 10287, Running: 4, Completed: 5729 [ 7h 7m ] +INFO: Idle: 10284, Running: 4, Completed: 5732 [ 7h 7m ] +INFO: Idle: 10283, Running: 4, Completed: 5733 [ 7h 8m ] +INFO: Idle: 10281, Running: 4, Completed: 5735 [ 7h 8m ] +INFO: Idle: 10279, Running: 4, Completed: 5737 [ 7h 8m ] +INFO: Idle: 10277, Running: 4, Completed: 5739 [ 7h 8m ] +INFO: Idle: 10276, Running: 4, Completed: 5740 [ 7h 8m ] +INFO: Idle: 10275, Running: 4, Completed: 5741 [ 7h 8m ] +INFO: Idle: 10273, Running: 4, Completed: 5743 [ 7h 9m ] +INFO: Idle: 10272, Running: 4, Completed: 5744 [ 7h 9m ] +INFO: Idle: 10270, Running: 4, Completed: 5746 [ 7h 9m ] +INFO: Idle: 10269, Running: 4, Completed: 5747 [ 7h 9m ] +INFO: Idle: 10267, Running: 4, Completed: 5749 [ 7h 9m ] +INFO: Idle: 10266, Running: 4, Completed: 5750 [ 7h 9m ] +INFO: Idle: 10264, Running: 4, Completed: 5752 [ 7h 9m ] +INFO: Idle: 10263, Running: 4, Completed: 5753 [ 7h 9m ] +INFO: Idle: 10261, Running: 4, Completed: 5755 [ 7h 10m ] +INFO: Idle: 10260, Running: 4, Completed: 5756 [ 7h 10m ] +INFO: Idle: 10259, Running: 4, Completed: 5757 [ 7h 10m ] +INFO: Idle: 10256, Running: 4, Completed: 5760 [ 7h 10m ] +INFO: Idle: 10255, Running: 4, Completed: 5761 [ 7h 10m ] +INFO: Idle: 10253, Running: 4, Completed: 5763 [ 7h 10m ] +INFO: Idle: 10252, Running: 4, Completed: 5764 [ 7h 10m ] +INFO: Idle: 10251, Running: 4, Completed: 5765 [ 7h 10m ] +INFO: Idle: 10250, Running: 4, Completed: 5766 [ 7h 10m ] +INFO: Idle: 10248, Running: 4, Completed: 5768 [ 7h 11m ] +INFO: Idle: 10247, Running: 4, Completed: 5769 [ 7h 11m ] +INFO: Idle: 10244, Running: 4, Completed: 5772 [ 7h 11m ] +INFO: Idle: 10243, Running: 4, Completed: 5773 [ 7h 11m ] +INFO: Idle: 10242, Running: 4, Completed: 5774 [ 7h 11m ] +INFO: Idle: 10240, Running: 4, Completed: 5776 [ 7h 11m ] +INFO: Idle: 10238, Running: 4, Completed: 5778 [ 7h 12m ] +INFO: Idle: 10236, Running: 4, Completed: 5780 [ 7h 12m ] +INFO: Idle: 10234, Running: 4, Completed: 5782 [ 7h 12m ] +INFO: Idle: 10233, Running: 4, Completed: 5783 [ 7h 12m ] +INFO: Idle: 10232, Running: 4, Completed: 5784 [ 7h 12m ] +INFO: Idle: 10229, Running: 4, Completed: 5787 [ 7h 12m ] +INFO: Idle: 10228, Running: 4, Completed: 5788 [ 7h 12m ] +INFO: Idle: 10227, Running: 4, Completed: 5789 [ 7h 12m ] +INFO: Idle: 10225, Running: 4, Completed: 5791 [ 7h 12m ] +INFO: Idle: 10224, Running: 4, Completed: 5792 [ 7h 13m ] +INFO: Idle: 10223, Running: 4, Completed: 5793 [ 7h 13m ] +INFO: Idle: 10220, Running: 4, Completed: 5796 [ 7h 13m ] +INFO: Idle: 10219, Running: 4, Completed: 5797 [ 7h 13m ] +INFO: Idle: 10217, Running: 4, Completed: 5799 [ 7h 13m ] +INFO: Idle: 10215, Running: 4, Completed: 5801 [ 7h 13m ] +INFO: Idle: 10213, Running: 4, Completed: 5803 [ 7h 13m ] +INFO: Idle: 10211, Running: 4, Completed: 5805 [ 7h 13m ] +INFO: Idle: 10209, Running: 4, Completed: 5807 [ 7h 14m ] +INFO: Idle: 10208, Running: 4, Completed: 5808 [ 7h 14m ] +INFO: Idle: 10205, Running: 4, Completed: 5811 [ 7h 14m ] +INFO: Idle: 10204, Running: 4, Completed: 5812 [ 7h 14m ] +INFO: Idle: 10201, Running: 4, Completed: 5815 [ 7h 14m ] +INFO: Idle: 10198, Running: 4, Completed: 5818 [ 7h 15m ] +INFO: Idle: 10197, Running: 4, Completed: 5819 [ 7h 15m ] +INFO: Idle: 10195, Running: 4, Completed: 5821 [ 7h 15m ] +INFO: Idle: 10194, Running: 4, Completed: 5822 [ 7h 15m ] +INFO: Idle: 10193, Running: 4, Completed: 5823 [ 7h 15m ] +INFO: Idle: 10192, Running: 4, Completed: 5824 [ 7h 15m ] +INFO: Idle: 10191, Running: 4, Completed: 5825 [ 7h 15m ] +INFO: Idle: 10189, Running: 4, Completed: 5827 [ 7h 16m ] +INFO: Idle: 10188, Running: 4, Completed: 5828 [ 7h 16m ] +INFO: Idle: 10187, Running: 4, Completed: 5829 [ 7h 16m ] +INFO: Idle: 10185, Running: 4, Completed: 5831 [ 7h 16m ] +INFO: Idle: 10184, Running: 4, Completed: 5832 [ 7h 16m ] +INFO: Idle: 10182, Running: 4, Completed: 5834 [ 7h 16m ] +INFO: Idle: 10181, Running: 4, Completed: 5835 [ 7h 16m ] +INFO: Idle: 10179, Running: 4, Completed: 5837 [ 7h 16m ] +INFO: Idle: 10178, Running: 4, Completed: 5838 [ 7h 17m ] +INFO: Idle: 10177, Running: 4, Completed: 5839 [ 7h 17m ] +INFO: Idle: 10174, Running: 4, Completed: 5842 [ 7h 17m ] +INFO: Idle: 10173, Running: 4, Completed: 5843 [ 7h 17m ] +INFO: Idle: 10171, Running: 4, Completed: 5845 [ 7h 17m ] +INFO: Idle: 10170, Running: 4, Completed: 5846 [ 7h 17m ] +INFO: Idle: 10169, Running: 4, Completed: 5847 [ 7h 17m ] +INFO: Idle: 10168, Running: 4, Completed: 5848 [ 7h 17m ] +INFO: Idle: 10166, Running: 4, Completed: 5850 [ 7h 17m ] +INFO: Idle: 10165, Running: 4, Completed: 5851 [ 7h 18m ] +INFO: Idle: 10162, Running: 4, Completed: 5854 [ 7h 18m ] +INFO: Idle: 10161, Running: 4, Completed: 5855 [ 7h 18m ] +INFO: Idle: 10160, Running: 4, Completed: 5856 [ 7h 18m ] +INFO: Idle: 10158, Running: 4, Completed: 5858 [ 7h 18m ] +INFO: Idle: 10156, Running: 4, Completed: 5860 [ 7h 18m ] +INFO: Idle: 10154, Running: 4, Completed: 5862 [ 7h 19m ] +INFO: Idle: 10152, Running: 4, Completed: 5864 [ 7h 19m ] +INFO: Idle: 10151, Running: 4, Completed: 5865 [ 7h 19m ] +INFO: Idle: 10150, Running: 4, Completed: 5866 [ 7h 19m ] +INFO: Idle: 10148, Running: 4, Completed: 5868 [ 7h 19m ] +INFO: Idle: 10147, Running: 4, Completed: 5869 [ 7h 19m ] +INFO: Idle: 10146, Running: 4, Completed: 5870 [ 7h 19m ] +INFO: Idle: 10145, Running: 4, Completed: 5871 [ 7h 19m ] +INFO: Idle: 10144, Running: 4, Completed: 5872 [ 7h 19m ] +INFO: Idle: 10142, Running: 4, Completed: 5874 [ 7h 19m ] +INFO: Idle: 10141, Running: 4, Completed: 5875 [ 7h 20m ] +INFO: Idle: 10138, Running: 4, Completed: 5878 [ 7h 20m ] +INFO: Idle: 10137, Running: 4, Completed: 5879 [ 7h 20m ] +INFO: Idle: 10135, Running: 4, Completed: 5881 [ 7h 20m ] +INFO: Idle: 10133, Running: 4, Completed: 5883 [ 7h 20m ] +INFO: Idle: 10131, Running: 4, Completed: 5885 [ 7h 20m ] +INFO: Idle: 10129, Running: 4, Completed: 5887 [ 7h 20m ] +INFO: Idle: 10128, Running: 4, Completed: 5888 [ 7h 20m ] +INFO: Idle: 10127, Running: 4, Completed: 5889 [ 7h 21m ] +INFO: Idle: 10125, Running: 4, Completed: 5891 [ 7h 21m ] +INFO: Idle: 10123, Running: 4, Completed: 5893 [ 7h 21m ] +INFO: Idle: 10121, Running: 4, Completed: 5895 [ 7h 21m ] +INFO: Idle: 10119, Running: 4, Completed: 5897 [ 7h 21m ] +INFO: Idle: 10118, Running: 4, Completed: 5898 [ 7h 21m ] +INFO: Idle: 10117, Running: 4, Completed: 5899 [ 7h 22m ] +INFO: Idle: 10115, Running: 4, Completed: 5901 [ 7h 22m ] +INFO: Idle: 10114, Running: 4, Completed: 5902 [ 7h 22m ] +INFO: Idle: 10113, Running: 4, Completed: 5903 [ 7h 22m ] +INFO: Idle: 10111, Running: 4, Completed: 5905 [ 7h 22m ] +INFO: Idle: 10109, Running: 4, Completed: 5907 [ 7h 22m ] +INFO: Idle: 10107, Running: 4, Completed: 5909 [ 7h 22m ] +INFO: Idle: 10106, Running: 4, Completed: 5910 [ 7h 23m ] +INFO: Idle: 10104, Running: 4, Completed: 5912 [ 7h 23m ] +INFO: Idle: 10103, Running: 4, Completed: 5913 [ 7h 23m ] +INFO: Idle: 10100, Running: 4, Completed: 5916 [ 7h 23m ] +INFO: Idle: 10099, Running: 4, Completed: 5917 [ 7h 23m ] +INFO: Idle: 10096, Running: 4, Completed: 5920 [ 7h 23m ] +INFO: Idle: 10095, Running: 4, Completed: 5921 [ 7h 23m ] +INFO: Idle: 10094, Running: 4, Completed: 5922 [ 7h 24m ] +INFO: Idle: 10093, Running: 4, Completed: 5923 [ 7h 24m ] +INFO: Idle: 10091, Running: 4, Completed: 5925 [ 7h 24m ] +INFO: Idle: 10090, Running: 4, Completed: 5926 [ 7h 24m ] +INFO: Idle: 10089, Running: 4, Completed: 5927 [ 7h 24m ] +INFO: Idle: 10087, Running: 4, Completed: 5929 [ 7h 24m ] +INFO: Idle: 10085, Running: 4, Completed: 5931 [ 7h 24m ] +INFO: Idle: 10084, Running: 4, Completed: 5932 [ 7h 25m ] +INFO: Idle: 10083, Running: 4, Completed: 5933 [ 7h 25m ] +INFO: Idle: 10082, Running: 4, Completed: 5934 [ 7h 25m ] +INFO: Idle: 10081, Running: 4, Completed: 5935 [ 7h 25m ] +INFO: Idle: 10080, Running: 4, Completed: 5936 [ 7h 25m ] +INFO: Idle: 10078, Running: 4, Completed: 5938 [ 7h 25m ] +INFO: Idle: 10077, Running: 4, Completed: 5939 [ 7h 25m ] +INFO: Idle: 10074, Running: 4, Completed: 5942 [ 7h 25m ] +INFO: Idle: 10073, Running: 4, Completed: 5943 [ 7h 26m ] +INFO: Idle: 10070, Running: 4, Completed: 5946 [ 7h 26m ] +INFO: Idle: 10069, Running: 4, Completed: 5947 [ 7h 26m ] +INFO: Idle: 10067, Running: 4, Completed: 5949 [ 7h 26m ] +INFO: Idle: 10066, Running: 4, Completed: 5950 [ 7h 26m ] +INFO: Idle: 10065, Running: 4, Completed: 5951 [ 7h 27m ] +INFO: Idle: 10062, Running: 4, Completed: 5954 [ 7h 27m ] +INFO: Idle: 10061, Running: 4, Completed: 5955 [ 7h 27m ] +INFO: Idle: 10058, Running: 4, Completed: 5958 [ 7h 27m ] +INFO: Idle: 10057, Running: 4, Completed: 5959 [ 7h 27m ] +INFO: Idle: 10055, Running: 4, Completed: 5961 [ 7h 27m ] +INFO: Idle: 10054, Running: 4, Completed: 5962 [ 7h 27m ] +INFO: Idle: 10053, Running: 4, Completed: 5963 [ 7h 28m ] +INFO: Idle: 10052, Running: 4, Completed: 5964 [ 7h 28m ] +INFO: Idle: 10050, Running: 4, Completed: 5966 [ 7h 28m ] +INFO: Idle: 10049, Running: 4, Completed: 5967 [ 7h 28m ] +INFO: Idle: 10048, Running: 4, Completed: 5968 [ 7h 28m ] +INFO: Idle: 10046, Running: 4, Completed: 5970 [ 7h 28m ] +INFO: Idle: 10045, Running: 4, Completed: 5971 [ 7h 28m ] +INFO: Idle: 10043, Running: 4, Completed: 5973 [ 7h 28m ] +INFO: Idle: 10041, Running: 4, Completed: 5975 [ 7h 29m ] +INFO: Idle: 10037, Running: 4, Completed: 5979 [ 7h 29m ] +INFO: Idle: 10034, Running: 4, Completed: 5982 [ 7h 29m ] +INFO: Idle: 10033, Running: 4, Completed: 5983 [ 7h 29m ] +INFO: Idle: 10032, Running: 4, Completed: 5984 [ 7h 29m ] +INFO: Idle: 10030, Running: 4, Completed: 5986 [ 7h 29m ] +INFO: Idle: 10029, Running: 4, Completed: 5987 [ 7h 30m ] +INFO: Idle: 10027, Running: 4, Completed: 5989 [ 7h 30m ] +INFO: Idle: 10025, Running: 4, Completed: 5991 [ 7h 30m ] +INFO: Idle: 10022, Running: 4, Completed: 5994 [ 7h 30m ] +INFO: Idle: 10021, Running: 4, Completed: 5995 [ 7h 30m ] +INFO: Idle: 10018, Running: 4, Completed: 5998 [ 7h 30m ] +INFO: Idle: 10017, Running: 4, Completed: 5999 [ 7h 31m ] +INFO: Idle: 10014, Running: 4, Completed: 6002 [ 7h 31m ] +INFO: Idle: 10013, Running: 4, Completed: 6003 [ 7h 31m ] +INFO: Idle: 10011, Running: 4, Completed: 6005 [ 7h 31m ] +INFO: Idle: 10010, Running: 4, Completed: 6006 [ 7h 31m ] +INFO: Idle: 10009, Running: 4, Completed: 6007 [ 7h 31m ] +INFO: Idle: 10007, Running: 4, Completed: 6009 [ 7h 31m ] +INFO: Idle: 10005, Running: 4, Completed: 6011 [ 7h 32m ] +INFO: Idle: 10003, Running: 4, Completed: 6013 [ 7h 32m ] +INFO: Idle: 10001, Running: 4, Completed: 6015 [ 7h 32m ] +INFO: Idle: 9999, Running: 4, Completed: 6017 [ 7h 32m ] +INFO: Idle: 9997, Running: 4, Completed: 6019 [ 7h 32m ] +INFO: Idle: 9995, Running: 4, Completed: 6021 [ 7h 32m ] +INFO: Idle: 9993, Running: 4, Completed: 6023 [ 7h 32m ] +INFO: Idle: 9991, Running: 4, Completed: 6025 [ 7h 33m ] +INFO: Idle: 9990, Running: 4, Completed: 6026 [ 7h 33m ] +INFO: Idle: 9989, Running: 4, Completed: 6027 [ 7h 33m ] +INFO: Idle: 9987, Running: 4, Completed: 6029 [ 7h 33m ] +INFO: Idle: 9986, Running: 4, Completed: 6030 [ 7h 33m ] +INFO: Idle: 9985, Running: 4, Completed: 6031 [ 7h 33m ] +INFO: Idle: 9983, Running: 4, Completed: 6033 [ 7h 34m ] +INFO: Idle: 9980, Running: 4, Completed: 6036 [ 7h 34m ] +INFO: Idle: 9979, Running: 4, Completed: 6037 [ 7h 34m ] +INFO: Idle: 9978, Running: 4, Completed: 6038 [ 7h 34m ] +INFO: Idle: 9976, Running: 4, Completed: 6040 [ 7h 34m ] +INFO: Idle: 9975, Running: 4, Completed: 6041 [ 7h 34m ] +INFO: Idle: 9972, Running: 4, Completed: 6044 [ 7h 34m ] +INFO: Idle: 9971, Running: 4, Completed: 6045 [ 7h 35m ] +INFO: Idle: 9968, Running: 4, Completed: 6048 [ 7h 35m ] +INFO: Idle: 9967, Running: 4, Completed: 6049 [ 7h 35m ] +INFO: Idle: 9966, Running: 4, Completed: 6050 [ 7h 35m ] +INFO: Idle: 9965, Running: 4, Completed: 6051 [ 7h 35m ] +INFO: Idle: 9964, Running: 4, Completed: 6052 [ 7h 35m ] +INFO: Idle: 9963, Running: 4, Completed: 6053 [ 7h 35m ] +INFO: Idle: 9962, Running: 4, Completed: 6054 [ 7h 35m ] +INFO: Idle: 9961, Running: 4, Completed: 6055 [ 7h 35m ] +INFO: Idle: 9960, Running: 4, Completed: 6056 [ 7h 36m ] +INFO: Idle: 9959, Running: 4, Completed: 6057 [ 7h 36m ] +INFO: Idle: 9957, Running: 4, Completed: 6059 [ 7h 36m ] +INFO: Idle: 9955, Running: 4, Completed: 6061 [ 7h 36m ] +INFO: Idle: 9954, Running: 4, Completed: 6062 [ 7h 36m ] +INFO: Idle: 9953, Running: 4, Completed: 6063 [ 7h 36m ] +INFO: Idle: 9952, Running: 4, Completed: 6064 [ 7h 36m ] +INFO: Idle: 9950, Running: 4, Completed: 6066 [ 7h 37m ] +INFO: Idle: 9949, Running: 4, Completed: 6067 [ 7h 37m ] +INFO: Idle: 9947, Running: 4, Completed: 6069 [ 7h 37m ] +INFO: Idle: 9946, Running: 4, Completed: 6070 [ 7h 37m ] +INFO: Idle: 9945, Running: 4, Completed: 6071 [ 7h 37m ] +INFO: Idle: 9942, Running: 4, Completed: 6074 [ 7h 37m ] +INFO: Idle: 9941, Running: 4, Completed: 6075 [ 7h 38m ] +INFO: Idle: 9939, Running: 4, Completed: 6077 [ 7h 38m ] +INFO: Idle: 9938, Running: 4, Completed: 6078 [ 7h 38m ] +INFO: Idle: 9937, Running: 4, Completed: 6079 [ 7h 38m ] +INFO: Idle: 9934, Running: 4, Completed: 6082 [ 7h 38m ] +INFO: Idle: 9933, Running: 4, Completed: 6083 [ 7h 38m ] +INFO: Idle: 9930, Running: 4, Completed: 6086 [ 7h 38m ] +INFO: Idle: 9929, Running: 4, Completed: 6087 [ 7h 39m ] +INFO: Idle: 9927, Running: 4, Completed: 6089 [ 7h 39m ] +INFO: Idle: 9926, Running: 4, Completed: 6090 [ 7h 39m ] +INFO: Idle: 9925, Running: 4, Completed: 6091 [ 7h 39m ] +INFO: Idle: 9924, Running: 4, Completed: 6092 [ 7h 39m ] +INFO: Idle: 9922, Running: 4, Completed: 6094 [ 7h 39m ] +INFO: Idle: 9921, Running: 4, Completed: 6095 [ 7h 39m ] +INFO: Idle: 9919, Running: 4, Completed: 6097 [ 7h 40m ] +INFO: Idle: 9918, Running: 4, Completed: 6098 [ 7h 40m ] +INFO: Idle: 9917, Running: 4, Completed: 6099 [ 7h 40m ] +INFO: Idle: 9914, Running: 4, Completed: 6102 [ 7h 40m ] +INFO: Idle: 9913, Running: 4, Completed: 6103 [ 7h 40m ] +INFO: Idle: 9910, Running: 4, Completed: 6106 [ 7h 40m ] +INFO: Idle: 9909, Running: 4, Completed: 6107 [ 7h 41m ] +INFO: Idle: 9906, Running: 4, Completed: 6110 [ 7h 41m ] +INFO: Idle: 9905, Running: 4, Completed: 6111 [ 7h 41m ] +INFO: Idle: 9904, Running: 4, Completed: 6112 [ 7h 41m ] +INFO: Idle: 9902, Running: 4, Completed: 6114 [ 7h 41m ] +INFO: Idle: 9901, Running: 4, Completed: 6115 [ 7h 41m ] +INFO: Idle: 9899, Running: 4, Completed: 6117 [ 7h 41m ] +INFO: Idle: 9898, Running: 4, Completed: 6118 [ 7h 41m ] +INFO: Idle: 9897, Running: 4, Completed: 6119 [ 7h 41m ] +INFO: Idle: 9895, Running: 4, Completed: 6121 [ 7h 42m ] +INFO: Idle: 9894, Running: 4, Completed: 6122 [ 7h 42m ] +INFO: Idle: 9893, Running: 4, Completed: 6123 [ 7h 42m ] +INFO: Idle: 9890, Running: 4, Completed: 6126 [ 7h 42m ] +INFO: Idle: 9889, Running: 4, Completed: 6127 [ 7h 42m ] +INFO: Idle: 9886, Running: 4, Completed: 6130 [ 7h 42m ] +INFO: Idle: 9885, Running: 4, Completed: 6131 [ 7h 42m ] +INFO: Idle: 9883, Running: 4, Completed: 6133 [ 7h 43m ] +INFO: Idle: 9881, Running: 4, Completed: 6135 [ 7h 43m ] +INFO: Idle: 9879, Running: 4, Completed: 6137 [ 7h 43m ] +INFO: Idle: 9877, Running: 4, Completed: 6139 [ 7h 43m ] +INFO: Idle: 9875, Running: 4, Completed: 6141 [ 7h 43m ] +INFO: Idle: 9873, Running: 4, Completed: 6143 [ 7h 43m ] +INFO: Idle: 9871, Running: 4, Completed: 6145 [ 7h 44m ] +INFO: Idle: 9869, Running: 4, Completed: 6147 [ 7h 44m ] +INFO: Idle: 9867, Running: 4, Completed: 6149 [ 7h 44m ] +INFO: Idle: 9866, Running: 4, Completed: 6150 [ 7h 44m ] +INFO: Idle: 9864, Running: 4, Completed: 6152 [ 7h 44m ] +INFO: Idle: 9863, Running: 4, Completed: 6153 [ 7h 44m ] +INFO: Idle: 9862, Running: 4, Completed: 6154 [ 7h 44m ] +INFO: Idle: 9859, Running: 4, Completed: 6157 [ 7h 45m ] +INFO: Idle: 9858, Running: 4, Completed: 6158 [ 7h 45m ] +INFO: Idle: 9855, Running: 4, Completed: 6161 [ 7h 45m ] +INFO: Idle: 9854, Running: 4, Completed: 6162 [ 7h 45m ] +INFO: Idle: 9851, Running: 4, Completed: 6165 [ 7h 45m ] +INFO: Idle: 9849, Running: 4, Completed: 6167 [ 7h 45m ] +INFO: Idle: 9848, Running: 4, Completed: 6168 [ 7h 46m ] +INFO: Idle: 9847, Running: 4, Completed: 6169 [ 7h 46m ] +INFO: Idle: 9845, Running: 4, Completed: 6171 [ 7h 46m ] +INFO: Idle: 9844, Running: 4, Completed: 6172 [ 7h 46m ] +INFO: Idle: 9843, Running: 4, Completed: 6173 [ 7h 46m ] +INFO: Idle: 9841, Running: 4, Completed: 6175 [ 7h 46m ] +INFO: Idle: 9839, Running: 4, Completed: 6177 [ 7h 46m ] +INFO: Idle: 9837, Running: 4, Completed: 6179 [ 7h 47m ] +INFO: Idle: 9835, Running: 4, Completed: 6181 [ 7h 47m ] +INFO: Idle: 9834, Running: 4, Completed: 6182 [ 7h 47m ] +INFO: Idle: 9833, Running: 4, Completed: 6183 [ 7h 47m ] +INFO: Idle: 9831, Running: 4, Completed: 6185 [ 7h 47m ] +INFO: Idle: 9830, Running: 4, Completed: 6186 [ 7h 47m ] +INFO: Idle: 9828, Running: 4, Completed: 6188 [ 7h 47m ] +INFO: Idle: 9827, Running: 4, Completed: 6189 [ 7h 48m ] +INFO: Idle: 9826, Running: 4, Completed: 6190 [ 7h 48m ] +INFO: Idle: 9825, Running: 4, Completed: 6191 [ 7h 48m ] +INFO: Idle: 9824, Running: 4, Completed: 6192 [ 7h 48m ] +INFO: Idle: 9822, Running: 4, Completed: 6194 [ 7h 48m ] +INFO: Idle: 9820, Running: 4, Completed: 6196 [ 7h 48m ] +INFO: Idle: 9819, Running: 4, Completed: 6197 [ 7h 48m ] +INFO: Idle: 9818, Running: 4, Completed: 6198 [ 7h 48m ] +INFO: Idle: 9817, Running: 4, Completed: 6199 [ 7h 49m ] +INFO: Idle: 9816, Running: 4, Completed: 6200 [ 7h 49m ] +INFO: Idle: 9813, Running: 4, Completed: 6203 [ 7h 49m ] +INFO: Idle: 9811, Running: 4, Completed: 6205 [ 7h 49m ] +INFO: Idle: 9810, Running: 4, Completed: 6206 [ 7h 49m ] +INFO: Idle: 9809, Running: 4, Completed: 6207 [ 7h 49m ] +INFO: Idle: 9807, Running: 4, Completed: 6209 [ 7h 49m ] +INFO: Idle: 9806, Running: 4, Completed: 6210 [ 7h 50m ] +INFO: Idle: 9805, Running: 4, Completed: 6211 [ 7h 50m ] +INFO: Idle: 9804, Running: 4, Completed: 6212 [ 7h 50m ] +INFO: Idle: 9803, Running: 4, Completed: 6213 [ 7h 50m ] +INFO: Idle: 9801, Running: 4, Completed: 6215 [ 7h 50m ] +INFO: Idle: 9800, Running: 4, Completed: 6216 [ 7h 50m ] +INFO: Idle: 9799, Running: 4, Completed: 6217 [ 7h 50m ] +INFO: Idle: 9798, Running: 4, Completed: 6218 [ 7h 50m ] +INFO: Idle: 9796, Running: 4, Completed: 6220 [ 7h 50m ] +INFO: Idle: 9794, Running: 4, Completed: 6222 [ 7h 51m ] +INFO: Idle: 9793, Running: 4, Completed: 6223 [ 7h 51m ] +INFO: Idle: 9791, Running: 4, Completed: 6225 [ 7h 51m ] +INFO: Idle: 9790, Running: 4, Completed: 6226 [ 7h 51m ] +INFO: Idle: 9788, Running: 4, Completed: 6228 [ 7h 51m ] +INFO: Idle: 9786, Running: 4, Completed: 6230 [ 7h 51m ] +INFO: Idle: 9785, Running: 4, Completed: 6231 [ 7h 52m ] +INFO: Idle: 9783, Running: 4, Completed: 6233 [ 7h 52m ] +INFO: Idle: 9782, Running: 4, Completed: 6234 [ 7h 52m ] +INFO: Idle: 9781, Running: 4, Completed: 6235 [ 7h 52m ] +INFO: Idle: 9778, Running: 4, Completed: 6238 [ 7h 52m ] +INFO: Idle: 9777, Running: 4, Completed: 6239 [ 7h 52m ] +INFO: Idle: 9775, Running: 4, Completed: 6241 [ 7h 52m ] +INFO: Idle: 9774, Running: 4, Completed: 6242 [ 7h 52m ] +INFO: Idle: 9773, Running: 4, Completed: 6243 [ 7h 53m ] +INFO: Idle: 9772, Running: 4, Completed: 6244 [ 7h 53m ] +INFO: Idle: 9771, Running: 4, Completed: 6245 [ 7h 53m ] +INFO: Idle: 9769, Running: 4, Completed: 6247 [ 7h 53m ] +INFO: Idle: 9768, Running: 4, Completed: 6248 [ 7h 53m ] +INFO: Idle: 9766, Running: 4, Completed: 6250 [ 7h 53m ] +INFO: Idle: 9765, Running: 4, Completed: 6251 [ 7h 53m ] +INFO: Idle: 9764, Running: 4, Completed: 6252 [ 7h 53m ] +INFO: Idle: 9761, Running: 4, Completed: 6255 [ 7h 54m ] +INFO: Idle: 9760, Running: 4, Completed: 6256 [ 7h 54m ] +INFO: Idle: 9759, Running: 4, Completed: 6257 [ 7h 54m ] +INFO: Idle: 9757, Running: 4, Completed: 6259 [ 7h 54m ] +INFO: Idle: 9755, Running: 4, Completed: 6261 [ 7h 54m ] +INFO: Idle: 9753, Running: 4, Completed: 6263 [ 7h 54m ] +INFO: Idle: 9751, Running: 4, Completed: 6265 [ 7h 54m ] +INFO: Idle: 9749, Running: 4, Completed: 6267 [ 7h 54m ] +INFO: Idle: 9747, Running: 4, Completed: 6269 [ 7h 55m ] +INFO: Idle: 9745, Running: 4, Completed: 6271 [ 7h 55m ] +INFO: Idle: 9743, Running: 4, Completed: 6273 [ 7h 55m ] +INFO: Idle: 9741, Running: 4, Completed: 6275 [ 7h 55m ] +INFO: Idle: 9739, Running: 4, Completed: 6277 [ 7h 55m ] +INFO: Idle: 9738, Running: 4, Completed: 6278 [ 7h 55m ] +INFO: Idle: 9737, Running: 4, Completed: 6279 [ 7h 55m ] +INFO: Idle: 9735, Running: 4, Completed: 6281 [ 7h 56m ] +INFO: Idle: 9734, Running: 4, Completed: 6282 [ 7h 56m ] +INFO: Idle: 9733, Running: 4, Completed: 6283 [ 7h 56m ] +INFO: Idle: 9731, Running: 4, Completed: 6285 [ 7h 56m ] +INFO: Idle: 9730, Running: 4, Completed: 6286 [ 7h 56m ] +INFO: Idle: 9729, Running: 4, Completed: 6287 [ 7h 56m ] +INFO: Idle: 9727, Running: 4, Completed: 6289 [ 7h 56m ] +INFO: Idle: 9725, Running: 4, Completed: 6291 [ 7h 57m ] +INFO: Idle: 9723, Running: 4, Completed: 6293 [ 7h 57m ] +INFO: Idle: 9722, Running: 4, Completed: 6294 [ 7h 57m ] +INFO: Idle: 9720, Running: 4, Completed: 6296 [ 7h 57m ] +INFO: Idle: 9719, Running: 4, Completed: 6297 [ 7h 57m ] +INFO: Idle: 9717, Running: 4, Completed: 6299 [ 7h 57m ] +INFO: Idle: 9716, Running: 4, Completed: 6300 [ 7h 57m ] +INFO: Idle: 9715, Running: 4, Completed: 6301 [ 7h 58m ] +INFO: Idle: 9714, Running: 4, Completed: 6302 [ 7h 58m ] +INFO: Idle: 9712, Running: 4, Completed: 6304 [ 7h 58m ] +INFO: Idle: 9711, Running: 4, Completed: 6305 [ 7h 58m ] +INFO: Idle: 9709, Running: 4, Completed: 6307 [ 7h 58m ] +INFO: Idle: 9707, Running: 4, Completed: 6309 [ 7h 58m ] +INFO: Idle: 9706, Running: 4, Completed: 6310 [ 7h 58m ] +INFO: Idle: 9705, Running: 4, Completed: 6311 [ 7h 58m ] +INFO: Idle: 9703, Running: 4, Completed: 6313 [ 7h 59m ] +INFO: Idle: 9702, Running: 4, Completed: 6314 [ 7h 59m ] +INFO: Idle: 9700, Running: 4, Completed: 6316 [ 7h 59m ] +INFO: Idle: 9699, Running: 4, Completed: 6317 [ 7h 59m ] +INFO: Idle: 9698, Running: 4, Completed: 6318 [ 7h 59m ] +INFO: Idle: 9697, Running: 4, Completed: 6319 [ 7h 59m ] +INFO: Idle: 9696, Running: 4, Completed: 6320 [ 7h 59m ] +INFO: Idle: 9694, Running: 4, Completed: 6322 [ 7h 59m ] +INFO: Idle: 9692, Running: 4, Completed: 6324 [ 8h 0m ] +INFO: Idle: 9691, Running: 4, Completed: 6325 [ 8h 0m ] +INFO: Idle: 9689, Running: 4, Completed: 6327 [ 8h 0m ] +INFO: Idle: 9688, Running: 4, Completed: 6328 [ 8h 0m ] +INFO: Idle: 9685, Running: 4, Completed: 6331 [ 8h 0m ] +INFO: Idle: 9683, Running: 4, Completed: 6333 [ 8h 0m ] +INFO: Idle: 9682, Running: 4, Completed: 6334 [ 8h 1m ] +INFO: Idle: 9681, Running: 4, Completed: 6335 [ 8h 1m ] +INFO: Idle: 9679, Running: 4, Completed: 6337 [ 8h 1m ] +INFO: Idle: 9678, Running: 4, Completed: 6338 [ 8h 1m ] +INFO: Idle: 9677, Running: 4, Completed: 6339 [ 8h 1m ] +INFO: Idle: 9676, Running: 4, Completed: 6340 [ 8h 1m ] +INFO: Idle: 9674, Running: 4, Completed: 6342 [ 8h 1m ] +INFO: Idle: 9673, Running: 4, Completed: 6343 [ 8h 1m ] +INFO: Idle: 9671, Running: 4, Completed: 6345 [ 8h 2m ] +INFO: Idle: 9670, Running: 4, Completed: 6346 [ 8h 2m ] +INFO: Idle: 9668, Running: 4, Completed: 6348 [ 8h 2m ] +INFO: Idle: 9666, Running: 4, Completed: 6350 [ 8h 2m ] +INFO: Idle: 9665, Running: 4, Completed: 6351 [ 8h 2m ] +INFO: Idle: 9662, Running: 4, Completed: 6354 [ 8h 3m ] +INFO: Idle: 9660, Running: 4, Completed: 6356 [ 8h 3m ] +INFO: Idle: 9659, Running: 4, Completed: 6357 [ 8h 3m ] +INFO: Idle: 9658, Running: 4, Completed: 6358 [ 8h 3m ] +INFO: Idle: 9657, Running: 4, Completed: 6359 [ 8h 3m ] +INFO: Idle: 9654, Running: 4, Completed: 6362 [ 8h 3m ] +INFO: Idle: 9653, Running: 4, Completed: 6363 [ 8h 3m ] +INFO: Idle: 9650, Running: 4, Completed: 6366 [ 8h 3m ] +INFO: Idle: 9649, Running: 4, Completed: 6367 [ 8h 4m ] +INFO: Idle: 9647, Running: 4, Completed: 6369 [ 8h 4m ] +INFO: Idle: 9646, Running: 4, Completed: 6370 [ 8h 4m ] +INFO: Idle: 9645, Running: 4, Completed: 6371 [ 8h 4m ] +INFO: Idle: 9644, Running: 4, Completed: 6372 [ 8h 4m ] +INFO: Idle: 9643, Running: 4, Completed: 6373 [ 8h 4m ] +INFO: Idle: 9641, Running: 4, Completed: 6375 [ 8h 4m ] +INFO: Idle: 9640, Running: 4, Completed: 6376 [ 8h 4m ] +INFO: Idle: 9638, Running: 4, Completed: 6378 [ 8h 5m ] +INFO: Idle: 9637, Running: 4, Completed: 6379 [ 8h 5m ] +INFO: Idle: 9636, Running: 4, Completed: 6380 [ 8h 5m ] +INFO: Idle: 9633, Running: 4, Completed: 6383 [ 8h 5m ] +INFO: Idle: 9632, Running: 4, Completed: 6384 [ 8h 5m ] +INFO: Idle: 9629, Running: 4, Completed: 6387 [ 8h 5m ] +INFO: Idle: 9625, Running: 4, Completed: 6391 [ 8h 6m ] +INFO: Idle: 9623, Running: 4, Completed: 6393 [ 8h 6m ] +INFO: Idle: 9621, Running: 4, Completed: 6395 [ 8h 6m ] +INFO: Idle: 9618, Running: 4, Completed: 6398 [ 8h 6m ] +INFO: Idle: 9617, Running: 4, Completed: 6399 [ 8h 6m ] +INFO: Idle: 9615, Running: 4, Completed: 6401 [ 8h 6m ] +INFO: Idle: 9613, Running: 4, Completed: 6403 [ 8h 7m ] +INFO: Idle: 9611, Running: 4, Completed: 6405 [ 8h 7m ] +INFO: Idle: 9609, Running: 4, Completed: 6407 [ 8h 7m ] +INFO: Idle: 9607, Running: 4, Completed: 6409 [ 8h 7m ] +INFO: Idle: 9605, Running: 4, Completed: 6411 [ 8h 7m ] +INFO: Idle: 9603, Running: 4, Completed: 6413 [ 8h 7m ] +INFO: Idle: 9601, Running: 4, Completed: 6415 [ 8h 8m ] +INFO: Idle: 9599, Running: 4, Completed: 6417 [ 8h 8m ] +INFO: Idle: 9596, Running: 4, Completed: 6420 [ 8h 8m ] +INFO: Idle: 9595, Running: 4, Completed: 6421 [ 8h 8m ] +INFO: Idle: 9594, Running: 4, Completed: 6422 [ 8h 8m ] +INFO: Idle: 9592, Running: 4, Completed: 6424 [ 8h 9m ] +INFO: Idle: 9591, Running: 4, Completed: 6425 [ 8h 9m ] +INFO: Idle: 9589, Running: 4, Completed: 6427 [ 8h 9m ] +INFO: Idle: 9587, Running: 4, Completed: 6429 [ 8h 9m ] +INFO: Idle: 9585, Running: 4, Completed: 6431 [ 8h 9m ] +INFO: Idle: 9584, Running: 4, Completed: 6432 [ 8h 9m ] +INFO: Idle: 9583, Running: 4, Completed: 6433 [ 8h 9m ] +INFO: Idle: 9581, Running: 4, Completed: 6435 [ 8h 10m ] +INFO: Idle: 9579, Running: 4, Completed: 6437 [ 8h 10m ] +INFO: Idle: 9578, Running: 4, Completed: 6438 [ 8h 10m ] +INFO: Idle: 9577, Running: 4, Completed: 6439 [ 8h 10m ] +INFO: Idle: 9575, Running: 4, Completed: 6441 [ 8h 10m ] +INFO: Idle: 9573, Running: 4, Completed: 6443 [ 8h 10m ] +INFO: Idle: 9572, Running: 4, Completed: 6444 [ 8h 10m ] +INFO: Idle: 9571, Running: 4, Completed: 6445 [ 8h 11m ] +INFO: Idle: 9570, Running: 4, Completed: 6446 [ 8h 11m ] +INFO: Idle: 9569, Running: 4, Completed: 6447 [ 8h 11m ] +INFO: Idle: 9568, Running: 4, Completed: 6448 [ 8h 11m ] +INFO: Idle: 9566, Running: 4, Completed: 6450 [ 8h 11m ] +INFO: Idle: 9565, Running: 4, Completed: 6451 [ 8h 11m ] +INFO: Idle: 9563, Running: 4, Completed: 6453 [ 8h 11m ] +INFO: Idle: 9562, Running: 4, Completed: 6454 [ 8h 11m ] +INFO: Idle: 9561, Running: 4, Completed: 6455 [ 8h 12m ] +INFO: Idle: 9558, Running: 4, Completed: 6458 [ 8h 12m ] +INFO: Idle: 9557, Running: 4, Completed: 6459 [ 8h 12m ] +INFO: Idle: 9555, Running: 4, Completed: 6461 [ 8h 12m ] +INFO: Idle: 9554, Running: 4, Completed: 6462 [ 8h 12m ] +INFO: Idle: 9553, Running: 4, Completed: 6463 [ 8h 12m ] +INFO: Idle: 9550, Running: 4, Completed: 6466 [ 8h 13m ] +INFO: Idle: 9549, Running: 4, Completed: 6467 [ 8h 13m ] +INFO: Idle: 9546, Running: 4, Completed: 6470 [ 8h 13m ] +INFO: Idle: 9545, Running: 4, Completed: 6471 [ 8h 13m ] +INFO: Idle: 9543, Running: 4, Completed: 6473 [ 8h 13m ] +INFO: Idle: 9541, Running: 4, Completed: 6475 [ 8h 13m ] +INFO: Idle: 9540, Running: 4, Completed: 6476 [ 8h 13m ] +INFO: Idle: 9539, Running: 4, Completed: 6477 [ 8h 14m ] +INFO: Idle: 9537, Running: 4, Completed: 6479 [ 8h 14m ] +INFO: Idle: 9536, Running: 4, Completed: 6480 [ 8h 14m ] +INFO: Idle: 9533, Running: 4, Completed: 6483 [ 8h 14m ] +INFO: Idle: 9531, Running: 4, Completed: 6485 [ 8h 14m ] +INFO: Idle: 9529, Running: 4, Completed: 6487 [ 8h 15m ] +INFO: Idle: 9527, Running: 4, Completed: 6489 [ 8h 15m ] +INFO: Idle: 9526, Running: 4, Completed: 6490 [ 8h 15m ] +INFO: Idle: 9525, Running: 4, Completed: 6491 [ 8h 15m ] +INFO: Idle: 9524, Running: 4, Completed: 6492 [ 8h 15m ] +INFO: Idle: 9523, Running: 4, Completed: 6493 [ 8h 15m ] +INFO: Idle: 9522, Running: 4, Completed: 6494 [ 8h 15m ] +INFO: Idle: 9521, Running: 4, Completed: 6495 [ 8h 15m ] +INFO: Idle: 9519, Running: 4, Completed: 6497 [ 8h 15m ] +INFO: Idle: 9518, Running: 4, Completed: 6498 [ 8h 15m ] +INFO: Idle: 9517, Running: 4, Completed: 6499 [ 8h 15m ] +INFO: Idle: 9516, Running: 4, Completed: 6500 [ 8h 16m ] +INFO: Idle: 9515, Running: 4, Completed: 6501 [ 8h 16m ] +INFO: Idle: 9513, Running: 4, Completed: 6503 [ 8h 16m ] +INFO: Idle: 9512, Running: 4, Completed: 6504 [ 8h 16m ] +INFO: Idle: 9511, Running: 4, Completed: 6505 [ 8h 16m ] +INFO: Idle: 9509, Running: 4, Completed: 6507 [ 8h 16m ] +INFO: Idle: 9508, Running: 4, Completed: 6508 [ 8h 16m ] +INFO: Idle: 9507, Running: 4, Completed: 6509 [ 8h 16m ] +INFO: Idle: 9505, Running: 4, Completed: 6511 [ 8h 16m ] +INFO: Idle: 9504, Running: 4, Completed: 6512 [ 8h 17m ] +INFO: Idle: 9503, Running: 4, Completed: 6513 [ 8h 17m ] +INFO: Idle: 9501, Running: 4, Completed: 6515 [ 8h 17m ] +INFO: Idle: 9499, Running: 4, Completed: 6517 [ 8h 17m ] +INFO: Idle: 9497, Running: 4, Completed: 6519 [ 8h 17m ] +INFO: Idle: 9495, Running: 4, Completed: 6521 [ 8h 17m ] +INFO: Idle: 9493, Running: 4, Completed: 6523 [ 8h 17m ] +INFO: Idle: 9491, Running: 4, Completed: 6525 [ 8h 18m ] +INFO: Idle: 9488, Running: 4, Completed: 6528 [ 8h 18m ] +INFO: Idle: 9487, Running: 4, Completed: 6529 [ 8h 18m ] +INFO: Idle: 9484, Running: 4, Completed: 6532 [ 8h 18m ] +INFO: Idle: 9483, Running: 4, Completed: 6533 [ 8h 18m ] +INFO: Idle: 9480, Running: 4, Completed: 6536 [ 8h 18m ] +INFO: Idle: 9479, Running: 4, Completed: 6537 [ 8h 19m ] +INFO: Idle: 9478, Running: 4, Completed: 6538 [ 8h 19m ] +INFO: Idle: 9476, Running: 4, Completed: 6540 [ 8h 19m ] +INFO: Idle: 9475, Running: 4, Completed: 6541 [ 8h 19m ] +INFO: Idle: 9474, Running: 4, Completed: 6542 [ 8h 19m ] +INFO: Idle: 9472, Running: 4, Completed: 6544 [ 8h 19m ] +INFO: Idle: 9471, Running: 4, Completed: 6545 [ 8h 19m ] +INFO: Idle: 9469, Running: 4, Completed: 6547 [ 8h 19m ] +INFO: Idle: 9467, Running: 4, Completed: 6549 [ 8h 20m ] +INFO: Idle: 9464, Running: 4, Completed: 6552 [ 8h 20m ] +INFO: Idle: 9463, Running: 4, Completed: 6553 [ 8h 20m ] +INFO: Idle: 9461, Running: 4, Completed: 6555 [ 8h 20m ] +INFO: Idle: 9460, Running: 4, Completed: 6556 [ 8h 20m ] +INFO: Idle: 9459, Running: 4, Completed: 6557 [ 8h 20m ] +INFO: Idle: 9458, Running: 4, Completed: 6558 [ 8h 20m ] +INFO: Idle: 9456, Running: 4, Completed: 6560 [ 8h 20m ] +INFO: Idle: 9455, Running: 4, Completed: 6561 [ 8h 21m ] +INFO: Idle: 9454, Running: 4, Completed: 6562 [ 8h 21m ] +INFO: Idle: 9452, Running: 4, Completed: 6564 [ 8h 21m ] +INFO: Idle: 9451, Running: 4, Completed: 6565 [ 8h 21m ] +INFO: Idle: 9449, Running: 4, Completed: 6567 [ 8h 21m ] +INFO: Idle: 9448, Running: 4, Completed: 6568 [ 8h 21m ] +INFO: Idle: 9446, Running: 4, Completed: 6570 [ 8h 21m ] +INFO: Idle: 9445, Running: 4, Completed: 6571 [ 8h 22m ] +INFO: Idle: 9444, Running: 4, Completed: 6572 [ 8h 22m ] +INFO: Idle: 9443, Running: 4, Completed: 6573 [ 8h 22m ] +INFO: Idle: 9442, Running: 4, Completed: 6574 [ 8h 22m ] +INFO: Idle: 9441, Running: 4, Completed: 6575 [ 8h 22m ] +INFO: Idle: 9439, Running: 4, Completed: 6577 [ 8h 22m ] +INFO: Idle: 9438, Running: 4, Completed: 6578 [ 8h 22m ] +INFO: Idle: 9436, Running: 4, Completed: 6580 [ 8h 22m ] +INFO: Idle: 9434, Running: 4, Completed: 6582 [ 8h 23m ] +INFO: Idle: 9432, Running: 4, Completed: 6584 [ 8h 23m ] +INFO: Idle: 9430, Running: 4, Completed: 6586 [ 8h 23m ] +INFO: Idle: 9428, Running: 4, Completed: 6588 [ 8h 23m ] +INFO: Idle: 9426, Running: 4, Completed: 6590 [ 8h 23m ] +INFO: Idle: 9423, Running: 4, Completed: 6593 [ 8h 23m ] +INFO: Idle: 9422, Running: 4, Completed: 6594 [ 8h 24m ] +INFO: Idle: 9419, Running: 4, Completed: 6597 [ 8h 24m ] +INFO: Idle: 9418, Running: 4, Completed: 6598 [ 8h 24m ] +INFO: Idle: 9414, Running: 4, Completed: 6602 [ 8h 24m ] +INFO: Idle: 9413, Running: 4, Completed: 6603 [ 8h 24m ] +INFO: Idle: 9412, Running: 4, Completed: 6604 [ 8h 24m ] +INFO: Idle: 9409, Running: 4, Completed: 6607 [ 8h 24m ] +INFO: Idle: 9406, Running: 4, Completed: 6610 [ 8h 25m ] +INFO: Idle: 9405, Running: 4, Completed: 6611 [ 8h 25m ] +INFO: Idle: 9403, Running: 4, Completed: 6613 [ 8h 25m ] +INFO: Idle: 9401, Running: 4, Completed: 6615 [ 8h 25m ] +INFO: Idle: 9399, Running: 4, Completed: 6617 [ 8h 25m ] +INFO: Idle: 9397, Running: 4, Completed: 6619 [ 8h 26m ] +INFO: Idle: 9394, Running: 4, Completed: 6622 [ 8h 26m ] +INFO: Idle: 9393, Running: 4, Completed: 6623 [ 8h 26m ] +INFO: Idle: 9391, Running: 4, Completed: 6625 [ 8h 26m ] +INFO: Idle: 9390, Running: 4, Completed: 6626 [ 8h 26m ] +INFO: Idle: 9389, Running: 4, Completed: 6627 [ 8h 26m ] +INFO: Idle: 9388, Running: 4, Completed: 6628 [ 8h 26m ] +INFO: Idle: 9387, Running: 4, Completed: 6629 [ 8h 26m ] +INFO: Idle: 9386, Running: 4, Completed: 6630 [ 8h 27m ] +INFO: Idle: 9384, Running: 4, Completed: 6632 [ 8h 27m ] +INFO: Idle: 9383, Running: 4, Completed: 6633 [ 8h 27m ] +INFO: Idle: 9382, Running: 4, Completed: 6634 [ 8h 27m ] +INFO: Idle: 9381, Running: 4, Completed: 6635 [ 8h 27m ] +INFO: Idle: 9380, Running: 4, Completed: 6636 [ 8h 27m ] +INFO: Idle: 9378, Running: 4, Completed: 6638 [ 8h 27m ] +INFO: Idle: 9377, Running: 4, Completed: 6639 [ 8h 27m ] +INFO: Idle: 9376, Running: 4, Completed: 6640 [ 8h 27m ] +INFO: Idle: 9375, Running: 4, Completed: 6641 [ 8h 28m ] +INFO: Idle: 9373, Running: 4, Completed: 6643 [ 8h 28m ] +INFO: Idle: 9370, Running: 4, Completed: 6646 [ 8h 28m ] +INFO: Idle: 9369, Running: 4, Completed: 6647 [ 8h 28m ] +INFO: Idle: 9367, Running: 4, Completed: 6649 [ 8h 28m ] +INFO: Idle: 9366, Running: 4, Completed: 6650 [ 8h 28m ] +INFO: Idle: 9365, Running: 4, Completed: 6651 [ 8h 28m ] +INFO: Idle: 9364, Running: 4, Completed: 6652 [ 8h 28m ] +INFO: Idle: 9362, Running: 4, Completed: 6654 [ 8h 28m ] +INFO: Idle: 9361, Running: 4, Completed: 6655 [ 8h 29m ] +INFO: Idle: 9358, Running: 4, Completed: 6658 [ 8h 29m ] +INFO: Idle: 9357, Running: 4, Completed: 6659 [ 8h 29m ] +INFO: Idle: 9356, Running: 4, Completed: 6660 [ 8h 29m ] +INFO: Idle: 9354, Running: 4, Completed: 6662 [ 8h 29m ] +INFO: Idle: 9352, Running: 4, Completed: 6664 [ 8h 30m ] +INFO: Idle: 9350, Running: 4, Completed: 6666 [ 8h 30m ] +INFO: Idle: 9348, Running: 4, Completed: 6668 [ 8h 30m ] +INFO: Idle: 9347, Running: 4, Completed: 6669 [ 8h 30m ] +INFO: Idle: 9346, Running: 4, Completed: 6670 [ 8h 30m ] +INFO: Idle: 9343, Running: 4, Completed: 6673 [ 8h 30m ] +INFO: Idle: 9342, Running: 4, Completed: 6674 [ 8h 30m ] +INFO: Idle: 9341, Running: 4, Completed: 6675 [ 8h 30m ] +INFO: Idle: 9339, Running: 4, Completed: 6677 [ 8h 30m ] +INFO: Idle: 9338, Running: 4, Completed: 6678 [ 8h 30m ] +INFO: Idle: 9337, Running: 4, Completed: 6679 [ 8h 31m ] +INFO: Idle: 9334, Running: 4, Completed: 6682 [ 8h 31m ] +INFO: Idle: 9333, Running: 4, Completed: 6683 [ 8h 31m ] +INFO: Idle: 9331, Running: 4, Completed: 6685 [ 8h 31m ] +INFO: Idle: 9329, Running: 4, Completed: 6687 [ 8h 31m ] +INFO: Idle: 9327, Running: 4, Completed: 6689 [ 8h 31m ] +INFO: Idle: 9325, Running: 4, Completed: 6691 [ 8h 31m ] +INFO: Idle: 9324, Running: 4, Completed: 6692 [ 8h 31m ] +INFO: Idle: 9323, Running: 4, Completed: 6693 [ 8h 32m ] +INFO: Idle: 9321, Running: 4, Completed: 6695 [ 8h 32m ] +INFO: Idle: 9319, Running: 4, Completed: 6697 [ 8h 32m ] +INFO: Idle: 9317, Running: 4, Completed: 6699 [ 8h 32m ] +INFO: Idle: 9315, Running: 4, Completed: 6701 [ 8h 32m ] +INFO: Idle: 9314, Running: 4, Completed: 6702 [ 8h 32m ] +INFO: Idle: 9313, Running: 4, Completed: 6703 [ 8h 33m ] +INFO: Idle: 9311, Running: 4, Completed: 6705 [ 8h 33m ] +INFO: Idle: 9310, Running: 4, Completed: 6706 [ 8h 33m ] +INFO: Idle: 9309, Running: 4, Completed: 6707 [ 8h 33m ] +INFO: Idle: 9307, Running: 4, Completed: 6709 [ 8h 33m ] +INFO: Idle: 9304, Running: 4, Completed: 6712 [ 8h 33m ] +INFO: Idle: 9303, Running: 4, Completed: 6713 [ 8h 34m ] +INFO: Idle: 9302, Running: 4, Completed: 6714 [ 8h 34m ] +INFO: Idle: 9300, Running: 4, Completed: 6716 [ 8h 34m ] +INFO: Idle: 9299, Running: 4, Completed: 6717 [ 8h 34m ] +INFO: Idle: 9296, Running: 4, Completed: 6720 [ 8h 34m ] +INFO: Idle: 9295, Running: 4, Completed: 6721 [ 8h 34m ] +INFO: Idle: 9292, Running: 4, Completed: 6724 [ 8h 34m ] +INFO: Idle: 9291, Running: 4, Completed: 6725 [ 8h 35m ] +INFO: Idle: 9290, Running: 4, Completed: 6726 [ 8h 35m ] +INFO: Idle: 9289, Running: 4, Completed: 6727 [ 8h 35m ] +INFO: Idle: 9287, Running: 4, Completed: 6729 [ 8h 35m ] +INFO: Idle: 9286, Running: 4, Completed: 6730 [ 8h 35m ] +INFO: Idle: 9285, Running: 4, Completed: 6731 [ 8h 35m ] +INFO: Idle: 9283, Running: 4, Completed: 6733 [ 8h 35m ] +INFO: Idle: 9281, Running: 4, Completed: 6735 [ 8h 36m ] +INFO: Idle: 9279, Running: 4, Completed: 6737 [ 8h 36m ] +INFO: Idle: 9278, Running: 4, Completed: 6738 [ 8h 36m ] +INFO: Idle: 9277, Running: 4, Completed: 6739 [ 8h 36m ] +INFO: Idle: 9276, Running: 4, Completed: 6740 [ 8h 36m ] +INFO: Idle: 9274, Running: 4, Completed: 6742 [ 8h 36m ] +INFO: Idle: 9273, Running: 4, Completed: 6743 [ 8h 36m ] +INFO: Idle: 9270, Running: 4, Completed: 6746 [ 8h 37m ] +INFO: Idle: 9269, Running: 4, Completed: 6747 [ 8h 37m ] +INFO: Idle: 9266, Running: 4, Completed: 6750 [ 8h 37m ] +INFO: Idle: 9265, Running: 4, Completed: 6751 [ 8h 37m ] +INFO: Idle: 9263, Running: 4, Completed: 6753 [ 8h 37m ] +INFO: Idle: 9262, Running: 4, Completed: 6754 [ 8h 38m ] +INFO: Idle: 9261, Running: 4, Completed: 6755 [ 8h 38m ] +INFO: Idle: 9259, Running: 4, Completed: 6757 [ 8h 38m ] +INFO: Idle: 9258, Running: 4, Completed: 6758 [ 8h 38m ] +INFO: Idle: 9257, Running: 4, Completed: 6759 [ 8h 38m ] +INFO: Idle: 9254, Running: 4, Completed: 6762 [ 8h 38m ] +INFO: Idle: 9253, Running: 4, Completed: 6763 [ 8h 38m ] +INFO: Idle: 9251, Running: 4, Completed: 6765 [ 8h 39m ] +INFO: Idle: 9250, Running: 4, Completed: 6766 [ 8h 39m ] +INFO: Idle: 9249, Running: 4, Completed: 6767 [ 8h 39m ] +INFO: Idle: 9248, Running: 4, Completed: 6768 [ 8h 39m ] +INFO: Idle: 9246, Running: 4, Completed: 6770 [ 8h 39m ] +INFO: Idle: 9245, Running: 4, Completed: 6771 [ 8h 39m ] +INFO: Idle: 9243, Running: 4, Completed: 6773 [ 8h 39m ] +INFO: Idle: 9242, Running: 4, Completed: 6774 [ 8h 39m ] +INFO: Idle: 9241, Running: 4, Completed: 6775 [ 8h 39m ] +INFO: Idle: 9239, Running: 4, Completed: 6777 [ 8h 40m ] +INFO: Idle: 9237, Running: 4, Completed: 6779 [ 8h 40m ] +INFO: Idle: 9233, Running: 4, Completed: 6783 [ 8h 40m ] +INFO: Idle: 9230, Running: 4, Completed: 6786 [ 8h 40m ] +INFO: Idle: 9229, Running: 4, Completed: 6787 [ 8h 40m ] +INFO: Idle: 9228, Running: 4, Completed: 6788 [ 8h 41m ] +INFO: Idle: 9226, Running: 4, Completed: 6790 [ 8h 41m ] +INFO: Idle: 9225, Running: 4, Completed: 6791 [ 8h 41m ] +INFO: Idle: 9223, Running: 4, Completed: 6793 [ 8h 41m ] +INFO: Idle: 9221, Running: 4, Completed: 6795 [ 8h 41m ] +INFO: Idle: 9218, Running: 4, Completed: 6798 [ 8h 41m ] +INFO: Idle: 9217, Running: 4, Completed: 6799 [ 8h 41m ] +INFO: Idle: 9214, Running: 4, Completed: 6802 [ 8h 42m ] +INFO: Idle: 9213, Running: 4, Completed: 6803 [ 8h 42m ] +INFO: Idle: 9210, Running: 4, Completed: 6806 [ 8h 42m ] +INFO: Idle: 9209, Running: 4, Completed: 6807 [ 8h 42m ] +INFO: Idle: 9207, Running: 4, Completed: 6809 [ 8h 42m ] +INFO: Idle: 9206, Running: 4, Completed: 6810 [ 8h 42m ] +INFO: Idle: 9205, Running: 4, Completed: 6811 [ 8h 42m ] +INFO: Idle: 9203, Running: 4, Completed: 6813 [ 8h 43m ] +INFO: Idle: 9202, Running: 4, Completed: 6814 [ 8h 43m ] +INFO: Idle: 9201, Running: 4, Completed: 6815 [ 8h 43m ] +INFO: Idle: 9198, Running: 4, Completed: 6818 [ 8h 43m ] +INFO: Idle: 9197, Running: 4, Completed: 6819 [ 8h 43m ] +INFO: Idle: 9195, Running: 4, Completed: 6821 [ 8h 43m ] +INFO: Idle: 9194, Running: 4, Completed: 6822 [ 8h 43m ] +INFO: Idle: 9193, Running: 4, Completed: 6823 [ 8h 43m ] +INFO: Idle: 9191, Running: 4, Completed: 6825 [ 8h 43m ] +INFO: Idle: 9190, Running: 4, Completed: 6826 [ 8h 44m ] +INFO: Idle: 9189, Running: 4, Completed: 6827 [ 8h 44m ] +INFO: Idle: 9187, Running: 4, Completed: 6829 [ 8h 44m ] +INFO: Idle: 9186, Running: 4, Completed: 6830 [ 8h 44m ] +INFO: Idle: 9185, Running: 4, Completed: 6831 [ 8h 44m ] +INFO: Idle: 9184, Running: 4, Completed: 6832 [ 8h 44m ] +INFO: Idle: 9183, Running: 4, Completed: 6833 [ 8h 44m ] +INFO: Idle: 9182, Running: 4, Completed: 6834 [ 8h 44m ] +INFO: Idle: 9181, Running: 4, Completed: 6835 [ 8h 44m ] +INFO: Idle: 9179, Running: 4, Completed: 6837 [ 8h 45m ] +INFO: Idle: 9177, Running: 4, Completed: 6839 [ 8h 45m ] +INFO: Idle: 9176, Running: 4, Completed: 6840 [ 8h 45m ] +INFO: Idle: 9175, Running: 4, Completed: 6841 [ 8h 45m ] +INFO: Idle: 9174, Running: 4, Completed: 6842 [ 8h 45m ] +INFO: Idle: 9173, Running: 4, Completed: 6843 [ 8h 45m ] +INFO: Idle: 9172, Running: 4, Completed: 6844 [ 8h 45m ] +INFO: Idle: 9171, Running: 4, Completed: 6845 [ 8h 46m ] +INFO: Idle: 9170, Running: 4, Completed: 6846 [ 8h 46m ] +INFO: Idle: 9168, Running: 4, Completed: 6848 [ 8h 46m ] +INFO: Idle: 9167, Running: 4, Completed: 6849 [ 8h 46m ] +INFO: Idle: 9166, Running: 4, Completed: 6850 [ 8h 46m ] +INFO: Idle: 9163, Running: 4, Completed: 6853 [ 8h 46m ] +INFO: Idle: 9162, Running: 4, Completed: 6854 [ 8h 46m ] +INFO: Idle: 9161, Running: 4, Completed: 6855 [ 8h 46m ] +INFO: Idle: 9160, Running: 4, Completed: 6856 [ 8h 47m ] +INFO: Idle: 9159, Running: 4, Completed: 6857 [ 8h 47m ] +INFO: Idle: 9158, Running: 4, Completed: 6858 [ 8h 47m ] +INFO: Idle: 9156, Running: 4, Completed: 6860 [ 8h 47m ] +INFO: Idle: 9155, Running: 4, Completed: 6861 [ 8h 47m ] +INFO: Idle: 9153, Running: 4, Completed: 6863 [ 8h 47m ] +INFO: Idle: 9152, Running: 4, Completed: 6864 [ 8h 47m ] +INFO: Idle: 9151, Running: 4, Completed: 6865 [ 8h 47m ] +INFO: Idle: 9150, Running: 4, Completed: 6866 [ 8h 47m ] +INFO: Idle: 9149, Running: 4, Completed: 6867 [ 8h 48m ] +INFO: Idle: 9148, Running: 4, Completed: 6868 [ 8h 48m ] +INFO: Idle: 9146, Running: 4, Completed: 6870 [ 8h 48m ] +INFO: Idle: 9145, Running: 4, Completed: 6871 [ 8h 48m ] +INFO: Idle: 9142, Running: 4, Completed: 6874 [ 8h 48m ] +INFO: Idle: 9141, Running: 4, Completed: 6875 [ 8h 48m ] +INFO: Idle: 9138, Running: 4, Completed: 6878 [ 8h 49m ] +INFO: Idle: 9137, Running: 4, Completed: 6879 [ 8h 49m ] +INFO: Idle: 9135, Running: 4, Completed: 6881 [ 8h 49m ] +INFO: Idle: 9134, Running: 4, Completed: 6882 [ 8h 49m ] +INFO: Idle: 9133, Running: 4, Completed: 6883 [ 8h 49m ] +INFO: Idle: 9132, Running: 4, Completed: 6884 [ 8h 49m ] +INFO: Idle: 9130, Running: 4, Completed: 6886 [ 8h 49m ] +INFO: Idle: 9129, Running: 4, Completed: 6887 [ 8h 50m ] +INFO: Idle: 9126, Running: 4, Completed: 6890 [ 8h 50m ] +INFO: Idle: 9125, Running: 4, Completed: 6891 [ 8h 50m ] +INFO: Idle: 9124, Running: 4, Completed: 6892 [ 8h 50m ] +INFO: Idle: 9123, Running: 4, Completed: 6893 [ 8h 50m ] +INFO: Idle: 9122, Running: 4, Completed: 6894 [ 8h 50m ] +INFO: Idle: 9121, Running: 4, Completed: 6895 [ 8h 50m ] +INFO: Idle: 9120, Running: 4, Completed: 6896 [ 8h 50m ] +INFO: Idle: 9119, Running: 4, Completed: 6897 [ 8h 50m ] +INFO: Idle: 9118, Running: 4, Completed: 6898 [ 8h 50m ] +INFO: Idle: 9117, Running: 4, Completed: 6899 [ 8h 51m ] +INFO: Idle: 9115, Running: 4, Completed: 6901 [ 8h 51m ] +INFO: Idle: 9113, Running: 4, Completed: 6903 [ 8h 51m ] +INFO: Idle: 9111, Running: 4, Completed: 6905 [ 8h 51m ] +INFO: Idle: 9109, Running: 4, Completed: 6907 [ 8h 51m ] +INFO: Idle: 9105, Running: 4, Completed: 6911 [ 8h 52m ] +INFO: Idle: 9102, Running: 4, Completed: 6914 [ 8h 52m ] +INFO: Idle: 9101, Running: 4, Completed: 6915 [ 8h 52m ] +INFO: Idle: 9100, Running: 4, Completed: 6916 [ 8h 52m ] +INFO: Idle: 9099, Running: 4, Completed: 6917 [ 8h 52m ] +INFO: Idle: 9097, Running: 4, Completed: 6919 [ 8h 52m ] +INFO: Idle: 9095, Running: 4, Completed: 6921 [ 8h 52m ] +INFO: Idle: 9093, Running: 4, Completed: 6923 [ 8h 53m ] +INFO: Idle: 9090, Running: 4, Completed: 6926 [ 8h 53m ] +INFO: Idle: 9089, Running: 4, Completed: 6927 [ 8h 53m ] +INFO: Idle: 9086, Running: 4, Completed: 6930 [ 8h 53m ] +INFO: Idle: 9085, Running: 4, Completed: 6931 [ 8h 53m ] +INFO: Idle: 9084, Running: 4, Completed: 6932 [ 8h 53m ] +INFO: Idle: 9082, Running: 4, Completed: 6934 [ 8h 53m ] +INFO: Idle: 9081, Running: 4, Completed: 6935 [ 8h 54m ] +INFO: Idle: 9079, Running: 4, Completed: 6937 [ 8h 54m ] +INFO: Idle: 9077, Running: 4, Completed: 6939 [ 8h 54m ] +INFO: Idle: 9076, Running: 4, Completed: 6940 [ 8h 54m ] +INFO: Idle: 9075, Running: 4, Completed: 6941 [ 8h 54m ] +INFO: Idle: 9074, Running: 4, Completed: 6942 [ 8h 54m ] +INFO: Idle: 9073, Running: 4, Completed: 6943 [ 8h 54m ] +INFO: Idle: 9072, Running: 4, Completed: 6944 [ 8h 54m ] +INFO: Idle: 9070, Running: 4, Completed: 6946 [ 8h 54m ] +INFO: Idle: 9069, Running: 4, Completed: 6947 [ 8h 54m ] +INFO: Idle: 9067, Running: 4, Completed: 6949 [ 8h 55m ] +INFO: Idle: 9066, Running: 4, Completed: 6950 [ 8h 55m ] +INFO: Idle: 9065, Running: 4, Completed: 6951 [ 8h 55m ] +INFO: Idle: 9063, Running: 4, Completed: 6953 [ 8h 55m ] +INFO: Idle: 9062, Running: 4, Completed: 6954 [ 8h 55m ] +INFO: Idle: 9061, Running: 4, Completed: 6955 [ 8h 55m ] +INFO: Idle: 9059, Running: 4, Completed: 6957 [ 8h 55m ] +INFO: Idle: 9058, Running: 4, Completed: 6958 [ 8h 56m ] +INFO: Idle: 9057, Running: 4, Completed: 6959 [ 8h 56m ] +INFO: Idle: 9055, Running: 4, Completed: 6961 [ 8h 56m ] +INFO: Idle: 9054, Running: 4, Completed: 6962 [ 8h 56m ] +INFO: Idle: 9053, Running: 4, Completed: 6963 [ 8h 56m ] +INFO: Idle: 9051, Running: 4, Completed: 6965 [ 8h 56m ] +INFO: Idle: 9049, Running: 4, Completed: 6967 [ 8h 56m ] +INFO: Idle: 9048, Running: 4, Completed: 6968 [ 8h 56m ] +INFO: Idle: 9047, Running: 4, Completed: 6969 [ 8h 57m ] +INFO: Idle: 9046, Running: 4, Completed: 6970 [ 8h 57m ] +INFO: Idle: 9045, Running: 4, Completed: 6971 [ 8h 57m ] +INFO: Idle: 9044, Running: 4, Completed: 6972 [ 8h 57m ] +INFO: Idle: 9043, Running: 4, Completed: 6973 [ 8h 57m ] +INFO: Idle: 9042, Running: 4, Completed: 6974 [ 8h 57m ] +INFO: Idle: 9040, Running: 4, Completed: 6976 [ 8h 57m ] +INFO: Idle: 9039, Running: 4, Completed: 6977 [ 8h 57m ] +INFO: Idle: 9038, Running: 4, Completed: 6978 [ 8h 57m ] +INFO: Idle: 9036, Running: 4, Completed: 6980 [ 8h 57m ] +INFO: Idle: 9035, Running: 4, Completed: 6981 [ 8h 58m ] +INFO: Idle: 9034, Running: 4, Completed: 6982 [ 8h 58m ] +INFO: Idle: 9033, Running: 4, Completed: 6983 [ 8h 58m ] +INFO: Idle: 9032, Running: 4, Completed: 6984 [ 8h 58m ] +INFO: Idle: 9030, Running: 4, Completed: 6986 [ 8h 58m ] +INFO: Idle: 9028, Running: 4, Completed: 6988 [ 8h 58m ] +INFO: Idle: 9027, Running: 4, Completed: 6989 [ 8h 58m ] +INFO: Idle: 9025, Running: 4, Completed: 6991 [ 8h 59m ] +INFO: Idle: 9024, Running: 4, Completed: 6992 [ 8h 59m ] +INFO: Idle: 9023, Running: 4, Completed: 6993 [ 8h 59m ] +INFO: Idle: 9022, Running: 4, Completed: 6994 [ 8h 59m ] +INFO: Idle: 9021, Running: 4, Completed: 6995 [ 8h 59m ] +INFO: Idle: 9020, Running: 4, Completed: 6996 [ 8h 59m ] +INFO: Idle: 9018, Running: 4, Completed: 6998 [ 8h 59m ] +INFO: Idle: 9017, Running: 4, Completed: 6999 [ 8h 59m ] +INFO: Idle: 9015, Running: 4, Completed: 7001 [ 8h 59m ] +INFO: Idle: 9014, Running: 4, Completed: 7002 [ 9h 0m ] +INFO: Idle: 9013, Running: 4, Completed: 7003 [ 9h 0m ] +INFO: Idle: 9010, Running: 4, Completed: 7006 [ 9h 0m ] +INFO: Idle: 9009, Running: 4, Completed: 7007 [ 9h 0m ] +INFO: Idle: 9007, Running: 4, Completed: 7009 [ 9h 0m ] +INFO: Idle: 9006, Running: 4, Completed: 7010 [ 9h 1m ] +INFO: Idle: 9005, Running: 4, Completed: 7011 [ 9h 1m ] +INFO: Idle: 9004, Running: 4, Completed: 7012 [ 9h 1m ] +INFO: Idle: 9002, Running: 4, Completed: 7014 [ 9h 1m ] +INFO: Idle: 9001, Running: 4, Completed: 7015 [ 9h 1m ] +INFO: Idle: 8999, Running: 4, Completed: 7017 [ 9h 1m ] +INFO: Idle: 8998, Running: 4, Completed: 7018 [ 9h 1m ] +INFO: Idle: 8997, Running: 4, Completed: 7019 [ 9h 1m ] +INFO: Idle: 8996, Running: 4, Completed: 7020 [ 9h 1m ] +INFO: Idle: 8995, Running: 4, Completed: 7021 [ 9h 1m ] +INFO: Idle: 8994, Running: 4, Completed: 7022 [ 9h 2m ] +INFO: Idle: 8993, Running: 4, Completed: 7023 [ 9h 2m ] +INFO: Idle: 8992, Running: 4, Completed: 7024 [ 9h 2m ] +INFO: Idle: 8991, Running: 4, Completed: 7025 [ 9h 2m ] +INFO: Idle: 8990, Running: 4, Completed: 7026 [ 9h 2m ] +INFO: Idle: 8989, Running: 4, Completed: 7027 [ 9h 2m ] +INFO: Idle: 8988, Running: 4, Completed: 7028 [ 9h 2m ] +INFO: Idle: 8985, Running: 4, Completed: 7031 [ 9h 2m ] +INFO: Idle: 8983, Running: 4, Completed: 7033 [ 9h 2m ] +INFO: Idle: 8981, Running: 4, Completed: 7035 [ 9h 3m ] +INFO: Idle: 8978, Running: 4, Completed: 7038 [ 9h 3m ] +INFO: Idle: 8977, Running: 4, Completed: 7039 [ 9h 3m ] +INFO: Idle: 8974, Running: 4, Completed: 7042 [ 9h 3m ] +INFO: Idle: 8973, Running: 4, Completed: 7043 [ 9h 3m ] +INFO: Idle: 8972, Running: 4, Completed: 7044 [ 9h 3m ] +INFO: Idle: 8969, Running: 4, Completed: 7047 [ 9h 4m ] +INFO: Idle: 8967, Running: 4, Completed: 7049 [ 9h 4m ] +INFO: Idle: 8966, Running: 4, Completed: 7050 [ 9h 4m ] +INFO: Idle: 8965, Running: 4, Completed: 7051 [ 9h 4m ] +INFO: Idle: 8962, Running: 4, Completed: 7054 [ 9h 4m ] +INFO: Idle: 8961, Running: 4, Completed: 7055 [ 9h 4m ] +INFO: Idle: 8958, Running: 4, Completed: 7058 [ 9h 5m ] +INFO: Idle: 8957, Running: 4, Completed: 7059 [ 9h 5m ] +INFO: Idle: 8954, Running: 4, Completed: 7062 [ 9h 5m ] +INFO: Idle: 8953, Running: 4, Completed: 7063 [ 9h 5m ] +INFO: Idle: 8951, Running: 4, Completed: 7065 [ 9h 5m ] +INFO: Idle: 8949, Running: 4, Completed: 7067 [ 9h 5m ] +INFO: Idle: 8947, Running: 4, Completed: 7069 [ 9h 5m ] +INFO: Idle: 8945, Running: 4, Completed: 7071 [ 9h 6m ] +INFO: Idle: 8944, Running: 4, Completed: 7072 [ 9h 6m ] +INFO: Idle: 8941, Running: 4, Completed: 7075 [ 9h 6m ] +INFO: Idle: 8939, Running: 4, Completed: 7077 [ 9h 6m ] +INFO: Idle: 8937, Running: 4, Completed: 7079 [ 9h 6m ] +INFO: Idle: 8935, Running: 4, Completed: 7081 [ 9h 6m ] +INFO: Idle: 8933, Running: 4, Completed: 7083 [ 9h 7m ] +INFO: Idle: 8931, Running: 4, Completed: 7085 [ 9h 7m ] +INFO: Idle: 8930, Running: 4, Completed: 7086 [ 9h 7m ] +INFO: Idle: 8929, Running: 4, Completed: 7087 [ 9h 7m ] +INFO: Idle: 8927, Running: 4, Completed: 7089 [ 9h 7m ] +INFO: Idle: 8925, Running: 4, Completed: 7091 [ 9h 7m ] +INFO: Idle: 8923, Running: 4, Completed: 7093 [ 9h 8m ] +INFO: Idle: 8921, Running: 4, Completed: 7095 [ 9h 8m ] +INFO: Idle: 8920, Running: 4, Completed: 7096 [ 9h 8m ] +INFO: Idle: 8919, Running: 4, Completed: 7097 [ 9h 8m ] +INFO: Idle: 8918, Running: 4, Completed: 7098 [ 9h 8m ] +INFO: Idle: 8917, Running: 4, Completed: 7099 [ 9h 8m ] +INFO: Idle: 8916, Running: 4, Completed: 7100 [ 9h 8m ] +INFO: Idle: 8915, Running: 4, Completed: 7101 [ 9h 8m ] +INFO: Idle: 8914, Running: 4, Completed: 7102 [ 9h 9m ] +INFO: Idle: 8912, Running: 4, Completed: 7104 [ 9h 9m ] +INFO: Idle: 8911, Running: 4, Completed: 7105 [ 9h 9m ] +INFO: Idle: 8910, Running: 4, Completed: 7106 [ 9h 9m ] +INFO: Idle: 8907, Running: 4, Completed: 7109 [ 9h 9m ] +INFO: Idle: 8906, Running: 4, Completed: 7110 [ 9h 9m ] +INFO: Idle: 8905, Running: 4, Completed: 7111 [ 9h 9m ] +INFO: Idle: 8904, Running: 4, Completed: 7112 [ 9h 9m ] +INFO: Idle: 8903, Running: 4, Completed: 7113 [ 9h 9m ] +INFO: Idle: 8902, Running: 4, Completed: 7114 [ 9h 10m ] +INFO: Idle: 8901, Running: 4, Completed: 7115 [ 9h 10m ] +INFO: Idle: 8900, Running: 4, Completed: 7116 [ 9h 10m ] +INFO: Idle: 8899, Running: 4, Completed: 7117 [ 9h 10m ] +INFO: Idle: 8897, Running: 4, Completed: 7119 [ 9h 10m ] +INFO: Idle: 8896, Running: 4, Completed: 7120 [ 9h 10m ] +INFO: Idle: 8895, Running: 4, Completed: 7121 [ 9h 10m ] +INFO: Idle: 8894, Running: 4, Completed: 7122 [ 9h 10m ] +INFO: Idle: 8893, Running: 4, Completed: 7123 [ 9h 11m ] +INFO: Idle: 8892, Running: 4, Completed: 7124 [ 9h 11m ] +INFO: Idle: 8890, Running: 4, Completed: 7126 [ 9h 11m ] +INFO: Idle: 8888, Running: 4, Completed: 7128 [ 9h 11m ] +INFO: Idle: 8886, Running: 4, Completed: 7130 [ 9h 11m ] +INFO: Idle: 8885, Running: 4, Completed: 7131 [ 9h 11m ] +INFO: Idle: 8882, Running: 4, Completed: 7134 [ 9h 11m ] +INFO: Idle: 8881, Running: 4, Completed: 7135 [ 9h 12m ] +INFO: Idle: 8879, Running: 4, Completed: 7137 [ 9h 12m ] +INFO: Idle: 8878, Running: 4, Completed: 7138 [ 9h 12m ] +INFO: Idle: 8877, Running: 4, Completed: 7139 [ 9h 12m ] +INFO: Idle: 8876, Running: 4, Completed: 7140 [ 9h 12m ] +INFO: Idle: 8874, Running: 4, Completed: 7142 [ 9h 12m ] +INFO: Idle: 8873, Running: 4, Completed: 7143 [ 9h 12m ] +INFO: Idle: 8871, Running: 4, Completed: 7145 [ 9h 13m ] +INFO: Idle: 8870, Running: 4, Completed: 7146 [ 9h 13m ] +INFO: Idle: 8869, Running: 4, Completed: 7147 [ 9h 13m ] +INFO: Idle: 8868, Running: 4, Completed: 7148 [ 9h 13m ] +INFO: Idle: 8867, Running: 4, Completed: 7149 [ 9h 13m ] +INFO: Idle: 8866, Running: 4, Completed: 7150 [ 9h 13m ] +INFO: Idle: 8865, Running: 4, Completed: 7151 [ 9h 13m ] +INFO: Idle: 8864, Running: 4, Completed: 7152 [ 9h 13m ] +INFO: Idle: 8863, Running: 4, Completed: 7153 [ 9h 13m ] +INFO: Idle: 8861, Running: 4, Completed: 7155 [ 9h 14m ] +INFO: Idle: 8860, Running: 4, Completed: 7156 [ 9h 14m ] +INFO: Idle: 8857, Running: 4, Completed: 7159 [ 9h 14m ] +INFO: Idle: 8855, Running: 4, Completed: 7161 [ 9h 14m ] +INFO: Idle: 8853, Running: 4, Completed: 7163 [ 9h 14m ] +INFO: Idle: 8851, Running: 4, Completed: 7165 [ 9h 14m ] +INFO: Idle: 8849, Running: 4, Completed: 7167 [ 9h 15m ] +INFO: Idle: 8846, Running: 4, Completed: 7170 [ 9h 15m ] +INFO: Idle: 8845, Running: 4, Completed: 7171 [ 9h 15m ] +INFO: Idle: 8844, Running: 4, Completed: 7172 [ 9h 15m ] +INFO: Idle: 8841, Running: 4, Completed: 7175 [ 9h 15m ] +INFO: Idle: 8839, Running: 4, Completed: 7177 [ 9h 15m ] +INFO: Idle: 8837, Running: 4, Completed: 7179 [ 9h 16m ] +INFO: Idle: 8834, Running: 4, Completed: 7182 [ 9h 16m ] +INFO: Idle: 8833, Running: 4, Completed: 7183 [ 9h 16m ] +INFO: Idle: 8830, Running: 4, Completed: 7186 [ 9h 16m ] +INFO: Idle: 8829, Running: 4, Completed: 7187 [ 9h 16m ] +INFO: Idle: 8828, Running: 4, Completed: 7188 [ 9h 16m ] +INFO: Idle: 8826, Running: 4, Completed: 7190 [ 9h 16m ] +INFO: Idle: 8825, Running: 4, Completed: 7191 [ 9h 16m ] +INFO: Idle: 8823, Running: 4, Completed: 7193 [ 9h 17m ] +INFO: Idle: 8821, Running: 4, Completed: 7195 [ 9h 17m ] +INFO: Idle: 8820, Running: 4, Completed: 7196 [ 9h 17m ] +INFO: Idle: 8819, Running: 4, Completed: 7197 [ 9h 17m ] +INFO: Idle: 8818, Running: 4, Completed: 7198 [ 9h 17m ] +INFO: Idle: 8817, Running: 4, Completed: 7199 [ 9h 17m ] +INFO: Idle: 8816, Running: 4, Completed: 7200 [ 9h 17m ] +INFO: Idle: 8813, Running: 4, Completed: 7203 [ 9h 17m ] +INFO: Idle: 8811, Running: 4, Completed: 7205 [ 9h 18m ] +INFO: Idle: 8809, Running: 4, Completed: 7207 [ 9h 18m ] +INFO: Idle: 8807, Running: 4, Completed: 7209 [ 9h 18m ] +INFO: Idle: 8805, Running: 4, Completed: 7211 [ 9h 18m ] +INFO: Idle: 8803, Running: 4, Completed: 7213 [ 9h 18m ] +INFO: Idle: 8802, Running: 4, Completed: 7214 [ 9h 18m ] +INFO: Idle: 8801, Running: 4, Completed: 7215 [ 9h 19m ] +INFO: Idle: 8799, Running: 4, Completed: 7217 [ 9h 19m ] +INFO: Idle: 8797, Running: 4, Completed: 7219 [ 9h 19m ] +INFO: Idle: 8795, Running: 4, Completed: 7221 [ 9h 19m ] +INFO: Idle: 8793, Running: 4, Completed: 7223 [ 9h 19m ] +INFO: Idle: 8792, Running: 4, Completed: 7224 [ 9h 19m ] +INFO: Idle: 8791, Running: 4, Completed: 7225 [ 9h 19m ] +INFO: Idle: 8790, Running: 4, Completed: 7226 [ 9h 20m ] +INFO: Idle: 8789, Running: 4, Completed: 7227 [ 9h 20m ] +INFO: Idle: 8788, Running: 4, Completed: 7228 [ 9h 20m ] +INFO: Idle: 8787, Running: 4, Completed: 7229 [ 9h 20m ] +INFO: Idle: 8786, Running: 4, Completed: 7230 [ 9h 20m ] +INFO: Idle: 8785, Running: 4, Completed: 7231 [ 9h 20m ] +INFO: Idle: 8784, Running: 4, Completed: 7232 [ 9h 20m ] +INFO: Idle: 8783, Running: 4, Completed: 7233 [ 9h 20m ] +INFO: Idle: 8782, Running: 4, Completed: 7234 [ 9h 20m ] +INFO: Idle: 8780, Running: 4, Completed: 7236 [ 9h 20m ] +INFO: Idle: 8779, Running: 4, Completed: 7237 [ 9h 20m ] +INFO: Idle: 8778, Running: 4, Completed: 7238 [ 9h 21m ] +INFO: Idle: 8777, Running: 4, Completed: 7239 [ 9h 21m ] +INFO: Idle: 8776, Running: 4, Completed: 7240 [ 9h 21m ] +INFO: Idle: 8775, Running: 4, Completed: 7241 [ 9h 21m ] +INFO: Idle: 8774, Running: 4, Completed: 7242 [ 9h 21m ] +INFO: Idle: 8773, Running: 4, Completed: 7243 [ 9h 21m ] +INFO: Idle: 8772, Running: 4, Completed: 7244 [ 9h 21m ] +INFO: Idle: 8771, Running: 4, Completed: 7245 [ 9h 21m ] +INFO: Idle: 8769, Running: 4, Completed: 7247 [ 9h 21m ] +INFO: Idle: 8768, Running: 4, Completed: 7248 [ 9h 22m ] +INFO: Idle: 8767, Running: 4, Completed: 7249 [ 9h 22m ] +INFO: Idle: 8766, Running: 4, Completed: 7250 [ 9h 22m ] +INFO: Idle: 8765, Running: 4, Completed: 7251 [ 9h 22m ] +INFO: Idle: 8764, Running: 4, Completed: 7252 [ 9h 22m ] +INFO: Idle: 8762, Running: 4, Completed: 7254 [ 9h 22m ] +INFO: Idle: 8761, Running: 4, Completed: 7255 [ 9h 22m ] +INFO: Idle: 8759, Running: 4, Completed: 7257 [ 9h 22m ] +INFO: Idle: 8758, Running: 4, Completed: 7258 [ 9h 23m ] +INFO: Idle: 8757, Running: 4, Completed: 7259 [ 9h 23m ] +INFO: Idle: 8754, Running: 4, Completed: 7262 [ 9h 23m ] +INFO: Idle: 8753, Running: 4, Completed: 7263 [ 9h 23m ] +INFO: Idle: 8751, Running: 4, Completed: 7265 [ 9h 23m ] +INFO: Idle: 8750, Running: 4, Completed: 7266 [ 9h 23m ] +INFO: Idle: 8749, Running: 4, Completed: 7267 [ 9h 24m ] +INFO: Idle: 8748, Running: 4, Completed: 7268 [ 9h 24m ] +INFO: Idle: 8746, Running: 4, Completed: 7270 [ 9h 24m ] +INFO: Idle: 8745, Running: 4, Completed: 7271 [ 9h 24m ] +INFO: Idle: 8743, Running: 4, Completed: 7273 [ 9h 24m ] +INFO: Idle: 8742, Running: 4, Completed: 7274 [ 9h 24m ] +INFO: Idle: 8741, Running: 4, Completed: 7275 [ 9h 24m ] +INFO: Idle: 8740, Running: 4, Completed: 7276 [ 9h 24m ] +INFO: Idle: 8739, Running: 4, Completed: 7277 [ 9h 24m ] +INFO: Idle: 8738, Running: 4, Completed: 7278 [ 9h 24m ] +INFO: Idle: 8737, Running: 4, Completed: 7279 [ 9h 25m ] +INFO: Idle: 8735, Running: 4, Completed: 7281 [ 9h 25m ] +INFO: Idle: 8733, Running: 4, Completed: 7283 [ 9h 25m ] +INFO: Idle: 8732, Running: 4, Completed: 7284 [ 9h 25m ] +INFO: Idle: 8730, Running: 4, Completed: 7286 [ 9h 25m ] +INFO: Idle: 8729, Running: 4, Completed: 7287 [ 9h 25m ] +INFO: Idle: 8727, Running: 4, Completed: 7289 [ 9h 25m ] +INFO: Idle: 8725, Running: 4, Completed: 7291 [ 9h 26m ] +INFO: Idle: 8723, Running: 4, Completed: 7293 [ 9h 26m ] +INFO: Idle: 8721, Running: 4, Completed: 7295 [ 9h 26m ] +INFO: Idle: 8718, Running: 4, Completed: 7298 [ 9h 26m ] +INFO: Idle: 8717, Running: 4, Completed: 7299 [ 9h 26m ] +INFO: Idle: 8716, Running: 4, Completed: 7300 [ 9h 26m ] +INFO: Idle: 8713, Running: 4, Completed: 7303 [ 9h 27m ] +INFO: Idle: 8711, Running: 4, Completed: 7305 [ 9h 27m ] +INFO: Idle: 8709, Running: 4, Completed: 7307 [ 9h 27m ] +INFO: Idle: 8706, Running: 4, Completed: 7310 [ 9h 27m ] +INFO: Idle: 8705, Running: 4, Completed: 7311 [ 9h 27m ] +INFO: Idle: 8702, Running: 4, Completed: 7314 [ 9h 27m ] +INFO: Idle: 8701, Running: 4, Completed: 7315 [ 9h 28m ] +INFO: Idle: 8700, Running: 4, Completed: 7316 [ 9h 28m ] +INFO: Idle: 8698, Running: 4, Completed: 7318 [ 9h 28m ] +INFO: Idle: 8697, Running: 4, Completed: 7319 [ 9h 28m ] +INFO: Idle: 8695, Running: 4, Completed: 7321 [ 9h 28m ] +INFO: Idle: 8693, Running: 4, Completed: 7323 [ 9h 28m ] +INFO: Idle: 8692, Running: 4, Completed: 7324 [ 9h 28m ] +INFO: Idle: 8691, Running: 4, Completed: 7325 [ 9h 28m ] +INFO: Idle: 8689, Running: 4, Completed: 7327 [ 9h 29m ] +INFO: Idle: 8688, Running: 4, Completed: 7328 [ 9h 29m ] +INFO: Idle: 8687, Running: 4, Completed: 7329 [ 9h 29m ] +INFO: Idle: 8685, Running: 4, Completed: 7331 [ 9h 29m ] +INFO: Idle: 8684, Running: 4, Completed: 7332 [ 9h 29m ] +INFO: Idle: 8683, Running: 4, Completed: 7333 [ 9h 29m ] +INFO: Idle: 8681, Running: 4, Completed: 7335 [ 9h 29m ] +INFO: Idle: 8679, Running: 4, Completed: 7337 [ 9h 29m ] +INFO: Idle: 8677, Running: 4, Completed: 7339 [ 9h 30m ] +INFO: Idle: 8675, Running: 4, Completed: 7341 [ 9h 30m ] +INFO: Idle: 8673, Running: 4, Completed: 7343 [ 9h 30m ] +INFO: Idle: 8671, Running: 4, Completed: 7345 [ 9h 30m ] +INFO: Idle: 8669, Running: 4, Completed: 7347 [ 9h 30m ] +INFO: Idle: 8667, Running: 4, Completed: 7349 [ 9h 31m ] +INFO: Idle: 8665, Running: 4, Completed: 7351 [ 9h 31m ] +INFO: Idle: 8663, Running: 4, Completed: 7353 [ 9h 31m ] +INFO: Idle: 8660, Running: 4, Completed: 7356 [ 9h 31m ] +INFO: Idle: 8659, Running: 4, Completed: 7357 [ 9h 31m ] +INFO: Idle: 8657, Running: 4, Completed: 7359 [ 9h 32m ] +INFO: Idle: 8656, Running: 4, Completed: 7360 [ 9h 32m ] +INFO: Idle: 8655, Running: 4, Completed: 7361 [ 9h 32m ] +INFO: Idle: 8653, Running: 4, Completed: 7363 [ 9h 32m ] +INFO: Idle: 8651, Running: 4, Completed: 7365 [ 9h 32m ] +INFO: Idle: 8649, Running: 4, Completed: 7367 [ 9h 33m ] +INFO: Idle: 8648, Running: 4, Completed: 7368 [ 9h 33m ] +INFO: Idle: 8646, Running: 4, Completed: 7370 [ 9h 33m ] +INFO: Idle: 8645, Running: 4, Completed: 7371 [ 9h 33m ] +INFO: Idle: 8644, Running: 4, Completed: 7372 [ 9h 33m ] +INFO: Idle: 8643, Running: 4, Completed: 7373 [ 9h 33m ] +INFO: Idle: 8641, Running: 4, Completed: 7375 [ 9h 33m ] +INFO: Idle: 8639, Running: 4, Completed: 7377 [ 9h 33m ] +INFO: Idle: 8637, Running: 4, Completed: 7379 [ 9h 34m ] +INFO: Idle: 8635, Running: 4, Completed: 7381 [ 9h 34m ] +INFO: Idle: 8633, Running: 4, Completed: 7383 [ 9h 34m ] +INFO: Idle: 8631, Running: 4, Completed: 7385 [ 9h 34m ] +INFO: Idle: 8629, Running: 4, Completed: 7387 [ 9h 34m ] +INFO: Idle: 8627, Running: 4, Completed: 7389 [ 9h 35m ] +INFO: Idle: 8626, Running: 4, Completed: 7390 [ 9h 35m ] +INFO: Idle: 8625, Running: 4, Completed: 7391 [ 9h 35m ] +INFO: Idle: 8623, Running: 4, Completed: 7393 [ 9h 35m ] +INFO: Idle: 8621, Running: 4, Completed: 7395 [ 9h 35m ] +INFO: Idle: 8620, Running: 4, Completed: 7396 [ 9h 35m ] +INFO: Idle: 8619, Running: 4, Completed: 7397 [ 9h 35m ] +INFO: Idle: 8618, Running: 4, Completed: 7398 [ 9h 36m ] +INFO: Idle: 8616, Running: 4, Completed: 7400 [ 9h 36m ] +INFO: Idle: 8615, Running: 4, Completed: 7401 [ 9h 36m ] +INFO: Idle: 8612, Running: 4, Completed: 7404 [ 9h 36m ] +INFO: Idle: 8611, Running: 4, Completed: 7405 [ 9h 36m ] +INFO: Idle: 8610, Running: 4, Completed: 7406 [ 9h 36m ] +INFO: Idle: 8607, Running: 4, Completed: 7409 [ 9h 37m ] +INFO: Idle: 8606, Running: 4, Completed: 7410 [ 9h 37m ] +INFO: Idle: 8605, Running: 4, Completed: 7411 [ 9h 37m ] +INFO: Idle: 8604, Running: 4, Completed: 7412 [ 9h 37m ] +INFO: Idle: 8602, Running: 4, Completed: 7414 [ 9h 37m ] +INFO: Idle: 8599, Running: 4, Completed: 7417 [ 9h 37m ] +INFO: Idle: 8598, Running: 4, Completed: 7418 [ 9h 37m ] +INFO: Idle: 8597, Running: 4, Completed: 7419 [ 9h 38m ] +INFO: Idle: 8595, Running: 4, Completed: 7421 [ 9h 38m ] +INFO: Idle: 8593, Running: 4, Completed: 7423 [ 9h 38m ] +INFO: Idle: 8591, Running: 4, Completed: 7425 [ 9h 38m ] +INFO: Idle: 8589, Running: 4, Completed: 7427 [ 9h 38m ] +INFO: Idle: 8587, Running: 4, Completed: 7429 [ 9h 39m ] +INFO: Idle: 8586, Running: 4, Completed: 7430 [ 9h 39m ] +INFO: Idle: 8585, Running: 4, Completed: 7431 [ 9h 39m ] +INFO: Idle: 8584, Running: 4, Completed: 7432 [ 9h 39m ] +INFO: Idle: 8583, Running: 4, Completed: 7433 [ 9h 39m ] +INFO: Idle: 8582, Running: 4, Completed: 7434 [ 9h 39m ] +INFO: Idle: 8581, Running: 4, Completed: 7435 [ 9h 39m ] +INFO: Idle: 8580, Running: 4, Completed: 7436 [ 9h 39m ] +INFO: Idle: 8578, Running: 4, Completed: 7438 [ 9h 39m ] +INFO: Idle: 8577, Running: 4, Completed: 7439 [ 9h 39m ] +INFO: Idle: 8576, Running: 4, Completed: 7440 [ 9h 40m ] +INFO: Idle: 8575, Running: 4, Completed: 7441 [ 9h 40m ] +INFO: Idle: 8573, Running: 4, Completed: 7443 [ 9h 40m ] +INFO: Idle: 8572, Running: 4, Completed: 7444 [ 9h 40m ] +INFO: Idle: 8571, Running: 4, Completed: 7445 [ 9h 40m ] +INFO: Idle: 8570, Running: 3, Completed: 7447 [ 9h 40m ] +INFO: Idle: 8566, Running: 4, Completed: 7450 [ 9h 40m ] +INFO: Idle: 8565, Running: 4, Completed: 7451 [ 9h 40m ] +INFO: Idle: 8563, Running: 4, Completed: 7453 [ 9h 41m ] +INFO: Idle: 8561, Running: 4, Completed: 7455 [ 9h 41m ] +INFO: Idle: 8559, Running: 4, Completed: 7457 [ 9h 41m ] +INFO: Idle: 8557, Running: 4, Completed: 7459 [ 9h 41m ] +INFO: Idle: 8555, Running: 4, Completed: 7461 [ 9h 41m ] +INFO: Idle: 8553, Running: 4, Completed: 7463 [ 9h 41m ] +INFO: Idle: 8551, Running: 4, Completed: 7465 [ 9h 42m ] +INFO: Idle: 8549, Running: 4, Completed: 7467 [ 9h 42m ] +INFO: Idle: 8547, Running: 4, Completed: 7469 [ 9h 42m ] +INFO: Idle: 8545, Running: 4, Completed: 7471 [ 9h 42m ] +INFO: Idle: 8543, Running: 4, Completed: 7473 [ 9h 42m ] +INFO: Idle: 8541, Running: 4, Completed: 7475 [ 9h 43m ] +INFO: Idle: 8539, Running: 4, Completed: 7477 [ 9h 43m ] +INFO: Idle: 8537, Running: 4, Completed: 7479 [ 9h 43m ] +INFO: Idle: 8535, Running: 4, Completed: 7481 [ 9h 43m ] +INFO: Idle: 8532, Running: 4, Completed: 7484 [ 9h 43m ] +INFO: Idle: 8531, Running: 4, Completed: 7485 [ 9h 44m ] +INFO: Idle: 8529, Running: 4, Completed: 7487 [ 9h 44m ] +INFO: Idle: 8527, Running: 4, Completed: 7489 [ 9h 44m ] +INFO: Idle: 8525, Running: 4, Completed: 7491 [ 9h 44m ] +INFO: Idle: 8523, Running: 4, Completed: 7493 [ 9h 44m ] +INFO: Idle: 8521, Running: 4, Completed: 7495 [ 9h 45m ] +INFO: Idle: 8519, Running: 4, Completed: 7497 [ 9h 45m ] +INFO: Idle: 8517, Running: 4, Completed: 7499 [ 9h 45m ] +INFO: Idle: 8515, Running: 4, Completed: 7501 [ 9h 45m ] +INFO: Idle: 8514, Running: 4, Completed: 7502 [ 9h 45m ] +INFO: Idle: 8511, Running: 4, Completed: 7505 [ 9h 46m ] +INFO: Idle: 8510, Running: 4, Completed: 7506 [ 9h 46m ] +INFO: Idle: 8509, Running: 4, Completed: 7507 [ 9h 46m ] +INFO: Idle: 8508, Running: 4, Completed: 7508 [ 9h 46m ] +INFO: Idle: 8506, Running: 4, Completed: 7510 [ 9h 46m ] +INFO: Idle: 8504, Running: 4, Completed: 7512 [ 9h 46m ] +INFO: Idle: 8502, Running: 4, Completed: 7514 [ 9h 47m ] +INFO: Idle: 8501, Running: 4, Completed: 7515 [ 9h 47m ] +INFO: Idle: 8498, Running: 4, Completed: 7518 [ 9h 47m ] +INFO: Idle: 8497, Running: 4, Completed: 7519 [ 9h 47m ] +INFO: Idle: 8495, Running: 4, Completed: 7521 [ 9h 47m ] +INFO: Idle: 8494, Running: 4, Completed: 7522 [ 9h 47m ] +INFO: Idle: 8493, Running: 4, Completed: 7523 [ 9h 47m ] +INFO: Idle: 8492, Running: 4, Completed: 7524 [ 9h 48m ] +INFO: Idle: 8491, Running: 4, Completed: 7525 [ 9h 48m ] +INFO: Idle: 8490, Running: 4, Completed: 7526 [ 9h 48m ] +INFO: Idle: 8489, Running: 4, Completed: 7527 [ 9h 48m ] +INFO: Idle: 8488, Running: 4, Completed: 7528 [ 9h 48m ] +INFO: Idle: 8487, Running: 4, Completed: 7529 [ 9h 48m ] +INFO: Idle: 8485, Running: 4, Completed: 7531 [ 9h 48m ] +INFO: Idle: 8483, Running: 4, Completed: 7533 [ 9h 49m ] +INFO: Idle: 8481, Running: 4, Completed: 7535 [ 9h 49m ] +INFO: Idle: 8480, Running: 4, Completed: 7536 [ 9h 49m ] +INFO: Idle: 8479, Running: 4, Completed: 7537 [ 9h 49m ] +INFO: Idle: 8478, Running: 4, Completed: 7538 [ 9h 49m ] +INFO: Idle: 8477, Running: 4, Completed: 7539 [ 9h 49m ] +INFO: Idle: 8476, Running: 4, Completed: 7540 [ 9h 49m ] +INFO: Idle: 8474, Running: 4, Completed: 7542 [ 9h 49m ] +INFO: Idle: 8473, Running: 4, Completed: 7543 [ 9h 50m ] +INFO: Idle: 8471, Running: 4, Completed: 7545 [ 9h 50m ] +INFO: Idle: 8469, Running: 4, Completed: 7547 [ 9h 50m ] +INFO: Idle: 8468, Running: 4, Completed: 7548 [ 9h 50m ] +INFO: Idle: 8466, Running: 4, Completed: 7550 [ 9h 50m ] +INFO: Idle: 8465, Running: 4, Completed: 7551 [ 9h 50m ] +INFO: Idle: 8464, Running: 4, Completed: 7552 [ 9h 50m ] +INFO: Idle: 8461, Running: 4, Completed: 7555 [ 9h 51m ] +INFO: Idle: 8460, Running: 4, Completed: 7556 [ 9h 51m ] +INFO: Idle: 8458, Running: 4, Completed: 7558 [ 9h 51m ] +INFO: Idle: 8457, Running: 4, Completed: 7559 [ 9h 51m ] +INFO: Idle: 8456, Running: 4, Completed: 7560 [ 9h 51m ] +INFO: Idle: 8455, Running: 4, Completed: 7561 [ 9h 51m ] +INFO: Idle: 8453, Running: 4, Completed: 7563 [ 9h 51m ] +INFO: Idle: 8451, Running: 4, Completed: 7565 [ 9h 51m ] +INFO: Idle: 8450, Running: 4, Completed: 7566 [ 9h 51m ] +INFO: Idle: 8449, Running: 4, Completed: 7567 [ 9h 52m ] +INFO: Idle: 8447, Running: 4, Completed: 7569 [ 9h 52m ] +INFO: Idle: 8446, Running: 4, Completed: 7570 [ 9h 52m ] +INFO: Idle: 8445, Running: 4, Completed: 7571 [ 9h 52m ] +INFO: Idle: 8443, Running: 4, Completed: 7573 [ 9h 52m ] +INFO: Idle: 8442, Running: 4, Completed: 7574 [ 9h 52m ] +INFO: Idle: 8441, Running: 4, Completed: 7575 [ 9h 52m ] +INFO: Idle: 8440, Running: 4, Completed: 7576 [ 9h 53m ] +INFO: Idle: 8439, Running: 4, Completed: 7577 [ 9h 53m ] +INFO: Idle: 8438, Running: 4, Completed: 7578 [ 9h 53m ] +INFO: Idle: 8436, Running: 4, Completed: 7580 [ 9h 53m ] +INFO: Idle: 8435, Running: 4, Completed: 7581 [ 9h 53m ] +INFO: Idle: 8434, Running: 4, Completed: 7582 [ 9h 53m ] +INFO: Idle: 8432, Running: 4, Completed: 7584 [ 9h 53m ] +INFO: Idle: 8431, Running: 4, Completed: 7585 [ 9h 53m ] +INFO: Idle: 8429, Running: 4, Completed: 7587 [ 9h 53m ] +INFO: Idle: 8428, Running: 4, Completed: 7588 [ 9h 54m ] +INFO: Idle: 8427, Running: 4, Completed: 7589 [ 9h 54m ] +INFO: Idle: 8426, Running: 4, Completed: 7590 [ 9h 54m ] +INFO: Idle: 8425, Running: 4, Completed: 7591 [ 9h 54m ] +INFO: Idle: 8424, Running: 4, Completed: 7592 [ 9h 54m ] +INFO: Idle: 8422, Running: 4, Completed: 7594 [ 9h 54m ] +INFO: Idle: 8421, Running: 4, Completed: 7595 [ 9h 54m ] +INFO: Idle: 8420, Running: 4, Completed: 7596 [ 9h 55m ] +INFO: Idle: 8418, Running: 4, Completed: 7598 [ 9h 55m ] +INFO: Idle: 8416, Running: 4, Completed: 7600 [ 9h 55m ] +INFO: Idle: 8415, Running: 4, Completed: 7601 [ 9h 55m ] +INFO: Idle: 8414, Running: 4, Completed: 7602 [ 9h 55m ] +INFO: Idle: 8413, Running: 4, Completed: 7603 [ 9h 55m ] +INFO: Idle: 8412, Running: 4, Completed: 7604 [ 9h 55m ] +INFO: Idle: 8411, Running: 4, Completed: 7605 [ 9h 55m ] +INFO: Idle: 8410, Running: 4, Completed: 7606 [ 9h 55m ] +INFO: Idle: 8408, Running: 4, Completed: 7608 [ 9h 56m ] +INFO: Idle: 8406, Running: 4, Completed: 7610 [ 9h 56m ] +INFO: Idle: 8403, Running: 4, Completed: 7613 [ 9h 56m ] +INFO: Idle: 8402, Running: 4, Completed: 7614 [ 9h 56m ] +INFO: Idle: 8399, Running: 4, Completed: 7617 [ 9h 56m ] +INFO: Idle: 8398, Running: 4, Completed: 7618 [ 9h 57m ] +INFO: Idle: 8397, Running: 4, Completed: 7619 [ 9h 57m ] +INFO: Idle: 8394, Running: 4, Completed: 7622 [ 9h 57m ] +INFO: Idle: 8393, Running: 4, Completed: 7623 [ 9h 57m ] +INFO: Idle: 8392, Running: 4, Completed: 7624 [ 9h 57m ] +INFO: Idle: 8390, Running: 4, Completed: 7626 [ 9h 57m ] +INFO: Idle: 8388, Running: 4, Completed: 7628 [ 9h 57m ] +INFO: Idle: 8386, Running: 4, Completed: 7630 [ 9h 57m ] +INFO: Idle: 8384, Running: 4, Completed: 7632 [ 9h 58m ] +INFO: Idle: 8383, Running: 4, Completed: 7633 [ 9h 58m ] +INFO: Idle: 8382, Running: 4, Completed: 7634 [ 9h 58m ] +INFO: Idle: 8380, Running: 4, Completed: 7636 [ 9h 58m ] +INFO: Idle: 8379, Running: 4, Completed: 7637 [ 9h 58m ] +INFO: Idle: 8376, Running: 4, Completed: 7640 [ 9h 58m ] +INFO: Idle: 8375, Running: 4, Completed: 7641 [ 9h 58m ] +INFO: Idle: 8373, Running: 4, Completed: 7643 [ 9h 59m ] +INFO: Idle: 8372, Running: 4, Completed: 7644 [ 9h 59m ] +INFO: Idle: 8371, Running: 4, Completed: 7645 [ 9h 59m ] +INFO: Idle: 8369, Running: 4, Completed: 7647 [ 9h 59m ] +INFO: Idle: 8367, Running: 4, Completed: 7649 [ 9h 59m ] +INFO: Idle: 8364, Running: 4, Completed: 7652 [ 9h 59m ] +INFO: Idle: 8363, Running: 4, Completed: 7653 [ 10h 0m ] +INFO: Idle: 8360, Running: 4, Completed: 7656 [ 10h 0m ] +INFO: Idle: 8359, Running: 4, Completed: 7657 [ 10h 0m ] +INFO: Idle: 8355, Running: 4, Completed: 7661 [ 10h 0m ] +INFO: Idle: 8353, Running: 4, Completed: 7663 [ 10h 0m ] +INFO: Idle: 8351, Running: 4, Completed: 7665 [ 10h 0m ] +INFO: Idle: 8349, Running: 4, Completed: 7667 [ 10h 1m ] +INFO: Idle: 8347, Running: 4, Completed: 7669 [ 10h 1m ] +INFO: Idle: 8345, Running: 4, Completed: 7671 [ 10h 1m ] +INFO: Idle: 8343, Running: 4, Completed: 7673 [ 10h 1m ] +INFO: Idle: 8341, Running: 4, Completed: 7675 [ 10h 1m ] +INFO: Idle: 8340, Running: 4, Completed: 7676 [ 10h 1m ] +INFO: Idle: 8339, Running: 4, Completed: 7677 [ 10h 2m ] +INFO: Idle: 8337, Running: 4, Completed: 7679 [ 10h 2m ] +INFO: Idle: 8335, Running: 4, Completed: 7681 [ 10h 2m ] +INFO: Idle: 8333, Running: 4, Completed: 7683 [ 10h 2m ] +INFO: Idle: 8331, Running: 4, Completed: 7685 [ 10h 2m ] +INFO: Idle: 8329, Running: 4, Completed: 7687 [ 10h 3m ] +INFO: Idle: 8328, Running: 4, Completed: 7688 [ 10h 3m ] +INFO: Idle: 8326, Running: 4, Completed: 7690 [ 10h 3m ] +INFO: Idle: 8325, Running: 4, Completed: 7691 [ 10h 3m ] +INFO: Idle: 8323, Running: 4, Completed: 7693 [ 10h 3m ] +INFO: Idle: 8321, Running: 4, Completed: 7695 [ 10h 3m ] +INFO: Idle: 8319, Running: 4, Completed: 7697 [ 10h 4m ] +INFO: Idle: 8318, Running: 4, Completed: 7698 [ 10h 4m ] +INFO: Idle: 8317, Running: 4, Completed: 7699 [ 10h 4m ] +INFO: Idle: 8315, Running: 4, Completed: 7701 [ 10h 4m ] +INFO: Idle: 8313, Running: 4, Completed: 7703 [ 10h 4m ] +INFO: Idle: 8311, Running: 4, Completed: 7705 [ 10h 5m ] +INFO: Idle: 8310, Running: 4, Completed: 7706 [ 10h 5m ] +INFO: Idle: 8307, Running: 4, Completed: 7709 [ 10h 5m ] +INFO: Idle: 8305, Running: 4, Completed: 7711 [ 10h 5m ] +INFO: Idle: 8304, Running: 4, Completed: 7712 [ 10h 5m ] +INFO: Idle: 8303, Running: 4, Completed: 7713 [ 10h 5m ] +INFO: Idle: 8301, Running: 4, Completed: 7715 [ 10h 5m ] +INFO: Idle: 8299, Running: 4, Completed: 7717 [ 10h 6m ] +INFO: Idle: 8297, Running: 4, Completed: 7719 [ 10h 6m ] +INFO: Idle: 8295, Running: 4, Completed: 7721 [ 10h 6m ] +INFO: Idle: 8293, Running: 4, Completed: 7723 [ 10h 6m ] +INFO: Idle: 8291, Running: 4, Completed: 7725 [ 10h 6m ] +INFO: Idle: 8289, Running: 4, Completed: 7727 [ 10h 7m ] +INFO: Idle: 8288, Running: 4, Completed: 7728 [ 10h 7m ] +INFO: Idle: 8287, Running: 4, Completed: 7729 [ 10h 7m ] +INFO: Idle: 8285, Running: 4, Completed: 7731 [ 10h 7m ] +INFO: Idle: 8283, Running: 4, Completed: 7733 [ 10h 7m ] +INFO: Idle: 8282, Running: 4, Completed: 7734 [ 10h 7m ] +INFO: Idle: 8281, Running: 4, Completed: 7735 [ 10h 7m ] +INFO: Idle: 8280, Running: 4, Completed: 7736 [ 10h 8m ] +INFO: Idle: 8279, Running: 4, Completed: 7737 [ 10h 8m ] +INFO: Idle: 8278, Running: 4, Completed: 7738 [ 10h 8m ] +INFO: Idle: 8276, Running: 4, Completed: 7740 [ 10h 8m ] +INFO: Idle: 8275, Running: 4, Completed: 7741 [ 10h 8m ] +INFO: Idle: 8274, Running: 4, Completed: 7742 [ 10h 8m ] +INFO: Idle: 8273, Running: 4, Completed: 7743 [ 10h 8m ] +INFO: Idle: 8272, Running: 4, Completed: 7744 [ 10h 8m ] +INFO: Idle: 8270, Running: 4, Completed: 7746 [ 10h 9m ] +INFO: Idle: 8268, Running: 4, Completed: 7748 [ 10h 9m ] +INFO: Idle: 8267, Running: 4, Completed: 7749 [ 10h 9m ] +INFO: Idle: 8264, Running: 4, Completed: 7752 [ 10h 9m ] +INFO: Idle: 8263, Running: 4, Completed: 7753 [ 10h 9m ] +INFO: Idle: 8261, Running: 4, Completed: 7755 [ 10h 9m ] +INFO: Idle: 8260, Running: 4, Completed: 7756 [ 10h 9m ] +INFO: Idle: 8259, Running: 4, Completed: 7757 [ 10h 10m ] +INFO: Idle: 8256, Running: 4, Completed: 7760 [ 10h 10m ] +INFO: Idle: 8255, Running: 4, Completed: 7761 [ 10h 10m ] +INFO: Idle: 8252, Running: 4, Completed: 7764 [ 10h 10m ] +INFO: Idle: 8251, Running: 4, Completed: 7765 [ 10h 10m ] +INFO: Idle: 8249, Running: 4, Completed: 7767 [ 10h 11m ] +INFO: Idle: 8247, Running: 4, Completed: 7769 [ 10h 11m ] +INFO: Idle: 8246, Running: 4, Completed: 7770 [ 10h 11m ] +INFO: Idle: 8245, Running: 4, Completed: 7771 [ 10h 11m ] +INFO: Idle: 8243, Running: 4, Completed: 7773 [ 10h 11m ] +INFO: Idle: 8242, Running: 4, Completed: 7774 [ 10h 11m ] +INFO: Idle: 8241, Running: 4, Completed: 7775 [ 10h 11m ] +INFO: Idle: 8240, Running: 4, Completed: 7776 [ 10h 11m ] +INFO: Idle: 8238, Running: 4, Completed: 7778 [ 10h 12m ] +INFO: Idle: 8237, Running: 4, Completed: 7779 [ 10h 12m ] +INFO: Idle: 8235, Running: 4, Completed: 7781 [ 10h 12m ] +INFO: Idle: 8234, Running: 4, Completed: 7782 [ 10h 12m ] +INFO: Idle: 8233, Running: 4, Completed: 7783 [ 10h 12m ] +INFO: Idle: 8232, Running: 4, Completed: 7784 [ 10h 12m ] +INFO: Idle: 8231, Running: 4, Completed: 7785 [ 10h 12m ] +INFO: Idle: 8228, Running: 4, Completed: 7788 [ 10h 12m ] +INFO: Idle: 8227, Running: 4, Completed: 7789 [ 10h 13m ] +INFO: Idle: 8225, Running: 4, Completed: 7791 [ 10h 13m ] +INFO: Idle: 8224, Running: 4, Completed: 7792 [ 10h 13m ] +INFO: Idle: 8223, Running: 4, Completed: 7793 [ 10h 13m ] +INFO: Idle: 8221, Running: 4, Completed: 7795 [ 10h 13m ] +INFO: Idle: 8220, Running: 4, Completed: 7796 [ 10h 13m ] +INFO: Idle: 8219, Running: 4, Completed: 7797 [ 10h 13m ] +INFO: Idle: 8217, Running: 4, Completed: 7799 [ 10h 13m ] +INFO: Idle: 8216, Running: 4, Completed: 7800 [ 10h 13m ] +INFO: Idle: 8215, Running: 4, Completed: 7801 [ 10h 13m ] +INFO: Idle: 8213, Running: 4, Completed: 7803 [ 10h 14m ] +INFO: Idle: 8211, Running: 4, Completed: 7805 [ 10h 14m ] +INFO: Idle: 8209, Running: 4, Completed: 7807 [ 10h 14m ] +INFO: Idle: 8206, Running: 4, Completed: 7810 [ 10h 14m ] +INFO: Idle: 8205, Running: 4, Completed: 7811 [ 10h 15m ] +INFO: Idle: 8203, Running: 4, Completed: 7813 [ 10h 15m ] +INFO: Idle: 8201, Running: 4, Completed: 7815 [ 10h 15m ] +INFO: Idle: 8199, Running: 4, Completed: 7817 [ 10h 15m ] +INFO: Idle: 8197, Running: 4, Completed: 7819 [ 10h 15m ] +INFO: Idle: 8194, Running: 4, Completed: 7822 [ 10h 15m ] +INFO: Idle: 8193, Running: 4, Completed: 7823 [ 10h 16m ] +INFO: Idle: 8191, Running: 4, Completed: 7825 [ 10h 16m ] +INFO: Idle: 8189, Running: 4, Completed: 7827 [ 10h 16m ] +INFO: Idle: 8188, Running: 4, Completed: 7828 [ 10h 16m ] +INFO: Idle: 8185, Running: 4, Completed: 7831 [ 10h 16m ] +INFO: Idle: 8183, Running: 4, Completed: 7833 [ 10h 17m ] +INFO: Idle: 8181, Running: 4, Completed: 7835 [ 10h 17m ] +INFO: Idle: 8179, Running: 4, Completed: 7837 [ 10h 17m ] +INFO: Idle: 8177, Running: 4, Completed: 7839 [ 10h 17m ] +INFO: Idle: 8174, Running: 4, Completed: 7842 [ 10h 17m ] +INFO: Idle: 8173, Running: 4, Completed: 7843 [ 10h 18m ] +INFO: Idle: 8172, Running: 4, Completed: 7844 [ 10h 18m ] +INFO: Idle: 8171, Running: 4, Completed: 7845 [ 10h 18m ] +INFO: Idle: 8169, Running: 4, Completed: 7847 [ 10h 18m ] +INFO: Idle: 8168, Running: 4, Completed: 7848 [ 10h 18m ] +INFO: Idle: 8164, Running: 4, Completed: 7852 [ 10h 19m ] +INFO: Idle: 8163, Running: 4, Completed: 7853 [ 10h 19m ] +INFO: Idle: 8162, Running: 4, Completed: 7854 [ 10h 19m ] +INFO: Idle: 8160, Running: 4, Completed: 7856 [ 10h 19m ] +INFO: Idle: 8159, Running: 4, Completed: 7857 [ 10h 19m ] +INFO: Idle: 8157, Running: 4, Completed: 7859 [ 10h 19m ] +INFO: Idle: 8156, Running: 4, Completed: 7860 [ 10h 19m ] +INFO: Idle: 8154, Running: 4, Completed: 7862 [ 10h 20m ] +INFO: Idle: 8153, Running: 4, Completed: 7863 [ 10h 20m ] +INFO: Idle: 8152, Running: 4, Completed: 7864 [ 10h 20m ] +INFO: Idle: 8151, Running: 4, Completed: 7865 [ 10h 20m ] +INFO: Idle: 8150, Running: 4, Completed: 7866 [ 10h 20m ] +INFO: Idle: 8149, Running: 4, Completed: 7867 [ 10h 20m ] +INFO: Idle: 8147, Running: 4, Completed: 7869 [ 10h 20m ] +INFO: Idle: 8145, Running: 4, Completed: 7871 [ 10h 21m ] +INFO: Idle: 8143, Running: 4, Completed: 7873 [ 10h 21m ] +INFO: Idle: 8142, Running: 4, Completed: 7874 [ 10h 21m ] +INFO: Idle: 8141, Running: 4, Completed: 7875 [ 10h 21m ] +INFO: Idle: 8140, Running: 4, Completed: 7876 [ 10h 21m ] +INFO: Idle: 8139, Running: 4, Completed: 7877 [ 10h 21m ] +INFO: Idle: 8138, Running: 4, Completed: 7878 [ 10h 21m ] +INFO: Idle: 8137, Running: 4, Completed: 7879 [ 10h 21m ] +INFO: Idle: 8136, Running: 4, Completed: 7880 [ 10h 21m ] +INFO: Idle: 8135, Running: 4, Completed: 7881 [ 10h 22m ] +INFO: Idle: 8133, Running: 4, Completed: 7883 [ 10h 22m ] +INFO: Idle: 8131, Running: 4, Completed: 7885 [ 10h 22m ] +INFO: Idle: 8130, Running: 4, Completed: 7886 [ 10h 22m ] +INFO: Idle: 8129, Running: 4, Completed: 7887 [ 10h 22m ] +INFO: Idle: 8127, Running: 4, Completed: 7889 [ 10h 22m ] +INFO: Idle: 8126, Running: 4, Completed: 7890 [ 10h 22m ] +INFO: Idle: 8124, Running: 4, Completed: 7892 [ 10h 23m ] +INFO: Idle: 8123, Running: 4, Completed: 7893 [ 10h 23m ] +INFO: Idle: 8122, Running: 4, Completed: 7894 [ 10h 23m ] +INFO: Idle: 8120, Running: 4, Completed: 7896 [ 10h 23m ] +INFO: Idle: 8119, Running: 4, Completed: 7897 [ 10h 23m ] +INFO: Idle: 8118, Running: 4, Completed: 7898 [ 10h 23m ] +INFO: Idle: 8117, Running: 4, Completed: 7899 [ 10h 23m ] +INFO: Idle: 8115, Running: 4, Completed: 7901 [ 10h 23m ] +INFO: Idle: 8113, Running: 4, Completed: 7903 [ 10h 23m ] +INFO: Idle: 8112, Running: 4, Completed: 7904 [ 10h 23m ] +INFO: Idle: 8111, Running: 4, Completed: 7905 [ 10h 24m ] +INFO: Idle: 8108, Running: 4, Completed: 7908 [ 10h 24m ] +INFO: Idle: 8107, Running: 4, Completed: 7909 [ 10h 24m ] +INFO: Idle: 8105, Running: 4, Completed: 7911 [ 10h 24m ] +INFO: Idle: 8104, Running: 4, Completed: 7912 [ 10h 24m ] +INFO: Idle: 8103, Running: 4, Completed: 7913 [ 10h 25m ] +INFO: Idle: 8102, Running: 4, Completed: 7914 [ 10h 25m ] +INFO: Idle: 8100, Running: 4, Completed: 7916 [ 10h 25m ] +INFO: Idle: 8098, Running: 4, Completed: 7918 [ 10h 25m ] +INFO: Idle: 8096, Running: 4, Completed: 7920 [ 10h 25m ] +INFO: Idle: 8094, Running: 4, Completed: 7922 [ 10h 25m ] +INFO: Idle: 8092, Running: 4, Completed: 7924 [ 10h 26m ] +INFO: Idle: 8090, Running: 4, Completed: 7926 [ 10h 26m ] +INFO: Idle: 8088, Running: 4, Completed: 7928 [ 10h 26m ] +INFO: Idle: 8087, Running: 4, Completed: 7929 [ 10h 26m ] +INFO: Idle: 8085, Running: 4, Completed: 7931 [ 10h 26m ] +INFO: Idle: 8084, Running: 4, Completed: 7932 [ 10h 26m ] +INFO: Idle: 8083, Running: 4, Completed: 7933 [ 10h 27m ] +INFO: Idle: 8082, Running: 4, Completed: 7934 [ 10h 27m ] +INFO: Idle: 8081, Running: 4, Completed: 7935 [ 10h 27m ] +INFO: Idle: 8080, Running: 4, Completed: 7936 [ 10h 27m ] +INFO: Idle: 8078, Running: 4, Completed: 7938 [ 10h 27m ] +INFO: Idle: 8077, Running: 4, Completed: 7939 [ 10h 27m ] +INFO: Idle: 8076, Running: 4, Completed: 7940 [ 10h 27m ] +INFO: Idle: 8074, Running: 4, Completed: 7942 [ 10h 27m ] +INFO: Idle: 8073, Running: 4, Completed: 7943 [ 10h 27m ] +INFO: Idle: 8072, Running: 4, Completed: 7944 [ 10h 28m ] +INFO: Idle: 8071, Running: 4, Completed: 7945 [ 10h 28m ] +INFO: Idle: 8070, Running: 4, Completed: 7946 [ 10h 28m ] +INFO: Idle: 8068, Running: 4, Completed: 7948 [ 10h 28m ] +INFO: Idle: 8067, Running: 4, Completed: 7949 [ 10h 28m ] +INFO: Idle: 8065, Running: 4, Completed: 7951 [ 10h 28m ] +INFO: Idle: 8064, Running: 4, Completed: 7952 [ 10h 28m ] +INFO: Idle: 8062, Running: 4, Completed: 7954 [ 10h 28m ] +INFO: Idle: 8060, Running: 4, Completed: 7956 [ 10h 29m ] +INFO: Idle: 8059, Running: 4, Completed: 7957 [ 10h 29m ] +INFO: Idle: 8057, Running: 4, Completed: 7959 [ 10h 29m ] +INFO: Idle: 8056, Running: 4, Completed: 7960 [ 10h 29m ] +INFO: Idle: 8055, Running: 4, Completed: 7961 [ 10h 29m ] +INFO: Idle: 8053, Running: 4, Completed: 7963 [ 10h 29m ] +INFO: Idle: 8052, Running: 4, Completed: 7964 [ 10h 29m ] +INFO: Idle: 8051, Running: 4, Completed: 7965 [ 10h 29m ] +INFO: Idle: 8050, Running: 4, Completed: 7966 [ 10h 30m ] +INFO: Idle: 8048, Running: 4, Completed: 7968 [ 10h 30m ] +INFO: Idle: 8047, Running: 4, Completed: 7969 [ 10h 30m ] +INFO: Idle: 8046, Running: 4, Completed: 7970 [ 10h 30m ] +INFO: Idle: 8045, Running: 4, Completed: 7971 [ 10h 30m ] +INFO: Idle: 8043, Running: 4, Completed: 7973 [ 10h 30m ] +INFO: Idle: 8042, Running: 4, Completed: 7974 [ 10h 30m ] +INFO: Idle: 8041, Running: 4, Completed: 7975 [ 10h 30m ] +INFO: Idle: 8040, Running: 4, Completed: 7976 [ 10h 30m ] +INFO: Idle: 8038, Running: 4, Completed: 7978 [ 10h 30m ] +INFO: Idle: 8037, Running: 4, Completed: 7979 [ 10h 31m ] +INFO: Idle: 8035, Running: 4, Completed: 7981 [ 10h 31m ] +INFO: Idle: 8034, Running: 4, Completed: 7982 [ 10h 31m ] +INFO: Idle: 8033, Running: 4, Completed: 7983 [ 10h 31m ] +INFO: Idle: 8031, Running: 4, Completed: 7985 [ 10h 31m ] +INFO: Idle: 8029, Running: 4, Completed: 7987 [ 10h 31m ] +INFO: Idle: 8027, Running: 4, Completed: 7989 [ 10h 31m ] +INFO: Idle: 8025, Running: 4, Completed: 7991 [ 10h 32m ] +INFO: Idle: 8023, Running: 4, Completed: 7993 [ 10h 32m ] +INFO: Idle: 8021, Running: 4, Completed: 7995 [ 10h 32m ] +INFO: Idle: 8019, Running: 4, Completed: 7997 [ 10h 32m ] +INFO: Idle: 8017, Running: 4, Completed: 7999 [ 10h 32m ] +INFO: Idle: 8015, Running: 4, Completed: 8001 [ 10h 32m ] +INFO: Idle: 8013, Running: 4, Completed: 8003 [ 10h 33m ] +INFO: Idle: 8011, Running: 4, Completed: 8005 [ 10h 33m ] +INFO: Idle: 8009, Running: 4, Completed: 8007 [ 10h 33m ] +INFO: Idle: 8008, Running: 4, Completed: 8008 [ 10h 33m ] +INFO: Idle: 8007, Running: 4, Completed: 8009 [ 10h 33m ] +INFO: Idle: 8006, Running: 4, Completed: 8010 [ 10h 33m ] +INFO: Idle: 8005, Running: 4, Completed: 8011 [ 10h 33m ] +INFO: Idle: 8004, Running: 4, Completed: 8012 [ 10h 33m ] +INFO: Idle: 8003, Running: 4, Completed: 8013 [ 10h 33m ] +INFO: Idle: 8002, Running: 4, Completed: 8014 [ 10h 33m ] +INFO: Idle: 8001, Running: 4, Completed: 8015 [ 10h 34m ] +INFO: Idle: 7999, Running: 4, Completed: 8017 [ 10h 34m ] +INFO: Idle: 7998, Running: 4, Completed: 8018 [ 10h 34m ] +INFO: Idle: 7997, Running: 4, Completed: 8019 [ 10h 34m ] +INFO: Idle: 7995, Running: 4, Completed: 8021 [ 10h 34m ] +INFO: Idle: 7994, Running: 4, Completed: 8022 [ 10h 34m ] +INFO: Idle: 7993, Running: 4, Completed: 8023 [ 10h 34m ] +INFO: Idle: 7991, Running: 4, Completed: 8025 [ 10h 35m ] +INFO: Idle: 7990, Running: 4, Completed: 8026 [ 10h 35m ] +INFO: Idle: 7989, Running: 4, Completed: 8027 [ 10h 35m ] +INFO: Idle: 7988, Running: 4, Completed: 8028 [ 10h 35m ] +INFO: Idle: 7987, Running: 4, Completed: 8029 [ 10h 35m ] +INFO: Idle: 7986, Running: 4, Completed: 8030 [ 10h 35m ] +INFO: Idle: 7984, Running: 4, Completed: 8032 [ 10h 35m ] +INFO: Idle: 7983, Running: 4, Completed: 8033 [ 10h 35m ] +INFO: Idle: 7982, Running: 4, Completed: 8034 [ 10h 36m ] +INFO: Idle: 7981, Running: 4, Completed: 8035 [ 10h 36m ] +INFO: Idle: 7980, Running: 4, Completed: 8036 [ 10h 36m ] +INFO: Idle: 7978, Running: 4, Completed: 8038 [ 10h 36m ] +INFO: Idle: 7977, Running: 4, Completed: 8039 [ 10h 36m ] +INFO: Idle: 7974, Running: 4, Completed: 8042 [ 10h 36m ] +INFO: Idle: 7973, Running: 4, Completed: 8043 [ 10h 37m ] +INFO: Idle: 7972, Running: 4, Completed: 8044 [ 10h 37m ] +INFO: Idle: 7971, Running: 4, Completed: 8045 [ 10h 37m ] +INFO: Idle: 7969, Running: 4, Completed: 8047 [ 10h 37m ] +INFO: Idle: 7968, Running: 4, Completed: 8048 [ 10h 37m ] +INFO: Idle: 7967, Running: 4, Completed: 8049 [ 10h 37m ] +INFO: Idle: 7965, Running: 4, Completed: 8051 [ 10h 37m ] +INFO: Idle: 7963, Running: 4, Completed: 8053 [ 10h 38m ] +INFO: Idle: 7961, Running: 4, Completed: 8055 [ 10h 38m ] +INFO: Idle: 7960, Running: 4, Completed: 8056 [ 10h 38m ] +INFO: Idle: 7957, Running: 4, Completed: 8059 [ 10h 38m ] +INFO: Idle: 7954, Running: 4, Completed: 8062 [ 10h 38m ] +INFO: Idle: 7953, Running: 4, Completed: 8063 [ 10h 39m ] +INFO: Idle: 7952, Running: 4, Completed: 8064 [ 10h 39m ] +INFO: Idle: 7949, Running: 4, Completed: 8067 [ 10h 39m ] +INFO: Idle: 7948, Running: 4, Completed: 8068 [ 10h 39m ] +INFO: Idle: 7947, Running: 4, Completed: 8069 [ 10h 39m ] +INFO: Idle: 7946, Running: 4, Completed: 8070 [ 10h 39m ] +INFO: Idle: 7945, Running: 4, Completed: 8071 [ 10h 40m ] +INFO: Idle: 7943, Running: 4, Completed: 8073 [ 10h 40m ] +INFO: Idle: 7942, Running: 4, Completed: 8074 [ 10h 40m ] +INFO: Idle: 7941, Running: 4, Completed: 8075 [ 10h 40m ] +INFO: Idle: 7940, Running: 4, Completed: 8076 [ 10h 40m ] +INFO: Idle: 7939, Running: 4, Completed: 8077 [ 10h 40m ] +INFO: Idle: 7938, Running: 4, Completed: 8078 [ 10h 40m ] +INFO: Idle: 7937, Running: 4, Completed: 8079 [ 10h 40m ] +INFO: Idle: 7936, Running: 4, Completed: 8080 [ 10h 40m ] +INFO: Idle: 7934, Running: 4, Completed: 8082 [ 10h 40m ] +INFO: Idle: 7933, Running: 4, Completed: 8083 [ 10h 41m ] +INFO: Idle: 7932, Running: 4, Completed: 8084 [ 10h 41m ] +INFO: Idle: 7930, Running: 4, Completed: 8086 [ 10h 41m ] +INFO: Idle: 7928, Running: 4, Completed: 8088 [ 10h 41m ] +INFO: Idle: 7927, Running: 4, Completed: 8089 [ 10h 41m ] +INFO: Idle: 7926, Running: 4, Completed: 8090 [ 10h 41m ] +INFO: Idle: 7924, Running: 4, Completed: 8092 [ 10h 41m ] +INFO: Idle: 7923, Running: 4, Completed: 8093 [ 10h 41m ] +INFO: Idle: 7922, Running: 4, Completed: 8094 [ 10h 42m ] +INFO: Idle: 7921, Running: 4, Completed: 8095 [ 10h 42m ] +INFO: Idle: 7919, Running: 4, Completed: 8097 [ 10h 42m ] +INFO: Idle: 7918, Running: 4, Completed: 8098 [ 10h 42m ] +INFO: Idle: 7917, Running: 4, Completed: 8099 [ 10h 42m ] +INFO: Idle: 7915, Running: 4, Completed: 8101 [ 10h 42m ] +INFO: Idle: 7914, Running: 4, Completed: 8102 [ 10h 42m ] +INFO: Idle: 7912, Running: 4, Completed: 8104 [ 10h 42m ] +INFO: Idle: 7911, Running: 4, Completed: 8105 [ 10h 43m ] +INFO: Idle: 7909, Running: 4, Completed: 8107 [ 10h 43m ] +INFO: Idle: 7907, Running: 4, Completed: 8109 [ 10h 43m ] +INFO: Idle: 7905, Running: 4, Completed: 8111 [ 10h 43m ] +INFO: Idle: 7903, Running: 4, Completed: 8113 [ 10h 43m ] +INFO: Idle: 7902, Running: 4, Completed: 8114 [ 10h 43m ] +INFO: Idle: 7899, Running: 4, Completed: 8117 [ 10h 44m ] +INFO: Idle: 7898, Running: 4, Completed: 8118 [ 10h 44m ] +INFO: Idle: 7897, Running: 4, Completed: 8119 [ 10h 44m ] +INFO: Idle: 7896, Running: 4, Completed: 8120 [ 10h 44m ] +INFO: Idle: 7895, Running: 4, Completed: 8121 [ 10h 44m ] +INFO: Idle: 7894, Running: 4, Completed: 8122 [ 10h 44m ] +INFO: Idle: 7893, Running: 4, Completed: 8123 [ 10h 44m ] +INFO: Idle: 7892, Running: 4, Completed: 8124 [ 10h 44m ] +INFO: Idle: 7891, Running: 4, Completed: 8125 [ 10h 44m ] +INFO: Idle: 7890, Running: 4, Completed: 8126 [ 10h 44m ] +INFO: Idle: 7888, Running: 4, Completed: 8128 [ 10h 44m ] +INFO: Idle: 7887, Running: 4, Completed: 8129 [ 10h 45m ] +INFO: Idle: 7886, Running: 4, Completed: 8130 [ 10h 45m ] +INFO: Idle: 7884, Running: 4, Completed: 8132 [ 10h 45m ] +INFO: Idle: 7882, Running: 4, Completed: 8134 [ 10h 45m ] +INFO: Idle: 7881, Running: 4, Completed: 8135 [ 10h 45m ] +INFO: Idle: 7880, Running: 4, Completed: 8136 [ 10h 45m ] +INFO: Idle: 7879, Running: 4, Completed: 8137 [ 10h 45m ] +INFO: Idle: 7878, Running: 4, Completed: 8138 [ 10h 45m ] +INFO: Idle: 7877, Running: 4, Completed: 8139 [ 10h 45m ] +INFO: Idle: 7876, Running: 4, Completed: 8140 [ 10h 46m ] +INFO: Idle: 7875, Running: 4, Completed: 8141 [ 10h 46m ] +INFO: Idle: 7874, Running: 4, Completed: 8142 [ 10h 46m ] +INFO: Idle: 7873, Running: 4, Completed: 8143 [ 10h 46m ] +INFO: Idle: 7872, Running: 4, Completed: 8144 [ 10h 46m ] +INFO: Idle: 7871, Running: 4, Completed: 8145 [ 10h 46m ] +INFO: Idle: 7870, Running: 4, Completed: 8146 [ 10h 46m ] +INFO: Idle: 7869, Running: 4, Completed: 8147 [ 10h 46m ] +INFO: Idle: 7868, Running: 4, Completed: 8148 [ 10h 46m ] +INFO: Idle: 7867, Running: 4, Completed: 8149 [ 10h 46m ] +INFO: Idle: 7866, Running: 4, Completed: 8150 [ 10h 47m ] +INFO: Idle: 7865, Running: 4, Completed: 8151 [ 10h 47m ] +INFO: Idle: 7864, Running: 4, Completed: 8152 [ 10h 47m ] +INFO: Idle: 7863, Running: 4, Completed: 8153 [ 10h 47m ] +INFO: Idle: 7861, Running: 4, Completed: 8155 [ 10h 47m ] +INFO: Idle: 7859, Running: 4, Completed: 8157 [ 10h 47m ] +INFO: Idle: 7856, Running: 4, Completed: 8160 [ 10h 48m ] +INFO: Idle: 7855, Running: 4, Completed: 8161 [ 10h 48m ] +INFO: Idle: 7853, Running: 4, Completed: 8163 [ 10h 48m ] +INFO: Idle: 7852, Running: 4, Completed: 8164 [ 10h 48m ] +INFO: Idle: 7851, Running: 4, Completed: 8165 [ 10h 48m ] +INFO: Idle: 7849, Running: 4, Completed: 8167 [ 10h 48m ] +INFO: Idle: 7848, Running: 4, Completed: 8168 [ 10h 48m ] +INFO: Idle: 7846, Running: 4, Completed: 8170 [ 10h 48m ] +INFO: Idle: 7845, Running: 4, Completed: 8171 [ 10h 49m ] +INFO: Idle: 7843, Running: 4, Completed: 8173 [ 10h 49m ] +INFO: Idle: 7842, Running: 4, Completed: 8174 [ 10h 49m ] +INFO: Idle: 7841, Running: 4, Completed: 8175 [ 10h 49m ] +INFO: Idle: 7840, Running: 4, Completed: 8176 [ 10h 49m ] +INFO: Idle: 7839, Running: 4, Completed: 8177 [ 10h 49m ] +INFO: Idle: 7838, Running: 4, Completed: 8178 [ 10h 49m ] +INFO: Idle: 7837, Running: 4, Completed: 8179 [ 10h 50m ] +INFO: Idle: 7835, Running: 4, Completed: 8181 [ 10h 50m ] +INFO: Idle: 7833, Running: 4, Completed: 8183 [ 10h 50m ] +INFO: Idle: 7832, Running: 4, Completed: 8184 [ 10h 50m ] +INFO: Idle: 7829, Running: 4, Completed: 8187 [ 10h 50m ] +INFO: Idle: 7827, Running: 4, Completed: 8189 [ 10h 50m ] +INFO: Idle: 7825, Running: 4, Completed: 8191 [ 10h 51m ] +INFO: Idle: 7824, Running: 4, Completed: 8192 [ 10h 51m ] +INFO: Idle: 7822, Running: 4, Completed: 8194 [ 10h 51m ] +INFO: Idle: 7821, Running: 4, Completed: 8195 [ 10h 51m ] +INFO: Idle: 7820, Running: 4, Completed: 8196 [ 10h 51m ] +INFO: Idle: 7819, Running: 4, Completed: 8197 [ 10h 51m ] +INFO: Idle: 7818, Running: 4, Completed: 8198 [ 10h 52m ] +INFO: Idle: 7817, Running: 4, Completed: 8199 [ 10h 52m ] +INFO: Idle: 7815, Running: 4, Completed: 8201 [ 10h 52m ] +INFO: Idle: 7813, Running: 4, Completed: 8203 [ 10h 52m ] +INFO: Idle: 7811, Running: 4, Completed: 8205 [ 10h 52m ] +INFO: Idle: 7810, Running: 4, Completed: 8206 [ 10h 52m ] +INFO: Idle: 7809, Running: 4, Completed: 8207 [ 10h 52m ] +INFO: Idle: 7808, Running: 4, Completed: 8208 [ 10h 53m ] +INFO: Idle: 7806, Running: 4, Completed: 8210 [ 10h 53m ] +INFO: Idle: 7805, Running: 4, Completed: 8211 [ 10h 53m ] +INFO: Idle: 7804, Running: 4, Completed: 8212 [ 10h 53m ] +INFO: Idle: 7802, Running: 4, Completed: 8214 [ 10h 53m ] +INFO: Idle: 7800, Running: 4, Completed: 8216 [ 10h 53m ] +INFO: Idle: 7799, Running: 4, Completed: 8217 [ 10h 53m ] +INFO: Idle: 7798, Running: 4, Completed: 8218 [ 10h 53m ] +INFO: Idle: 7796, Running: 4, Completed: 8220 [ 10h 53m ] +INFO: Idle: 7795, Running: 4, Completed: 8221 [ 10h 54m ] +INFO: Idle: 7794, Running: 4, Completed: 8222 [ 10h 54m ] +INFO: Idle: 7793, Running: 4, Completed: 8223 [ 10h 54m ] +INFO: Idle: 7792, Running: 4, Completed: 8224 [ 10h 54m ] +INFO: Idle: 7791, Running: 4, Completed: 8225 [ 10h 54m ] +INFO: Idle: 7789, Running: 4, Completed: 8227 [ 10h 54m ] +INFO: Idle: 7788, Running: 4, Completed: 8228 [ 10h 54m ] +INFO: Idle: 7787, Running: 4, Completed: 8229 [ 10h 54m ] +INFO: Idle: 7786, Running: 4, Completed: 8230 [ 10h 55m ] +INFO: Idle: 7785, Running: 4, Completed: 8231 [ 10h 55m ] +INFO: Idle: 7784, Running: 4, Completed: 8232 [ 10h 55m ] +INFO: Idle: 7783, Running: 4, Completed: 8233 [ 10h 55m ] +INFO: Idle: 7782, Running: 4, Completed: 8234 [ 10h 55m ] +INFO: Idle: 7780, Running: 4, Completed: 8236 [ 10h 55m ] +INFO: Idle: 7779, Running: 4, Completed: 8237 [ 10h 55m ] +INFO: Idle: 7778, Running: 4, Completed: 8238 [ 10h 55m ] +INFO: Idle: 7775, Running: 4, Completed: 8241 [ 10h 56m ] +INFO: Idle: 7773, Running: 4, Completed: 8243 [ 10h 56m ] +INFO: Idle: 7771, Running: 4, Completed: 8245 [ 10h 56m ] +INFO: Idle: 7770, Running: 4, Completed: 8246 [ 10h 56m ] +INFO: Idle: 7769, Running: 4, Completed: 8247 [ 10h 56m ] +INFO: Idle: 7767, Running: 4, Completed: 8249 [ 10h 56m ] +INFO: Idle: 7766, Running: 4, Completed: 8250 [ 10h 56m ] +INFO: Idle: 7765, Running: 4, Completed: 8251 [ 10h 56m ] +INFO: Idle: 7764, Running: 4, Completed: 8252 [ 10h 56m ] +INFO: Idle: 7763, Running: 4, Completed: 8253 [ 10h 56m ] +INFO: Idle: 7761, Running: 4, Completed: 8255 [ 10h 57m ] +INFO: Idle: 7760, Running: 4, Completed: 8256 [ 10h 57m ] +INFO: Idle: 7759, Running: 4, Completed: 8257 [ 10h 57m ] +INFO: Idle: 7757, Running: 4, Completed: 8259 [ 10h 57m ] +INFO: Idle: 7756, Running: 4, Completed: 8260 [ 10h 57m ] +INFO: Idle: 7754, Running: 4, Completed: 8262 [ 10h 57m ] +INFO: Idle: 7752, Running: 4, Completed: 8264 [ 10h 57m ] +INFO: Idle: 7751, Running: 4, Completed: 8265 [ 10h 57m ] +INFO: Idle: 7749, Running: 4, Completed: 8267 [ 10h 58m ] +INFO: Idle: 7748, Running: 4, Completed: 8268 [ 10h 58m ] +INFO: Idle: 7747, Running: 4, Completed: 8269 [ 10h 58m ] +INFO: Idle: 7745, Running: 4, Completed: 8271 [ 10h 58m ] +INFO: Idle: 7744, Running: 4, Completed: 8272 [ 10h 58m ] +INFO: Idle: 7743, Running: 4, Completed: 8273 [ 10h 58m ] +INFO: Idle: 7742, Running: 4, Completed: 8274 [ 10h 58m ] +INFO: Idle: 7740, Running: 4, Completed: 8276 [ 10h 58m ] +INFO: Idle: 7739, Running: 4, Completed: 8277 [ 10h 59m ] +INFO: Idle: 7738, Running: 4, Completed: 8278 [ 10h 59m ] +INFO: Idle: 7736, Running: 4, Completed: 8280 [ 10h 59m ] +INFO: Idle: 7735, Running: 4, Completed: 8281 [ 10h 59m ] +INFO: Idle: 7733, Running: 4, Completed: 8283 [ 10h 59m ] +INFO: Idle: 7731, Running: 4, Completed: 8285 [ 10h 59m ] +INFO: Idle: 7730, Running: 4, Completed: 8286 [ 10h 59m ] +INFO: Idle: 7728, Running: 4, Completed: 8288 [ 11h 0m ] +INFO: Idle: 7727, Running: 4, Completed: 8289 [ 11h 0m ] +INFO: Idle: 7725, Running: 4, Completed: 8291 [ 11h 0m ] +INFO: Idle: 7724, Running: 4, Completed: 8292 [ 11h 0m ] +INFO: Idle: 7723, Running: 4, Completed: 8293 [ 11h 0m ] +INFO: Idle: 7720, Running: 4, Completed: 8296 [ 11h 0m ] +INFO: Idle: 7719, Running: 4, Completed: 8297 [ 11h 0m ] +INFO: Idle: 7717, Running: 4, Completed: 8299 [ 11h 1m ] +INFO: Idle: 7716, Running: 4, Completed: 8300 [ 11h 1m ] +INFO: Idle: 7715, Running: 4, Completed: 8301 [ 11h 1m ] +INFO: Idle: 7714, Running: 4, Completed: 8302 [ 11h 1m ] +INFO: Idle: 7713, Running: 4, Completed: 8303 [ 11h 1m ] +INFO: Idle: 7711, Running: 4, Completed: 8305 [ 11h 1m ] +INFO: Idle: 7710, Running: 4, Completed: 8306 [ 11h 1m ] +INFO: Idle: 7707, Running: 4, Completed: 8309 [ 11h 1m ] +INFO: Idle: 7706, Running: 4, Completed: 8310 [ 11h 2m ] +INFO: Idle: 7705, Running: 4, Completed: 8311 [ 11h 2m ] +INFO: Idle: 7704, Running: 4, Completed: 8312 [ 11h 2m ] +INFO: Idle: 7702, Running: 4, Completed: 8314 [ 11h 2m ] +INFO: Idle: 7700, Running: 4, Completed: 8316 [ 11h 2m ] +INFO: Idle: 7699, Running: 4, Completed: 8317 [ 11h 2m ] +INFO: Idle: 7698, Running: 4, Completed: 8318 [ 11h 2m ] +INFO: Idle: 7697, Running: 4, Completed: 8319 [ 11h 2m ] +INFO: Idle: 7696, Running: 4, Completed: 8320 [ 11h 2m ] +INFO: Idle: 7693, Running: 4, Completed: 8323 [ 11h 3m ] +INFO: Idle: 7691, Running: 4, Completed: 8325 [ 11h 3m ] +INFO: Idle: 7690, Running: 4, Completed: 8326 [ 11h 3m ] +INFO: Idle: 7689, Running: 4, Completed: 8327 [ 11h 3m ] +INFO: Idle: 7688, Running: 4, Completed: 8328 [ 11h 3m ] +INFO: Idle: 7687, Running: 4, Completed: 8329 [ 11h 3m ] +INFO: Idle: 7686, Running: 4, Completed: 8330 [ 11h 3m ] +INFO: Idle: 7685, Running: 4, Completed: 8331 [ 11h 3m ] +INFO: Idle: 7684, Running: 4, Completed: 8332 [ 11h 4m ] +INFO: Idle: 7682, Running: 4, Completed: 8334 [ 11h 4m ] +INFO: Idle: 7681, Running: 4, Completed: 8335 [ 11h 4m ] +INFO: Idle: 7680, Running: 4, Completed: 8336 [ 11h 4m ] +INFO: Idle: 7679, Running: 4, Completed: 8337 [ 11h 4m ] +INFO: Idle: 7678, Running: 4, Completed: 8338 [ 11h 4m ] +INFO: Idle: 7676, Running: 4, Completed: 8340 [ 11h 4m ] +INFO: Idle: 7674, Running: 4, Completed: 8342 [ 11h 4m ] +INFO: Idle: 7673, Running: 4, Completed: 8343 [ 11h 5m ] +INFO: Idle: 7671, Running: 4, Completed: 8345 [ 11h 5m ] +INFO: Idle: 7670, Running: 4, Completed: 8346 [ 11h 5m ] +INFO: Idle: 7668, Running: 4, Completed: 8348 [ 11h 5m ] +INFO: Idle: 7666, Running: 4, Completed: 8350 [ 11h 5m ] +INFO: Idle: 7665, Running: 4, Completed: 8351 [ 11h 5m ] +INFO: Idle: 7662, Running: 4, Completed: 8354 [ 11h 6m ] +INFO: Idle: 7661, Running: 4, Completed: 8355 [ 11h 6m ] +INFO: Idle: 7659, Running: 4, Completed: 8357 [ 11h 6m ] +INFO: Idle: 7657, Running: 4, Completed: 8359 [ 11h 6m ] +INFO: Idle: 7656, Running: 4, Completed: 8360 [ 11h 6m ] +INFO: Idle: 7654, Running: 4, Completed: 8362 [ 11h 6m ] +INFO: Idle: 7653, Running: 4, Completed: 8363 [ 11h 6m ] +INFO: Idle: 7652, Running: 4, Completed: 8364 [ 11h 6m ] +INFO: Idle: 7651, Running: 4, Completed: 8365 [ 11h 6m ] +INFO: Idle: 7650, Running: 4, Completed: 8366 [ 11h 7m ] +INFO: Idle: 7649, Running: 4, Completed: 8367 [ 11h 7m ] +INFO: Idle: 7648, Running: 4, Completed: 8368 [ 11h 7m ] +INFO: Idle: 7646, Running: 4, Completed: 8370 [ 11h 7m ] +INFO: Idle: 7645, Running: 4, Completed: 8371 [ 11h 7m ] +INFO: Idle: 7644, Running: 4, Completed: 8372 [ 11h 7m ] +INFO: Idle: 7641, Running: 4, Completed: 8375 [ 11h 7m ] +INFO: Idle: 7640, Running: 4, Completed: 8376 [ 11h 7m ] +INFO: Idle: 7639, Running: 4, Completed: 8377 [ 11h 7m ] +INFO: Idle: 7637, Running: 4, Completed: 8379 [ 11h 8m ] +INFO: Idle: 7633, Running: 4, Completed: 8383 [ 11h 8m ] +INFO: Idle: 7631, Running: 4, Completed: 8385 [ 11h 8m ] +INFO: Idle: 7629, Running: 4, Completed: 8387 [ 11h 8m ] +INFO: Idle: 7627, Running: 4, Completed: 8389 [ 11h 8m ] +INFO: Idle: 7625, Running: 4, Completed: 8391 [ 11h 8m ] +INFO: Idle: 7623, Running: 4, Completed: 8393 [ 11h 9m ] +INFO: Idle: 7621, Running: 4, Completed: 8395 [ 11h 9m ] +INFO: Idle: 7619, Running: 4, Completed: 8397 [ 11h 9m ] +INFO: Idle: 7618, Running: 4, Completed: 8398 [ 11h 9m ] +INFO: Idle: 7617, Running: 4, Completed: 8399 [ 11h 9m ] +INFO: Idle: 7615, Running: 4, Completed: 8401 [ 11h 9m ] +INFO: Idle: 7614, Running: 4, Completed: 8402 [ 11h 9m ] +INFO: Idle: 7613, Running: 4, Completed: 8403 [ 11h 10m ] +INFO: Idle: 7611, Running: 4, Completed: 8405 [ 11h 10m ] +INFO: Idle: 7610, Running: 4, Completed: 8406 [ 11h 10m ] +INFO: Idle: 7609, Running: 4, Completed: 8407 [ 11h 10m ] +INFO: Idle: 7607, Running: 4, Completed: 8409 [ 11h 10m ] +INFO: Idle: 7604, Running: 4, Completed: 8412 [ 11h 10m ] +INFO: Idle: 7603, Running: 4, Completed: 8413 [ 11h 10m ] +INFO: Idle: 7602, Running: 4, Completed: 8414 [ 11h 11m ] +INFO: Idle: 7601, Running: 4, Completed: 8415 [ 11h 11m ] +INFO: Idle: 7600, Running: 4, Completed: 8416 [ 11h 11m ] +INFO: Idle: 7599, Running: 4, Completed: 8417 [ 11h 11m ] +INFO: Idle: 7598, Running: 4, Completed: 8418 [ 11h 11m ] +INFO: Idle: 7596, Running: 4, Completed: 8420 [ 11h 11m ] +INFO: Idle: 7595, Running: 4, Completed: 8421 [ 11h 11m ] +INFO: Idle: 7594, Running: 4, Completed: 8422 [ 11h 11m ] +INFO: Idle: 7591, Running: 4, Completed: 8425 [ 11h 11m ] +INFO: Idle: 7589, Running: 4, Completed: 8427 [ 11h 12m ] +INFO: Idle: 7587, Running: 4, Completed: 8429 [ 11h 12m ] +INFO: Idle: 7586, Running: 4, Completed: 8430 [ 11h 12m ] +INFO: Idle: 7585, Running: 4, Completed: 8431 [ 11h 12m ] +INFO: Idle: 7583, Running: 4, Completed: 8433 [ 11h 12m ] +INFO: Idle: 7582, Running: 4, Completed: 8434 [ 11h 12m ] +INFO: Idle: 7579, Running: 4, Completed: 8437 [ 11h 13m ] +INFO: Idle: 7577, Running: 4, Completed: 8439 [ 11h 13m ] +INFO: Idle: 7576, Running: 4, Completed: 8440 [ 11h 13m ] +INFO: Idle: 7575, Running: 4, Completed: 8441 [ 11h 13m ] +INFO: Idle: 7574, Running: 4, Completed: 8442 [ 11h 13m ] +INFO: Idle: 7572, Running: 4, Completed: 8444 [ 11h 13m ] +INFO: Idle: 7571, Running: 4, Completed: 8445 [ 11h 13m ] +INFO: Idle: 7569, Running: 4, Completed: 8447 [ 11h 13m ] +INFO: Idle: 7568, Running: 4, Completed: 8448 [ 11h 14m ] +INFO: Idle: 7565, Running: 4, Completed: 8451 [ 11h 14m ] +INFO: Idle: 7564, Running: 4, Completed: 8452 [ 11h 14m ] +INFO: Idle: 7563, Running: 4, Completed: 8453 [ 11h 14m ] +INFO: Idle: 7562, Running: 4, Completed: 8454 [ 11h 14m ] +INFO: Idle: 7561, Running: 4, Completed: 8455 [ 11h 14m ] +INFO: Idle: 7560, Running: 4, Completed: 8456 [ 11h 14m ] +INFO: Idle: 7559, Running: 4, Completed: 8457 [ 11h 14m ] +INFO: Idle: 7558, Running: 4, Completed: 8458 [ 11h 15m ] +INFO: Idle: 7557, Running: 4, Completed: 8459 [ 11h 15m ] +INFO: Idle: 7556, Running: 4, Completed: 8460 [ 11h 15m ] +INFO: Idle: 7554, Running: 4, Completed: 8462 [ 11h 15m ] +INFO: Idle: 7553, Running: 4, Completed: 8463 [ 11h 15m ] +INFO: Idle: 7552, Running: 4, Completed: 8464 [ 11h 15m ] +INFO: Idle: 7551, Running: 4, Completed: 8465 [ 11h 15m ] +INFO: Idle: 7550, Running: 4, Completed: 8466 [ 11h 15m ] +INFO: Idle: 7548, Running: 4, Completed: 8468 [ 11h 15m ] +INFO: Idle: 7547, Running: 4, Completed: 8469 [ 11h 15m ] +INFO: Idle: 7546, Running: 4, Completed: 8470 [ 11h 16m ] +INFO: Idle: 7545, Running: 4, Completed: 8471 [ 11h 16m ] +INFO: Idle: 7543, Running: 4, Completed: 8473 [ 11h 16m ] +INFO: Idle: 7542, Running: 4, Completed: 8474 [ 11h 16m ] +INFO: Idle: 7540, Running: 4, Completed: 8476 [ 11h 16m ] +INFO: Idle: 7538, Running: 4, Completed: 8478 [ 11h 16m ] +INFO: Idle: 7537, Running: 4, Completed: 8479 [ 11h 17m ] +INFO: Idle: 7536, Running: 4, Completed: 8480 [ 11h 17m ] +INFO: Idle: 7534, Running: 4, Completed: 8482 [ 11h 17m ] +INFO: Idle: 7533, Running: 4, Completed: 8483 [ 11h 17m ] +INFO: Idle: 7531, Running: 4, Completed: 8485 [ 11h 17m ] +INFO: Idle: 7529, Running: 4, Completed: 8487 [ 11h 17m ] +INFO: Idle: 7527, Running: 4, Completed: 8489 [ 11h 17m ] +INFO: Idle: 7526, Running: 4, Completed: 8490 [ 11h 17m ] +INFO: Idle: 7525, Running: 4, Completed: 8491 [ 11h 17m ] +INFO: Idle: 7524, Running: 4, Completed: 8492 [ 11h 18m ] +INFO: Idle: 7523, Running: 4, Completed: 8493 [ 11h 18m ] +INFO: Idle: 7522, Running: 4, Completed: 8494 [ 11h 18m ] +INFO: Idle: 7521, Running: 4, Completed: 8495 [ 11h 18m ] +INFO: Idle: 7520, Running: 4, Completed: 8496 [ 11h 18m ] +INFO: Idle: 7517, Running: 4, Completed: 8499 [ 11h 18m ] +INFO: Idle: 7516, Running: 4, Completed: 8500 [ 11h 18m ] +INFO: Idle: 7513, Running: 4, Completed: 8503 [ 11h 18m ] +INFO: Idle: 7512, Running: 4, Completed: 8504 [ 11h 19m ] +INFO: Idle: 7510, Running: 4, Completed: 8506 [ 11h 19m ] +INFO: Idle: 7509, Running: 4, Completed: 8507 [ 11h 19m ] +INFO: Idle: 7505, Running: 4, Completed: 8511 [ 11h 19m ] +INFO: Idle: 7503, Running: 4, Completed: 8513 [ 11h 19m ] +INFO: Idle: 7501, Running: 4, Completed: 8515 [ 11h 19m ] +INFO: Idle: 7499, Running: 4, Completed: 8517 [ 11h 19m ] +INFO: Idle: 7496, Running: 4, Completed: 8520 [ 11h 20m ] +INFO: Idle: 7495, Running: 4, Completed: 8521 [ 11h 20m ] +INFO: Idle: 7492, Running: 4, Completed: 8524 [ 11h 20m ] +INFO: Idle: 7491, Running: 4, Completed: 8525 [ 11h 20m ] +INFO: Idle: 7488, Running: 4, Completed: 8528 [ 11h 20m ] +INFO: Idle: 7487, Running: 4, Completed: 8529 [ 11h 20m ] +INFO: Idle: 7484, Running: 4, Completed: 8532 [ 11h 21m ] +INFO: Idle: 7483, Running: 4, Completed: 8533 [ 11h 21m ] +INFO: Idle: 7480, Running: 4, Completed: 8536 [ 11h 21m ] +INFO: Idle: 7479, Running: 4, Completed: 8537 [ 11h 21m ] +INFO: Idle: 7477, Running: 4, Completed: 8539 [ 11h 21m ] +INFO: Idle: 7476, Running: 4, Completed: 8540 [ 11h 22m ] +INFO: Idle: 7475, Running: 4, Completed: 8541 [ 11h 22m ] +INFO: Idle: 7474, Running: 4, Completed: 8542 [ 11h 22m ] +INFO: Idle: 7473, Running: 4, Completed: 8543 [ 11h 22m ] +INFO: Idle: 7472, Running: 4, Completed: 8544 [ 11h 22m ] +INFO: Idle: 7471, Running: 4, Completed: 8545 [ 11h 22m ] +INFO: Idle: 7470, Running: 4, Completed: 8546 [ 11h 22m ] +INFO: Idle: 7468, Running: 4, Completed: 8548 [ 11h 22m ] +INFO: Idle: 7467, Running: 4, Completed: 8549 [ 11h 22m ] +INFO: Idle: 7465, Running: 4, Completed: 8551 [ 11h 22m ] +INFO: Idle: 7464, Running: 4, Completed: 8552 [ 11h 23m ] +INFO: Idle: 7463, Running: 4, Completed: 8553 [ 11h 23m ] +INFO: Idle: 7461, Running: 4, Completed: 8555 [ 11h 23m ] +INFO: Idle: 7460, Running: 4, Completed: 8556 [ 11h 23m ] +INFO: Idle: 7459, Running: 4, Completed: 8557 [ 11h 23m ] +INFO: Idle: 7457, Running: 4, Completed: 8559 [ 11h 23m ] +INFO: Idle: 7456, Running: 4, Completed: 8560 [ 11h 23m ] +INFO: Idle: 7455, Running: 4, Completed: 8561 [ 11h 23m ] +INFO: Idle: 7454, Running: 4, Completed: 8562 [ 11h 23m ] +INFO: Idle: 7452, Running: 4, Completed: 8564 [ 11h 23m ] +INFO: Idle: 7451, Running: 4, Completed: 8565 [ 11h 24m ] +INFO: Idle: 7448, Running: 4, Completed: 8568 [ 11h 24m ] +INFO: Idle: 7447, Running: 4, Completed: 8569 [ 11h 24m ] +INFO: Idle: 7446, Running: 4, Completed: 8570 [ 11h 24m ] +INFO: Idle: 7444, Running: 4, Completed: 8572 [ 11h 24m ] +INFO: Idle: 7442, Running: 4, Completed: 8574 [ 11h 24m ] +INFO: Idle: 7440, Running: 4, Completed: 8576 [ 11h 25m ] +INFO: Idle: 7438, Running: 4, Completed: 8578 [ 11h 25m ] +INFO: Idle: 7437, Running: 4, Completed: 8579 [ 11h 25m ] +INFO: Idle: 7435, Running: 4, Completed: 8581 [ 11h 25m ] +INFO: Idle: 7434, Running: 4, Completed: 8582 [ 11h 25m ] +INFO: Idle: 7433, Running: 4, Completed: 8583 [ 11h 25m ] +INFO: Idle: 7431, Running: 4, Completed: 8585 [ 11h 25m ] +INFO: Idle: 7430, Running: 4, Completed: 8586 [ 11h 25m ] +INFO: Idle: 7429, Running: 4, Completed: 8587 [ 11h 25m ] +INFO: Idle: 7427, Running: 4, Completed: 8589 [ 11h 25m ] +INFO: Idle: 7423, Running: 4, Completed: 8593 [ 11h 26m ] +INFO: Idle: 7421, Running: 4, Completed: 8595 [ 11h 26m ] +INFO: Idle: 7419, Running: 4, Completed: 8597 [ 11h 26m ] +INFO: Idle: 7417, Running: 4, Completed: 8599 [ 11h 26m ] +INFO: Idle: 7413, Running: 4, Completed: 8603 [ 11h 27m ] +INFO: Idle: 7410, Running: 4, Completed: 8606 [ 11h 27m ] +INFO: Idle: 7409, Running: 4, Completed: 8607 [ 11h 27m ] +INFO: Idle: 7407, Running: 4, Completed: 8609 [ 11h 27m ] +INFO: Idle: 7405, Running: 4, Completed: 8611 [ 11h 27m ] +INFO: Idle: 7403, Running: 4, Completed: 8613 [ 11h 27m ] +INFO: Idle: 7401, Running: 4, Completed: 8615 [ 11h 28m ] +INFO: Idle: 7399, Running: 4, Completed: 8617 [ 11h 28m ] +INFO: Idle: 7398, Running: 4, Completed: 8618 [ 11h 28m ] +INFO: Idle: 7397, Running: 4, Completed: 8619 [ 11h 28m ] +INFO: Idle: 7396, Running: 4, Completed: 8620 [ 11h 28m ] +INFO: Idle: 7395, Running: 4, Completed: 8621 [ 11h 28m ] +INFO: Idle: 7394, Running: 4, Completed: 8622 [ 11h 28m ] +INFO: Idle: 7393, Running: 4, Completed: 8623 [ 11h 28m ] +INFO: Idle: 7392, Running: 4, Completed: 8624 [ 11h 28m ] +INFO: Idle: 7391, Running: 4, Completed: 8625 [ 11h 29m ] +INFO: Idle: 7390, Running: 4, Completed: 8626 [ 11h 29m ] +INFO: Idle: 7389, Running: 4, Completed: 8627 [ 11h 29m ] +INFO: Idle: 7388, Running: 4, Completed: 8628 [ 11h 29m ] +INFO: Idle: 7386, Running: 4, Completed: 8630 [ 11h 29m ] +INFO: Idle: 7385, Running: 4, Completed: 8631 [ 11h 29m ] +INFO: Idle: 7383, Running: 4, Completed: 8633 [ 11h 29m ] +INFO: Idle: 7381, Running: 4, Completed: 8635 [ 11h 30m ] +INFO: Idle: 7378, Running: 4, Completed: 8638 [ 11h 30m ] +INFO: Idle: 7377, Running: 4, Completed: 8639 [ 11h 30m ] +INFO: Idle: 7375, Running: 4, Completed: 8641 [ 11h 30m ] +INFO: Idle: 7374, Running: 4, Completed: 8642 [ 11h 30m ] +INFO: Idle: 7373, Running: 4, Completed: 8643 [ 11h 30m ] +INFO: Idle: 7372, Running: 4, Completed: 8644 [ 11h 30m ] +INFO: Idle: 7370, Running: 4, Completed: 8646 [ 11h 30m ] +INFO: Idle: 7369, Running: 4, Completed: 8647 [ 11h 30m ] +INFO: Idle: 7366, Running: 4, Completed: 8650 [ 11h 31m ] +INFO: Idle: 7365, Running: 4, Completed: 8651 [ 11h 31m ] +INFO: Idle: 7364, Running: 4, Completed: 8652 [ 11h 31m ] +INFO: Idle: 7362, Running: 4, Completed: 8654 [ 11h 31m ] +INFO: Idle: 7360, Running: 4, Completed: 8656 [ 11h 31m ] +INFO: Idle: 7358, Running: 4, Completed: 8658 [ 11h 31m ] +INFO: Idle: 7356, Running: 4, Completed: 8660 [ 11h 31m ] +INFO: Idle: 7355, Running: 4, Completed: 8661 [ 11h 32m ] +INFO: Idle: 7354, Running: 4, Completed: 8662 [ 11h 32m ] +INFO: Idle: 7352, Running: 4, Completed: 8664 [ 11h 32m ] +INFO: Idle: 7351, Running: 4, Completed: 8665 [ 11h 32m ] +INFO: Idle: 7350, Running: 4, Completed: 8666 [ 11h 32m ] +INFO: Idle: 7349, Running: 4, Completed: 8667 [ 11h 32m ] +INFO: Idle: 7347, Running: 4, Completed: 8669 [ 11h 32m ] +INFO: Idle: 7345, Running: 4, Completed: 8671 [ 11h 32m ] +INFO: Idle: 7342, Running: 4, Completed: 8674 [ 11h 32m ] +INFO: Idle: 7341, Running: 4, Completed: 8675 [ 11h 32m ] +INFO: Idle: 7339, Running: 4, Completed: 8677 [ 11h 33m ] +INFO: Idle: 7337, Running: 4, Completed: 8679 [ 11h 33m ] +INFO: Idle: 7336, Running: 4, Completed: 8680 [ 11h 33m ] +INFO: Idle: 7335, Running: 4, Completed: 8681 [ 11h 33m ] +INFO: Idle: 7333, Running: 4, Completed: 8683 [ 11h 33m ] +INFO: Idle: 7331, Running: 4, Completed: 8685 [ 11h 33m ] +INFO: Idle: 7329, Running: 4, Completed: 8687 [ 11h 33m ] +INFO: Idle: 7327, Running: 4, Completed: 8689 [ 11h 34m ] +INFO: Idle: 7325, Running: 4, Completed: 8691 [ 11h 34m ] +INFO: Idle: 7323, Running: 4, Completed: 8693 [ 11h 34m ] +INFO: Idle: 7322, Running: 4, Completed: 8694 [ 11h 34m ] +INFO: Idle: 7321, Running: 4, Completed: 8695 [ 11h 34m ] +INFO: Idle: 7319, Running: 4, Completed: 8697 [ 11h 34m ] +INFO: Idle: 7318, Running: 4, Completed: 8698 [ 11h 34m ] +INFO: Idle: 7317, Running: 4, Completed: 8699 [ 11h 35m ] +INFO: Idle: 7315, Running: 4, Completed: 8701 [ 11h 35m ] +INFO: Idle: 7313, Running: 4, Completed: 8703 [ 11h 35m ] +INFO: Idle: 7311, Running: 4, Completed: 8705 [ 11h 35m ] +INFO: Idle: 7310, Running: 4, Completed: 8706 [ 11h 35m ] +INFO: Idle: 7308, Running: 4, Completed: 8708 [ 11h 35m ] +INFO: Idle: 7307, Running: 4, Completed: 8709 [ 11h 36m ] +INFO: Idle: 7304, Running: 4, Completed: 8712 [ 11h 36m ] +INFO: Idle: 7303, Running: 4, Completed: 8713 [ 11h 36m ] +INFO: Idle: 7300, Running: 4, Completed: 8716 [ 11h 36m ] +INFO: Idle: 7299, Running: 4, Completed: 8717 [ 11h 36m ] +INFO: Idle: 7298, Running: 4, Completed: 8718 [ 11h 36m ] +INFO: Idle: 7297, Running: 4, Completed: 8719 [ 11h 36m ] +INFO: Idle: 7295, Running: 4, Completed: 8721 [ 11h 36m ] +INFO: Idle: 7294, Running: 4, Completed: 8722 [ 11h 37m ] +INFO: Idle: 7293, Running: 4, Completed: 8723 [ 11h 37m ] +INFO: Idle: 7291, Running: 4, Completed: 8725 [ 11h 37m ] +INFO: Idle: 7289, Running: 4, Completed: 8727 [ 11h 37m ] +INFO: Idle: 7287, Running: 4, Completed: 8729 [ 11h 37m ] +INFO: Idle: 7286, Running: 4, Completed: 8730 [ 11h 37m ] +INFO: Idle: 7285, Running: 4, Completed: 8731 [ 11h 37m ] +INFO: Idle: 7284, Running: 4, Completed: 8732 [ 11h 38m ] +INFO: Idle: 7282, Running: 4, Completed: 8734 [ 11h 38m ] +INFO: Idle: 7281, Running: 4, Completed: 8735 [ 11h 38m ] +INFO: Idle: 7279, Running: 4, Completed: 8737 [ 11h 38m ] +INFO: Idle: 7278, Running: 4, Completed: 8738 [ 11h 38m ] +INFO: Idle: 7277, Running: 4, Completed: 8739 [ 11h 38m ] +INFO: Idle: 7274, Running: 4, Completed: 8742 [ 11h 38m ] +INFO: Idle: 7273, Running: 4, Completed: 8743 [ 11h 39m ] +INFO: Idle: 7271, Running: 4, Completed: 8745 [ 11h 39m ] +INFO: Idle: 7270, Running: 4, Completed: 8746 [ 11h 39m ] +INFO: Idle: 7269, Running: 4, Completed: 8747 [ 11h 39m ] +INFO: Idle: 7266, Running: 4, Completed: 8750 [ 11h 39m ] +INFO: Idle: 7265, Running: 4, Completed: 8751 [ 11h 39m ] +INFO: Idle: 7262, Running: 4, Completed: 8754 [ 11h 40m ] +INFO: Idle: 7261, Running: 4, Completed: 8755 [ 11h 40m ] +INFO: Idle: 7260, Running: 4, Completed: 8756 [ 11h 40m ] +INFO: Idle: 7259, Running: 4, Completed: 8757 [ 11h 40m ] +INFO: Idle: 7258, Running: 4, Completed: 8758 [ 11h 40m ] +INFO: Idle: 7257, Running: 4, Completed: 8759 [ 11h 40m ] +INFO: Idle: 7256, Running: 4, Completed: 8760 [ 11h 40m ] +INFO: Idle: 7254, Running: 4, Completed: 8762 [ 11h 40m ] +INFO: Idle: 7253, Running: 4, Completed: 8763 [ 11h 41m ] +INFO: Idle: 7252, Running: 4, Completed: 8764 [ 11h 41m ] +INFO: Idle: 7250, Running: 4, Completed: 8766 [ 11h 41m ] +INFO: Idle: 7249, Running: 4, Completed: 8767 [ 11h 41m ] +INFO: Idle: 7247, Running: 4, Completed: 8769 [ 11h 41m ] +INFO: Idle: 7245, Running: 4, Completed: 8771 [ 11h 41m ] +INFO: Idle: 7242, Running: 4, Completed: 8774 [ 11h 41m ] +INFO: Idle: 7241, Running: 4, Completed: 8775 [ 11h 42m ] +INFO: Idle: 7238, Running: 4, Completed: 8778 [ 11h 42m ] +INFO: Idle: 7237, Running: 4, Completed: 8779 [ 11h 42m ] +INFO: Idle: 7236, Running: 4, Completed: 8780 [ 11h 42m ] +INFO: Idle: 7235, Running: 4, Completed: 8781 [ 11h 42m ] +INFO: Idle: 7233, Running: 4, Completed: 8783 [ 11h 42m ] +INFO: Idle: 7232, Running: 4, Completed: 8784 [ 11h 42m ] +INFO: Idle: 7231, Running: 4, Completed: 8785 [ 11h 42m ] +INFO: Idle: 7229, Running: 4, Completed: 8787 [ 11h 43m ] +INFO: Idle: 7226, Running: 4, Completed: 8790 [ 11h 43m ] +INFO: Idle: 7225, Running: 4, Completed: 8791 [ 11h 43m ] +INFO: Idle: 7222, Running: 4, Completed: 8794 [ 11h 43m ] +INFO: Idle: 7221, Running: 4, Completed: 8795 [ 11h 43m ] +INFO: Idle: 7218, Running: 4, Completed: 8798 [ 11h 43m ] +INFO: Idle: 7217, Running: 4, Completed: 8799 [ 11h 43m ] +INFO: Idle: 7215, Running: 4, Completed: 8801 [ 11h 44m ] +INFO: Idle: 7214, Running: 4, Completed: 8802 [ 11h 44m ] +INFO: Idle: 7213, Running: 4, Completed: 8803 [ 11h 44m ] +INFO: Idle: 7211, Running: 4, Completed: 8805 [ 11h 44m ] +INFO: Idle: 7210, Running: 4, Completed: 8806 [ 11h 44m ] +INFO: Idle: 7209, Running: 4, Completed: 8807 [ 11h 44m ] +INFO: Idle: 7208, Running: 4, Completed: 8808 [ 11h 44m ] +INFO: Idle: 7207, Running: 4, Completed: 8809 [ 11h 44m ] +INFO: Idle: 7205, Running: 4, Completed: 8811 [ 11h 44m ] +INFO: Idle: 7204, Running: 4, Completed: 8812 [ 11h 44m ] +INFO: Idle: 7203, Running: 4, Completed: 8813 [ 11h 44m ] +INFO: Idle: 7201, Running: 4, Completed: 8815 [ 11h 45m ] +INFO: Idle: 7199, Running: 4, Completed: 8817 [ 11h 45m ] +INFO: Idle: 7197, Running: 4, Completed: 8819 [ 11h 45m ] +INFO: Idle: 7195, Running: 4, Completed: 8821 [ 11h 45m ] +INFO: Idle: 7194, Running: 4, Completed: 8822 [ 11h 45m ] +INFO: Idle: 7193, Running: 4, Completed: 8823 [ 11h 45m ] +INFO: Idle: 7191, Running: 4, Completed: 8825 [ 11h 46m ] +INFO: Idle: 7190, Running: 4, Completed: 8826 [ 11h 46m ] +INFO: Idle: 7189, Running: 4, Completed: 8827 [ 11h 46m ] +INFO: Idle: 7187, Running: 4, Completed: 8829 [ 11h 46m ] +INFO: Idle: 7184, Running: 4, Completed: 8832 [ 11h 46m ] +INFO: Idle: 7183, Running: 4, Completed: 8833 [ 11h 46m ] +INFO: Idle: 7182, Running: 4, Completed: 8834 [ 11h 46m ] +INFO: Idle: 7180, Running: 4, Completed: 8836 [ 11h 47m ] +INFO: Idle: 7179, Running: 4, Completed: 8837 [ 11h 47m ] +INFO: Idle: 7176, Running: 4, Completed: 8840 [ 11h 47m ] +INFO: Idle: 7175, Running: 4, Completed: 8841 [ 11h 47m ] +INFO: Idle: 7172, Running: 4, Completed: 8844 [ 11h 47m ] +INFO: Idle: 7171, Running: 4, Completed: 8845 [ 11h 47m ] +INFO: Idle: 7170, Running: 4, Completed: 8846 [ 11h 47m ] +INFO: Idle: 7169, Running: 4, Completed: 8847 [ 11h 48m ] +INFO: Idle: 7168, Running: 4, Completed: 8848 [ 11h 48m ] +INFO: Idle: 7167, Running: 4, Completed: 8849 [ 11h 48m ] +INFO: Idle: 7166, Running: 4, Completed: 8850 [ 11h 48m ] +INFO: Idle: 7165, Running: 4, Completed: 8851 [ 11h 48m ] +INFO: Idle: 7164, Running: 4, Completed: 8852 [ 11h 48m ] +INFO: Idle: 7163, Running: 4, Completed: 8853 [ 11h 48m ] +INFO: Idle: 7161, Running: 4, Completed: 8855 [ 11h 48m ] +INFO: Idle: 7159, Running: 4, Completed: 8857 [ 11h 48m ] +INFO: Idle: 7158, Running: 4, Completed: 8858 [ 11h 49m ] +INFO: Idle: 7157, Running: 4, Completed: 8859 [ 11h 49m ] +INFO: Idle: 7156, Running: 4, Completed: 8860 [ 11h 49m ] +INFO: Idle: 7154, Running: 4, Completed: 8862 [ 11h 49m ] +INFO: Idle: 7153, Running: 4, Completed: 8863 [ 11h 49m ] +INFO: Idle: 7151, Running: 4, Completed: 8865 [ 11h 49m ] +INFO: Idle: 7150, Running: 4, Completed: 8866 [ 11h 49m ] +INFO: Idle: 7149, Running: 4, Completed: 8867 [ 11h 50m ] +INFO: Idle: 7146, Running: 4, Completed: 8870 [ 11h 50m ] +INFO: Idle: 7145, Running: 4, Completed: 8871 [ 11h 50m ] +INFO: Idle: 7143, Running: 4, Completed: 8873 [ 11h 50m ] +INFO: Idle: 7142, Running: 4, Completed: 8874 [ 11h 50m ] +INFO: Idle: 7141, Running: 4, Completed: 8875 [ 11h 50m ] +INFO: Idle: 7138, Running: 4, Completed: 8878 [ 11h 50m ] +INFO: Idle: 7137, Running: 4, Completed: 8879 [ 11h 51m ] +INFO: Idle: 7134, Running: 4, Completed: 8882 [ 11h 51m ] +INFO: Idle: 7133, Running: 4, Completed: 8883 [ 11h 51m ] +INFO: Idle: 7131, Running: 4, Completed: 8885 [ 11h 51m ] +INFO: Idle: 7130, Running: 4, Completed: 8886 [ 11h 51m ] +INFO: Idle: 7129, Running: 4, Completed: 8887 [ 11h 51m ] +INFO: Idle: 7128, Running: 4, Completed: 8888 [ 11h 51m ] +INFO: Idle: 7126, Running: 4, Completed: 8890 [ 11h 52m ] +INFO: Idle: 7125, Running: 4, Completed: 8891 [ 11h 52m ] +INFO: Idle: 7123, Running: 4, Completed: 8893 [ 11h 52m ] +INFO: Idle: 7122, Running: 4, Completed: 8894 [ 11h 52m ] +INFO: Idle: 7121, Running: 4, Completed: 8895 [ 11h 52m ] +INFO: Idle: 7119, Running: 4, Completed: 8897 [ 11h 52m ] +INFO: Idle: 7117, Running: 4, Completed: 8899 [ 11h 52m ] +INFO: Idle: 7115, Running: 4, Completed: 8901 [ 11h 53m ] +INFO: Idle: 7113, Running: 4, Completed: 8903 [ 11h 53m ] +INFO: Idle: 7110, Running: 4, Completed: 8906 [ 11h 53m ] +INFO: Idle: 7109, Running: 4, Completed: 8907 [ 11h 53m ] +INFO: Idle: 7108, Running: 4, Completed: 8908 [ 11h 53m ] +INFO: Idle: 7105, Running: 4, Completed: 8911 [ 11h 53m ] +INFO: Idle: 7103, Running: 4, Completed: 8913 [ 11h 53m ] +INFO: Idle: 7102, Running: 4, Completed: 8914 [ 11h 54m ] +INFO: Idle: 7101, Running: 4, Completed: 8915 [ 11h 54m ] +INFO: Idle: 7098, Running: 4, Completed: 8918 [ 11h 54m ] +INFO: Idle: 7097, Running: 4, Completed: 8919 [ 11h 54m ] +INFO: Idle: 7094, Running: 4, Completed: 8922 [ 11h 54m ] +INFO: Idle: 7093, Running: 4, Completed: 8923 [ 11h 54m ] +INFO: Idle: 7090, Running: 4, Completed: 8926 [ 11h 54m ] +INFO: Idle: 7089, Running: 4, Completed: 8927 [ 11h 55m ] +INFO: Idle: 7087, Running: 4, Completed: 8929 [ 11h 55m ] +INFO: Idle: 7085, Running: 4, Completed: 8931 [ 11h 55m ] +INFO: Idle: 7083, Running: 4, Completed: 8933 [ 11h 55m ] +INFO: Idle: 7082, Running: 4, Completed: 8934 [ 11h 55m ] +INFO: Idle: 7081, Running: 4, Completed: 8935 [ 11h 55m ] +INFO: Idle: 7079, Running: 4, Completed: 8937 [ 11h 55m ] +INFO: Idle: 7077, Running: 4, Completed: 8939 [ 11h 55m ] +INFO: Idle: 7075, Running: 4, Completed: 8941 [ 11h 56m ] +INFO: Idle: 7073, Running: 4, Completed: 8943 [ 11h 56m ] +INFO: Idle: 7071, Running: 4, Completed: 8945 [ 11h 56m ] +INFO: Idle: 7070, Running: 4, Completed: 8946 [ 11h 56m ] +INFO: Idle: 7069, Running: 4, Completed: 8947 [ 11h 56m ] +INFO: Idle: 7067, Running: 4, Completed: 8949 [ 11h 56m ] +INFO: Idle: 7066, Running: 4, Completed: 8950 [ 11h 57m ] +INFO: Idle: 7065, Running: 4, Completed: 8951 [ 11h 57m ] +INFO: Idle: 7063, Running: 4, Completed: 8953 [ 11h 57m ] +INFO: Idle: 7062, Running: 4, Completed: 8954 [ 11h 57m ] +INFO: Idle: 7061, Running: 4, Completed: 8955 [ 11h 57m ] +INFO: Idle: 7059, Running: 4, Completed: 8957 [ 11h 57m ] +INFO: Idle: 7056, Running: 4, Completed: 8960 [ 11h 57m ] +INFO: Idle: 7055, Running: 4, Completed: 8961 [ 11h 58m ] +INFO: Idle: 7054, Running: 4, Completed: 8962 [ 11h 58m ] +INFO: Idle: 7052, Running: 4, Completed: 8964 [ 11h 58m ] +INFO: Idle: 7051, Running: 4, Completed: 8965 [ 11h 58m ] +INFO: Idle: 7049, Running: 4, Completed: 8967 [ 11h 58m ] +INFO: Idle: 7048, Running: 4, Completed: 8968 [ 11h 58m ] +INFO: Idle: 7047, Running: 4, Completed: 8969 [ 11h 58m ] +INFO: Idle: 7046, Running: 4, Completed: 8970 [ 11h 58m ] +INFO: Idle: 7043, Running: 4, Completed: 8973 [ 11h 59m ] +INFO: Idle: 7041, Running: 4, Completed: 8975 [ 11h 59m ] +INFO: Idle: 7039, Running: 4, Completed: 8977 [ 11h 59m ] +INFO: Idle: 7038, Running: 4, Completed: 8978 [ 11h 59m ] +INFO: Idle: 7037, Running: 4, Completed: 8979 [ 11h 59m ] +INFO: Idle: 7035, Running: 4, Completed: 8981 [ 11h 59m ] +INFO: Idle: 7034, Running: 4, Completed: 8982 [ 11h 59m ] +INFO: Idle: 7031, Running: 4, Completed: 8985 [ 12h 0m ] +INFO: Idle: 7030, Running: 4, Completed: 8986 [ 12h 0m ] +INFO: Idle: 7029, Running: 4, Completed: 8987 [ 12h 0m ] +INFO: Idle: 7028, Running: 4, Completed: 8988 [ 12h 0m ] +INFO: Idle: 7027, Running: 4, Completed: 8989 [ 12h 0m ] +INFO: Idle: 7026, Running: 4, Completed: 8990 [ 12h 0m ] +INFO: Idle: 7024, Running: 4, Completed: 8992 [ 12h 0m ] +INFO: Idle: 7023, Running: 4, Completed: 8993 [ 12h 0m ] +INFO: Idle: 7022, Running: 4, Completed: 8994 [ 12h 0m ] +INFO: Idle: 7021, Running: 4, Completed: 8995 [ 12h 1m ] +INFO: Idle: 7020, Running: 4, Completed: 8996 [ 12h 1m ] +INFO: Idle: 7019, Running: 4, Completed: 8997 [ 12h 1m ] +INFO: Idle: 7017, Running: 4, Completed: 8999 [ 12h 1m ] +INFO: Idle: 7015, Running: 4, Completed: 9001 [ 12h 1m ] +INFO: Idle: 7014, Running: 4, Completed: 9002 [ 12h 1m ] +INFO: Idle: 7013, Running: 4, Completed: 9003 [ 12h 1m ] +INFO: Idle: 7012, Running: 4, Completed: 9004 [ 12h 1m ] +INFO: Idle: 7010, Running: 4, Completed: 9006 [ 12h 2m ] +INFO: Idle: 7009, Running: 4, Completed: 9007 [ 12h 2m ] +INFO: Idle: 7008, Running: 4, Completed: 9008 [ 12h 2m ] +INFO: Idle: 7006, Running: 4, Completed: 9010 [ 12h 2m ] +INFO: Idle: 7005, Running: 4, Completed: 9011 [ 12h 2m ] +INFO: Idle: 7004, Running: 4, Completed: 9012 [ 12h 2m ] +INFO: Idle: 7003, Running: 4, Completed: 9013 [ 12h 2m ] +INFO: Idle: 7002, Running: 4, Completed: 9014 [ 12h 2m ] +INFO: Idle: 7001, Running: 4, Completed: 9015 [ 12h 2m ] +INFO: Idle: 7000, Running: 4, Completed: 9016 [ 12h 2m ] +INFO: Idle: 6998, Running: 4, Completed: 9018 [ 12h 3m ] +INFO: Idle: 6997, Running: 4, Completed: 9019 [ 12h 3m ] +INFO: Idle: 6995, Running: 4, Completed: 9021 [ 12h 3m ] +INFO: Idle: 6994, Running: 4, Completed: 9022 [ 12h 3m ] +INFO: Idle: 6993, Running: 4, Completed: 9023 [ 12h 3m ] +INFO: Idle: 6992, Running: 4, Completed: 9024 [ 12h 3m ] +INFO: Idle: 6990, Running: 4, Completed: 9026 [ 12h 3m ] +INFO: Idle: 6989, Running: 4, Completed: 9027 [ 12h 4m ] +INFO: Idle: 6987, Running: 4, Completed: 9029 [ 12h 4m ] +INFO: Idle: 6986, Running: 4, Completed: 9030 [ 12h 4m ] +INFO: Idle: 6985, Running: 4, Completed: 9031 [ 12h 4m ] +INFO: Idle: 6982, Running: 4, Completed: 9034 [ 12h 4m ] +INFO: Idle: 6981, Running: 4, Completed: 9035 [ 12h 4m ] +INFO: Idle: 6980, Running: 4, Completed: 9036 [ 12h 4m ] +INFO: Idle: 6978, Running: 4, Completed: 9038 [ 12h 4m ] +INFO: Idle: 6977, Running: 4, Completed: 9039 [ 12h 5m ] +INFO: Idle: 6976, Running: 4, Completed: 9040 [ 12h 5m ] +INFO: Idle: 6975, Running: 4, Completed: 9041 [ 12h 5m ] +INFO: Idle: 6973, Running: 4, Completed: 9043 [ 12h 5m ] +INFO: Idle: 6972, Running: 4, Completed: 9044 [ 12h 5m ] +INFO: Idle: 6970, Running: 4, Completed: 9046 [ 12h 5m ] +INFO: Idle: 6969, Running: 4, Completed: 9047 [ 12h 5m ] +INFO: Idle: 6968, Running: 4, Completed: 9048 [ 12h 5m ] +INFO: Idle: 6965, Running: 4, Completed: 9051 [ 12h 6m ] +INFO: Idle: 6964, Running: 4, Completed: 9052 [ 12h 6m ] +INFO: Idle: 6962, Running: 4, Completed: 9054 [ 12h 6m ] +INFO: Idle: 6961, Running: 4, Completed: 9055 [ 12h 6m ] +INFO: Idle: 6957, Running: 4, Completed: 9059 [ 12h 6m ] +INFO: Idle: 6955, Running: 4, Completed: 9061 [ 12h 6m ] +INFO: Idle: 6953, Running: 4, Completed: 9063 [ 12h 6m ] +INFO: Idle: 6951, Running: 4, Completed: 9065 [ 12h 6m ] +INFO: Idle: 6949, Running: 4, Completed: 9067 [ 12h 7m ] +INFO: Idle: 6947, Running: 4, Completed: 9069 [ 12h 7m ] +INFO: Idle: 6945, Running: 4, Completed: 9071 [ 12h 7m ] +INFO: Idle: 6943, Running: 4, Completed: 9073 [ 12h 7m ] +INFO: Idle: 6942, Running: 4, Completed: 9074 [ 12h 7m ] +INFO: Idle: 6941, Running: 4, Completed: 9075 [ 12h 7m ] +INFO: Idle: 6939, Running: 4, Completed: 9077 [ 12h 8m ] +INFO: Idle: 6938, Running: 4, Completed: 9078 [ 12h 8m ] +INFO: Idle: 6937, Running: 4, Completed: 9079 [ 12h 8m ] +INFO: Idle: 6935, Running: 4, Completed: 9081 [ 12h 8m ] +INFO: Idle: 6934, Running: 4, Completed: 9082 [ 12h 8m ] +INFO: Idle: 6933, Running: 4, Completed: 9083 [ 12h 8m ] +INFO: Idle: 6931, Running: 4, Completed: 9085 [ 12h 8m ] +INFO: Idle: 6929, Running: 4, Completed: 9087 [ 12h 8m ] +INFO: Idle: 6928, Running: 4, Completed: 9088 [ 12h 8m ] +INFO: Idle: 6927, Running: 4, Completed: 9089 [ 12h 9m ] +INFO: Idle: 6926, Running: 4, Completed: 9090 [ 12h 9m ] +INFO: Idle: 6925, Running: 4, Completed: 9091 [ 12h 9m ] +INFO: Idle: 6924, Running: 4, Completed: 9092 [ 12h 9m ] +INFO: Idle: 6923, Running: 4, Completed: 9093 [ 12h 9m ] +INFO: Idle: 6922, Running: 4, Completed: 9094 [ 12h 9m ] +INFO: Idle: 6920, Running: 4, Completed: 9096 [ 12h 9m ] +INFO: Idle: 6919, Running: 4, Completed: 9097 [ 12h 9m ] +INFO: Idle: 6918, Running: 4, Completed: 9098 [ 12h 10m ] +INFO: Idle: 6916, Running: 4, Completed: 9100 [ 12h 10m ] +INFO: Idle: 6915, Running: 4, Completed: 9101 [ 12h 10m ] +INFO: Idle: 6913, Running: 4, Completed: 9103 [ 12h 10m ] +INFO: Idle: 6911, Running: 4, Completed: 9105 [ 12h 10m ] +INFO: Idle: 6910, Running: 4, Completed: 9106 [ 12h 10m ] +INFO: Idle: 6909, Running: 4, Completed: 9107 [ 12h 10m ] +INFO: Idle: 6907, Running: 4, Completed: 9109 [ 12h 11m ] +INFO: Idle: 6906, Running: 4, Completed: 9110 [ 12h 11m ] +INFO: Idle: 6903, Running: 4, Completed: 9113 [ 12h 11m ] +INFO: Idle: 6901, Running: 4, Completed: 9115 [ 12h 11m ] +INFO: Idle: 6900, Running: 4, Completed: 9116 [ 12h 11m ] +INFO: Idle: 6898, Running: 4, Completed: 9118 [ 12h 11m ] +INFO: Idle: 6896, Running: 4, Completed: 9120 [ 12h 11m ] +INFO: Idle: 6895, Running: 4, Completed: 9121 [ 12h 12m ] +INFO: Idle: 6893, Running: 4, Completed: 9123 [ 12h 12m ] +INFO: Idle: 6892, Running: 4, Completed: 9124 [ 12h 12m ] +INFO: Idle: 6889, Running: 4, Completed: 9127 [ 12h 12m ] +INFO: Idle: 6888, Running: 4, Completed: 9128 [ 12h 12m ] +INFO: Idle: 6887, Running: 4, Completed: 9129 [ 12h 12m ] +INFO: Idle: 6886, Running: 4, Completed: 9130 [ 12h 12m ] +INFO: Idle: 6885, Running: 4, Completed: 9131 [ 12h 13m ] +INFO: Idle: 6884, Running: 4, Completed: 9132 [ 12h 13m ] +INFO: Idle: 6883, Running: 4, Completed: 9133 [ 12h 13m ] +INFO: Idle: 6882, Running: 4, Completed: 9134 [ 12h 13m ] +INFO: Idle: 6881, Running: 4, Completed: 9135 [ 12h 13m ] +INFO: Idle: 6880, Running: 4, Completed: 9136 [ 12h 13m ] +INFO: Idle: 6877, Running: 4, Completed: 9139 [ 12h 13m ] +INFO: Idle: 6876, Running: 4, Completed: 9140 [ 12h 13m ] +INFO: Idle: 6875, Running: 4, Completed: 9141 [ 12h 13m ] +INFO: Idle: 6874, Running: 4, Completed: 9142 [ 12h 14m ] +INFO: Idle: 6872, Running: 4, Completed: 9144 [ 12h 14m ] +INFO: Idle: 6871, Running: 4, Completed: 9145 [ 12h 14m ] +INFO: Idle: 6870, Running: 4, Completed: 9146 [ 12h 14m ] +INFO: Idle: 6869, Running: 4, Completed: 9147 [ 12h 14m ] +INFO: Idle: 6867, Running: 4, Completed: 9149 [ 12h 14m ] +INFO: Idle: 6866, Running: 4, Completed: 9150 [ 12h 14m ] +INFO: Idle: 6864, Running: 4, Completed: 9152 [ 12h 14m ] +INFO: Idle: 6862, Running: 4, Completed: 9154 [ 12h 15m ] +INFO: Idle: 6861, Running: 4, Completed: 9155 [ 12h 15m ] +INFO: Idle: 6859, Running: 4, Completed: 9157 [ 12h 15m ] +INFO: Idle: 6857, Running: 4, Completed: 9159 [ 12h 15m ] +INFO: Idle: 6854, Running: 4, Completed: 9162 [ 12h 15m ] +INFO: Idle: 6853, Running: 4, Completed: 9163 [ 12h 15m ] +INFO: Idle: 6852, Running: 4, Completed: 9164 [ 12h 16m ] +INFO: Idle: 6850, Running: 4, Completed: 9166 [ 12h 16m ] +INFO: Idle: 6849, Running: 4, Completed: 9167 [ 12h 16m ] +INFO: Idle: 6848, Running: 4, Completed: 9168 [ 12h 16m ] +INFO: Idle: 6847, Running: 4, Completed: 9169 [ 12h 16m ] +INFO: Idle: 6845, Running: 4, Completed: 9171 [ 12h 16m ] +INFO: Idle: 6844, Running: 4, Completed: 9172 [ 12h 16m ] +INFO: Idle: 6842, Running: 4, Completed: 9174 [ 12h 16m ] +INFO: Idle: 6841, Running: 4, Completed: 9175 [ 12h 16m ] +INFO: Idle: 6840, Running: 4, Completed: 9176 [ 12h 16m ] +INFO: Idle: 6837, Running: 4, Completed: 9179 [ 12h 17m ] +INFO: Idle: 6836, Running: 4, Completed: 9180 [ 12h 17m ] +INFO: Idle: 6834, Running: 4, Completed: 9182 [ 12h 17m ] +INFO: Idle: 6833, Running: 4, Completed: 9183 [ 12h 17m ] +INFO: Idle: 6829, Running: 4, Completed: 9187 [ 12h 17m ] +INFO: Idle: 6827, Running: 4, Completed: 9189 [ 12h 17m ] +INFO: Idle: 6825, Running: 4, Completed: 9191 [ 12h 18m ] +INFO: Idle: 6822, Running: 4, Completed: 9194 [ 12h 18m ] +INFO: Idle: 6821, Running: 4, Completed: 9195 [ 12h 18m ] +INFO: Idle: 6819, Running: 4, Completed: 9197 [ 12h 18m ] +INFO: Idle: 6817, Running: 4, Completed: 9199 [ 12h 18m ] +INFO: Idle: 6815, Running: 4, Completed: 9201 [ 12h 18m ] +INFO: Idle: 6813, Running: 4, Completed: 9203 [ 12h 19m ] +INFO: Idle: 6811, Running: 4, Completed: 9205 [ 12h 19m ] +INFO: Idle: 6809, Running: 4, Completed: 9207 [ 12h 19m ] +INFO: Idle: 6807, Running: 4, Completed: 9209 [ 12h 19m ] +INFO: Idle: 6805, Running: 4, Completed: 9211 [ 12h 19m ] +INFO: Idle: 6803, Running: 4, Completed: 9213 [ 12h 20m ] +INFO: Idle: 6800, Running: 4, Completed: 9216 [ 12h 20m ] +INFO: Idle: 6799, Running: 4, Completed: 9217 [ 12h 20m ] +INFO: Idle: 6798, Running: 4, Completed: 9218 [ 12h 20m ] +INFO: Idle: 6796, Running: 4, Completed: 9220 [ 12h 20m ] +INFO: Idle: 6795, Running: 4, Completed: 9221 [ 12h 20m ] +INFO: Idle: 6793, Running: 4, Completed: 9223 [ 12h 21m ] +INFO: Idle: 6791, Running: 4, Completed: 9225 [ 12h 21m ] +INFO: Idle: 6790, Running: 4, Completed: 9226 [ 12h 21m ] +INFO: Idle: 6789, Running: 4, Completed: 9227 [ 12h 21m ] +INFO: Idle: 6788, Running: 4, Completed: 9228 [ 12h 21m ] +INFO: Idle: 6786, Running: 4, Completed: 9230 [ 12h 21m ] +INFO: Idle: 6785, Running: 4, Completed: 9231 [ 12h 21m ] +INFO: Idle: 6783, Running: 4, Completed: 9233 [ 12h 21m ] +INFO: Idle: 6782, Running: 4, Completed: 9234 [ 12h 22m ] +INFO: Idle: 6781, Running: 4, Completed: 9235 [ 12h 22m ] +INFO: Idle: 6779, Running: 4, Completed: 9237 [ 12h 22m ] +INFO: Idle: 6777, Running: 4, Completed: 9239 [ 12h 22m ] +INFO: Idle: 6776, Running: 4, Completed: 9240 [ 12h 22m ] +INFO: Idle: 6775, Running: 4, Completed: 9241 [ 12h 22m ] +INFO: Idle: 6774, Running: 4, Completed: 9242 [ 12h 22m ] +INFO: Idle: 6773, Running: 4, Completed: 9243 [ 12h 23m ] +INFO: Idle: 6772, Running: 4, Completed: 9244 [ 12h 23m ] +INFO: Idle: 6770, Running: 4, Completed: 9246 [ 12h 23m ] +INFO: Idle: 6769, Running: 4, Completed: 9247 [ 12h 23m ] +INFO: Idle: 6767, Running: 4, Completed: 9249 [ 12h 23m ] +INFO: Idle: 6765, Running: 4, Completed: 9251 [ 12h 23m ] +INFO: Idle: 6762, Running: 4, Completed: 9254 [ 12h 23m ] +INFO: Idle: 6761, Running: 4, Completed: 9255 [ 12h 24m ] +INFO: Idle: 6759, Running: 4, Completed: 9257 [ 12h 24m ] +INFO: Idle: 6758, Running: 4, Completed: 9258 [ 12h 24m ] +INFO: Idle: 6757, Running: 4, Completed: 9259 [ 12h 24m ] +INFO: Idle: 6755, Running: 4, Completed: 9261 [ 12h 24m ] +INFO: Idle: 6754, Running: 4, Completed: 9262 [ 12h 24m ] +INFO: Idle: 6753, Running: 4, Completed: 9263 [ 12h 24m ] +INFO: Idle: 6751, Running: 4, Completed: 9265 [ 12h 24m ] +INFO: Idle: 6749, Running: 4, Completed: 9267 [ 12h 25m ] +INFO: Idle: 6747, Running: 4, Completed: 9269 [ 12h 25m ] +INFO: Idle: 6745, Running: 4, Completed: 9271 [ 12h 25m ] +INFO: Idle: 6744, Running: 4, Completed: 9272 [ 12h 25m ] +INFO: Idle: 6743, Running: 4, Completed: 9273 [ 12h 25m ] +INFO: Idle: 6741, Running: 4, Completed: 9275 [ 12h 25m ] +INFO: Idle: 6740, Running: 4, Completed: 9276 [ 12h 25m ] +INFO: Idle: 6737, Running: 4, Completed: 9279 [ 12h 26m ] +INFO: Idle: 6735, Running: 4, Completed: 9281 [ 12h 26m ] +INFO: Idle: 6733, Running: 4, Completed: 9283 [ 12h 26m ] +INFO: Idle: 6731, Running: 4, Completed: 9285 [ 12h 26m ] +INFO: Idle: 6730, Running: 4, Completed: 9286 [ 12h 26m ] +INFO: Idle: 6729, Running: 4, Completed: 9287 [ 12h 26m ] +INFO: Idle: 6728, Running: 4, Completed: 9288 [ 12h 26m ] +INFO: Idle: 6727, Running: 4, Completed: 9289 [ 12h 27m ] +INFO: Idle: 6725, Running: 4, Completed: 9291 [ 12h 27m ] +INFO: Idle: 6723, Running: 4, Completed: 9293 [ 12h 27m ] +INFO: Idle: 6722, Running: 4, Completed: 9294 [ 12h 27m ] +INFO: Idle: 6721, Running: 4, Completed: 9295 [ 12h 27m ] +INFO: Idle: 6720, Running: 4, Completed: 9296 [ 12h 27m ] +INFO: Idle: 6719, Running: 4, Completed: 9297 [ 12h 27m ] +INFO: Idle: 6717, Running: 4, Completed: 9299 [ 12h 27m ] +INFO: Idle: 6716, Running: 4, Completed: 9300 [ 12h 27m ] +INFO: Idle: 6714, Running: 4, Completed: 9302 [ 12h 28m ] +INFO: Idle: 6712, Running: 4, Completed: 9304 [ 12h 28m ] +INFO: Idle: 6710, Running: 4, Completed: 9306 [ 12h 28m ] +INFO: Idle: 6709, Running: 4, Completed: 9307 [ 12h 28m ] +INFO: Idle: 6708, Running: 4, Completed: 9308 [ 12h 28m ] +INFO: Idle: 6707, Running: 4, Completed: 9309 [ 12h 28m ] +INFO: Idle: 6705, Running: 4, Completed: 9311 [ 12h 28m ] +INFO: Idle: 6703, Running: 4, Completed: 9313 [ 12h 28m ] +INFO: Idle: 6701, Running: 4, Completed: 9315 [ 12h 29m ] +INFO: Idle: 6699, Running: 4, Completed: 9317 [ 12h 29m ] +INFO: Idle: 6696, Running: 4, Completed: 9320 [ 12h 29m ] +INFO: Idle: 6695, Running: 4, Completed: 9321 [ 12h 29m ] +INFO: Idle: 6692, Running: 4, Completed: 9324 [ 12h 29m ] +INFO: Idle: 6691, Running: 4, Completed: 9325 [ 12h 29m ] +INFO: Idle: 6688, Running: 4, Completed: 9328 [ 12h 29m ] +INFO: Idle: 6687, Running: 4, Completed: 9329 [ 12h 30m ] +INFO: Idle: 6684, Running: 4, Completed: 9332 [ 12h 30m ] +INFO: Idle: 6683, Running: 4, Completed: 9333 [ 12h 30m ] +INFO: Idle: 6682, Running: 4, Completed: 9334 [ 12h 30m ] +INFO: Idle: 6679, Running: 4, Completed: 9337 [ 12h 30m ] +INFO: Idle: 6678, Running: 4, Completed: 9338 [ 12h 30m ] +INFO: Idle: 6676, Running: 4, Completed: 9340 [ 12h 31m ] +INFO: Idle: 6675, Running: 4, Completed: 9341 [ 12h 31m ] +INFO: Idle: 6673, Running: 4, Completed: 9343 [ 12h 31m ] +INFO: Idle: 6671, Running: 4, Completed: 9345 [ 12h 31m ] +INFO: Idle: 6668, Running: 4, Completed: 9348 [ 12h 31m ] +INFO: Idle: 6667, Running: 4, Completed: 9349 [ 12h 31m ] +INFO: Idle: 6665, Running: 4, Completed: 9351 [ 12h 31m ] +INFO: Idle: 6664, Running: 4, Completed: 9352 [ 12h 32m ] +INFO: Idle: 6663, Running: 4, Completed: 9353 [ 12h 32m ] +INFO: Idle: 6662, Running: 4, Completed: 9354 [ 12h 32m ] +INFO: Idle: 6660, Running: 4, Completed: 9356 [ 12h 32m ] +INFO: Idle: 6659, Running: 4, Completed: 9357 [ 12h 32m ] +INFO: Idle: 6658, Running: 4, Completed: 9358 [ 12h 32m ] +INFO: Idle: 6656, Running: 4, Completed: 9360 [ 12h 32m ] +INFO: Idle: 6655, Running: 4, Completed: 9361 [ 12h 32m ] +INFO: Idle: 6653, Running: 4, Completed: 9363 [ 12h 33m ] +INFO: Idle: 6652, Running: 4, Completed: 9364 [ 12h 33m ] +INFO: Idle: 6650, Running: 4, Completed: 9366 [ 12h 33m ] +INFO: Idle: 6649, Running: 4, Completed: 9367 [ 12h 33m ] +INFO: Idle: 6648, Running: 4, Completed: 9368 [ 12h 33m ] +INFO: Idle: 6647, Running: 4, Completed: 9369 [ 12h 33m ] +INFO: Idle: 6646, Running: 4, Completed: 9370 [ 12h 33m ] +INFO: Idle: 6645, Running: 4, Completed: 9371 [ 12h 33m ] +INFO: Idle: 6643, Running: 4, Completed: 9373 [ 12h 33m ] +INFO: Idle: 6642, Running: 4, Completed: 9374 [ 12h 34m ] +INFO: Idle: 6641, Running: 4, Completed: 9375 [ 12h 34m ] +INFO: Idle: 6638, Running: 4, Completed: 9378 [ 12h 34m ] +INFO: Idle: 6637, Running: 4, Completed: 9379 [ 12h 34m ] +INFO: Idle: 6636, Running: 4, Completed: 9380 [ 12h 34m ] +INFO: Idle: 6634, Running: 4, Completed: 9382 [ 12h 34m ] +INFO: Idle: 6633, Running: 4, Completed: 9383 [ 12h 34m ] +INFO: Idle: 6632, Running: 4, Completed: 9384 [ 12h 34m ] +INFO: Idle: 6630, Running: 4, Completed: 9386 [ 12h 35m ] +INFO: Idle: 6628, Running: 4, Completed: 9388 [ 12h 35m ] +INFO: Idle: 6626, Running: 4, Completed: 9390 [ 12h 35m ] +INFO: Idle: 6624, Running: 4, Completed: 9392 [ 12h 35m ] +INFO: Idle: 6622, Running: 4, Completed: 9394 [ 12h 35m ] +INFO: Idle: 6618, Running: 4, Completed: 9398 [ 12h 35m ] +INFO: Idle: 6617, Running: 4, Completed: 9399 [ 12h 35m ] +INFO: Idle: 6616, Running: 4, Completed: 9400 [ 12h 36m ] +INFO: Idle: 6614, Running: 4, Completed: 9402 [ 12h 36m ] +INFO: Idle: 6613, Running: 4, Completed: 9403 [ 12h 36m ] +INFO: Idle: 6609, Running: 4, Completed: 9407 [ 12h 36m ] +INFO: Idle: 6606, Running: 4, Completed: 9410 [ 12h 36m ] +INFO: Idle: 6605, Running: 4, Completed: 9411 [ 12h 36m ] +INFO: Idle: 6601, Running: 4, Completed: 9415 [ 12h 37m ] +INFO: Idle: 6599, Running: 4, Completed: 9417 [ 12h 37m ] +INFO: Idle: 6597, Running: 4, Completed: 9419 [ 12h 37m ] +INFO: Idle: 6595, Running: 4, Completed: 9421 [ 12h 37m ] +INFO: Idle: 6594, Running: 4, Completed: 9422 [ 12h 37m ] +INFO: Idle: 6593, Running: 4, Completed: 9423 [ 12h 38m ] +INFO: Idle: 6591, Running: 4, Completed: 9425 [ 12h 38m ] +INFO: Idle: 6590, Running: 4, Completed: 9426 [ 12h 38m ] +INFO: Idle: 6589, Running: 4, Completed: 9427 [ 12h 38m ] +INFO: Idle: 6588, Running: 4, Completed: 9428 [ 12h 38m ] +INFO: Idle: 6587, Running: 4, Completed: 9429 [ 12h 38m ] +INFO: Idle: 6586, Running: 4, Completed: 9430 [ 12h 38m ] +INFO: Idle: 6585, Running: 4, Completed: 9431 [ 12h 38m ] +INFO: Idle: 6584, Running: 4, Completed: 9432 [ 12h 38m ] +INFO: Idle: 6582, Running: 4, Completed: 9434 [ 12h 38m ] +INFO: Idle: 6581, Running: 4, Completed: 9435 [ 12h 39m ] +INFO: Idle: 6579, Running: 4, Completed: 9437 [ 12h 39m ] +INFO: Idle: 6577, Running: 4, Completed: 9439 [ 12h 39m ] +INFO: Idle: 6574, Running: 4, Completed: 9442 [ 12h 39m ] +INFO: Idle: 6573, Running: 4, Completed: 9443 [ 12h 39m ] +INFO: Idle: 6571, Running: 4, Completed: 9445 [ 12h 39m ] +INFO: Idle: 6570, Running: 4, Completed: 9446 [ 12h 39m ] +INFO: Idle: 6569, Running: 4, Completed: 9447 [ 12h 40m ] +INFO: Idle: 6568, Running: 4, Completed: 9448 [ 12h 40m ] +INFO: Idle: 6566, Running: 4, Completed: 9450 [ 12h 40m ] +INFO: Idle: 6565, Running: 4, Completed: 9451 [ 12h 40m ] +INFO: Idle: 6562, Running: 4, Completed: 9454 [ 12h 40m ] +INFO: Idle: 6561, Running: 4, Completed: 9455 [ 12h 40m ] +INFO: Idle: 6560, Running: 4, Completed: 9456 [ 12h 40m ] +INFO: Idle: 6558, Running: 4, Completed: 9458 [ 12h 41m ] +INFO: Idle: 6556, Running: 4, Completed: 9460 [ 12h 41m ] +INFO: Idle: 6554, Running: 4, Completed: 9462 [ 12h 41m ] +INFO: Idle: 6552, Running: 4, Completed: 9464 [ 12h 41m ] +INFO: Idle: 6551, Running: 4, Completed: 9465 [ 12h 41m ] +INFO: Idle: 6549, Running: 4, Completed: 9467 [ 12h 41m ] +INFO: Idle: 6548, Running: 4, Completed: 9468 [ 12h 41m ] +INFO: Idle: 6547, Running: 4, Completed: 9469 [ 12h 41m ] +INFO: Idle: 6545, Running: 4, Completed: 9471 [ 12h 41m ] +INFO: Idle: 6544, Running: 4, Completed: 9472 [ 12h 42m ] +INFO: Idle: 6542, Running: 4, Completed: 9474 [ 12h 42m ] +INFO: Idle: 6541, Running: 4, Completed: 9475 [ 12h 42m ] +INFO: Idle: 6538, Running: 4, Completed: 9478 [ 12h 42m ] +INFO: Idle: 6537, Running: 4, Completed: 9479 [ 12h 42m ] +INFO: Idle: 6535, Running: 4, Completed: 9481 [ 12h 42m ] +INFO: Idle: 6533, Running: 4, Completed: 9483 [ 12h 42m ] +INFO: Idle: 6531, Running: 4, Completed: 9485 [ 12h 42m ] +INFO: Idle: 6529, Running: 4, Completed: 9487 [ 12h 43m ] +INFO: Idle: 6527, Running: 4, Completed: 9489 [ 12h 43m ] +INFO: Idle: 6525, Running: 4, Completed: 9491 [ 12h 43m ] +INFO: Idle: 6523, Running: 4, Completed: 9493 [ 12h 43m ] +INFO: Idle: 6521, Running: 4, Completed: 9495 [ 12h 43m ] +INFO: Idle: 6519, Running: 4, Completed: 9497 [ 12h 44m ] +INFO: Idle: 6518, Running: 4, Completed: 9498 [ 12h 44m ] +INFO: Idle: 6517, Running: 4, Completed: 9499 [ 12h 44m ] +INFO: Idle: 6515, Running: 4, Completed: 9501 [ 12h 44m ] +INFO: Idle: 6514, Running: 4, Completed: 9502 [ 12h 44m ] +INFO: Idle: 6513, Running: 4, Completed: 9503 [ 12h 44m ] +INFO: Idle: 6511, Running: 4, Completed: 9505 [ 12h 44m ] +INFO: Idle: 6510, Running: 4, Completed: 9506 [ 12h 44m ] +INFO: Idle: 6507, Running: 4, Completed: 9509 [ 12h 45m ] +INFO: Idle: 6505, Running: 4, Completed: 9511 [ 12h 45m ] +INFO: Idle: 6504, Running: 4, Completed: 9512 [ 12h 45m ] +INFO: Idle: 6503, Running: 4, Completed: 9513 [ 12h 45m ] +INFO: Idle: 6501, Running: 4, Completed: 9515 [ 12h 45m ] +INFO: Idle: 6500, Running: 4, Completed: 9516 [ 12h 45m ] +INFO: Idle: 6499, Running: 4, Completed: 9517 [ 12h 45m ] +INFO: Idle: 6497, Running: 4, Completed: 9519 [ 12h 45m ] +INFO: Idle: 6495, Running: 4, Completed: 9521 [ 12h 46m ] +INFO: Idle: 6494, Running: 4, Completed: 9522 [ 12h 46m ] +INFO: Idle: 6493, Running: 4, Completed: 9523 [ 12h 46m ] +INFO: Idle: 6492, Running: 4, Completed: 9524 [ 12h 46m ] +INFO: Idle: 6491, Running: 4, Completed: 9525 [ 12h 46m ] +INFO: Idle: 6490, Running: 4, Completed: 9526 [ 12h 46m ] +INFO: Idle: 6489, Running: 4, Completed: 9527 [ 12h 46m ] +INFO: Idle: 6487, Running: 4, Completed: 9529 [ 12h 46m ] +INFO: Idle: 6486, Running: 4, Completed: 9530 [ 12h 47m ] +INFO: Idle: 6485, Running: 4, Completed: 9531 [ 12h 47m ] +INFO: Idle: 6483, Running: 4, Completed: 9533 [ 12h 47m ] +INFO: Idle: 6482, Running: 4, Completed: 9534 [ 12h 47m ] +INFO: Idle: 6481, Running: 4, Completed: 9535 [ 12h 47m ] +INFO: Idle: 6480, Running: 4, Completed: 9536 [ 12h 47m ] +INFO: Idle: 6478, Running: 4, Completed: 9538 [ 12h 47m ] +INFO: Idle: 6477, Running: 4, Completed: 9539 [ 12h 47m ] +INFO: Idle: 6475, Running: 4, Completed: 9541 [ 12h 47m ] +INFO: Idle: 6474, Running: 4, Completed: 9542 [ 12h 48m ] +INFO: Idle: 6473, Running: 4, Completed: 9543 [ 12h 48m ] +INFO: Idle: 6470, Running: 4, Completed: 9546 [ 12h 48m ] +INFO: Idle: 6469, Running: 4, Completed: 9547 [ 12h 48m ] +INFO: Idle: 6467, Running: 4, Completed: 9549 [ 12h 48m ] +INFO: Idle: 6466, Running: 4, Completed: 9550 [ 12h 48m ] +INFO: Idle: 6465, Running: 4, Completed: 9551 [ 12h 49m ] +INFO: Idle: 6462, Running: 4, Completed: 9554 [ 12h 49m ] +INFO: Idle: 6461, Running: 4, Completed: 9555 [ 12h 49m ] +INFO: Idle: 6458, Running: 4, Completed: 9558 [ 12h 49m ] +INFO: Idle: 6457, Running: 4, Completed: 9559 [ 12h 49m ] +INFO: Idle: 6455, Running: 4, Completed: 9561 [ 12h 49m ] +INFO: Idle: 6454, Running: 4, Completed: 9562 [ 12h 50m ] +INFO: Idle: 6453, Running: 4, Completed: 9563 [ 12h 50m ] +INFO: Idle: 6452, Running: 4, Completed: 9564 [ 12h 50m ] +INFO: Idle: 6450, Running: 4, Completed: 9566 [ 12h 50m ] +INFO: Idle: 6449, Running: 4, Completed: 9567 [ 12h 50m ] +INFO: Idle: 6448, Running: 4, Completed: 9568 [ 12h 50m ] +INFO: Idle: 6446, Running: 4, Completed: 9570 [ 12h 50m ] +INFO: Idle: 6445, Running: 4, Completed: 9571 [ 12h 50m ] +INFO: Idle: 6443, Running: 4, Completed: 9573 [ 12h 50m ] +INFO: Idle: 6441, Running: 4, Completed: 9575 [ 12h 51m ] +INFO: Idle: 6437, Running: 4, Completed: 9579 [ 12h 51m ] +INFO: Idle: 6435, Running: 4, Completed: 9581 [ 12h 51m ] +INFO: Idle: 6433, Running: 4, Completed: 9583 [ 12h 51m ] +INFO: Idle: 6432, Running: 4, Completed: 9584 [ 12h 51m ] +INFO: Idle: 6430, Running: 4, Completed: 9586 [ 12h 52m ] +INFO: Idle: 6429, Running: 4, Completed: 9587 [ 12h 52m ] +INFO: Idle: 6427, Running: 4, Completed: 9589 [ 12h 52m ] +INFO: Idle: 6426, Running: 4, Completed: 9590 [ 12h 52m ] +INFO: Idle: 6425, Running: 4, Completed: 9591 [ 12h 52m ] +INFO: Idle: 6423, Running: 4, Completed: 9593 [ 12h 52m ] +INFO: Idle: 6422, Running: 4, Completed: 9594 [ 12h 52m ] +INFO: Idle: 6421, Running: 4, Completed: 9595 [ 12h 52m ] +INFO: Idle: 6418, Running: 4, Completed: 9598 [ 12h 52m ] +INFO: Idle: 6417, Running: 4, Completed: 9599 [ 12h 53m ] +INFO: Idle: 6414, Running: 4, Completed: 9602 [ 12h 53m ] +INFO: Idle: 6413, Running: 4, Completed: 9603 [ 12h 53m ] +INFO: Idle: 6411, Running: 4, Completed: 9605 [ 12h 53m ] +INFO: Idle: 6409, Running: 4, Completed: 9607 [ 12h 53m ] +INFO: Idle: 6407, Running: 4, Completed: 9609 [ 12h 53m ] +INFO: Idle: 6405, Running: 4, Completed: 9611 [ 12h 54m ] +INFO: Idle: 6402, Running: 4, Completed: 9614 [ 12h 54m ] +INFO: Idle: 6401, Running: 4, Completed: 9615 [ 12h 54m ] +INFO: Idle: 6399, Running: 4, Completed: 9617 [ 12h 54m ] +INFO: Idle: 6398, Running: 4, Completed: 9618 [ 12h 54m ] +INFO: Idle: 6397, Running: 4, Completed: 9619 [ 12h 54m ] +INFO: Idle: 6395, Running: 4, Completed: 9621 [ 12h 54m ] +INFO: Idle: 6394, Running: 4, Completed: 9622 [ 12h 54m ] +INFO: Idle: 6393, Running: 4, Completed: 9623 [ 12h 55m ] +INFO: Idle: 6392, Running: 4, Completed: 9624 [ 12h 55m ] +INFO: Idle: 6391, Running: 4, Completed: 9625 [ 12h 55m ] +INFO: Idle: 6390, Running: 4, Completed: 9626 [ 12h 55m ] +INFO: Idle: 6389, Running: 4, Completed: 9627 [ 12h 55m ] +INFO: Idle: 6388, Running: 4, Completed: 9628 [ 12h 55m ] +INFO: Idle: 6387, Running: 4, Completed: 9629 [ 12h 55m ] +INFO: Idle: 6386, Running: 4, Completed: 9630 [ 12h 55m ] +INFO: Idle: 6385, Running: 4, Completed: 9631 [ 12h 55m ] +INFO: Idle: 6383, Running: 4, Completed: 9633 [ 12h 56m ] +INFO: Idle: 6382, Running: 4, Completed: 9634 [ 12h 56m ] +INFO: Idle: 6381, Running: 4, Completed: 9635 [ 12h 56m ] +INFO: Idle: 6379, Running: 4, Completed: 9637 [ 12h 56m ] +INFO: Idle: 6378, Running: 4, Completed: 9638 [ 12h 56m ] +INFO: Idle: 6377, Running: 4, Completed: 9639 [ 12h 56m ] +INFO: Idle: 6376, Running: 4, Completed: 9640 [ 12h 56m ] +INFO: Idle: 6375, Running: 4, Completed: 9641 [ 12h 56m ] +INFO: Idle: 6374, Running: 4, Completed: 9642 [ 12h 56m ] +INFO: Idle: 6372, Running: 4, Completed: 9644 [ 12h 57m ] +INFO: Idle: 6371, Running: 4, Completed: 9645 [ 12h 57m ] +INFO: Idle: 6370, Running: 4, Completed: 9646 [ 12h 57m ] +INFO: Idle: 6368, Running: 4, Completed: 9648 [ 12h 57m ] +INFO: Idle: 6367, Running: 4, Completed: 9649 [ 12h 57m ] +INFO: Idle: 6366, Running: 4, Completed: 9650 [ 12h 57m ] +INFO: Idle: 6365, Running: 4, Completed: 9651 [ 12h 57m ] +INFO: Idle: 6364, Running: 4, Completed: 9652 [ 12h 57m ] +INFO: Idle: 6362, Running: 4, Completed: 9654 [ 12h 57m ] +INFO: Idle: 6361, Running: 4, Completed: 9655 [ 12h 58m ] +INFO: Idle: 6360, Running: 4, Completed: 9656 [ 12h 58m ] +INFO: Idle: 6359, Running: 4, Completed: 9657 [ 12h 58m ] +INFO: Idle: 6357, Running: 4, Completed: 9659 [ 12h 58m ] +INFO: Idle: 6356, Running: 4, Completed: 9660 [ 12h 58m ] +INFO: Idle: 6355, Running: 4, Completed: 9661 [ 12h 58m ] +INFO: Idle: 6354, Running: 4, Completed: 9662 [ 12h 58m ] +INFO: Idle: 6353, Running: 4, Completed: 9663 [ 12h 58m ] +INFO: Idle: 6351, Running: 4, Completed: 9665 [ 12h 59m ] +INFO: Idle: 6350, Running: 4, Completed: 9666 [ 12h 59m ] +INFO: Idle: 6349, Running: 4, Completed: 9667 [ 12h 59m ] +INFO: Idle: 6347, Running: 4, Completed: 9669 [ 12h 59m ] +INFO: Idle: 6346, Running: 4, Completed: 9670 [ 12h 59m ] +INFO: Idle: 6345, Running: 4, Completed: 9671 [ 12h 59m ] +INFO: Idle: 6342, Running: 4, Completed: 9674 [ 12h 59m ] +INFO: Idle: 6341, Running: 4, Completed: 9675 [ 13h 0m ] +INFO: Idle: 6339, Running: 4, Completed: 9677 [ 13h 0m ] +INFO: Idle: 6338, Running: 4, Completed: 9678 [ 13h 0m ] +INFO: Idle: 6337, Running: 4, Completed: 9679 [ 13h 0m ] +INFO: Idle: 6336, Running: 4, Completed: 9680 [ 13h 0m ] +INFO: Idle: 6335, Running: 4, Completed: 9681 [ 13h 0m ] +INFO: Idle: 6333, Running: 4, Completed: 9683 [ 13h 0m ] +INFO: Idle: 6331, Running: 4, Completed: 9685 [ 13h 0m ] +INFO: Idle: 6330, Running: 4, Completed: 9686 [ 13h 0m ] +INFO: Idle: 6329, Running: 4, Completed: 9687 [ 13h 1m ] +INFO: Idle: 6328, Running: 4, Completed: 9688 [ 13h 1m ] +INFO: Idle: 6327, Running: 4, Completed: 9689 [ 13h 1m ] +INFO: Idle: 6326, Running: 4, Completed: 9690 [ 13h 1m ] +INFO: Idle: 6324, Running: 4, Completed: 9692 [ 13h 1m ] +INFO: Idle: 6323, Running: 4, Completed: 9693 [ 13h 1m ] +INFO: Idle: 6322, Running: 4, Completed: 9694 [ 13h 1m ] +INFO: Idle: 6321, Running: 4, Completed: 9695 [ 13h 1m ] +INFO: Idle: 6320, Running: 4, Completed: 9696 [ 13h 1m ] +INFO: Idle: 6318, Running: 4, Completed: 9698 [ 13h 2m ] +INFO: Idle: 6317, Running: 4, Completed: 9699 [ 13h 2m ] +INFO: Idle: 6315, Running: 4, Completed: 9701 [ 13h 2m ] +INFO: Idle: 6313, Running: 4, Completed: 9703 [ 13h 2m ] +INFO: Idle: 6310, Running: 4, Completed: 9706 [ 13h 2m ] +INFO: Idle: 6309, Running: 4, Completed: 9707 [ 13h 2m ] +INFO: Idle: 6306, Running: 4, Completed: 9710 [ 13h 2m ] +INFO: Idle: 6305, Running: 4, Completed: 9711 [ 13h 3m ] +INFO: Idle: 6304, Running: 4, Completed: 9712 [ 13h 3m ] +INFO: Idle: 6303, Running: 4, Completed: 9713 [ 13h 3m ] +INFO: Idle: 6301, Running: 4, Completed: 9715 [ 13h 3m ] +INFO: Idle: 6300, Running: 4, Completed: 9716 [ 13h 3m ] +INFO: Idle: 6299, Running: 4, Completed: 9717 [ 13h 3m ] +INFO: Idle: 6297, Running: 4, Completed: 9719 [ 13h 3m ] +INFO: Idle: 6294, Running: 4, Completed: 9722 [ 13h 4m ] +INFO: Idle: 6293, Running: 4, Completed: 9723 [ 13h 4m ] +INFO: Idle: 6290, Running: 4, Completed: 9726 [ 13h 4m ] +INFO: Idle: 6289, Running: 4, Completed: 9727 [ 13h 4m ] +INFO: Idle: 6288, Running: 4, Completed: 9728 [ 13h 4m ] +INFO: Idle: 6286, Running: 4, Completed: 9730 [ 13h 4m ] +INFO: Idle: 6285, Running: 4, Completed: 9731 [ 13h 4m ] +INFO: Idle: 6283, Running: 4, Completed: 9733 [ 13h 4m ] +INFO: Idle: 6281, Running: 4, Completed: 9735 [ 13h 4m ] +INFO: Idle: 6280, Running: 4, Completed: 9736 [ 13h 5m ] +INFO: Idle: 6279, Running: 4, Completed: 9737 [ 13h 5m ] +INFO: Idle: 6278, Running: 4, Completed: 9738 [ 13h 5m ] +INFO: Idle: 6277, Running: 4, Completed: 9739 [ 13h 5m ] +INFO: Idle: 6275, Running: 4, Completed: 9741 [ 13h 5m ] +INFO: Idle: 6273, Running: 4, Completed: 9743 [ 13h 5m ] +INFO: Idle: 6271, Running: 4, Completed: 9745 [ 13h 5m ] +INFO: Idle: 6269, Running: 4, Completed: 9747 [ 13h 6m ] +INFO: Idle: 6267, Running: 4, Completed: 9749 [ 13h 6m ] +INFO: Idle: 6266, Running: 4, Completed: 9750 [ 13h 6m ] +INFO: Idle: 6265, Running: 4, Completed: 9751 [ 13h 6m ] +INFO: Idle: 6263, Running: 4, Completed: 9753 [ 13h 6m ] +INFO: Idle: 6262, Running: 4, Completed: 9754 [ 13h 6m ] +INFO: Idle: 6261, Running: 4, Completed: 9755 [ 13h 6m ] +INFO: Idle: 6260, Running: 4, Completed: 9756 [ 13h 6m ] +INFO: Idle: 6259, Running: 4, Completed: 9757 [ 13h 6m ] +INFO: Idle: 6258, Running: 4, Completed: 9758 [ 13h 6m ] +INFO: Idle: 6257, Running: 4, Completed: 9759 [ 13h 7m ] +INFO: Idle: 6255, Running: 4, Completed: 9761 [ 13h 7m ] +INFO: Idle: 6253, Running: 4, Completed: 9763 [ 13h 7m ] +INFO: Idle: 6251, Running: 4, Completed: 9765 [ 13h 7m ] +INFO: Idle: 6250, Running: 4, Completed: 9766 [ 13h 7m ] +INFO: Idle: 6249, Running: 4, Completed: 9767 [ 13h 7m ] +INFO: Idle: 6248, Running: 4, Completed: 9768 [ 13h 7m ] +INFO: Idle: 6247, Running: 4, Completed: 9769 [ 13h 8m ] +INFO: Idle: 6246, Running: 4, Completed: 9770 [ 13h 8m ] +INFO: Idle: 6244, Running: 4, Completed: 9772 [ 13h 8m ] +INFO: Idle: 6243, Running: 4, Completed: 9773 [ 13h 8m ] +INFO: Idle: 6242, Running: 4, Completed: 9774 [ 13h 8m ] +INFO: Idle: 6239, Running: 4, Completed: 9777 [ 13h 8m ] +INFO: Idle: 6238, Running: 4, Completed: 9778 [ 13h 8m ] +INFO: Idle: 6237, Running: 4, Completed: 9779 [ 13h 8m ] +INFO: Idle: 6236, Running: 4, Completed: 9780 [ 13h 9m ] +INFO: Idle: 6235, Running: 4, Completed: 9781 [ 13h 9m ] +INFO: Idle: 6233, Running: 4, Completed: 9783 [ 13h 9m ] +INFO: Idle: 6232, Running: 4, Completed: 9784 [ 13h 9m ] +INFO: Idle: 6231, Running: 4, Completed: 9785 [ 13h 9m ] +INFO: Idle: 6229, Running: 4, Completed: 9787 [ 13h 9m ] +INFO: Idle: 6228, Running: 4, Completed: 9788 [ 13h 9m ] +INFO: Idle: 6227, Running: 4, Completed: 9789 [ 13h 9m ] +INFO: Idle: 6226, Running: 4, Completed: 9790 [ 13h 10m ] +INFO: Idle: 6225, Running: 4, Completed: 9791 [ 13h 10m ] +INFO: Idle: 6224, Running: 4, Completed: 9792 [ 13h 10m ] +INFO: Idle: 6223, Running: 4, Completed: 9793 [ 13h 10m ] +INFO: Idle: 6222, Running: 4, Completed: 9794 [ 13h 10m ] +INFO: Idle: 6221, Running: 4, Completed: 9795 [ 13h 10m ] +INFO: Idle: 6219, Running: 4, Completed: 9797 [ 13h 10m ] +INFO: Idle: 6218, Running: 4, Completed: 9798 [ 13h 10m ] +INFO: Idle: 6217, Running: 4, Completed: 9799 [ 13h 10m ] +INFO: Idle: 6214, Running: 4, Completed: 9802 [ 13h 11m ] +INFO: Idle: 6213, Running: 4, Completed: 9803 [ 13h 11m ] +INFO: Idle: 6211, Running: 4, Completed: 9805 [ 13h 11m ] +INFO: Idle: 6210, Running: 4, Completed: 9806 [ 13h 11m ] +INFO: Idle: 6209, Running: 4, Completed: 9807 [ 13h 11m ] +INFO: Idle: 6208, Running: 4, Completed: 9808 [ 13h 11m ] +INFO: Idle: 6206, Running: 4, Completed: 9810 [ 13h 11m ] +INFO: Idle: 6205, Running: 4, Completed: 9811 [ 13h 11m ] +INFO: Idle: 6203, Running: 4, Completed: 9813 [ 13h 12m ] +INFO: Idle: 6201, Running: 4, Completed: 9815 [ 13h 12m ] +INFO: Idle: 6200, Running: 4, Completed: 9816 [ 13h 12m ] +INFO: Idle: 6199, Running: 4, Completed: 9817 [ 13h 12m ] +INFO: Idle: 6198, Running: 4, Completed: 9818 [ 13h 12m ] +INFO: Idle: 6197, Running: 4, Completed: 9819 [ 13h 12m ] +INFO: Idle: 6196, Running: 4, Completed: 9820 [ 13h 12m ] +INFO: Idle: 6195, Running: 4, Completed: 9821 [ 13h 12m ] +INFO: Idle: 6193, Running: 4, Completed: 9823 [ 13h 13m ] +INFO: Idle: 6192, Running: 4, Completed: 9824 [ 13h 13m ] +INFO: Idle: 6189, Running: 4, Completed: 9827 [ 13h 13m ] +INFO: Idle: 6187, Running: 4, Completed: 9829 [ 13h 13m ] +INFO: Idle: 6185, Running: 4, Completed: 9831 [ 13h 13m ] +INFO: Idle: 6181, Running: 4, Completed: 9835 [ 13h 14m ] +INFO: Idle: 6178, Running: 4, Completed: 9838 [ 13h 14m ] +INFO: Idle: 6177, Running: 4, Completed: 9839 [ 13h 14m ] +INFO: Idle: 6176, Running: 4, Completed: 9840 [ 13h 14m ] +INFO: Idle: 6173, Running: 4, Completed: 9843 [ 13h 14m ] +INFO: Idle: 6171, Running: 4, Completed: 9845 [ 13h 14m ] +INFO: Idle: 6169, Running: 4, Completed: 9847 [ 13h 15m ] +INFO: Idle: 6166, Running: 4, Completed: 9850 [ 13h 15m ] +INFO: Idle: 6165, Running: 4, Completed: 9851 [ 13h 15m ] +INFO: Idle: 6162, Running: 4, Completed: 9854 [ 13h 15m ] +INFO: Idle: 6161, Running: 4, Completed: 9855 [ 13h 15m ] +INFO: Idle: 6159, Running: 4, Completed: 9857 [ 13h 15m ] +INFO: Idle: 6158, Running: 4, Completed: 9858 [ 13h 15m ] +INFO: Idle: 6157, Running: 4, Completed: 9859 [ 13h 15m ] +INFO: Idle: 6155, Running: 4, Completed: 9861 [ 13h 16m ] +INFO: Idle: 6153, Running: 4, Completed: 9863 [ 13h 16m ] +INFO: Idle: 6152, Running: 4, Completed: 9864 [ 13h 16m ] +INFO: Idle: 6151, Running: 4, Completed: 9865 [ 13h 16m ] +INFO: Idle: 6149, Running: 4, Completed: 9867 [ 13h 16m ] +INFO: Idle: 6148, Running: 4, Completed: 9868 [ 13h 16m ] +INFO: Idle: 6145, Running: 4, Completed: 9871 [ 13h 16m ] +INFO: Idle: 6143, Running: 4, Completed: 9873 [ 13h 16m ] +INFO: Idle: 6141, Running: 4, Completed: 9875 [ 13h 17m ] +INFO: Idle: 6139, Running: 4, Completed: 9877 [ 13h 17m ] +INFO: Idle: 6138, Running: 4, Completed: 9878 [ 13h 17m ] +INFO: Idle: 6137, Running: 4, Completed: 9879 [ 13h 17m ] +INFO: Idle: 6135, Running: 4, Completed: 9881 [ 13h 17m ] +INFO: Idle: 6134, Running: 4, Completed: 9882 [ 13h 17m ] +INFO: Idle: 6133, Running: 4, Completed: 9883 [ 13h 17m ] +INFO: Idle: 6132, Running: 4, Completed: 9884 [ 13h 17m ] +INFO: Idle: 6131, Running: 4, Completed: 9885 [ 13h 18m ] +INFO: Idle: 6130, Running: 4, Completed: 9886 [ 13h 18m ] +INFO: Idle: 6129, Running: 4, Completed: 9887 [ 13h 18m ] +INFO: Idle: 6127, Running: 4, Completed: 9889 [ 13h 18m ] +INFO: Idle: 6126, Running: 4, Completed: 9890 [ 13h 18m ] +INFO: Idle: 6125, Running: 4, Completed: 9891 [ 13h 18m ] +INFO: Idle: 6123, Running: 4, Completed: 9893 [ 13h 18m ] +INFO: Idle: 6122, Running: 4, Completed: 9894 [ 13h 18m ] +INFO: Idle: 6121, Running: 4, Completed: 9895 [ 13h 19m ] +INFO: Idle: 6120, Running: 4, Completed: 9896 [ 13h 19m ] +INFO: Idle: 6119, Running: 4, Completed: 9897 [ 13h 19m ] +INFO: Idle: 6118, Running: 4, Completed: 9898 [ 13h 19m ] +INFO: Idle: 6116, Running: 4, Completed: 9900 [ 13h 19m ] +INFO: Idle: 6115, Running: 4, Completed: 9901 [ 13h 19m ] +INFO: Idle: 6114, Running: 4, Completed: 9902 [ 13h 19m ] +INFO: Idle: 6112, Running: 4, Completed: 9904 [ 13h 19m ] +INFO: Idle: 6111, Running: 4, Completed: 9905 [ 13h 19m ] +INFO: Idle: 6110, Running: 4, Completed: 9906 [ 13h 19m ] +INFO: Idle: 6109, Running: 4, Completed: 9907 [ 13h 20m ] +INFO: Idle: 6108, Running: 4, Completed: 9908 [ 13h 20m ] +INFO: Idle: 6106, Running: 4, Completed: 9910 [ 13h 20m ] +INFO: Idle: 6104, Running: 4, Completed: 9912 [ 13h 20m ] +INFO: Idle: 6103, Running: 4, Completed: 9913 [ 13h 20m ] +INFO: Idle: 6101, Running: 4, Completed: 9915 [ 13h 20m ] +INFO: Idle: 6100, Running: 4, Completed: 9916 [ 13h 20m ] +INFO: Idle: 6099, Running: 4, Completed: 9917 [ 13h 21m ] +INFO: Idle: 6098, Running: 4, Completed: 9918 [ 13h 21m ] +INFO: Idle: 6097, Running: 4, Completed: 9919 [ 13h 21m ] +INFO: Idle: 6096, Running: 4, Completed: 9920 [ 13h 21m ] +INFO: Idle: 6094, Running: 4, Completed: 9922 [ 13h 21m ] +INFO: Idle: 6093, Running: 4, Completed: 9923 [ 13h 21m ] +INFO: Idle: 6091, Running: 4, Completed: 9925 [ 13h 21m ] +INFO: Idle: 6090, Running: 4, Completed: 9926 [ 13h 21m ] +INFO: Idle: 6089, Running: 4, Completed: 9927 [ 13h 22m ] +INFO: Idle: 6086, Running: 4, Completed: 9930 [ 13h 22m ] +INFO: Idle: 6085, Running: 4, Completed: 9931 [ 13h 22m ] +INFO: Idle: 6083, Running: 4, Completed: 9933 [ 13h 22m ] +INFO: Idle: 6082, Running: 4, Completed: 9934 [ 13h 22m ] +INFO: Idle: 6081, Running: 4, Completed: 9935 [ 13h 22m ] +INFO: Idle: 6080, Running: 4, Completed: 9936 [ 13h 22m ] +INFO: Idle: 6077, Running: 4, Completed: 9939 [ 13h 23m ] +INFO: Idle: 6075, Running: 4, Completed: 9941 [ 13h 23m ] +INFO: Idle: 6074, Running: 4, Completed: 9942 [ 13h 23m ] +INFO: Idle: 6073, Running: 4, Completed: 9943 [ 13h 23m ] +INFO: Idle: 6072, Running: 4, Completed: 9944 [ 13h 23m ] +INFO: Idle: 6071, Running: 4, Completed: 9945 [ 13h 23m ] +INFO: Idle: 6070, Running: 4, Completed: 9946 [ 13h 23m ] +INFO: Idle: 6069, Running: 4, Completed: 9947 [ 13h 23m ] +INFO: Idle: 6067, Running: 4, Completed: 9949 [ 13h 24m ] +INFO: Idle: 6066, Running: 4, Completed: 9950 [ 13h 24m ] +INFO: Idle: 6065, Running: 4, Completed: 9951 [ 13h 24m ] +INFO: Idle: 6063, Running: 4, Completed: 9953 [ 13h 24m ] +INFO: Idle: 6061, Running: 4, Completed: 9955 [ 13h 24m ] +INFO: Idle: 6059, Running: 4, Completed: 9957 [ 13h 24m ] +INFO: Idle: 6057, Running: 4, Completed: 9959 [ 13h 24m ] +INFO: Idle: 6054, Running: 4, Completed: 9962 [ 13h 25m ] +INFO: Idle: 6053, Running: 4, Completed: 9963 [ 13h 25m ] +INFO: Idle: 6050, Running: 4, Completed: 9966 [ 13h 25m ] +INFO: Idle: 6049, Running: 4, Completed: 9967 [ 13h 25m ] +INFO: Idle: 6048, Running: 4, Completed: 9968 [ 13h 25m ] +INFO: Idle: 6045, Running: 4, Completed: 9971 [ 13h 25m ] +INFO: Idle: 6043, Running: 4, Completed: 9973 [ 13h 26m ] +INFO: Idle: 6041, Running: 4, Completed: 9975 [ 13h 26m ] +INFO: Idle: 6038, Running: 4, Completed: 9978 [ 13h 26m ] +INFO: Idle: 6037, Running: 4, Completed: 9979 [ 13h 26m ] +INFO: Idle: 6034, Running: 4, Completed: 9982 [ 13h 26m ] +INFO: Idle: 6033, Running: 4, Completed: 9983 [ 13h 26m ] +INFO: Idle: 6032, Running: 4, Completed: 9984 [ 13h 26m ] +INFO: Idle: 6030, Running: 4, Completed: 9986 [ 13h 27m ] +INFO: Idle: 6029, Running: 4, Completed: 9987 [ 13h 27m ] +INFO: Idle: 6027, Running: 4, Completed: 9989 [ 13h 27m ] +INFO: Idle: 6025, Running: 4, Completed: 9991 [ 13h 27m ] +INFO: Idle: 6023, Running: 4, Completed: 9993 [ 13h 27m ] +INFO: Idle: 6021, Running: 4, Completed: 9995 [ 13h 27m ] +INFO: Idle: 6019, Running: 4, Completed: 9997 [ 13h 27m ] +INFO: Idle: 6017, Running: 4, Completed: 9999 [ 13h 28m ] +INFO: Idle: 6015, Running: 4, Completed: 10001 [ 13h 28m ] +INFO: Idle: 6013, Running: 4, Completed: 10003 [ 13h 28m ] +INFO: Idle: 6011, Running: 4, Completed: 10005 [ 13h 28m ] +INFO: Idle: 6009, Running: 4, Completed: 10007 [ 13h 28m ] +INFO: Idle: 6007, Running: 4, Completed: 10009 [ 13h 28m ] +INFO: Idle: 6006, Running: 4, Completed: 10010 [ 13h 29m ] +INFO: Idle: 6005, Running: 4, Completed: 10011 [ 13h 29m ] +INFO: Idle: 6004, Running: 4, Completed: 10012 [ 13h 29m ] +INFO: Idle: 6003, Running: 4, Completed: 10013 [ 13h 29m ] +INFO: Idle: 6001, Running: 4, Completed: 10015 [ 13h 29m ] +INFO: Idle: 5999, Running: 4, Completed: 10017 [ 13h 29m ] +INFO: Idle: 5997, Running: 4, Completed: 10019 [ 13h 29m ] +INFO: Idle: 5995, Running: 4, Completed: 10021 [ 13h 30m ] +INFO: Idle: 5994, Running: 4, Completed: 10022 [ 13h 30m ] +INFO: Idle: 5993, Running: 4, Completed: 10023 [ 13h 30m ] +INFO: Idle: 5992, Running: 4, Completed: 10024 [ 13h 30m ] +INFO: Idle: 5991, Running: 4, Completed: 10025 [ 13h 30m ] +INFO: Idle: 5990, Running: 4, Completed: 10026 [ 13h 30m ] +INFO: Idle: 5988, Running: 4, Completed: 10028 [ 13h 30m ] +INFO: Idle: 5987, Running: 4, Completed: 10029 [ 13h 30m ] +INFO: Idle: 5986, Running: 4, Completed: 10030 [ 13h 30m ] +INFO: Idle: 5983, Running: 4, Completed: 10033 [ 13h 31m ] +INFO: Idle: 5982, Running: 4, Completed: 10034 [ 13h 31m ] +INFO: Idle: 5981, Running: 4, Completed: 10035 [ 13h 31m ] +INFO: Idle: 5980, Running: 4, Completed: 10036 [ 13h 31m ] +INFO: Idle: 5979, Running: 4, Completed: 10037 [ 13h 31m ] +INFO: Idle: 5978, Running: 4, Completed: 10038 [ 13h 31m ] +INFO: Idle: 5976, Running: 4, Completed: 10040 [ 13h 31m ] +INFO: Idle: 5975, Running: 4, Completed: 10041 [ 13h 31m ] +INFO: Idle: 5973, Running: 4, Completed: 10043 [ 13h 32m ] +INFO: Idle: 5972, Running: 4, Completed: 10044 [ 13h 32m ] +INFO: Idle: 5971, Running: 4, Completed: 10045 [ 13h 32m ] +INFO: Idle: 5970, Running: 4, Completed: 10046 [ 13h 32m ] +INFO: Idle: 5969, Running: 4, Completed: 10047 [ 13h 32m ] +INFO: Idle: 5968, Running: 4, Completed: 10048 [ 13h 32m ] +INFO: Idle: 5966, Running: 4, Completed: 10050 [ 13h 32m ] +INFO: Idle: 5965, Running: 4, Completed: 10051 [ 13h 32m ] +INFO: Idle: 5963, Running: 4, Completed: 10053 [ 13h 32m ] +INFO: Idle: 5962, Running: 4, Completed: 10054 [ 13h 33m ] +INFO: Idle: 5961, Running: 4, Completed: 10055 [ 13h 33m ] +INFO: Idle: 5958, Running: 4, Completed: 10058 [ 13h 33m ] +INFO: Idle: 5957, Running: 4, Completed: 10059 [ 13h 33m ] +INFO: Idle: 5955, Running: 4, Completed: 10061 [ 13h 33m ] +INFO: Idle: 5954, Running: 4, Completed: 10062 [ 13h 33m ] +INFO: Idle: 5953, Running: 4, Completed: 10063 [ 13h 34m ] +INFO: Idle: 5952, Running: 4, Completed: 10064 [ 13h 34m ] +INFO: Idle: 5950, Running: 4, Completed: 10066 [ 13h 34m ] +INFO: Idle: 5949, Running: 4, Completed: 10067 [ 13h 34m ] +INFO: Idle: 5947, Running: 4, Completed: 10069 [ 13h 34m ] +INFO: Idle: 5946, Running: 4, Completed: 10070 [ 13h 34m ] +INFO: Idle: 5945, Running: 4, Completed: 10071 [ 13h 34m ] +INFO: Idle: 5944, Running: 4, Completed: 10072 [ 13h 34m ] +INFO: Idle: 5943, Running: 4, Completed: 10073 [ 13h 34m ] +INFO: Idle: 5942, Running: 4, Completed: 10074 [ 13h 35m ] +INFO: Idle: 5940, Running: 4, Completed: 10076 [ 13h 35m ] +INFO: Idle: 5939, Running: 4, Completed: 10077 [ 13h 35m ] +INFO: Idle: 5938, Running: 4, Completed: 10078 [ 13h 35m ] +INFO: Idle: 5937, Running: 4, Completed: 10079 [ 13h 35m ] +INFO: Idle: 5936, Running: 4, Completed: 10080 [ 13h 35m ] +INFO: Idle: 5933, Running: 4, Completed: 10083 [ 13h 35m ] +INFO: Idle: 5931, Running: 4, Completed: 10085 [ 13h 35m ] +INFO: Idle: 5929, Running: 4, Completed: 10087 [ 13h 36m ] +INFO: Idle: 5926, Running: 4, Completed: 10090 [ 13h 36m ] +INFO: Idle: 5925, Running: 4, Completed: 10091 [ 13h 36m ] +INFO: Idle: 5922, Running: 4, Completed: 10094 [ 13h 36m ] +INFO: Idle: 5921, Running: 4, Completed: 10095 [ 13h 36m ] +INFO: Idle: 5920, Running: 4, Completed: 10096 [ 13h 36m ] +INFO: Idle: 5917, Running: 4, Completed: 10099 [ 13h 37m ] +INFO: Idle: 5915, Running: 4, Completed: 10101 [ 13h 37m ] +INFO: Idle: 5913, Running: 4, Completed: 10103 [ 13h 37m ] +INFO: Idle: 5910, Running: 4, Completed: 10106 [ 13h 37m ] +INFO: Idle: 5909, Running: 4, Completed: 10107 [ 13h 37m ] +INFO: Idle: 5906, Running: 4, Completed: 10110 [ 13h 37m ] +INFO: Idle: 5905, Running: 4, Completed: 10111 [ 13h 38m ] +INFO: Idle: 5904, Running: 4, Completed: 10112 [ 13h 38m ] +INFO: Idle: 5902, Running: 4, Completed: 10114 [ 13h 38m ] +INFO: Idle: 5901, Running: 4, Completed: 10115 [ 13h 38m ] +INFO: Idle: 5899, Running: 4, Completed: 10117 [ 13h 38m ] +INFO: Idle: 5897, Running: 4, Completed: 10119 [ 13h 38m ] +INFO: Idle: 5896, Running: 4, Completed: 10120 [ 13h 38m ] +INFO: Idle: 5895, Running: 4, Completed: 10121 [ 13h 38m ] +INFO: Idle: 5893, Running: 4, Completed: 10123 [ 13h 39m ] +INFO: Idle: 5891, Running: 4, Completed: 10125 [ 13h 39m ] +INFO: Idle: 5889, Running: 4, Completed: 10127 [ 13h 39m ] +INFO: Idle: 5887, Running: 4, Completed: 10129 [ 13h 39m ] +INFO: Idle: 5885, Running: 4, Completed: 10131 [ 13h 39m ] +INFO: Idle: 5884, Running: 4, Completed: 10132 [ 13h 39m ] +INFO: Idle: 5883, Running: 4, Completed: 10133 [ 13h 40m ] +INFO: Idle: 5880, Running: 4, Completed: 10136 [ 13h 40m ] +INFO: Idle: 5879, Running: 4, Completed: 10137 [ 13h 40m ] +INFO: Idle: 5878, Running: 4, Completed: 10138 [ 13h 40m ] +INFO: Idle: 5876, Running: 4, Completed: 10140 [ 13h 40m ] +INFO: Idle: 5875, Running: 4, Completed: 10141 [ 13h 40m ] +INFO: Idle: 5874, Running: 4, Completed: 10142 [ 13h 40m ] +INFO: Idle: 5872, Running: 4, Completed: 10144 [ 13h 41m ] +INFO: Idle: 5871, Running: 4, Completed: 10145 [ 13h 41m ] +INFO: Idle: 5870, Running: 4, Completed: 10146 [ 13h 41m ] +INFO: Idle: 5869, Running: 4, Completed: 10147 [ 13h 41m ] +INFO: Idle: 5867, Running: 4, Completed: 10149 [ 13h 41m ] +INFO: Idle: 5866, Running: 4, Completed: 10150 [ 13h 41m ] +INFO: Idle: 5865, Running: 4, Completed: 10151 [ 13h 41m ] +INFO: Idle: 5863, Running: 4, Completed: 10153 [ 13h 41m ] +INFO: Idle: 5862, Running: 4, Completed: 10154 [ 13h 41m ] +INFO: Idle: 5861, Running: 4, Completed: 10155 [ 13h 41m ] +INFO: Idle: 5859, Running: 4, Completed: 10157 [ 13h 42m ] +INFO: Idle: 5858, Running: 4, Completed: 10158 [ 13h 42m ] +INFO: Idle: 5855, Running: 4, Completed: 10161 [ 13h 42m ] +INFO: Idle: 5854, Running: 4, Completed: 10162 [ 13h 42m ] +INFO: Idle: 5853, Running: 4, Completed: 10163 [ 13h 42m ] +INFO: Idle: 5850, Running: 4, Completed: 10166 [ 13h 43m ] +INFO: Idle: 5849, Running: 4, Completed: 10167 [ 13h 43m ] +INFO: Idle: 5847, Running: 4, Completed: 10169 [ 13h 43m ] +INFO: Idle: 5846, Running: 4, Completed: 10170 [ 13h 43m ] +INFO: Idle: 5845, Running: 4, Completed: 10171 [ 13h 43m ] +INFO: Idle: 5843, Running: 4, Completed: 10173 [ 13h 43m ] +INFO: Idle: 5842, Running: 4, Completed: 10174 [ 13h 43m ] +INFO: Idle: 5840, Running: 4, Completed: 10176 [ 13h 43m ] +INFO: Idle: 5838, Running: 4, Completed: 10178 [ 13h 44m ] +INFO: Idle: 5835, Running: 4, Completed: 10181 [ 13h 44m ] +INFO: Idle: 5834, Running: 4, Completed: 10182 [ 13h 44m ] +INFO: Idle: 5833, Running: 4, Completed: 10183 [ 13h 44m ] +INFO: Idle: 5830, Running: 4, Completed: 10186 [ 13h 44m ] +INFO: Idle: 5829, Running: 4, Completed: 10187 [ 13h 44m ] +INFO: Idle: 5827, Running: 4, Completed: 10189 [ 13h 45m ] +INFO: Idle: 5826, Running: 4, Completed: 10190 [ 13h 45m ] +INFO: Idle: 5824, Running: 4, Completed: 10192 [ 13h 45m ] +INFO: Idle: 5823, Running: 4, Completed: 10193 [ 13h 45m ] +INFO: Idle: 5822, Running: 4, Completed: 10194 [ 13h 45m ] +INFO: Idle: 5819, Running: 4, Completed: 10197 [ 13h 45m ] +INFO: Idle: 5818, Running: 4, Completed: 10198 [ 13h 45m ] +INFO: Idle: 5816, Running: 4, Completed: 10200 [ 13h 45m ] +INFO: Idle: 5814, Running: 4, Completed: 10202 [ 13h 46m ] +INFO: Idle: 5812, Running: 4, Completed: 10204 [ 13h 46m ] +INFO: Idle: 5810, Running: 4, Completed: 10206 [ 13h 46m ] +INFO: Idle: 5809, Running: 4, Completed: 10207 [ 13h 46m ] +INFO: Idle: 5808, Running: 4, Completed: 10208 [ 13h 46m ] +INFO: Idle: 5807, Running: 4, Completed: 10209 [ 13h 46m ] +INFO: Idle: 5805, Running: 4, Completed: 10211 [ 13h 47m ] +INFO: Idle: 5803, Running: 4, Completed: 10213 [ 13h 47m ] +INFO: Idle: 5802, Running: 4, Completed: 10214 [ 13h 47m ] +INFO: Idle: 5801, Running: 4, Completed: 10215 [ 13h 47m ] +INFO: Idle: 5799, Running: 4, Completed: 10217 [ 13h 47m ] +INFO: Idle: 5798, Running: 4, Completed: 10218 [ 13h 47m ] +INFO: Idle: 5797, Running: 4, Completed: 10219 [ 13h 47m ] +INFO: Idle: 5795, Running: 4, Completed: 10221 [ 13h 47m ] +INFO: Idle: 5794, Running: 4, Completed: 10222 [ 13h 48m ] +INFO: Idle: 5793, Running: 4, Completed: 10223 [ 13h 48m ] +INFO: Idle: 5792, Running: 4, Completed: 10224 [ 13h 48m ] +INFO: Idle: 5791, Running: 4, Completed: 10225 [ 13h 48m ] +INFO: Idle: 5789, Running: 4, Completed: 10227 [ 13h 48m ] +INFO: Idle: 5788, Running: 4, Completed: 10228 [ 13h 48m ] +INFO: Idle: 5787, Running: 4, Completed: 10229 [ 13h 48m ] +INFO: Idle: 5785, Running: 4, Completed: 10231 [ 13h 48m ] +INFO: Idle: 5784, Running: 4, Completed: 10232 [ 13h 49m ] +INFO: Idle: 5783, Running: 4, Completed: 10233 [ 13h 49m ] +INFO: Idle: 5781, Running: 4, Completed: 10235 [ 13h 49m ] +INFO: Idle: 5780, Running: 4, Completed: 10236 [ 13h 49m ] +INFO: Idle: 5779, Running: 4, Completed: 10237 [ 13h 49m ] +INFO: Idle: 5777, Running: 4, Completed: 10239 [ 13h 49m ] +INFO: Idle: 5776, Running: 4, Completed: 10240 [ 13h 49m ] +INFO: Idle: 5775, Running: 4, Completed: 10241 [ 13h 49m ] +INFO: Idle: 5774, Running: 4, Completed: 10242 [ 13h 49m ] +INFO: Idle: 5773, Running: 4, Completed: 10243 [ 13h 49m ] +INFO: Idle: 5771, Running: 4, Completed: 10245 [ 13h 49m ] +INFO: Idle: 5770, Running: 4, Completed: 10246 [ 13h 50m ] +INFO: Idle: 5769, Running: 4, Completed: 10247 [ 13h 50m ] +INFO: Idle: 5768, Running: 4, Completed: 10248 [ 13h 50m ] +INFO: Idle: 5767, Running: 4, Completed: 10249 [ 13h 50m ] +INFO: Idle: 5765, Running: 4, Completed: 10251 [ 13h 50m ] +INFO: Idle: 5764, Running: 4, Completed: 10252 [ 13h 50m ] +INFO: Idle: 5763, Running: 4, Completed: 10253 [ 13h 50m ] +INFO: Idle: 5762, Running: 4, Completed: 10254 [ 13h 50m ] +INFO: Idle: 5761, Running: 4, Completed: 10255 [ 13h 51m ] +INFO: Idle: 5760, Running: 4, Completed: 10256 [ 13h 51m ] +INFO: Idle: 5759, Running: 4, Completed: 10257 [ 13h 51m ] +INFO: Idle: 5757, Running: 4, Completed: 10259 [ 13h 51m ] +INFO: Idle: 5756, Running: 4, Completed: 10260 [ 13h 51m ] +INFO: Idle: 5754, Running: 4, Completed: 10262 [ 13h 51m ] +INFO: Idle: 5752, Running: 4, Completed: 10264 [ 13h 51m ] +INFO: Idle: 5751, Running: 4, Completed: 10265 [ 13h 51m ] +INFO: Idle: 5748, Running: 4, Completed: 10268 [ 13h 52m ] +INFO: Idle: 5747, Running: 4, Completed: 10269 [ 13h 52m ] +INFO: Idle: 5744, Running: 4, Completed: 10272 [ 13h 52m ] +INFO: Idle: 5743, Running: 4, Completed: 10273 [ 13h 52m ] +INFO: Idle: 5742, Running: 4, Completed: 10274 [ 13h 52m ] +INFO: Idle: 5740, Running: 4, Completed: 10276 [ 13h 52m ] +INFO: Idle: 5739, Running: 4, Completed: 10277 [ 13h 53m ] +INFO: Idle: 5738, Running: 4, Completed: 10278 [ 13h 53m ] +INFO: Idle: 5736, Running: 4, Completed: 10280 [ 13h 53m ] +INFO: Idle: 5734, Running: 4, Completed: 10282 [ 13h 53m ] +INFO: Idle: 5732, Running: 4, Completed: 10284 [ 13h 53m ] +INFO: Idle: 5731, Running: 4, Completed: 10285 [ 13h 53m ] +INFO: Idle: 5728, Running: 4, Completed: 10288 [ 13h 53m ] +INFO: Idle: 5726, Running: 4, Completed: 10290 [ 13h 54m ] +INFO: Idle: 5725, Running: 4, Completed: 10291 [ 13h 54m ] +INFO: Idle: 5724, Running: 4, Completed: 10292 [ 13h 54m ] +INFO: Idle: 5723, Running: 4, Completed: 10293 [ 13h 54m ] +INFO: Idle: 5722, Running: 4, Completed: 10294 [ 13h 54m ] +INFO: Idle: 5721, Running: 4, Completed: 10295 [ 13h 54m ] +INFO: Idle: 5719, Running: 4, Completed: 10297 [ 13h 54m ] +INFO: Idle: 5718, Running: 4, Completed: 10298 [ 13h 54m ] +INFO: Idle: 5717, Running: 4, Completed: 10299 [ 13h 55m ] +INFO: Idle: 5716, Running: 4, Completed: 10300 [ 13h 55m ] +INFO: Idle: 5715, Running: 4, Completed: 10301 [ 13h 55m ] +INFO: Idle: 5714, Running: 4, Completed: 10302 [ 13h 55m ] +INFO: Idle: 5713, Running: 4, Completed: 10303 [ 13h 55m ] +INFO: Idle: 5712, Running: 4, Completed: 10304 [ 13h 55m ] +INFO: Idle: 5711, Running: 4, Completed: 10305 [ 13h 55m ] +INFO: Idle: 5708, Running: 4, Completed: 10308 [ 13h 55m ] +INFO: Idle: 5707, Running: 4, Completed: 10309 [ 13h 56m ] +INFO: Idle: 5706, Running: 4, Completed: 10310 [ 13h 56m ] +INFO: Idle: 5704, Running: 4, Completed: 10312 [ 13h 56m ] +INFO: Idle: 5703, Running: 4, Completed: 10313 [ 13h 56m ] +INFO: Idle: 5701, Running: 4, Completed: 10315 [ 13h 56m ] +INFO: Idle: 5700, Running: 4, Completed: 10316 [ 13h 56m ] +INFO: Idle: 5699, Running: 4, Completed: 10317 [ 13h 56m ] +INFO: Idle: 5697, Running: 4, Completed: 10319 [ 13h 56m ] +INFO: Idle: 5695, Running: 4, Completed: 10321 [ 13h 57m ] +INFO: Idle: 5694, Running: 4, Completed: 10322 [ 13h 57m ] +INFO: Idle: 5691, Running: 4, Completed: 10325 [ 13h 57m ] +INFO: Idle: 5690, Running: 4, Completed: 10326 [ 13h 57m ] +INFO: Idle: 5687, Running: 4, Completed: 10329 [ 13h 57m ] +INFO: Idle: 5684, Running: 4, Completed: 10332 [ 13h 57m ] +INFO: Idle: 5683, Running: 4, Completed: 10333 [ 13h 57m ] +INFO: Idle: 5682, Running: 4, Completed: 10334 [ 13h 58m ] +INFO: Idle: 5679, Running: 4, Completed: 10337 [ 13h 58m ] +INFO: Idle: 5678, Running: 4, Completed: 10338 [ 13h 58m ] +INFO: Idle: 5675, Running: 4, Completed: 10341 [ 13h 58m ] +INFO: Idle: 5674, Running: 4, Completed: 10342 [ 13h 58m ] +INFO: Idle: 5673, Running: 4, Completed: 10343 [ 13h 58m ] +INFO: Idle: 5671, Running: 4, Completed: 10345 [ 13h 59m ] +INFO: Idle: 5669, Running: 4, Completed: 10347 [ 13h 59m ] +INFO: Idle: 5668, Running: 4, Completed: 10348 [ 13h 59m ] +INFO: Idle: 5667, Running: 4, Completed: 10349 [ 13h 59m ] +INFO: Idle: 5665, Running: 4, Completed: 10351 [ 13h 59m ] +INFO: Idle: 5664, Running: 4, Completed: 10352 [ 13h 59m ] +INFO: Idle: 5663, Running: 4, Completed: 10353 [ 13h 59m ] +INFO: Idle: 5660, Running: 4, Completed: 10356 [ 14h 0m ] +INFO: Idle: 5659, Running: 4, Completed: 10357 [ 14h 0m ] +INFO: Idle: 5658, Running: 4, Completed: 10358 [ 14h 0m ] +INFO: Idle: 5656, Running: 4, Completed: 10360 [ 14h 0m ] +INFO: Idle: 5655, Running: 4, Completed: 10361 [ 14h 0m ] +INFO: Idle: 5653, Running: 4, Completed: 10363 [ 14h 0m ] +INFO: Idle: 5652, Running: 4, Completed: 10364 [ 14h 0m ] +INFO: Idle: 5651, Running: 4, Completed: 10365 [ 14h 0m ] +INFO: Idle: 5649, Running: 4, Completed: 10367 [ 14h 1m ] +INFO: Idle: 5648, Running: 4, Completed: 10368 [ 14h 1m ] +INFO: Idle: 5647, Running: 4, Completed: 10369 [ 14h 1m ] +INFO: Idle: 5644, Running: 4, Completed: 10372 [ 14h 1m ] +INFO: Idle: 5643, Running: 4, Completed: 10373 [ 14h 1m ] +INFO: Idle: 5641, Running: 4, Completed: 10375 [ 14h 1m ] +INFO: Idle: 5639, Running: 4, Completed: 10377 [ 14h 1m ] +INFO: Idle: 5637, Running: 4, Completed: 10379 [ 14h 2m ] +INFO: Idle: 5635, Running: 4, Completed: 10381 [ 14h 2m ] +INFO: Idle: 5633, Running: 4, Completed: 10383 [ 14h 2m ] +INFO: Idle: 5632, Running: 4, Completed: 10384 [ 14h 2m ] +INFO: Idle: 5631, Running: 4, Completed: 10385 [ 14h 2m ] +INFO: Idle: 5630, Running: 4, Completed: 10386 [ 14h 2m ] +INFO: Idle: 5627, Running: 4, Completed: 10389 [ 14h 2m ] +INFO: Idle: 5626, Running: 4, Completed: 10390 [ 14h 2m ] +INFO: Idle: 5624, Running: 4, Completed: 10392 [ 14h 3m ] +INFO: Idle: 5623, Running: 4, Completed: 10393 [ 14h 3m ] +INFO: Idle: 5622, Running: 4, Completed: 10394 [ 14h 3m ] +INFO: Idle: 5619, Running: 4, Completed: 10397 [ 14h 3m ] +INFO: Idle: 5617, Running: 4, Completed: 10399 [ 14h 3m ] +INFO: Idle: 5616, Running: 4, Completed: 10400 [ 14h 3m ] +INFO: Idle: 5615, Running: 4, Completed: 10401 [ 14h 3m ] +INFO: Idle: 5612, Running: 4, Completed: 10404 [ 14h 3m ] +INFO: Idle: 5611, Running: 4, Completed: 10405 [ 14h 4m ] +INFO: Idle: 5608, Running: 4, Completed: 10408 [ 14h 4m ] +INFO: Idle: 5607, Running: 4, Completed: 10409 [ 14h 4m ] +INFO: Idle: 5605, Running: 4, Completed: 10411 [ 14h 4m ] +INFO: Idle: 5604, Running: 4, Completed: 10412 [ 14h 4m ] +INFO: Idle: 5603, Running: 4, Completed: 10413 [ 14h 4m ] +INFO: Idle: 5602, Running: 4, Completed: 10414 [ 14h 4m ] +INFO: Idle: 5601, Running: 4, Completed: 10415 [ 14h 5m ] +INFO: Idle: 5599, Running: 4, Completed: 10417 [ 14h 5m ] +INFO: Idle: 5598, Running: 4, Completed: 10418 [ 14h 5m ] +INFO: Idle: 5596, Running: 4, Completed: 10420 [ 14h 5m ] +INFO: Idle: 5595, Running: 4, Completed: 10421 [ 14h 5m ] +INFO: Idle: 5594, Running: 4, Completed: 10422 [ 14h 5m ] +INFO: Idle: 5593, Running: 4, Completed: 10423 [ 14h 5m ] +INFO: Idle: 5591, Running: 4, Completed: 10425 [ 14h 5m ] +INFO: Idle: 5590, Running: 4, Completed: 10426 [ 14h 6m ] +INFO: Idle: 5588, Running: 4, Completed: 10428 [ 14h 6m ] +INFO: Idle: 5587, Running: 4, Completed: 10429 [ 14h 6m ] +INFO: Idle: 5586, Running: 4, Completed: 10430 [ 14h 6m ] +INFO: Idle: 5585, Running: 4, Completed: 10431 [ 14h 6m ] +INFO: Idle: 5583, Running: 4, Completed: 10433 [ 14h 6m ] +INFO: Idle: 5581, Running: 4, Completed: 10435 [ 14h 6m ] +INFO: Idle: 5580, Running: 4, Completed: 10436 [ 14h 6m ] +INFO: Idle: 5579, Running: 4, Completed: 10437 [ 14h 6m ] +INFO: Idle: 5577, Running: 4, Completed: 10439 [ 14h 7m ] +INFO: Idle: 5575, Running: 4, Completed: 10441 [ 14h 7m ] +INFO: Idle: 5574, Running: 4, Completed: 10442 [ 14h 7m ] +INFO: Idle: 5572, Running: 4, Completed: 10444 [ 14h 7m ] +INFO: Idle: 5571, Running: 4, Completed: 10445 [ 14h 7m ] +INFO: Idle: 5569, Running: 4, Completed: 10447 [ 14h 7m ] +INFO: Idle: 5568, Running: 4, Completed: 10448 [ 14h 7m ] +INFO: Idle: 5567, Running: 4, Completed: 10449 [ 14h 7m ] +INFO: Idle: 5566, Running: 4, Completed: 10450 [ 14h 8m ] +INFO: Idle: 5565, Running: 4, Completed: 10451 [ 14h 8m ] +INFO: Idle: 5563, Running: 4, Completed: 10453 [ 14h 8m ] +INFO: Idle: 5562, Running: 4, Completed: 10454 [ 14h 8m ] +INFO: Idle: 5561, Running: 4, Completed: 10455 [ 14h 8m ] +INFO: Idle: 5559, Running: 4, Completed: 10457 [ 14h 8m ] +INFO: Idle: 5558, Running: 4, Completed: 10458 [ 14h 8m ] +INFO: Idle: 5557, Running: 4, Completed: 10459 [ 14h 8m ] +INFO: Idle: 5555, Running: 4, Completed: 10461 [ 14h 9m ] +INFO: Idle: 5554, Running: 4, Completed: 10462 [ 14h 9m ] +INFO: Idle: 5553, Running: 4, Completed: 10463 [ 14h 9m ] +INFO: Idle: 5551, Running: 4, Completed: 10465 [ 14h 9m ] +INFO: Idle: 5550, Running: 4, Completed: 10466 [ 14h 9m ] +INFO: Idle: 5549, Running: 4, Completed: 10467 [ 14h 9m ] +INFO: Idle: 5547, Running: 4, Completed: 10469 [ 14h 9m ] +INFO: Idle: 5546, Running: 4, Completed: 10470 [ 14h 9m ] +INFO: Idle: 5545, Running: 4, Completed: 10471 [ 14h 9m ] +INFO: Idle: 5544, Running: 4, Completed: 10472 [ 14h 9m ] +INFO: Idle: 5542, Running: 4, Completed: 10474 [ 14h 10m ] +INFO: Idle: 5541, Running: 4, Completed: 10475 [ 14h 10m ] +INFO: Idle: 5540, Running: 4, Completed: 10476 [ 14h 10m ] +INFO: Idle: 5539, Running: 4, Completed: 10477 [ 14h 10m ] +INFO: Idle: 5537, Running: 4, Completed: 10479 [ 14h 10m ] +INFO: Idle: 5536, Running: 4, Completed: 10480 [ 14h 10m ] +INFO: Idle: 5535, Running: 4, Completed: 10481 [ 14h 10m ] +INFO: Idle: 5534, Running: 4, Completed: 10482 [ 14h 10m ] +INFO: Idle: 5531, Running: 4, Completed: 10485 [ 14h 11m ] +INFO: Idle: 5530, Running: 4, Completed: 10486 [ 14h 11m ] +INFO: Idle: 5528, Running: 4, Completed: 10488 [ 14h 11m ] +INFO: Idle: 5526, Running: 4, Completed: 10490 [ 14h 11m ] +INFO: Idle: 5525, Running: 4, Completed: 10491 [ 14h 11m ] +INFO: Idle: 5523, Running: 4, Completed: 10493 [ 14h 11m ] +INFO: Idle: 5522, Running: 4, Completed: 10494 [ 14h 11m ] +INFO: Idle: 5521, Running: 4, Completed: 10495 [ 14h 11m ] +INFO: Idle: 5519, Running: 4, Completed: 10497 [ 14h 12m ] +INFO: Idle: 5518, Running: 4, Completed: 10498 [ 14h 12m ] +INFO: Idle: 5516, Running: 4, Completed: 10500 [ 14h 12m ] +INFO: Idle: 5515, Running: 4, Completed: 10501 [ 14h 12m ] +INFO: Idle: 5512, Running: 4, Completed: 10504 [ 14h 12m ] +INFO: Idle: 5511, Running: 4, Completed: 10505 [ 14h 12m ] +INFO: Idle: 5510, Running: 4, Completed: 10506 [ 14h 12m ] +INFO: Idle: 5509, Running: 4, Completed: 10507 [ 14h 12m ] +INFO: Idle: 5507, Running: 4, Completed: 10509 [ 14h 13m ] +INFO: Idle: 5504, Running: 4, Completed: 10512 [ 14h 13m ] +INFO: Idle: 5503, Running: 4, Completed: 10513 [ 14h 13m ] +INFO: Idle: 5500, Running: 4, Completed: 10516 [ 14h 13m ] +INFO: Idle: 5499, Running: 4, Completed: 10517 [ 14h 13m ] +INFO: Idle: 5496, Running: 4, Completed: 10520 [ 14h 14m ] +INFO: Idle: 5495, Running: 4, Completed: 10521 [ 14h 14m ] +INFO: Idle: 5494, Running: 4, Completed: 10522 [ 14h 14m ] +INFO: Idle: 5492, Running: 4, Completed: 10524 [ 14h 14m ] +INFO: Idle: 5491, Running: 4, Completed: 10525 [ 14h 14m ] +INFO: Idle: 5490, Running: 4, Completed: 10526 [ 14h 14m ] +INFO: Idle: 5487, Running: 4, Completed: 10529 [ 14h 14m ] +INFO: Idle: 5486, Running: 4, Completed: 10530 [ 14h 15m ] +INFO: Idle: 5485, Running: 4, Completed: 10531 [ 14h 15m ] +INFO: Idle: 5483, Running: 4, Completed: 10533 [ 14h 15m ] +INFO: Idle: 5482, Running: 4, Completed: 10534 [ 14h 15m ] +INFO: Idle: 5480, Running: 4, Completed: 10536 [ 14h 15m ] +INFO: Idle: 5479, Running: 4, Completed: 10537 [ 14h 15m ] +INFO: Idle: 5478, Running: 4, Completed: 10538 [ 14h 15m ] +INFO: Idle: 5476, Running: 4, Completed: 10540 [ 14h 15m ] +INFO: Idle: 5475, Running: 4, Completed: 10541 [ 14h 15m ] +INFO: Idle: 5474, Running: 4, Completed: 10542 [ 14h 15m ] +INFO: Idle: 5472, Running: 4, Completed: 10544 [ 14h 16m ] +INFO: Idle: 5471, Running: 4, Completed: 10545 [ 14h 16m ] +INFO: Idle: 5470, Running: 4, Completed: 10546 [ 14h 16m ] +INFO: Idle: 5468, Running: 4, Completed: 10548 [ 14h 16m ] +INFO: Idle: 5466, Running: 4, Completed: 10550 [ 14h 16m ] +INFO: Idle: 5465, Running: 4, Completed: 10551 [ 14h 16m ] +INFO: Idle: 5464, Running: 4, Completed: 10552 [ 14h 16m ] +INFO: Idle: 5463, Running: 4, Completed: 10553 [ 14h 16m ] +INFO: Idle: 5462, Running: 4, Completed: 10554 [ 14h 17m ] +INFO: Idle: 5459, Running: 4, Completed: 10557 [ 14h 17m ] +INFO: Idle: 5458, Running: 4, Completed: 10558 [ 14h 17m ] +INFO: Idle: 5456, Running: 4, Completed: 10560 [ 14h 17m ] +INFO: Idle: 5455, Running: 4, Completed: 10561 [ 14h 17m ] +INFO: Idle: 5453, Running: 4, Completed: 10563 [ 14h 17m ] +INFO: Idle: 5451, Running: 4, Completed: 10565 [ 14h 17m ] +INFO: Idle: 5450, Running: 4, Completed: 10566 [ 14h 17m ] +INFO: Idle: 5449, Running: 4, Completed: 10567 [ 14h 18m ] +INFO: Idle: 5447, Running: 4, Completed: 10569 [ 14h 18m ] +INFO: Idle: 5446, Running: 4, Completed: 10570 [ 14h 18m ] +INFO: Idle: 5443, Running: 4, Completed: 10573 [ 14h 18m ] +INFO: Idle: 5440, Running: 4, Completed: 10576 [ 14h 18m ] +INFO: Idle: 5439, Running: 4, Completed: 10577 [ 14h 18m ] +INFO: Idle: 5438, Running: 4, Completed: 10578 [ 14h 19m ] +INFO: Idle: 5436, Running: 4, Completed: 10580 [ 14h 19m ] +INFO: Idle: 5435, Running: 4, Completed: 10581 [ 14h 19m ] +INFO: Idle: 5433, Running: 4, Completed: 10583 [ 14h 19m ] +INFO: Idle: 5431, Running: 4, Completed: 10585 [ 14h 19m ] +INFO: Idle: 5429, Running: 4, Completed: 10587 [ 14h 19m ] +INFO: Idle: 5428, Running: 4, Completed: 10588 [ 14h 19m ] +INFO: Idle: 5427, Running: 4, Completed: 10589 [ 14h 19m ] +INFO: Idle: 5425, Running: 4, Completed: 10591 [ 14h 20m ] +INFO: Idle: 5424, Running: 4, Completed: 10592 [ 14h 20m ] +INFO: Idle: 5423, Running: 4, Completed: 10593 [ 14h 20m ] +INFO: Idle: 5421, Running: 4, Completed: 10595 [ 14h 20m ] +INFO: Idle: 5419, Running: 4, Completed: 10597 [ 14h 20m ] +INFO: Idle: 5418, Running: 4, Completed: 10598 [ 14h 20m ] +INFO: Idle: 5417, Running: 4, Completed: 10599 [ 14h 20m ] +INFO: Idle: 5415, Running: 4, Completed: 10601 [ 14h 21m ] +INFO: Idle: 5414, Running: 4, Completed: 10602 [ 14h 21m ] +INFO: Idle: 5413, Running: 4, Completed: 10603 [ 14h 21m ] +INFO: Idle: 5411, Running: 4, Completed: 10605 [ 14h 21m ] +INFO: Idle: 5410, Running: 4, Completed: 10606 [ 14h 21m ] +INFO: Idle: 5409, Running: 4, Completed: 10607 [ 14h 21m ] +INFO: Idle: 5407, Running: 4, Completed: 10609 [ 14h 21m ] +INFO: Idle: 5406, Running: 4, Completed: 10610 [ 14h 21m ] +INFO: Idle: 5405, Running: 4, Completed: 10611 [ 14h 21m ] +INFO: Idle: 5403, Running: 4, Completed: 10613 [ 14h 22m ] +INFO: Idle: 5402, Running: 4, Completed: 10614 [ 14h 22m ] +INFO: Idle: 5399, Running: 4, Completed: 10617 [ 14h 22m ] +INFO: Idle: 5398, Running: 4, Completed: 10618 [ 14h 22m ] +INFO: Idle: 5397, Running: 4, Completed: 10619 [ 14h 22m ] +INFO: Idle: 5394, Running: 4, Completed: 10622 [ 14h 22m ] +INFO: Idle: 5391, Running: 4, Completed: 10625 [ 14h 22m ] +INFO: Idle: 5390, Running: 4, Completed: 10626 [ 14h 23m ] +INFO: Idle: 5387, Running: 4, Completed: 10629 [ 14h 23m ] +INFO: Idle: 5386, Running: 4, Completed: 10630 [ 14h 23m ] +INFO: Idle: 5383, Running: 4, Completed: 10633 [ 14h 23m ] +INFO: Idle: 5382, Running: 4, Completed: 10634 [ 14h 23m ] +INFO: Idle: 5381, Running: 4, Completed: 10635 [ 14h 23m ] +INFO: Idle: 5379, Running: 4, Completed: 10637 [ 14h 23m ] +INFO: Idle: 5378, Running: 4, Completed: 10638 [ 14h 24m ] +INFO: Idle: 5377, Running: 4, Completed: 10639 [ 14h 24m ] +INFO: Idle: 5375, Running: 4, Completed: 10641 [ 14h 24m ] +INFO: Idle: 5374, Running: 4, Completed: 10642 [ 14h 24m ] +INFO: Idle: 5373, Running: 4, Completed: 10643 [ 14h 24m ] +INFO: Idle: 5372, Running: 4, Completed: 10644 [ 14h 24m ] +INFO: Idle: 5370, Running: 4, Completed: 10646 [ 14h 24m ] +INFO: Idle: 5369, Running: 4, Completed: 10647 [ 14h 24m ] +INFO: Idle: 5367, Running: 4, Completed: 10649 [ 14h 25m ] +INFO: Idle: 5366, Running: 4, Completed: 10650 [ 14h 25m ] +INFO: Idle: 5365, Running: 4, Completed: 10651 [ 14h 25m ] +INFO: Idle: 5363, Running: 4, Completed: 10653 [ 14h 25m ] +INFO: Idle: 5362, Running: 4, Completed: 10654 [ 14h 25m ] +INFO: Idle: 5361, Running: 4, Completed: 10655 [ 14h 25m ] +INFO: Idle: 5359, Running: 4, Completed: 10657 [ 14h 25m ] +INFO: Idle: 5358, Running: 4, Completed: 10658 [ 14h 25m ] +INFO: Idle: 5355, Running: 4, Completed: 10661 [ 14h 25m ] +INFO: Idle: 5354, Running: 4, Completed: 10662 [ 14h 26m ] +INFO: Idle: 5353, Running: 4, Completed: 10663 [ 14h 26m ] +INFO: Idle: 5350, Running: 4, Completed: 10666 [ 14h 26m ] +INFO: Idle: 5349, Running: 4, Completed: 10667 [ 14h 26m ] +INFO: Idle: 5348, Running: 4, Completed: 10668 [ 14h 26m ] +INFO: Idle: 5347, Running: 4, Completed: 10669 [ 14h 26m ] +INFO: Idle: 5346, Running: 4, Completed: 10670 [ 14h 26m ] +INFO: Idle: 5344, Running: 4, Completed: 10672 [ 14h 26m ] +INFO: Idle: 5340, Running: 4, Completed: 10676 [ 14h 27m ] +INFO: Idle: 5337, Running: 4, Completed: 10679 [ 14h 27m ] +INFO: Idle: 5336, Running: 4, Completed: 10680 [ 14h 27m ] +INFO: Idle: 5333, Running: 4, Completed: 10683 [ 14h 27m ] +INFO: Idle: 5332, Running: 4, Completed: 10684 [ 14h 27m ] +INFO: Idle: 5331, Running: 4, Completed: 10685 [ 14h 28m ] +INFO: Idle: 5328, Running: 4, Completed: 10688 [ 14h 28m ] +INFO: Idle: 5327, Running: 4, Completed: 10689 [ 14h 28m ] +INFO: Idle: 5324, Running: 4, Completed: 10692 [ 14h 28m ] +INFO: Idle: 5323, Running: 4, Completed: 10693 [ 14h 28m ] +INFO: Idle: 5321, Running: 4, Completed: 10695 [ 14h 28m ] +INFO: Idle: 5320, Running: 4, Completed: 10696 [ 14h 29m ] +INFO: Idle: 5319, Running: 4, Completed: 10697 [ 14h 29m ] +INFO: Idle: 5317, Running: 4, Completed: 10699 [ 14h 29m ] +INFO: Idle: 5316, Running: 4, Completed: 10700 [ 14h 29m ] +INFO: Idle: 5315, Running: 4, Completed: 10701 [ 14h 29m ] +INFO: Idle: 5313, Running: 4, Completed: 10703 [ 14h 29m ] +INFO: Idle: 5312, Running: 4, Completed: 10704 [ 14h 29m ] +INFO: Idle: 5311, Running: 4, Completed: 10705 [ 14h 29m ] +INFO: Idle: 5309, Running: 4, Completed: 10707 [ 14h 29m ] +INFO: Idle: 5308, Running: 4, Completed: 10708 [ 14h 29m ] +INFO: Idle: 5305, Running: 4, Completed: 10711 [ 14h 30m ] +INFO: Idle: 5304, Running: 4, Completed: 10712 [ 14h 30m ] +INFO: Idle: 5303, Running: 4, Completed: 10713 [ 14h 30m ] +INFO: Idle: 5300, Running: 4, Completed: 10716 [ 14h 30m ] +INFO: Idle: 5299, Running: 4, Completed: 10717 [ 14h 30m ] +INFO: Idle: 5298, Running: 4, Completed: 10718 [ 14h 30m ] +INFO: Idle: 5296, Running: 4, Completed: 10720 [ 14h 30m ] +INFO: Idle: 5294, Running: 4, Completed: 10722 [ 14h 31m ] +INFO: Idle: 5291, Running: 4, Completed: 10725 [ 14h 31m ] +INFO: Idle: 5290, Running: 4, Completed: 10726 [ 14h 31m ] +INFO: Idle: 5288, Running: 4, Completed: 10728 [ 14h 31m ] +INFO: Idle: 5286, Running: 4, Completed: 10730 [ 14h 31m ] +INFO: Idle: 5284, Running: 4, Completed: 10732 [ 14h 31m ] +INFO: Idle: 5280, Running: 4, Completed: 10736 [ 14h 32m ] +INFO: Idle: 5278, Running: 4, Completed: 10738 [ 14h 32m ] +INFO: Idle: 5276, Running: 4, Completed: 10740 [ 14h 32m ] +INFO: Idle: 5275, Running: 4, Completed: 10741 [ 14h 32m ] +INFO: Idle: 5273, Running: 4, Completed: 10743 [ 14h 32m ] +INFO: Idle: 5272, Running: 4, Completed: 10744 [ 14h 32m ] +INFO: Idle: 5269, Running: 4, Completed: 10747 [ 14h 32m ] +INFO: Idle: 5268, Running: 4, Completed: 10748 [ 14h 33m ] +INFO: Idle: 5266, Running: 4, Completed: 10750 [ 14h 33m ] +INFO: Idle: 5265, Running: 4, Completed: 10751 [ 14h 33m ] +INFO: Idle: 5264, Running: 4, Completed: 10752 [ 14h 33m ] +INFO: Idle: 5263, Running: 4, Completed: 10753 [ 14h 33m ] +INFO: Idle: 5262, Running: 4, Completed: 10754 [ 14h 33m ] +INFO: Idle: 5259, Running: 4, Completed: 10757 [ 14h 33m ] +INFO: Idle: 5258, Running: 4, Completed: 10758 [ 14h 33m ] +INFO: Idle: 5256, Running: 4, Completed: 10760 [ 14h 34m ] +INFO: Idle: 5255, Running: 4, Completed: 10761 [ 14h 34m ] +INFO: Idle: 5254, Running: 4, Completed: 10762 [ 14h 34m ] +INFO: Idle: 5253, Running: 4, Completed: 10763 [ 14h 34m ] +INFO: Idle: 5251, Running: 4, Completed: 10765 [ 14h 34m ] +INFO: Idle: 5250, Running: 4, Completed: 10766 [ 14h 34m ] +INFO: Idle: 5248, Running: 4, Completed: 10768 [ 14h 34m ] +INFO: Idle: 5247, Running: 4, Completed: 10769 [ 14h 34m ] +INFO: Idle: 5246, Running: 4, Completed: 10770 [ 14h 34m ] +INFO: Idle: 5243, Running: 4, Completed: 10773 [ 14h 35m ] +INFO: Idle: 5240, Running: 4, Completed: 10776 [ 14h 35m ] +INFO: Idle: 5239, Running: 4, Completed: 10777 [ 14h 35m ] +INFO: Idle: 5238, Running: 4, Completed: 10778 [ 14h 35m ] +INFO: Idle: 5236, Running: 4, Completed: 10780 [ 14h 35m ] +INFO: Idle: 5235, Running: 4, Completed: 10781 [ 14h 36m ] +INFO: Idle: 5234, Running: 4, Completed: 10782 [ 14h 36m ] +INFO: Idle: 5232, Running: 4, Completed: 10784 [ 14h 36m ] +INFO: Idle: 5231, Running: 4, Completed: 10785 [ 14h 36m ] +INFO: Idle: 5230, Running: 4, Completed: 10786 [ 14h 36m ] +INFO: Idle: 5228, Running: 4, Completed: 10788 [ 14h 36m ] +INFO: Idle: 5227, Running: 4, Completed: 10789 [ 14h 36m ] +INFO: Idle: 5225, Running: 4, Completed: 10791 [ 14h 36m ] +INFO: Idle: 5224, Running: 4, Completed: 10792 [ 14h 36m ] +INFO: Idle: 5223, Running: 4, Completed: 10793 [ 14h 36m ] +INFO: Idle: 5222, Running: 4, Completed: 10794 [ 14h 37m ] +INFO: Idle: 5220, Running: 4, Completed: 10796 [ 14h 37m ] +INFO: Idle: 5219, Running: 4, Completed: 10797 [ 14h 37m ] +INFO: Idle: 5218, Running: 4, Completed: 10798 [ 14h 37m ] +INFO: Idle: 5215, Running: 4, Completed: 10801 [ 14h 37m ] +INFO: Idle: 5214, Running: 4, Completed: 10802 [ 14h 37m ] +INFO: Idle: 5213, Running: 4, Completed: 10803 [ 14h 37m ] +INFO: Idle: 5211, Running: 4, Completed: 10805 [ 14h 37m ] +INFO: Idle: 5210, Running: 4, Completed: 10806 [ 14h 38m ] +INFO: Idle: 5209, Running: 4, Completed: 10807 [ 14h 38m ] +INFO: Idle: 5206, Running: 4, Completed: 10810 [ 14h 38m ] +INFO: Idle: 5205, Running: 4, Completed: 10811 [ 14h 38m ] +INFO: Idle: 5203, Running: 4, Completed: 10813 [ 14h 38m ] +INFO: Idle: 5202, Running: 4, Completed: 10814 [ 14h 38m ] +INFO: Idle: 5201, Running: 4, Completed: 10815 [ 14h 38m ] +INFO: Idle: 5200, Running: 4, Completed: 10816 [ 14h 38m ] +INFO: Idle: 5198, Running: 4, Completed: 10818 [ 14h 38m ] +INFO: Idle: 5197, Running: 4, Completed: 10819 [ 14h 39m ] +INFO: Idle: 5196, Running: 4, Completed: 10820 [ 14h 39m ] +INFO: Idle: 5194, Running: 4, Completed: 10822 [ 14h 39m ] +INFO: Idle: 5193, Running: 4, Completed: 10823 [ 14h 39m ] +INFO: Idle: 5192, Running: 4, Completed: 10824 [ 14h 39m ] +INFO: Idle: 5191, Running: 4, Completed: 10825 [ 14h 39m ] +INFO: Idle: 5190, Running: 4, Completed: 10826 [ 14h 39m ] +INFO: Idle: 5187, Running: 4, Completed: 10829 [ 14h 40m ] +INFO: Idle: 5186, Running: 4, Completed: 10830 [ 14h 40m ] +INFO: Idle: 5184, Running: 4, Completed: 10832 [ 14h 40m ] +INFO: Idle: 5182, Running: 4, Completed: 10834 [ 14h 40m ] +INFO: Idle: 5181, Running: 4, Completed: 10835 [ 14h 40m ] +INFO: Idle: 5179, Running: 4, Completed: 10837 [ 14h 40m ] +INFO: Idle: 5178, Running: 4, Completed: 10838 [ 14h 40m ] +INFO: Idle: 5177, Running: 4, Completed: 10839 [ 14h 40m ] +INFO: Idle: 5175, Running: 4, Completed: 10841 [ 14h 41m ] +INFO: Idle: 5174, Running: 4, Completed: 10842 [ 14h 41m ] +INFO: Idle: 5172, Running: 4, Completed: 10844 [ 14h 41m ] +INFO: Idle: 5171, Running: 4, Completed: 10845 [ 14h 41m ] +INFO: Idle: 5169, Running: 4, Completed: 10847 [ 14h 41m ] +INFO: Idle: 5168, Running: 4, Completed: 10848 [ 14h 41m ] +INFO: Idle: 5167, Running: 4, Completed: 10849 [ 14h 41m ] +INFO: Idle: 5166, Running: 4, Completed: 10850 [ 14h 41m ] +INFO: Idle: 5165, Running: 4, Completed: 10851 [ 14h 41m ] +INFO: Idle: 5163, Running: 4, Completed: 10853 [ 14h 42m ] +INFO: Idle: 5162, Running: 4, Completed: 10854 [ 14h 42m ] +INFO: Idle: 5161, Running: 4, Completed: 10855 [ 14h 42m ] +INFO: Idle: 5160, Running: 4, Completed: 10856 [ 14h 42m ] +INFO: Idle: 5159, Running: 4, Completed: 10857 [ 14h 42m ] +INFO: Idle: 5158, Running: 4, Completed: 10858 [ 14h 42m ] +INFO: Idle: 5156, Running: 4, Completed: 10860 [ 14h 42m ] +INFO: Idle: 5155, Running: 4, Completed: 10861 [ 14h 42m ] +INFO: Idle: 5153, Running: 4, Completed: 10863 [ 14h 42m ] +INFO: Idle: 5152, Running: 4, Completed: 10864 [ 14h 42m ] +INFO: Idle: 5151, Running: 4, Completed: 10865 [ 14h 43m ] +INFO: Idle: 5150, Running: 4, Completed: 10866 [ 14h 43m ] +INFO: Idle: 5147, Running: 4, Completed: 10869 [ 14h 43m ] +INFO: Idle: 5146, Running: 4, Completed: 10870 [ 14h 43m ] +INFO: Idle: 5143, Running: 4, Completed: 10873 [ 14h 43m ] +INFO: Idle: 5142, Running: 4, Completed: 10874 [ 14h 43m ] +INFO: Idle: 5140, Running: 4, Completed: 10876 [ 14h 44m ] +INFO: Idle: 5139, Running: 4, Completed: 10877 [ 14h 44m ] +INFO: Idle: 5138, Running: 4, Completed: 10878 [ 14h 44m ] +INFO: Idle: 5137, Running: 4, Completed: 10879 [ 14h 44m ] +INFO: Idle: 5134, Running: 4, Completed: 10882 [ 14h 44m ] +INFO: Idle: 5133, Running: 4, Completed: 10883 [ 14h 44m ] +INFO: Idle: 5131, Running: 4, Completed: 10885 [ 14h 44m ] +INFO: Idle: 5130, Running: 4, Completed: 10886 [ 14h 44m ] +INFO: Idle: 5129, Running: 4, Completed: 10887 [ 14h 44m ] +INFO: Idle: 5127, Running: 4, Completed: 10889 [ 14h 45m ] +INFO: Idle: 5126, Running: 4, Completed: 10890 [ 14h 45m ] +INFO: Idle: 5125, Running: 4, Completed: 10891 [ 14h 45m ] +INFO: Idle: 5123, Running: 4, Completed: 10893 [ 14h 45m ] +INFO: Idle: 5122, Running: 4, Completed: 10894 [ 14h 45m ] +INFO: Idle: 5119, Running: 4, Completed: 10897 [ 14h 45m ] +INFO: Idle: 5118, Running: 4, Completed: 10898 [ 14h 45m ] +INFO: Idle: 5115, Running: 4, Completed: 10901 [ 14h 46m ] +INFO: Idle: 5114, Running: 4, Completed: 10902 [ 14h 46m ] +INFO: Idle: 5111, Running: 4, Completed: 10905 [ 14h 46m ] +INFO: Idle: 5110, Running: 4, Completed: 10906 [ 14h 46m ] +INFO: Idle: 5107, Running: 4, Completed: 10909 [ 14h 46m ] +INFO: Idle: 5106, Running: 4, Completed: 10910 [ 14h 46m ] +INFO: Idle: 5105, Running: 4, Completed: 10911 [ 14h 46m ] +INFO: Idle: 5104, Running: 4, Completed: 10912 [ 14h 46m ] +INFO: Idle: 5103, Running: 4, Completed: 10913 [ 14h 47m ] +INFO: Idle: 5102, Running: 4, Completed: 10914 [ 14h 47m ] +INFO: Idle: 5100, Running: 4, Completed: 10916 [ 14h 47m ] +INFO: Idle: 5099, Running: 4, Completed: 10917 [ 14h 47m ] +INFO: Idle: 5096, Running: 4, Completed: 10920 [ 14h 47m ] +INFO: Idle: 5095, Running: 4, Completed: 10921 [ 14h 47m ] +INFO: Idle: 5094, Running: 4, Completed: 10922 [ 14h 47m ] +INFO: Idle: 5093, Running: 4, Completed: 10923 [ 14h 48m ] +INFO: Idle: 5092, Running: 4, Completed: 10924 [ 14h 48m ] +INFO: Idle: 5091, Running: 4, Completed: 10925 [ 14h 48m ] +INFO: Idle: 5087, Running: 4, Completed: 10929 [ 14h 48m ] +INFO: Idle: 5084, Running: 4, Completed: 10932 [ 14h 48m ] +INFO: Idle: 5083, Running: 4, Completed: 10933 [ 14h 48m ] +INFO: Idle: 5080, Running: 4, Completed: 10936 [ 14h 49m ] +INFO: Idle: 5079, Running: 4, Completed: 10937 [ 14h 49m ] +INFO: Idle: 5078, Running: 4, Completed: 10938 [ 14h 49m ] +INFO: Idle: 5075, Running: 4, Completed: 10941 [ 14h 49m ] +INFO: Idle: 5074, Running: 4, Completed: 10942 [ 14h 49m ] +INFO: Idle: 5071, Running: 4, Completed: 10945 [ 14h 49m ] +INFO: Idle: 5070, Running: 4, Completed: 10946 [ 14h 50m ] +INFO: Idle: 5067, Running: 4, Completed: 10949 [ 14h 50m ] +INFO: Idle: 5066, Running: 4, Completed: 10950 [ 14h 50m ] +INFO: Idle: 5063, Running: 4, Completed: 10953 [ 14h 50m ] +INFO: Idle: 5062, Running: 4, Completed: 10954 [ 14h 50m ] +INFO: Idle: 5061, Running: 4, Completed: 10955 [ 14h 50m ] +INFO: Idle: 5059, Running: 4, Completed: 10957 [ 14h 50m ] +INFO: Idle: 5058, Running: 4, Completed: 10958 [ 14h 51m ] +INFO: Idle: 5056, Running: 4, Completed: 10960 [ 14h 51m ] +INFO: Idle: 5054, Running: 4, Completed: 10962 [ 14h 51m ] +INFO: Idle: 5052, Running: 4, Completed: 10964 [ 14h 51m ] +INFO: Idle: 5050, Running: 4, Completed: 10966 [ 14h 51m ] +INFO: Idle: 5049, Running: 4, Completed: 10967 [ 14h 51m ] +INFO: Idle: 5048, Running: 4, Completed: 10968 [ 14h 51m ] +INFO: Idle: 5047, Running: 4, Completed: 10969 [ 14h 52m ] +INFO: Idle: 5045, Running: 4, Completed: 10971 [ 14h 52m ] +INFO: Idle: 5043, Running: 4, Completed: 10973 [ 14h 52m ] +INFO: Idle: 5041, Running: 4, Completed: 10975 [ 14h 52m ] +INFO: Idle: 5039, Running: 4, Completed: 10977 [ 14h 52m ] +INFO: Idle: 5037, Running: 4, Completed: 10979 [ 14h 52m ] +INFO: Idle: 5035, Running: 4, Completed: 10981 [ 14h 53m ] +INFO: Idle: 5033, Running: 4, Completed: 10983 [ 14h 53m ] +INFO: Idle: 5031, Running: 4, Completed: 10985 [ 14h 53m ] +INFO: Idle: 5029, Running: 4, Completed: 10987 [ 14h 53m ] +INFO: Idle: 5027, Running: 4, Completed: 10989 [ 14h 53m ] +INFO: Idle: 5025, Running: 4, Completed: 10991 [ 14h 53m ] +INFO: Idle: 5023, Running: 4, Completed: 10993 [ 14h 54m ] +INFO: Idle: 5022, Running: 4, Completed: 10994 [ 14h 54m ] +INFO: Idle: 5021, Running: 4, Completed: 10995 [ 14h 54m ] +INFO: Idle: 5020, Running: 4, Completed: 10996 [ 14h 54m ] +INFO: Idle: 5019, Running: 4, Completed: 10997 [ 14h 54m ] +INFO: Idle: 5017, Running: 4, Completed: 10999 [ 14h 54m ] +INFO: Idle: 5015, Running: 4, Completed: 11001 [ 14h 54m ] +INFO: Idle: 5013, Running: 4, Completed: 11003 [ 14h 54m ] +INFO: Idle: 5012, Running: 4, Completed: 11004 [ 14h 55m ] +INFO: Idle: 5011, Running: 4, Completed: 11005 [ 14h 55m ] +INFO: Idle: 5008, Running: 4, Completed: 11008 [ 14h 55m ] +INFO: Idle: 5007, Running: 4, Completed: 11009 [ 14h 55m ] +INFO: Idle: 5006, Running: 4, Completed: 11010 [ 14h 55m ] +INFO: Idle: 5004, Running: 4, Completed: 11012 [ 14h 55m ] +INFO: Idle: 5003, Running: 4, Completed: 11013 [ 14h 55m ] +INFO: Idle: 5002, Running: 4, Completed: 11014 [ 14h 55m ] +INFO: Idle: 5000, Running: 4, Completed: 11016 [ 14h 56m ] +INFO: Idle: 4999, Running: 4, Completed: 11017 [ 14h 56m ] +INFO: Idle: 4998, Running: 4, Completed: 11018 [ 14h 56m ] +INFO: Idle: 4997, Running: 4, Completed: 11019 [ 14h 56m ] +INFO: Idle: 4995, Running: 4, Completed: 11021 [ 14h 56m ] +INFO: Idle: 4994, Running: 4, Completed: 11022 [ 14h 56m ] +INFO: Idle: 4992, Running: 4, Completed: 11024 [ 14h 56m ] +INFO: Idle: 4991, Running: 4, Completed: 11025 [ 14h 56m ] +INFO: Idle: 4990, Running: 4, Completed: 11026 [ 14h 56m ] +INFO: Idle: 4989, Running: 4, Completed: 11027 [ 14h 56m ] +INFO: Idle: 4987, Running: 4, Completed: 11029 [ 14h 57m ] +INFO: Idle: 4986, Running: 4, Completed: 11030 [ 14h 57m ] +INFO: Idle: 4985, Running: 4, Completed: 11031 [ 14h 57m ] +INFO: Idle: 4984, Running: 4, Completed: 11032 [ 14h 57m ] +INFO: Idle: 4982, Running: 4, Completed: 11034 [ 14h 57m ] +INFO: Idle: 4981, Running: 4, Completed: 11035 [ 14h 57m ] +INFO: Idle: 4980, Running: 4, Completed: 11036 [ 14h 57m ] +INFO: Idle: 4979, Running: 4, Completed: 11037 [ 14h 57m ] +INFO: Idle: 4977, Running: 4, Completed: 11039 [ 14h 57m ] +INFO: Idle: 4976, Running: 4, Completed: 11040 [ 14h 57m ] +INFO: Idle: 4975, Running: 4, Completed: 11041 [ 14h 58m ] +INFO: Idle: 4974, Running: 4, Completed: 11042 [ 14h 58m ] +INFO: Idle: 4971, Running: 4, Completed: 11045 [ 14h 58m ] +INFO: Idle: 4970, Running: 4, Completed: 11046 [ 14h 58m ] +INFO: Idle: 4968, Running: 4, Completed: 11048 [ 14h 58m ] +INFO: Idle: 4967, Running: 4, Completed: 11049 [ 14h 58m ] +INFO: Idle: 4966, Running: 4, Completed: 11050 [ 14h 58m ] +INFO: Idle: 4965, Running: 4, Completed: 11051 [ 14h 58m ] +INFO: Idle: 4963, Running: 4, Completed: 11053 [ 14h 59m ] +INFO: Idle: 4962, Running: 4, Completed: 11054 [ 14h 59m ] +INFO: Idle: 4960, Running: 4, Completed: 11056 [ 14h 59m ] +INFO: Idle: 4959, Running: 4, Completed: 11057 [ 14h 59m ] +INFO: Idle: 4958, Running: 4, Completed: 11058 [ 14h 59m ] +INFO: Idle: 4956, Running: 4, Completed: 11060 [ 14h 59m ] +INFO: Idle: 4955, Running: 4, Completed: 11061 [ 14h 59m ] +INFO: Idle: 4952, Running: 4, Completed: 11064 [ 14h 59m ] +INFO: Idle: 4951, Running: 4, Completed: 11065 [ 15h 0m ] +INFO: Idle: 4950, Running: 4, Completed: 11066 [ 15h 0m ] +INFO: Idle: 4948, Running: 4, Completed: 11068 [ 15h 0m ] +INFO: Idle: 4947, Running: 4, Completed: 11069 [ 15h 0m ] +INFO: Idle: 4945, Running: 4, Completed: 11071 [ 15h 0m ] +INFO: Idle: 4943, Running: 4, Completed: 11073 [ 15h 0m ] +INFO: Idle: 4941, Running: 4, Completed: 11075 [ 15h 1m ] +INFO: Idle: 4940, Running: 4, Completed: 11076 [ 15h 1m ] +INFO: Idle: 4939, Running: 4, Completed: 11077 [ 15h 1m ] +INFO: Idle: 4937, Running: 4, Completed: 11079 [ 15h 1m ] +INFO: Idle: 4936, Running: 4, Completed: 11080 [ 15h 1m ] +INFO: Idle: 4935, Running: 4, Completed: 11081 [ 15h 1m ] +INFO: Idle: 4933, Running: 4, Completed: 11083 [ 15h 1m ] +INFO: Idle: 4932, Running: 4, Completed: 11084 [ 15h 1m ] +INFO: Idle: 4931, Running: 4, Completed: 11085 [ 15h 2m ] +INFO: Idle: 4930, Running: 4, Completed: 11086 [ 15h 2m ] +INFO: Idle: 4929, Running: 4, Completed: 11087 [ 15h 2m ] +INFO: Idle: 4927, Running: 4, Completed: 11089 [ 15h 2m ] +INFO: Idle: 4926, Running: 4, Completed: 11090 [ 15h 2m ] +INFO: Idle: 4925, Running: 4, Completed: 11091 [ 15h 2m ] +INFO: Idle: 4923, Running: 4, Completed: 11093 [ 15h 2m ] +INFO: Idle: 4922, Running: 4, Completed: 11094 [ 15h 2m ] +INFO: Idle: 4921, Running: 4, Completed: 11095 [ 15h 2m ] +INFO: Idle: 4919, Running: 4, Completed: 11097 [ 15h 2m ] +INFO: Idle: 4917, Running: 4, Completed: 11099 [ 15h 3m ] +INFO: Idle: 4916, Running: 4, Completed: 11100 [ 15h 3m ] +INFO: Idle: 4915, Running: 4, Completed: 11101 [ 15h 3m ] +INFO: Idle: 4913, Running: 4, Completed: 11103 [ 15h 3m ] +INFO: Idle: 4912, Running: 4, Completed: 11104 [ 15h 3m ] +INFO: Idle: 4911, Running: 4, Completed: 11105 [ 15h 3m ] +INFO: Idle: 4910, Running: 4, Completed: 11106 [ 15h 3m ] +INFO: Idle: 4908, Running: 4, Completed: 11108 [ 15h 3m ] +INFO: Idle: 4907, Running: 4, Completed: 11109 [ 15h 3m ] +INFO: Idle: 4906, Running: 4, Completed: 11110 [ 15h 4m ] +INFO: Idle: 4905, Running: 4, Completed: 11111 [ 15h 4m ] +INFO: Idle: 4903, Running: 4, Completed: 11113 [ 15h 4m ] +INFO: Idle: 4900, Running: 4, Completed: 11116 [ 15h 4m ] +INFO: Idle: 4899, Running: 4, Completed: 11117 [ 15h 4m ] +INFO: Idle: 4896, Running: 4, Completed: 11120 [ 15h 4m ] +INFO: Idle: 4895, Running: 4, Completed: 11121 [ 15h 5m ] +INFO: Idle: 4893, Running: 4, Completed: 11123 [ 15h 5m ] +INFO: Idle: 4891, Running: 4, Completed: 11125 [ 15h 5m ] +INFO: Idle: 4887, Running: 4, Completed: 11129 [ 15h 5m ] +INFO: Idle: 4883, Running: 4, Completed: 11133 [ 15h 6m ] +INFO: Idle: 4881, Running: 4, Completed: 11135 [ 15h 6m ] +INFO: Idle: 4879, Running: 4, Completed: 11137 [ 15h 6m ] +INFO: Idle: 4877, Running: 4, Completed: 11139 [ 15h 6m ] +INFO: Idle: 4875, Running: 4, Completed: 11141 [ 15h 6m ] +INFO: Idle: 4873, Running: 4, Completed: 11143 [ 15h 6m ] +INFO: Idle: 4872, Running: 4, Completed: 11144 [ 15h 6m ] +INFO: Idle: 4871, Running: 4, Completed: 11145 [ 15h 6m ] +INFO: Idle: 4869, Running: 4, Completed: 11147 [ 15h 7m ] +INFO: Idle: 4868, Running: 4, Completed: 11148 [ 15h 7m ] +INFO: Idle: 4866, Running: 4, Completed: 11150 [ 15h 7m ] +INFO: Idle: 4865, Running: 4, Completed: 11151 [ 15h 7m ] +INFO: Idle: 4864, Running: 4, Completed: 11152 [ 15h 7m ] +INFO: Idle: 4862, Running: 4, Completed: 11154 [ 15h 7m ] +INFO: Idle: 4861, Running: 4, Completed: 11155 [ 15h 7m ] +INFO: Idle: 4860, Running: 4, Completed: 11156 [ 15h 8m ] +INFO: Idle: 4858, Running: 4, Completed: 11158 [ 15h 8m ] +INFO: Idle: 4857, Running: 4, Completed: 11159 [ 15h 8m ] +INFO: Idle: 4856, Running: 4, Completed: 11160 [ 15h 8m ] +INFO: Idle: 4855, Running: 4, Completed: 11161 [ 15h 8m ] +INFO: Idle: 4853, Running: 4, Completed: 11163 [ 15h 8m ] +INFO: Idle: 4852, Running: 4, Completed: 11164 [ 15h 8m ] +INFO: Idle: 4850, Running: 4, Completed: 11166 [ 15h 8m ] +INFO: Idle: 4849, Running: 4, Completed: 11167 [ 15h 8m ] +INFO: Idle: 4848, Running: 4, Completed: 11168 [ 15h 9m ] +INFO: Idle: 4846, Running: 4, Completed: 11170 [ 15h 9m ] +INFO: Idle: 4845, Running: 4, Completed: 11171 [ 15h 9m ] +INFO: Idle: 4844, Running: 4, Completed: 11172 [ 15h 9m ] +INFO: Idle: 4843, Running: 4, Completed: 11173 [ 15h 9m ] +INFO: Idle: 4841, Running: 4, Completed: 11175 [ 15h 9m ] +INFO: Idle: 4840, Running: 4, Completed: 11176 [ 15h 9m ] +INFO: Idle: 4838, Running: 4, Completed: 11178 [ 15h 9m ] +INFO: Idle: 4837, Running: 4, Completed: 11179 [ 15h 9m ] +INFO: Idle: 4836, Running: 4, Completed: 11180 [ 15h 9m ] +INFO: Idle: 4834, Running: 4, Completed: 11182 [ 15h 9m ] +INFO: Idle: 4833, Running: 4, Completed: 11183 [ 15h 10m ] +INFO: Idle: 4829, Running: 4, Completed: 11187 [ 15h 10m ] +INFO: Idle: 4827, Running: 4, Completed: 11189 [ 15h 10m ] +INFO: Idle: 4825, Running: 4, Completed: 11191 [ 15h 10m ] +INFO: Idle: 4824, Running: 4, Completed: 11192 [ 15h 10m ] +INFO: Idle: 4823, Running: 4, Completed: 11193 [ 15h 10m ] +INFO: Idle: 4821, Running: 4, Completed: 11195 [ 15h 10m ] +INFO: Idle: 4820, Running: 4, Completed: 11196 [ 15h 11m ] +INFO: Idle: 4818, Running: 4, Completed: 11198 [ 15h 11m ] +INFO: Idle: 4816, Running: 4, Completed: 11200 [ 15h 11m ] +INFO: Idle: 4813, Running: 4, Completed: 11203 [ 15h 11m ] +INFO: Idle: 4812, Running: 4, Completed: 11204 [ 15h 11m ] +INFO: Idle: 4809, Running: 4, Completed: 11207 [ 15h 11m ] +INFO: Idle: 4808, Running: 4, Completed: 11208 [ 15h 12m ] +INFO: Idle: 4805, Running: 4, Completed: 11211 [ 15h 12m ] +INFO: Idle: 4804, Running: 4, Completed: 11212 [ 15h 12m ] +INFO: Idle: 4802, Running: 4, Completed: 11214 [ 15h 12m ] +INFO: Idle: 4800, Running: 4, Completed: 11216 [ 15h 12m ] +INFO: Idle: 4798, Running: 4, Completed: 11218 [ 15h 12m ] +INFO: Idle: 4796, Running: 4, Completed: 11220 [ 15h 13m ] +INFO: Idle: 4793, Running: 4, Completed: 11223 [ 15h 13m ] +INFO: Idle: 4792, Running: 4, Completed: 11224 [ 15h 13m ] +INFO: Idle: 4789, Running: 4, Completed: 11227 [ 15h 13m ] +INFO: Idle: 4788, Running: 4, Completed: 11228 [ 15h 13m ] +INFO: Idle: 4784, Running: 4, Completed: 11232 [ 15h 14m ] +INFO: Idle: 4783, Running: 4, Completed: 11233 [ 15h 14m ] +INFO: Idle: 4780, Running: 4, Completed: 11236 [ 15h 14m ] +INFO: Idle: 4778, Running: 4, Completed: 11238 [ 15h 14m ] +INFO: Idle: 4777, Running: 4, Completed: 11239 [ 15h 14m ] +INFO: Idle: 4775, Running: 4, Completed: 11241 [ 15h 14m ] +INFO: Idle: 4774, Running: 4, Completed: 11242 [ 15h 14m ] +INFO: Idle: 4773, Running: 4, Completed: 11243 [ 15h 15m ] +INFO: Idle: 4771, Running: 4, Completed: 11245 [ 15h 15m ] +INFO: Idle: 4769, Running: 4, Completed: 11247 [ 15h 15m ] +INFO: Idle: 4767, Running: 4, Completed: 11249 [ 15h 15m ] +INFO: Idle: 4765, Running: 4, Completed: 11251 [ 15h 15m ] +INFO: Idle: 4764, Running: 4, Completed: 11252 [ 15h 15m ] +INFO: Idle: 4761, Running: 4, Completed: 11255 [ 15h 15m ] +INFO: Idle: 4760, Running: 4, Completed: 11256 [ 15h 15m ] +INFO: Idle: 4758, Running: 4, Completed: 11258 [ 15h 16m ] +INFO: Idle: 4756, Running: 4, Completed: 11260 [ 15h 16m ] +INFO: Idle: 4755, Running: 4, Completed: 11261 [ 15h 16m ] +INFO: Idle: 4754, Running: 4, Completed: 11262 [ 15h 16m ] +INFO: Idle: 4751, Running: 4, Completed: 11265 [ 15h 16m ] +INFO: Idle: 4750, Running: 4, Completed: 11266 [ 15h 16m ] +INFO: Idle: 4747, Running: 4, Completed: 11269 [ 15h 17m ] +INFO: Idle: 4746, Running: 4, Completed: 11270 [ 15h 17m ] +INFO: Idle: 4744, Running: 4, Completed: 11272 [ 15h 17m ] +INFO: Idle: 4743, Running: 4, Completed: 11273 [ 15h 17m ] +INFO: Idle: 4742, Running: 4, Completed: 11274 [ 15h 17m ] +INFO: Idle: 4741, Running: 4, Completed: 11275 [ 15h 17m ] +INFO: Idle: 4739, Running: 4, Completed: 11277 [ 15h 17m ] +INFO: Idle: 4738, Running: 4, Completed: 11278 [ 15h 17m ] +INFO: Idle: 4737, Running: 4, Completed: 11279 [ 15h 18m ] +INFO: Idle: 4734, Running: 4, Completed: 11282 [ 15h 18m ] +INFO: Idle: 4733, Running: 4, Completed: 11283 [ 15h 18m ] +INFO: Idle: 4732, Running: 4, Completed: 11284 [ 15h 18m ] +INFO: Idle: 4730, Running: 4, Completed: 11286 [ 15h 18m ] +INFO: Idle: 4729, Running: 4, Completed: 11287 [ 15h 18m ] +INFO: Idle: 4727, Running: 4, Completed: 11289 [ 15h 18m ] +INFO: Idle: 4726, Running: 4, Completed: 11290 [ 15h 19m ] +INFO: Idle: 4724, Running: 4, Completed: 11292 [ 15h 19m ] +INFO: Idle: 4722, Running: 4, Completed: 11294 [ 15h 19m ] +INFO: Idle: 4721, Running: 4, Completed: 11295 [ 15h 19m ] +INFO: Idle: 4719, Running: 4, Completed: 11297 [ 15h 19m ] +INFO: Idle: 4718, Running: 4, Completed: 11298 [ 15h 19m ] +INFO: Idle: 4715, Running: 4, Completed: 11301 [ 15h 19m ] +INFO: Idle: 4714, Running: 4, Completed: 11302 [ 15h 19m ] +INFO: Idle: 4713, Running: 4, Completed: 11303 [ 15h 19m ] +INFO: Idle: 4710, Running: 4, Completed: 11306 [ 15h 20m ] +INFO: Idle: 4709, Running: 4, Completed: 11307 [ 15h 20m ] +INFO: Idle: 4708, Running: 4, Completed: 11308 [ 15h 20m ] +INFO: Idle: 4707, Running: 4, Completed: 11309 [ 15h 20m ] +INFO: Idle: 4706, Running: 4, Completed: 11310 [ 15h 20m ] +INFO: Idle: 4704, Running: 4, Completed: 11312 [ 15h 20m ] +INFO: Idle: 4701, Running: 4, Completed: 11315 [ 15h 20m ] +INFO: Idle: 4700, Running: 4, Completed: 11316 [ 15h 21m ] +INFO: Idle: 4697, Running: 4, Completed: 11319 [ 15h 21m ] +INFO: Idle: 4696, Running: 4, Completed: 11320 [ 15h 21m ] +INFO: Idle: 4693, Running: 4, Completed: 11323 [ 15h 21m ] +INFO: Idle: 4692, Running: 4, Completed: 11324 [ 15h 21m ] +INFO: Idle: 4691, Running: 4, Completed: 11325 [ 15h 21m ] +INFO: Idle: 4689, Running: 4, Completed: 11327 [ 15h 21m ] +INFO: Idle: 4688, Running: 4, Completed: 11328 [ 15h 22m ] +INFO: Idle: 4687, Running: 4, Completed: 11329 [ 15h 22m ] +INFO: Idle: 4684, Running: 4, Completed: 11332 [ 15h 22m ] +INFO: Idle: 4683, Running: 4, Completed: 11333 [ 15h 22m ] +INFO: Idle: 4681, Running: 4, Completed: 11335 [ 15h 22m ] +INFO: Idle: 4680, Running: 4, Completed: 11336 [ 15h 22m ] +INFO: Idle: 4679, Running: 4, Completed: 11337 [ 15h 22m ] +INFO: Idle: 4677, Running: 4, Completed: 11339 [ 15h 22m ] +INFO: Idle: 4676, Running: 4, Completed: 11340 [ 15h 23m ] +INFO: Idle: 4674, Running: 4, Completed: 11342 [ 15h 23m ] +INFO: Idle: 4672, Running: 4, Completed: 11344 [ 15h 23m ] +INFO: Idle: 4671, Running: 4, Completed: 11345 [ 15h 23m ] +INFO: Idle: 4669, Running: 4, Completed: 11347 [ 15h 23m ] +INFO: Idle: 4668, Running: 4, Completed: 11348 [ 15h 23m ] +INFO: Idle: 4665, Running: 4, Completed: 11351 [ 15h 23m ] +INFO: Idle: 4664, Running: 4, Completed: 11352 [ 15h 23m ] +INFO: Idle: 4663, Running: 4, Completed: 11353 [ 15h 24m ] +INFO: Idle: 4660, Running: 4, Completed: 11356 [ 15h 24m ] +INFO: Idle: 4659, Running: 4, Completed: 11357 [ 15h 24m ] +INFO: Idle: 4658, Running: 4, Completed: 11358 [ 15h 24m ] +INFO: Idle: 4656, Running: 4, Completed: 11360 [ 15h 24m ] +INFO: Idle: 4655, Running: 4, Completed: 11361 [ 15h 24m ] +INFO: Idle: 4654, Running: 4, Completed: 11362 [ 15h 24m ] +INFO: Idle: 4651, Running: 4, Completed: 11365 [ 15h 24m ] +INFO: Idle: 4650, Running: 4, Completed: 11366 [ 15h 25m ] +INFO: Idle: 4649, Running: 4, Completed: 11367 [ 15h 25m ] +INFO: Idle: 4646, Running: 4, Completed: 11370 [ 15h 25m ] +INFO: Idle: 4642, Running: 4, Completed: 11374 [ 15h 25m ] +INFO: Idle: 4641, Running: 4, Completed: 11375 [ 15h 25m ] +INFO: Idle: 4639, Running: 4, Completed: 11377 [ 15h 25m ] +INFO: Idle: 4638, Running: 4, Completed: 11378 [ 15h 26m ] +INFO: Idle: 4637, Running: 4, Completed: 11379 [ 15h 26m ] +INFO: Idle: 4636, Running: 4, Completed: 11380 [ 15h 26m ] +INFO: Idle: 4633, Running: 4, Completed: 11383 [ 15h 26m ] +INFO: Idle: 4632, Running: 4, Completed: 11384 [ 15h 26m ] +INFO: Idle: 4630, Running: 4, Completed: 11386 [ 15h 26m ] +INFO: Idle: 4628, Running: 4, Completed: 11388 [ 15h 26m ] +INFO: Idle: 4627, Running: 4, Completed: 11389 [ 15h 26m ] +INFO: Idle: 4625, Running: 4, Completed: 11391 [ 15h 27m ] +INFO: Idle: 4624, Running: 4, Completed: 11392 [ 15h 27m ] +INFO: Idle: 4622, Running: 4, Completed: 11394 [ 15h 27m ] +INFO: Idle: 4621, Running: 4, Completed: 11395 [ 15h 27m ] +INFO: Idle: 4620, Running: 4, Completed: 11396 [ 15h 27m ] +INFO: Idle: 4617, Running: 4, Completed: 11399 [ 15h 27m ] +INFO: Idle: 4614, Running: 4, Completed: 11402 [ 15h 28m ] +INFO: Idle: 4613, Running: 4, Completed: 11403 [ 15h 28m ] +INFO: Idle: 4612, Running: 4, Completed: 11404 [ 15h 28m ] +INFO: Idle: 4610, Running: 4, Completed: 11406 [ 15h 28m ] +INFO: Idle: 4609, Running: 4, Completed: 11407 [ 15h 28m ] +INFO: Idle: 4608, Running: 4, Completed: 11408 [ 15h 28m ] +INFO: Idle: 4606, Running: 4, Completed: 11410 [ 15h 28m ] +INFO: Idle: 4605, Running: 4, Completed: 11411 [ 15h 29m ] +INFO: Idle: 4604, Running: 4, Completed: 11412 [ 15h 29m ] +INFO: Idle: 4602, Running: 4, Completed: 11414 [ 15h 29m ] +INFO: Idle: 4601, Running: 4, Completed: 11415 [ 15h 29m ] +INFO: Idle: 4599, Running: 4, Completed: 11417 [ 15h 29m ] +INFO: Idle: 4598, Running: 4, Completed: 11418 [ 15h 29m ] +INFO: Idle: 4597, Running: 4, Completed: 11419 [ 15h 29m ] +INFO: Idle: 4596, Running: 4, Completed: 11420 [ 15h 29m ] +INFO: Idle: 4593, Running: 4, Completed: 11423 [ 15h 30m ] +INFO: Idle: 4592, Running: 4, Completed: 11424 [ 15h 30m ] +INFO: Idle: 4589, Running: 4, Completed: 11427 [ 15h 30m ] +INFO: Idle: 4588, Running: 4, Completed: 11428 [ 15h 30m ] +INFO: Idle: 4586, Running: 4, Completed: 11430 [ 15h 30m ] +INFO: Idle: 4585, Running: 4, Completed: 11431 [ 15h 30m ] +INFO: Idle: 4584, Running: 4, Completed: 11432 [ 15h 30m ] +INFO: Idle: 4583, Running: 4, Completed: 11433 [ 15h 30m ] +INFO: Idle: 4581, Running: 4, Completed: 11435 [ 15h 30m ] +INFO: Idle: 4580, Running: 4, Completed: 11436 [ 15h 31m ] +INFO: Idle: 4579, Running: 4, Completed: 11437 [ 15h 31m ] +INFO: Idle: 4577, Running: 4, Completed: 11439 [ 15h 31m ] +INFO: Idle: 4576, Running: 4, Completed: 11440 [ 15h 31m ] +INFO: Idle: 4573, Running: 4, Completed: 11443 [ 15h 31m ] +INFO: Idle: 4572, Running: 4, Completed: 11444 [ 15h 31m ] +INFO: Idle: 4569, Running: 4, Completed: 11447 [ 15h 31m ] +INFO: Idle: 4568, Running: 4, Completed: 11448 [ 15h 32m ] +INFO: Idle: 4565, Running: 4, Completed: 11451 [ 15h 32m ] +INFO: Idle: 4564, Running: 4, Completed: 11452 [ 15h 32m ] +INFO: Idle: 4563, Running: 4, Completed: 11453 [ 15h 32m ] +INFO: Idle: 4561, Running: 4, Completed: 11455 [ 15h 32m ] +INFO: Idle: 4557, Running: 4, Completed: 11459 [ 15h 32m ] +INFO: Idle: 4553, Running: 4, Completed: 11463 [ 15h 33m ] +INFO: Idle: 4549, Running: 4, Completed: 11467 [ 15h 33m ] +INFO: Idle: 4545, Running: 4, Completed: 11471 [ 15h 33m ] +INFO: Idle: 4544, Running: 4, Completed: 11472 [ 15h 34m ] +INFO: Idle: 4542, Running: 4, Completed: 11474 [ 15h 34m ] +INFO: Idle: 4541, Running: 4, Completed: 11475 [ 15h 34m ] +INFO: Idle: 4539, Running: 4, Completed: 11477 [ 15h 34m ] +INFO: Idle: 4538, Running: 4, Completed: 11478 [ 15h 34m ] +INFO: Idle: 4536, Running: 4, Completed: 11480 [ 15h 34m ] +INFO: Idle: 4535, Running: 4, Completed: 11481 [ 15h 34m ] +INFO: Idle: 4534, Running: 4, Completed: 11482 [ 15h 34m ] +INFO: Idle: 4533, Running: 4, Completed: 11483 [ 15h 34m ] +INFO: Idle: 4530, Running: 4, Completed: 11486 [ 15h 35m ] +INFO: Idle: 4528, Running: 4, Completed: 11488 [ 15h 35m ] +INFO: Idle: 4526, Running: 4, Completed: 11490 [ 15h 35m ] +INFO: Idle: 4525, Running: 4, Completed: 11491 [ 15h 35m ] +INFO: Idle: 4522, Running: 4, Completed: 11494 [ 15h 35m ] +INFO: Idle: 4521, Running: 4, Completed: 11495 [ 15h 35m ] +INFO: Idle: 4520, Running: 4, Completed: 11496 [ 15h 36m ] +INFO: Idle: 4518, Running: 4, Completed: 11498 [ 15h 36m ] +INFO: Idle: 4517, Running: 4, Completed: 11499 [ 15h 36m ] +INFO: Idle: 4516, Running: 4, Completed: 11500 [ 15h 36m ] +INFO: Idle: 4514, Running: 4, Completed: 11502 [ 15h 36m ] +INFO: Idle: 4512, Running: 4, Completed: 11504 [ 15h 36m ] +INFO: Idle: 4510, Running: 4, Completed: 11506 [ 15h 36m ] +INFO: Idle: 4508, Running: 4, Completed: 11508 [ 15h 37m ] +INFO: Idle: 4506, Running: 4, Completed: 11510 [ 15h 37m ] +INFO: Idle: 4504, Running: 4, Completed: 11512 [ 15h 37m ] +INFO: Idle: 4502, Running: 4, Completed: 11514 [ 15h 37m ] +INFO: Idle: 4500, Running: 4, Completed: 11516 [ 15h 37m ] +INFO: Idle: 4497, Running: 4, Completed: 11519 [ 15h 37m ] +INFO: Idle: 4496, Running: 4, Completed: 11520 [ 15h 38m ] +INFO: Idle: 4493, Running: 4, Completed: 11523 [ 15h 38m ] +INFO: Idle: 4492, Running: 4, Completed: 11524 [ 15h 38m ] +INFO: Idle: 4491, Running: 4, Completed: 11525 [ 15h 38m ] +INFO: Idle: 4488, Running: 4, Completed: 11528 [ 15h 38m ] +INFO: Idle: 4486, Running: 4, Completed: 11530 [ 15h 38m ] +INFO: Idle: 4485, Running: 4, Completed: 11531 [ 15h 38m ] +INFO: Idle: 4483, Running: 4, Completed: 11533 [ 15h 38m ] +INFO: Idle: 4481, Running: 4, Completed: 11535 [ 15h 39m ] +INFO: Idle: 4478, Running: 4, Completed: 11538 [ 15h 39m ] +INFO: Idle: 4477, Running: 4, Completed: 11539 [ 15h 39m ] +INFO: Idle: 4475, Running: 4, Completed: 11541 [ 15h 39m ] +INFO: Idle: 4473, Running: 4, Completed: 11543 [ 15h 39m ] +INFO: Idle: 4472, Running: 4, Completed: 11544 [ 15h 39m ] +INFO: Idle: 4469, Running: 4, Completed: 11547 [ 15h 40m ] +INFO: Idle: 4468, Running: 4, Completed: 11548 [ 15h 40m ] +INFO: Idle: 4467, Running: 4, Completed: 11549 [ 15h 40m ] +INFO: Idle: 4465, Running: 4, Completed: 11551 [ 15h 40m ] +INFO: Idle: 4464, Running: 4, Completed: 11552 [ 15h 40m ] +INFO: Idle: 4463, Running: 4, Completed: 11553 [ 15h 40m ] +INFO: Idle: 4461, Running: 4, Completed: 11555 [ 15h 40m ] +INFO: Idle: 4459, Running: 4, Completed: 11557 [ 15h 41m ] +INFO: Idle: 4457, Running: 4, Completed: 11559 [ 15h 41m ] +INFO: Idle: 4455, Running: 4, Completed: 11561 [ 15h 41m ] +INFO: Idle: 4453, Running: 4, Completed: 11563 [ 15h 41m ] +INFO: Idle: 4451, Running: 4, Completed: 11565 [ 15h 41m ] +INFO: Idle: 4449, Running: 4, Completed: 11567 [ 15h 41m ] +INFO: Idle: 4447, Running: 4, Completed: 11569 [ 15h 42m ] +INFO: Idle: 4444, Running: 4, Completed: 11572 [ 15h 42m ] +INFO: Idle: 4443, Running: 4, Completed: 11573 [ 15h 42m ] +INFO: Idle: 4439, Running: 4, Completed: 11577 [ 15h 42m ] +INFO: Idle: 4438, Running: 4, Completed: 11578 [ 15h 42m ] +INFO: Idle: 4435, Running: 4, Completed: 11581 [ 15h 42m ] +INFO: Idle: 4433, Running: 4, Completed: 11583 [ 15h 43m ] +INFO: Idle: 4432, Running: 4, Completed: 11584 [ 15h 43m ] +INFO: Idle: 4430, Running: 4, Completed: 11586 [ 15h 43m ] +INFO: Idle: 4429, Running: 4, Completed: 11587 [ 15h 43m ] +INFO: Idle: 4428, Running: 4, Completed: 11588 [ 15h 43m ] +INFO: Idle: 4426, Running: 4, Completed: 11590 [ 15h 43m ] +INFO: Idle: 4424, Running: 4, Completed: 11592 [ 15h 43m ] +INFO: Idle: 4422, Running: 4, Completed: 11594 [ 15h 44m ] +INFO: Idle: 4420, Running: 4, Completed: 11596 [ 15h 44m ] +INFO: Idle: 4419, Running: 4, Completed: 11597 [ 15h 44m ] +INFO: Idle: 4418, Running: 4, Completed: 11598 [ 15h 44m ] +INFO: Idle: 4415, Running: 4, Completed: 11601 [ 15h 44m ] +INFO: Idle: 4414, Running: 4, Completed: 11602 [ 15h 44m ] +INFO: Idle: 4413, Running: 4, Completed: 11603 [ 15h 44m ] +INFO: Idle: 4410, Running: 4, Completed: 11606 [ 15h 45m ] +INFO: Idle: 4409, Running: 4, Completed: 11607 [ 15h 45m ] +INFO: Idle: 4407, Running: 4, Completed: 11609 [ 15h 45m ] +INFO: Idle: 4406, Running: 4, Completed: 11610 [ 15h 45m ] +INFO: Idle: 4405, Running: 4, Completed: 11611 [ 15h 45m ] +INFO: Idle: 4403, Running: 4, Completed: 11613 [ 15h 45m ] +INFO: Idle: 4402, Running: 4, Completed: 11614 [ 15h 45m ] +INFO: Idle: 4400, Running: 4, Completed: 11616 [ 15h 45m ] +INFO: Idle: 4398, Running: 4, Completed: 11618 [ 15h 46m ] +INFO: Idle: 4396, Running: 4, Completed: 11620 [ 15h 46m ] +INFO: Idle: 4394, Running: 4, Completed: 11622 [ 15h 46m ] +INFO: Idle: 4392, Running: 4, Completed: 11624 [ 15h 46m ] +INFO: Idle: 4390, Running: 4, Completed: 11626 [ 15h 46m ] +INFO: Idle: 4389, Running: 4, Completed: 11627 [ 15h 46m ] +INFO: Idle: 4387, Running: 4, Completed: 11629 [ 15h 47m ] +INFO: Idle: 4386, Running: 4, Completed: 11630 [ 15h 47m ] +INFO: Idle: 4385, Running: 4, Completed: 11631 [ 15h 47m ] +INFO: Idle: 4383, Running: 4, Completed: 11633 [ 15h 47m ] +INFO: Idle: 4382, Running: 4, Completed: 11634 [ 15h 47m ] +INFO: Idle: 4381, Running: 4, Completed: 11635 [ 15h 47m ] +INFO: Idle: 4379, Running: 4, Completed: 11637 [ 15h 47m ] +INFO: Idle: 4378, Running: 4, Completed: 11638 [ 15h 47m ] +INFO: Idle: 4376, Running: 4, Completed: 11640 [ 15h 47m ] +INFO: Idle: 4374, Running: 4, Completed: 11642 [ 15h 48m ] +INFO: Idle: 4372, Running: 4, Completed: 11644 [ 15h 48m ] +INFO: Idle: 4370, Running: 4, Completed: 11646 [ 15h 48m ] +INFO: Idle: 4368, Running: 4, Completed: 11648 [ 15h 48m ] +INFO: Idle: 4366, Running: 4, Completed: 11650 [ 15h 48m ] +INFO: Idle: 4362, Running: 4, Completed: 11654 [ 15h 49m ] +INFO: Idle: 4359, Running: 4, Completed: 11657 [ 15h 49m ] +INFO: Idle: 4358, Running: 4, Completed: 11658 [ 15h 49m ] +INFO: Idle: 4355, Running: 4, Completed: 11661 [ 15h 49m ] +INFO: Idle: 4354, Running: 4, Completed: 11662 [ 15h 49m ] +INFO: Idle: 4353, Running: 4, Completed: 11663 [ 15h 49m ] +INFO: Idle: 4350, Running: 4, Completed: 11666 [ 15h 50m ] +INFO: Idle: 4348, Running: 4, Completed: 11668 [ 15h 50m ] +INFO: Idle: 4346, Running: 4, Completed: 11670 [ 15h 50m ] +INFO: Idle: 4344, Running: 4, Completed: 11672 [ 15h 50m ] +INFO: Idle: 4342, Running: 4, Completed: 11674 [ 15h 50m ] +INFO: Idle: 4341, Running: 4, Completed: 11675 [ 15h 51m ] +INFO: Idle: 4339, Running: 4, Completed: 11677 [ 15h 51m ] +INFO: Idle: 4338, Running: 4, Completed: 11678 [ 15h 51m ] +INFO: Idle: 4337, Running: 4, Completed: 11679 [ 15h 51m ] +INFO: Idle: 4335, Running: 4, Completed: 11681 [ 15h 51m ] +INFO: Idle: 4334, Running: 4, Completed: 11682 [ 15h 51m ] +INFO: Idle: 4333, Running: 4, Completed: 11683 [ 15h 51m ] +INFO: Idle: 4331, Running: 4, Completed: 11685 [ 15h 51m ] +INFO: Idle: 4330, Running: 4, Completed: 11686 [ 15h 51m ] +INFO: Idle: 4328, Running: 4, Completed: 11688 [ 15h 51m ] +INFO: Idle: 4326, Running: 4, Completed: 11690 [ 15h 52m ] +INFO: Idle: 4325, Running: 4, Completed: 11691 [ 15h 52m ] +INFO: Idle: 4322, Running: 4, Completed: 11694 [ 15h 52m ] +INFO: Idle: 4320, Running: 4, Completed: 11696 [ 15h 52m ] +INFO: Idle: 4319, Running: 4, Completed: 11697 [ 15h 52m ] +INFO: Idle: 4318, Running: 4, Completed: 11698 [ 15h 52m ] +INFO: Idle: 4316, Running: 4, Completed: 11700 [ 15h 52m ] +INFO: Idle: 4315, Running: 4, Completed: 11701 [ 15h 52m ] +INFO: Idle: 4314, Running: 4, Completed: 11702 [ 15h 53m ] +INFO: Idle: 4313, Running: 4, Completed: 11703 [ 15h 53m ] +INFO: Idle: 4312, Running: 4, Completed: 11704 [ 15h 53m ] +INFO: Idle: 4310, Running: 4, Completed: 11706 [ 15h 53m ] +INFO: Idle: 4307, Running: 4, Completed: 11709 [ 15h 53m ] +INFO: Idle: 4306, Running: 4, Completed: 11710 [ 15h 53m ] +INFO: Idle: 4303, Running: 4, Completed: 11713 [ 15h 53m ] +INFO: Idle: 4302, Running: 4, Completed: 11714 [ 15h 54m ] +INFO: Idle: 4299, Running: 4, Completed: 11717 [ 15h 54m ] +INFO: Idle: 4298, Running: 4, Completed: 11718 [ 15h 54m ] +INFO: Idle: 4296, Running: 4, Completed: 11720 [ 15h 54m ] +INFO: Idle: 4295, Running: 4, Completed: 11721 [ 15h 54m ] +INFO: Idle: 4294, Running: 4, Completed: 11722 [ 15h 54m ] +INFO: Idle: 4291, Running: 4, Completed: 11725 [ 15h 54m ] +INFO: Idle: 4290, Running: 4, Completed: 11726 [ 15h 55m ] +INFO: Idle: 4289, Running: 4, Completed: 11727 [ 15h 55m ] +INFO: Idle: 4286, Running: 4, Completed: 11730 [ 15h 55m ] +INFO: Idle: 4285, Running: 4, Completed: 11731 [ 15h 55m ] +INFO: Idle: 4284, Running: 4, Completed: 11732 [ 15h 55m ] +INFO: Idle: 4282, Running: 4, Completed: 11734 [ 15h 55m ] +INFO: Idle: 4281, Running: 4, Completed: 11735 [ 15h 55m ] +INFO: Idle: 4280, Running: 4, Completed: 11736 [ 15h 55m ] +INFO: Idle: 4278, Running: 4, Completed: 11738 [ 15h 55m ] +INFO: Idle: 4277, Running: 4, Completed: 11739 [ 15h 55m ] +INFO: Idle: 4276, Running: 4, Completed: 11740 [ 15h 56m ] +INFO: Idle: 4275, Running: 4, Completed: 11741 [ 15h 56m ] +INFO: Idle: 4273, Running: 4, Completed: 11743 [ 15h 56m ] +INFO: Idle: 4272, Running: 4, Completed: 11744 [ 15h 56m ] +INFO: Idle: 4271, Running: 4, Completed: 11745 [ 15h 56m ] +INFO: Idle: 4269, Running: 4, Completed: 11747 [ 15h 56m ] +INFO: Idle: 4268, Running: 4, Completed: 11748 [ 15h 56m ] +INFO: Idle: 4267, Running: 4, Completed: 11749 [ 15h 57m ] +INFO: Idle: 4265, Running: 4, Completed: 11751 [ 15h 57m ] +INFO: Idle: 4263, Running: 4, Completed: 11753 [ 15h 57m ] +INFO: Idle: 4262, Running: 4, Completed: 11754 [ 15h 57m ] +INFO: Idle: 4261, Running: 4, Completed: 11755 [ 15h 57m ] +INFO: Idle: 4260, Running: 4, Completed: 11756 [ 15h 57m ] +INFO: Idle: 4258, Running: 4, Completed: 11758 [ 15h 57m ] +INFO: Idle: 4256, Running: 4, Completed: 11760 [ 15h 58m ] +INFO: Idle: 4253, Running: 4, Completed: 11763 [ 15h 58m ] +INFO: Idle: 4252, Running: 4, Completed: 11764 [ 15h 58m ] +INFO: Idle: 4249, Running: 4, Completed: 11767 [ 15h 58m ] +INFO: Idle: 4248, Running: 4, Completed: 11768 [ 15h 58m ] +INFO: Idle: 4245, Running: 4, Completed: 11771 [ 15h 58m ] +INFO: Idle: 4244, Running: 4, Completed: 11772 [ 15h 58m ] +INFO: Idle: 4240, Running: 4, Completed: 11776 [ 15h 59m ] +INFO: Idle: 4237, Running: 4, Completed: 11779 [ 15h 59m ] +INFO: Idle: 4236, Running: 4, Completed: 11780 [ 15h 59m ] +INFO: Idle: 4233, Running: 4, Completed: 11783 [ 15h 59m ] +INFO: Idle: 4232, Running: 4, Completed: 11784 [ 15h 59m ] +INFO: Idle: 4231, Running: 4, Completed: 11785 [ 16h 0m ] +INFO: Idle: 4230, Running: 4, Completed: 11786 [ 16h 0m ] +INFO: Idle: 4228, Running: 4, Completed: 11788 [ 16h 0m ] +INFO: Idle: 4225, Running: 4, Completed: 11791 [ 16h 0m ] +INFO: Idle: 4224, Running: 4, Completed: 11792 [ 16h 0m ] +INFO: Idle: 4221, Running: 4, Completed: 11795 [ 16h 0m ] +INFO: Idle: 4220, Running: 4, Completed: 11796 [ 16h 0m ] +INFO: Idle: 4217, Running: 4, Completed: 11799 [ 16h 1m ] +INFO: Idle: 4216, Running: 4, Completed: 11800 [ 16h 1m ] +INFO: Idle: 4214, Running: 4, Completed: 11802 [ 16h 1m ] +INFO: Idle: 4213, Running: 4, Completed: 11803 [ 16h 1m ] +INFO: Idle: 4212, Running: 4, Completed: 11804 [ 16h 1m ] +INFO: Idle: 4209, Running: 4, Completed: 11807 [ 16h 1m ] +INFO: Idle: 4208, Running: 4, Completed: 11808 [ 16h 2m ] +INFO: Idle: 4207, Running: 4, Completed: 11809 [ 16h 2m ] +INFO: Idle: 4205, Running: 4, Completed: 11811 [ 16h 2m ] +INFO: Idle: 4204, Running: 4, Completed: 11812 [ 16h 2m ] +INFO: Idle: 4203, Running: 4, Completed: 11813 [ 16h 2m ] +INFO: Idle: 4202, Running: 4, Completed: 11814 [ 16h 2m ] +INFO: Idle: 4201, Running: 4, Completed: 11815 [ 16h 2m ] +INFO: Idle: 4199, Running: 4, Completed: 11817 [ 16h 2m ] +INFO: Idle: 4198, Running: 4, Completed: 11818 [ 16h 2m ] +INFO: Idle: 4197, Running: 4, Completed: 11819 [ 16h 2m ] +INFO: Idle: 4195, Running: 4, Completed: 11821 [ 16h 2m ] +INFO: Idle: 4194, Running: 4, Completed: 11822 [ 16h 3m ] +INFO: Idle: 4193, Running: 4, Completed: 11823 [ 16h 3m ] +INFO: Idle: 4191, Running: 4, Completed: 11825 [ 16h 3m ] +INFO: Idle: 4190, Running: 4, Completed: 11826 [ 16h 3m ] +INFO: Idle: 4189, Running: 4, Completed: 11827 [ 16h 3m ] +INFO: Idle: 4187, Running: 4, Completed: 11829 [ 16h 3m ] +INFO: Idle: 4185, Running: 4, Completed: 11831 [ 16h 3m ] +INFO: Idle: 4183, Running: 4, Completed: 11833 [ 16h 4m ] +INFO: Idle: 4181, Running: 4, Completed: 11835 [ 16h 4m ] +INFO: Idle: 4179, Running: 4, Completed: 11837 [ 16h 4m ] +INFO: Idle: 4178, Running: 4, Completed: 11838 [ 16h 4m ] +INFO: Idle: 4176, Running: 4, Completed: 11840 [ 16h 4m ] +INFO: Idle: 4174, Running: 4, Completed: 11842 [ 16h 5m ] +INFO: Idle: 4171, Running: 4, Completed: 11845 [ 16h 5m ] +INFO: Idle: 4170, Running: 4, Completed: 11846 [ 16h 5m ] +INFO: Idle: 4167, Running: 4, Completed: 11849 [ 16h 5m ] +INFO: Idle: 4166, Running: 4, Completed: 11850 [ 16h 5m ] +INFO: Idle: 4163, Running: 4, Completed: 11853 [ 16h 5m ] +INFO: Idle: 4162, Running: 4, Completed: 11854 [ 16h 5m ] +INFO: Idle: 4158, Running: 4, Completed: 11858 [ 16h 6m ] +INFO: Idle: 4155, Running: 4, Completed: 11861 [ 16h 6m ] +INFO: Idle: 4154, Running: 4, Completed: 11862 [ 16h 6m ] +INFO: Idle: 4151, Running: 4, Completed: 11865 [ 16h 6m ] +INFO: Idle: 4150, Running: 4, Completed: 11866 [ 16h 7m ] +INFO: Idle: 4146, Running: 4, Completed: 11870 [ 16h 7m ] +INFO: Idle: 4142, Running: 4, Completed: 11874 [ 16h 7m ] +INFO: Idle: 4138, Running: 4, Completed: 11878 [ 16h 7m ] +INFO: Idle: 4134, Running: 4, Completed: 11882 [ 16h 8m ] +INFO: Idle: 4132, Running: 4, Completed: 11884 [ 16h 8m ] +INFO: Idle: 4131, Running: 4, Completed: 11885 [ 16h 8m ] +INFO: Idle: 4130, Running: 4, Completed: 11886 [ 16h 8m ] +INFO: Idle: 4129, Running: 4, Completed: 11887 [ 16h 8m ] +INFO: Idle: 4127, Running: 4, Completed: 11889 [ 16h 8m ] +INFO: Idle: 4126, Running: 4, Completed: 11890 [ 16h 8m ] +INFO: Idle: 4124, Running: 4, Completed: 11892 [ 16h 9m ] +INFO: Idle: 4123, Running: 4, Completed: 11893 [ 16h 9m ] +INFO: Idle: 4122, Running: 4, Completed: 11894 [ 16h 9m ] +INFO: Idle: 4121, Running: 4, Completed: 11895 [ 16h 9m ] +INFO: Idle: 4119, Running: 4, Completed: 11897 [ 16h 9m ] +INFO: Idle: 4118, Running: 4, Completed: 11898 [ 16h 9m ] +INFO: Idle: 4115, Running: 4, Completed: 11901 [ 16h 9m ] +INFO: Idle: 4114, Running: 4, Completed: 11902 [ 16h 10m ] +INFO: Idle: 4111, Running: 4, Completed: 11905 [ 16h 10m ] +INFO: Idle: 4110, Running: 4, Completed: 11906 [ 16h 10m ] +INFO: Idle: 4107, Running: 4, Completed: 11909 [ 16h 10m ] +INFO: Idle: 4106, Running: 4, Completed: 11910 [ 16h 10m ] +INFO: Idle: 4105, Running: 4, Completed: 11911 [ 16h 10m ] +INFO: Idle: 4102, Running: 4, Completed: 11914 [ 16h 11m ] +INFO: Idle: 4100, Running: 4, Completed: 11916 [ 16h 11m ] +INFO: Idle: 4098, Running: 4, Completed: 11918 [ 16h 11m ] +INFO: Idle: 4096, Running: 4, Completed: 11920 [ 16h 11m ] +INFO: Idle: 4094, Running: 4, Completed: 11922 [ 16h 11m ] +INFO: Idle: 4093, Running: 4, Completed: 11923 [ 16h 11m ] +INFO: Idle: 4091, Running: 4, Completed: 11925 [ 16h 11m ] +INFO: Idle: 4090, Running: 4, Completed: 11926 [ 16h 12m ] +INFO: Idle: 4089, Running: 4, Completed: 11927 [ 16h 12m ] +INFO: Idle: 4088, Running: 4, Completed: 11928 [ 16h 12m ] +INFO: Idle: 4087, Running: 4, Completed: 11929 [ 16h 12m ] +INFO: Idle: 4086, Running: 4, Completed: 11930 [ 16h 12m ] +INFO: Idle: 4084, Running: 4, Completed: 11932 [ 16h 12m ] +INFO: Idle: 4083, Running: 4, Completed: 11933 [ 16h 12m ] +INFO: Idle: 4082, Running: 4, Completed: 11934 [ 16h 12m ] +INFO: Idle: 4080, Running: 4, Completed: 11936 [ 16h 12m ] +INFO: Idle: 4078, Running: 4, Completed: 11938 [ 16h 12m ] +INFO: Idle: 4076, Running: 4, Completed: 11940 [ 16h 12m ] +INFO: Idle: 4074, Running: 4, Completed: 11942 [ 16h 13m ] +INFO: Idle: 4072, Running: 4, Completed: 11944 [ 16h 13m ] +INFO: Idle: 4071, Running: 4, Completed: 11945 [ 16h 13m ] +INFO: Idle: 4070, Running: 4, Completed: 11946 [ 16h 13m ] +INFO: Idle: 4067, Running: 4, Completed: 11949 [ 16h 13m ] +INFO: Idle: 4066, Running: 4, Completed: 11950 [ 16h 13m ] +INFO: Idle: 4065, Running: 4, Completed: 11951 [ 16h 13m ] +INFO: Idle: 4064, Running: 4, Completed: 11952 [ 16h 13m ] +INFO: Idle: 4063, Running: 4, Completed: 11953 [ 16h 14m ] +INFO: Idle: 4062, Running: 4, Completed: 11954 [ 16h 14m ] +INFO: Idle: 4061, Running: 4, Completed: 11955 [ 16h 14m ] +INFO: Idle: 4059, Running: 4, Completed: 11957 [ 16h 14m ] +INFO: Idle: 4058, Running: 4, Completed: 11958 [ 16h 14m ] +INFO: Idle: 4056, Running: 4, Completed: 11960 [ 16h 14m ] +INFO: Idle: 4055, Running: 4, Completed: 11961 [ 16h 14m ] +INFO: Idle: 4054, Running: 4, Completed: 11962 [ 16h 14m ] +INFO: Idle: 4051, Running: 4, Completed: 11965 [ 16h 15m ] +INFO: Idle: 4050, Running: 4, Completed: 11966 [ 16h 15m ] +INFO: Idle: 4048, Running: 4, Completed: 11968 [ 16h 15m ] +INFO: Idle: 4047, Running: 4, Completed: 11969 [ 16h 15m ] +INFO: Idle: 4046, Running: 4, Completed: 11970 [ 16h 15m ] +INFO: Idle: 4043, Running: 4, Completed: 11973 [ 16h 15m ] +INFO: Idle: 4042, Running: 4, Completed: 11974 [ 16h 15m ] +INFO: Idle: 4041, Running: 4, Completed: 11975 [ 16h 16m ] +INFO: Idle: 4038, Running: 4, Completed: 11978 [ 16h 16m ] +INFO: Idle: 4037, Running: 4, Completed: 11979 [ 16h 16m ] +INFO: Idle: 4036, Running: 4, Completed: 11980 [ 16h 16m ] +INFO: Idle: 4034, Running: 4, Completed: 11982 [ 16h 16m ] +INFO: Idle: 4033, Running: 4, Completed: 11983 [ 16h 16m ] +INFO: Idle: 4032, Running: 4, Completed: 11984 [ 16h 16m ] +INFO: Idle: 4030, Running: 4, Completed: 11986 [ 16h 16m ] +INFO: Idle: 4029, Running: 4, Completed: 11987 [ 16h 16m ] +INFO: Idle: 4028, Running: 4, Completed: 11988 [ 16h 17m ] +INFO: Idle: 4027, Running: 4, Completed: 11989 [ 16h 17m ] +INFO: Idle: 4025, Running: 4, Completed: 11991 [ 16h 17m ] +INFO: Idle: 4024, Running: 4, Completed: 11992 [ 16h 17m ] +INFO: Idle: 4023, Running: 4, Completed: 11993 [ 16h 17m ] +INFO: Idle: 4021, Running: 4, Completed: 11995 [ 16h 17m ] +INFO: Idle: 4020, Running: 4, Completed: 11996 [ 16h 17m ] +INFO: Idle: 4019, Running: 4, Completed: 11997 [ 16h 17m ] +INFO: Idle: 4017, Running: 4, Completed: 11999 [ 16h 18m ] +INFO: Idle: 4016, Running: 4, Completed: 12000 [ 16h 18m ] +INFO: Idle: 4014, Running: 4, Completed: 12002 [ 16h 18m ] +INFO: Idle: 4013, Running: 4, Completed: 12003 [ 16h 18m ] +INFO: Idle: 4012, Running: 4, Completed: 12004 [ 16h 18m ] +INFO: Idle: 4010, Running: 4, Completed: 12006 [ 16h 18m ] +INFO: Idle: 4008, Running: 4, Completed: 12008 [ 16h 18m ] +INFO: Idle: 4005, Running: 4, Completed: 12011 [ 16h 19m ] +INFO: Idle: 4004, Running: 4, Completed: 12012 [ 16h 19m ] +INFO: Idle: 4001, Running: 4, Completed: 12015 [ 16h 19m ] +INFO: Idle: 4000, Running: 4, Completed: 12016 [ 16h 19m ] +INFO: Idle: 3997, Running: 4, Completed: 12019 [ 16h 19m ] +INFO: Idle: 3996, Running: 4, Completed: 12020 [ 16h 19m ] +INFO: Idle: 3992, Running: 4, Completed: 12024 [ 16h 20m ] +INFO: Idle: 3989, Running: 4, Completed: 12027 [ 16h 20m ] +INFO: Idle: 3988, Running: 4, Completed: 12028 [ 16h 20m ] +INFO: Idle: 3985, Running: 4, Completed: 12031 [ 16h 20m ] +INFO: Idle: 3984, Running: 4, Completed: 12032 [ 16h 20m ] +INFO: Idle: 3983, Running: 4, Completed: 12033 [ 16h 20m ] +INFO: Idle: 3982, Running: 4, Completed: 12034 [ 16h 20m ] +INFO: Idle: 3980, Running: 4, Completed: 12036 [ 16h 21m ] +INFO: Idle: 3977, Running: 4, Completed: 12039 [ 16h 21m ] +INFO: Idle: 3976, Running: 4, Completed: 12040 [ 16h 21m ] +INFO: Idle: 3973, Running: 4, Completed: 12043 [ 16h 21m ] +INFO: Idle: 3972, Running: 4, Completed: 12044 [ 16h 21m ] +INFO: Idle: 3969, Running: 4, Completed: 12047 [ 16h 21m ] +INFO: Idle: 3968, Running: 4, Completed: 12048 [ 16h 22m ] +INFO: Idle: 3966, Running: 4, Completed: 12050 [ 16h 22m ] +INFO: Idle: 3965, Running: 4, Completed: 12051 [ 16h 22m ] +INFO: Idle: 3964, Running: 4, Completed: 12052 [ 16h 22m ] +INFO: Idle: 3961, Running: 4, Completed: 12055 [ 16h 22m ] +INFO: Idle: 3960, Running: 4, Completed: 12056 [ 16h 22m ] +INFO: Idle: 3959, Running: 4, Completed: 12057 [ 16h 22m ] +INFO: Idle: 3957, Running: 4, Completed: 12059 [ 16h 22m ] +INFO: Idle: 3956, Running: 4, Completed: 12060 [ 16h 23m ] +INFO: Idle: 3955, Running: 4, Completed: 12061 [ 16h 23m ] +INFO: Idle: 3954, Running: 4, Completed: 12062 [ 16h 23m ] +INFO: Idle: 3953, Running: 4, Completed: 12063 [ 16h 23m ] +INFO: Idle: 3951, Running: 4, Completed: 12065 [ 16h 23m ] +INFO: Idle: 3950, Running: 4, Completed: 12066 [ 16h 23m ] +INFO: Idle: 3949, Running: 4, Completed: 12067 [ 16h 23m ] +INFO: Idle: 3947, Running: 4, Completed: 12069 [ 16h 23m ] +INFO: Idle: 3946, Running: 4, Completed: 12070 [ 16h 23m ] +INFO: Idle: 3945, Running: 4, Completed: 12071 [ 16h 24m ] +INFO: Idle: 3943, Running: 4, Completed: 12073 [ 16h 24m ] +INFO: Idle: 3942, Running: 4, Completed: 12074 [ 16h 24m ] +INFO: Idle: 3941, Running: 4, Completed: 12075 [ 16h 24m ] +INFO: Idle: 3939, Running: 4, Completed: 12077 [ 16h 24m ] +INFO: Idle: 3937, Running: 4, Completed: 12079 [ 16h 24m ] +INFO: Idle: 3935, Running: 4, Completed: 12081 [ 16h 24m ] +INFO: Idle: 3933, Running: 4, Completed: 12083 [ 16h 25m ] +INFO: Idle: 3932, Running: 4, Completed: 12084 [ 16h 25m ] +INFO: Idle: 3931, Running: 4, Completed: 12085 [ 16h 25m ] +INFO: Idle: 3930, Running: 4, Completed: 12086 [ 16h 25m ] +INFO: Idle: 3928, Running: 4, Completed: 12088 [ 16h 25m ] +INFO: Idle: 3926, Running: 4, Completed: 12090 [ 16h 25m ] +INFO: Idle: 3923, Running: 4, Completed: 12093 [ 16h 25m ] +INFO: Idle: 3922, Running: 4, Completed: 12094 [ 16h 26m ] +INFO: Idle: 3919, Running: 4, Completed: 12097 [ 16h 26m ] +INFO: Idle: 3918, Running: 4, Completed: 12098 [ 16h 26m ] +INFO: Idle: 3915, Running: 4, Completed: 12101 [ 16h 26m ] +INFO: Idle: 3914, Running: 4, Completed: 12102 [ 16h 26m ] +INFO: Idle: 3910, Running: 4, Completed: 12106 [ 16h 27m ] +INFO: Idle: 3908, Running: 4, Completed: 12108 [ 16h 27m ] +INFO: Idle: 3906, Running: 4, Completed: 12110 [ 16h 27m ] +INFO: Idle: 3902, Running: 4, Completed: 12114 [ 16h 27m ] +INFO: Idle: 3899, Running: 4, Completed: 12117 [ 16h 27m ] +INFO: Idle: 3898, Running: 4, Completed: 12118 [ 16h 28m ] +INFO: Idle: 3895, Running: 4, Completed: 12121 [ 16h 28m ] +INFO: Idle: 3894, Running: 4, Completed: 12122 [ 16h 28m ] +INFO: Idle: 3893, Running: 4, Completed: 12123 [ 16h 28m ] +INFO: Idle: 3890, Running: 4, Completed: 12126 [ 16h 28m ] +INFO: Idle: 3889, Running: 4, Completed: 12127 [ 16h 28m ] +INFO: Idle: 3886, Running: 4, Completed: 12130 [ 16h 29m ] +INFO: Idle: 3885, Running: 4, Completed: 12131 [ 16h 29m ] +INFO: Idle: 3882, Running: 4, Completed: 12134 [ 16h 29m ] +INFO: Idle: 3881, Running: 4, Completed: 12135 [ 16h 29m ] +INFO: Idle: 3878, Running: 4, Completed: 12138 [ 16h 29m ] +INFO: Idle: 3877, Running: 4, Completed: 12139 [ 16h 29m ] +INFO: Idle: 3875, Running: 4, Completed: 12141 [ 16h 29m ] +INFO: Idle: 3874, Running: 4, Completed: 12142 [ 16h 30m ] +INFO: Idle: 3872, Running: 4, Completed: 12144 [ 16h 30m ] +INFO: Idle: 3871, Running: 4, Completed: 12145 [ 16h 30m ] +INFO: Idle: 3870, Running: 4, Completed: 12146 [ 16h 30m ] +INFO: Idle: 3869, Running: 4, Completed: 12147 [ 16h 30m ] +INFO: Idle: 3866, Running: 4, Completed: 12150 [ 16h 30m ] +INFO: Idle: 3865, Running: 4, Completed: 12151 [ 16h 30m ] +INFO: Idle: 3863, Running: 4, Completed: 12153 [ 16h 30m ] +INFO: Idle: 3862, Running: 4, Completed: 12154 [ 16h 31m ] +INFO: Idle: 3861, Running: 4, Completed: 12155 [ 16h 31m ] +INFO: Idle: 3859, Running: 4, Completed: 12157 [ 16h 31m ] +INFO: Idle: 3858, Running: 4, Completed: 12158 [ 16h 31m ] +INFO: Idle: 3856, Running: 4, Completed: 12160 [ 16h 31m ] +INFO: Idle: 3854, Running: 4, Completed: 12162 [ 16h 31m ] +INFO: Idle: 3853, Running: 4, Completed: 12163 [ 16h 31m ] +INFO: Idle: 3852, Running: 4, Completed: 12164 [ 16h 31m ] +INFO: Idle: 3850, Running: 4, Completed: 12166 [ 16h 32m ] +INFO: Idle: 3849, Running: 4, Completed: 12167 [ 16h 32m ] +INFO: Idle: 3848, Running: 4, Completed: 12168 [ 16h 32m ] +INFO: Idle: 3846, Running: 4, Completed: 12170 [ 16h 32m ] +INFO: Idle: 3845, Running: 4, Completed: 12171 [ 16h 32m ] +INFO: Idle: 3844, Running: 4, Completed: 12172 [ 16h 32m ] +INFO: Idle: 3842, Running: 4, Completed: 12174 [ 16h 32m ] +INFO: Idle: 3840, Running: 4, Completed: 12176 [ 16h 32m ] +INFO: Idle: 3838, Running: 4, Completed: 12178 [ 16h 33m ] +INFO: Idle: 3836, Running: 4, Completed: 12180 [ 16h 33m ] +INFO: Idle: 3835, Running: 4, Completed: 12181 [ 16h 33m ] +INFO: Idle: 3833, Running: 4, Completed: 12183 [ 16h 33m ] +INFO: Idle: 3832, Running: 4, Completed: 12184 [ 16h 33m ] +INFO: Idle: 3831, Running: 4, Completed: 12185 [ 16h 33m ] +INFO: Idle: 3830, Running: 4, Completed: 12186 [ 16h 33m ] +INFO: Idle: 3828, Running: 4, Completed: 12188 [ 16h 33m ] +INFO: Idle: 3827, Running: 4, Completed: 12189 [ 16h 33m ] +INFO: Idle: 3826, Running: 4, Completed: 12190 [ 16h 34m ] +INFO: Idle: 3825, Running: 4, Completed: 12191 [ 16h 34m ] +INFO: Idle: 3822, Running: 4, Completed: 12194 [ 16h 34m ] +INFO: Idle: 3821, Running: 4, Completed: 12195 [ 16h 34m ] +INFO: Idle: 3819, Running: 4, Completed: 12197 [ 16h 34m ] +INFO: Idle: 3817, Running: 4, Completed: 12199 [ 16h 34m ] +INFO: Idle: 3816, Running: 4, Completed: 12200 [ 16h 34m ] +INFO: Idle: 3814, Running: 4, Completed: 12202 [ 16h 35m ] +INFO: Idle: 3813, Running: 4, Completed: 12203 [ 16h 35m ] +INFO: Idle: 3812, Running: 4, Completed: 12204 [ 16h 35m ] +INFO: Idle: 3810, Running: 4, Completed: 12206 [ 16h 35m ] +INFO: Idle: 3809, Running: 4, Completed: 12207 [ 16h 35m ] +INFO: Idle: 3807, Running: 4, Completed: 12209 [ 16h 35m ] +INFO: Idle: 3806, Running: 4, Completed: 12210 [ 16h 35m ] +INFO: Idle: 3804, Running: 4, Completed: 12212 [ 16h 35m ] +INFO: Idle: 3803, Running: 4, Completed: 12213 [ 16h 35m ] +INFO: Idle: 3802, Running: 4, Completed: 12214 [ 16h 36m ] +INFO: Idle: 3799, Running: 4, Completed: 12217 [ 16h 36m ] +INFO: Idle: 3798, Running: 4, Completed: 12218 [ 16h 36m ] +INFO: Idle: 3797, Running: 4, Completed: 12219 [ 16h 36m ] +INFO: Idle: 3796, Running: 4, Completed: 12220 [ 16h 36m ] +INFO: Idle: 3794, Running: 4, Completed: 12222 [ 16h 36m ] +INFO: Idle: 3793, Running: 4, Completed: 12223 [ 16h 36m ] +INFO: Idle: 3791, Running: 4, Completed: 12225 [ 16h 36m ] +INFO: Idle: 3790, Running: 4, Completed: 12226 [ 16h 36m ] +INFO: Idle: 3789, Running: 4, Completed: 12227 [ 16h 36m ] +INFO: Idle: 3788, Running: 4, Completed: 12228 [ 16h 37m ] +INFO: Idle: 3785, Running: 4, Completed: 12231 [ 16h 37m ] +INFO: Idle: 3784, Running: 4, Completed: 12232 [ 16h 37m ] +INFO: Idle: 3783, Running: 4, Completed: 12233 [ 16h 37m ] +INFO: Idle: 3781, Running: 4, Completed: 12235 [ 16h 37m ] +INFO: Idle: 3780, Running: 4, Completed: 12236 [ 16h 37m ] +INFO: Idle: 3779, Running: 4, Completed: 12237 [ 16h 37m ] +INFO: Idle: 3776, Running: 4, Completed: 12240 [ 16h 37m ] +INFO: Idle: 3775, Running: 4, Completed: 12241 [ 16h 38m ] +INFO: Idle: 3773, Running: 4, Completed: 12243 [ 16h 38m ] +INFO: Idle: 3771, Running: 4, Completed: 12245 [ 16h 38m ] +INFO: Idle: 3769, Running: 4, Completed: 12247 [ 16h 38m ] +INFO: Idle: 3765, Running: 4, Completed: 12251 [ 16h 38m ] +INFO: Idle: 3763, Running: 4, Completed: 12253 [ 16h 38m ] +INFO: Idle: 3761, Running: 4, Completed: 12255 [ 16h 39m ] +INFO: Idle: 3760, Running: 4, Completed: 12256 [ 16h 39m ] +INFO: Idle: 3759, Running: 4, Completed: 12257 [ 16h 39m ] +INFO: Idle: 3756, Running: 4, Completed: 12260 [ 16h 39m ] +INFO: Idle: 3753, Running: 4, Completed: 12263 [ 16h 39m ] +INFO: Idle: 3752, Running: 4, Completed: 12264 [ 16h 39m ] +INFO: Idle: 3749, Running: 4, Completed: 12267 [ 16h 40m ] +INFO: Idle: 3748, Running: 4, Completed: 12268 [ 16h 40m ] +INFO: Idle: 3746, Running: 4, Completed: 12270 [ 16h 40m ] +INFO: Idle: 3745, Running: 4, Completed: 12271 [ 16h 40m ] +INFO: Idle: 3744, Running: 4, Completed: 12272 [ 16h 40m ] +INFO: Idle: 3743, Running: 4, Completed: 12273 [ 16h 40m ] +INFO: Idle: 3741, Running: 4, Completed: 12275 [ 16h 40m ] +INFO: Idle: 3740, Running: 4, Completed: 12276 [ 16h 40m ] +INFO: Idle: 3739, Running: 4, Completed: 12277 [ 16h 41m ] +INFO: Idle: 3737, Running: 4, Completed: 12279 [ 16h 41m ] +INFO: Idle: 3736, Running: 4, Completed: 12280 [ 16h 41m ] +INFO: Idle: 3735, Running: 4, Completed: 12281 [ 16h 41m ] +INFO: Idle: 3734, Running: 4, Completed: 12282 [ 16h 41m ] +INFO: Idle: 3732, Running: 4, Completed: 12284 [ 16h 41m ] +INFO: Idle: 3731, Running: 4, Completed: 12285 [ 16h 41m ] +INFO: Idle: 3730, Running: 4, Completed: 12286 [ 16h 41m ] +INFO: Idle: 3729, Running: 4, Completed: 12287 [ 16h 41m ] +INFO: Idle: 3727, Running: 4, Completed: 12289 [ 16h 41m ] +INFO: Idle: 3726, Running: 4, Completed: 12290 [ 16h 41m ] +INFO: Idle: 3724, Running: 4, Completed: 12292 [ 16h 42m ] +INFO: Idle: 3722, Running: 4, Completed: 12294 [ 16h 42m ] +INFO: Idle: 3721, Running: 4, Completed: 12295 [ 16h 42m ] +INFO: Idle: 3720, Running: 4, Completed: 12296 [ 16h 42m ] +INFO: Idle: 3717, Running: 4, Completed: 12299 [ 16h 42m ] +INFO: Idle: 3716, Running: 4, Completed: 12300 [ 16h 42m ] +INFO: Idle: 3714, Running: 4, Completed: 12302 [ 16h 43m ] +INFO: Idle: 3713, Running: 4, Completed: 12303 [ 16h 43m ] +INFO: Idle: 3712, Running: 4, Completed: 12304 [ 16h 43m ] +INFO: Idle: 3710, Running: 4, Completed: 12306 [ 16h 43m ] +INFO: Idle: 3709, Running: 4, Completed: 12307 [ 16h 43m ] +INFO: Idle: 3708, Running: 4, Completed: 12308 [ 16h 43m ] +INFO: Idle: 3707, Running: 4, Completed: 12309 [ 16h 43m ] +INFO: Idle: 3706, Running: 4, Completed: 12310 [ 16h 43m ] +INFO: Idle: 3704, Running: 4, Completed: 12312 [ 16h 43m ] +INFO: Idle: 3703, Running: 4, Completed: 12313 [ 16h 44m ] +INFO: Idle: 3702, Running: 4, Completed: 12314 [ 16h 44m ] +INFO: Idle: 3700, Running: 4, Completed: 12316 [ 16h 44m ] +INFO: Idle: 3699, Running: 4, Completed: 12317 [ 16h 44m ] +INFO: Idle: 3698, Running: 4, Completed: 12318 [ 16h 44m ] +INFO: Idle: 3696, Running: 4, Completed: 12320 [ 16h 44m ] +INFO: Idle: 3695, Running: 4, Completed: 12321 [ 16h 44m ] +INFO: Idle: 3694, Running: 4, Completed: 12322 [ 16h 44m ] +INFO: Idle: 3692, Running: 4, Completed: 12324 [ 16h 44m ] +INFO: Idle: 3691, Running: 4, Completed: 12325 [ 16h 44m ] +INFO: Idle: 3690, Running: 4, Completed: 12326 [ 16h 44m ] +INFO: Idle: 3688, Running: 4, Completed: 12328 [ 16h 45m ] +INFO: Idle: 3687, Running: 4, Completed: 12329 [ 16h 45m ] +INFO: Idle: 3686, Running: 4, Completed: 12330 [ 16h 45m ] +INFO: Idle: 3685, Running: 4, Completed: 12331 [ 16h 45m ] +INFO: Idle: 3683, Running: 4, Completed: 12333 [ 16h 45m ] +INFO: Idle: 3682, Running: 4, Completed: 12334 [ 16h 45m ] +INFO: Idle: 3681, Running: 4, Completed: 12335 [ 16h 45m ] +INFO: Idle: 3679, Running: 4, Completed: 12337 [ 16h 45m ] +INFO: Idle: 3678, Running: 4, Completed: 12338 [ 16h 45m ] +INFO: Idle: 3677, Running: 4, Completed: 12339 [ 16h 46m ] +INFO: Idle: 3676, Running: 4, Completed: 12340 [ 16h 46m ] +INFO: Idle: 3675, Running: 4, Completed: 12341 [ 16h 46m ] +INFO: Idle: 3674, Running: 4, Completed: 12342 [ 16h 46m ] +INFO: Idle: 3672, Running: 4, Completed: 12344 [ 16h 46m ] +INFO: Idle: 3671, Running: 4, Completed: 12345 [ 16h 46m ] +INFO: Idle: 3670, Running: 4, Completed: 12346 [ 16h 46m ] +INFO: Idle: 3668, Running: 4, Completed: 12348 [ 16h 46m ] +INFO: Idle: 3667, Running: 4, Completed: 12349 [ 16h 46m ] +INFO: Idle: 3666, Running: 4, Completed: 12350 [ 16h 46m ] +INFO: Idle: 3664, Running: 4, Completed: 12352 [ 16h 47m ] +INFO: Idle: 3663, Running: 4, Completed: 12353 [ 16h 47m ] +INFO: Idle: 3662, Running: 4, Completed: 12354 [ 16h 47m ] +INFO: Idle: 3660, Running: 4, Completed: 12356 [ 16h 47m ] +INFO: Idle: 3659, Running: 4, Completed: 12357 [ 16h 47m ] +INFO: Idle: 3658, Running: 4, Completed: 12358 [ 16h 47m ] +INFO: Idle: 3656, Running: 4, Completed: 12360 [ 16h 47m ] +INFO: Idle: 3653, Running: 4, Completed: 12363 [ 16h 48m ] +INFO: Idle: 3652, Running: 4, Completed: 12364 [ 16h 48m ] +INFO: Idle: 3650, Running: 4, Completed: 12366 [ 16h 48m ] +INFO: Idle: 3648, Running: 4, Completed: 12368 [ 16h 48m ] +INFO: Idle: 3647, Running: 4, Completed: 12369 [ 16h 48m ] +INFO: Idle: 3644, Running: 4, Completed: 12372 [ 16h 48m ] +INFO: Idle: 3643, Running: 4, Completed: 12373 [ 16h 49m ] +INFO: Idle: 3640, Running: 4, Completed: 12376 [ 16h 49m ] +INFO: Idle: 3639, Running: 4, Completed: 12377 [ 16h 49m ] +INFO: Idle: 3636, Running: 4, Completed: 12380 [ 16h 49m ] +INFO: Idle: 3635, Running: 4, Completed: 12381 [ 16h 49m ] +INFO: Idle: 3632, Running: 4, Completed: 12384 [ 16h 49m ] +INFO: Idle: 3631, Running: 4, Completed: 12385 [ 16h 49m ] +INFO: Idle: 3628, Running: 4, Completed: 12388 [ 16h 50m ] +INFO: Idle: 3626, Running: 4, Completed: 12390 [ 16h 50m ] +INFO: Idle: 3625, Running: 4, Completed: 12391 [ 16h 50m ] +INFO: Idle: 3624, Running: 4, Completed: 12392 [ 16h 50m ] +INFO: Idle: 3623, Running: 4, Completed: 12393 [ 16h 50m ] +INFO: Idle: 3621, Running: 4, Completed: 12395 [ 16h 50m ] +INFO: Idle: 3620, Running: 4, Completed: 12396 [ 16h 50m ] +INFO: Idle: 3617, Running: 4, Completed: 12399 [ 16h 51m ] +INFO: Idle: 3616, Running: 4, Completed: 12400 [ 16h 51m ] +INFO: Idle: 3615, Running: 4, Completed: 12401 [ 16h 51m ] +INFO: Idle: 3614, Running: 4, Completed: 12402 [ 16h 51m ] +INFO: Idle: 3612, Running: 4, Completed: 12404 [ 16h 51m ] +INFO: Idle: 3608, Running: 4, Completed: 12408 [ 16h 52m ] +INFO: Idle: 3605, Running: 4, Completed: 12411 [ 16h 52m ] +INFO: Idle: 3604, Running: 4, Completed: 12412 [ 16h 52m ] +INFO: Idle: 3601, Running: 4, Completed: 12415 [ 16h 52m ] +INFO: Idle: 3600, Running: 4, Completed: 12416 [ 16h 52m ] +INFO: Idle: 3599, Running: 4, Completed: 12417 [ 16h 52m ] +INFO: Idle: 3597, Running: 4, Completed: 12419 [ 16h 52m ] +INFO: Idle: 3596, Running: 4, Completed: 12420 [ 16h 53m ] +INFO: Idle: 3595, Running: 4, Completed: 12421 [ 16h 53m ] +INFO: Idle: 3592, Running: 4, Completed: 12424 [ 16h 53m ] +INFO: Idle: 3591, Running: 4, Completed: 12425 [ 16h 53m ] +INFO: Idle: 3590, Running: 4, Completed: 12426 [ 16h 53m ] +INFO: Idle: 3588, Running: 4, Completed: 12428 [ 16h 53m ] +INFO: Idle: 3587, Running: 4, Completed: 12429 [ 16h 53m ] +INFO: Idle: 3585, Running: 4, Completed: 12431 [ 16h 53m ] +INFO: Idle: 3583, Running: 4, Completed: 12433 [ 16h 53m ] +INFO: Idle: 3581, Running: 4, Completed: 12435 [ 16h 54m ] +INFO: Idle: 3580, Running: 4, Completed: 12436 [ 16h 54m ] +INFO: Idle: 3579, Running: 4, Completed: 12437 [ 16h 54m ] +INFO: Idle: 3578, Running: 4, Completed: 12438 [ 16h 54m ] +INFO: Idle: 3576, Running: 4, Completed: 12440 [ 16h 54m ] +INFO: Idle: 3575, Running: 4, Completed: 12441 [ 16h 54m ] +INFO: Idle: 3572, Running: 4, Completed: 12444 [ 16h 54m ] +INFO: Idle: 3571, Running: 4, Completed: 12445 [ 16h 55m ] +INFO: Idle: 3567, Running: 4, Completed: 12449 [ 16h 55m ] +INFO: Idle: 3564, Running: 4, Completed: 12452 [ 16h 55m ] +INFO: Idle: 3563, Running: 4, Completed: 12453 [ 16h 55m ] +INFO: Idle: 3560, Running: 4, Completed: 12456 [ 16h 55m ] +INFO: Idle: 3559, Running: 4, Completed: 12457 [ 16h 56m ] +INFO: Idle: 3556, Running: 4, Completed: 12460 [ 16h 56m ] +INFO: Idle: 3555, Running: 4, Completed: 12461 [ 16h 56m ] +INFO: Idle: 3553, Running: 4, Completed: 12463 [ 16h 56m ] +INFO: Idle: 3552, Running: 4, Completed: 12464 [ 16h 56m ] +INFO: Idle: 3551, Running: 4, Completed: 12465 [ 16h 56m ] +INFO: Idle: 3550, Running: 4, Completed: 12466 [ 16h 56m ] +INFO: Idle: 3548, Running: 4, Completed: 12468 [ 16h 56m ] +INFO: Idle: 3545, Running: 4, Completed: 12471 [ 16h 57m ] +INFO: Idle: 3544, Running: 4, Completed: 12472 [ 16h 57m ] +INFO: Idle: 3543, Running: 4, Completed: 12473 [ 16h 57m ] +INFO: Idle: 3542, Running: 4, Completed: 12474 [ 16h 57m ] +INFO: Idle: 3540, Running: 4, Completed: 12476 [ 16h 57m ] +INFO: Idle: 3536, Running: 4, Completed: 12480 [ 16h 58m ] +INFO: Idle: 3533, Running: 4, Completed: 12483 [ 16h 58m ] +INFO: Idle: 3532, Running: 4, Completed: 12484 [ 16h 58m ] +INFO: Idle: 3529, Running: 4, Completed: 12487 [ 16h 58m ] +INFO: Idle: 3528, Running: 4, Completed: 12488 [ 16h 58m ] +INFO: Idle: 3527, Running: 4, Completed: 12489 [ 16h 58m ] +INFO: Idle: 3524, Running: 4, Completed: 12492 [ 16h 59m ] +INFO: Idle: 3523, Running: 4, Completed: 12493 [ 16h 59m ] +INFO: Idle: 3520, Running: 4, Completed: 12496 [ 16h 59m ] +INFO: Idle: 3519, Running: 4, Completed: 12497 [ 16h 59m ] +INFO: Idle: 3517, Running: 4, Completed: 12499 [ 16h 59m ] +INFO: Idle: 3516, Running: 4, Completed: 12500 [ 16h 59m ] +INFO: Idle: 3515, Running: 4, Completed: 12501 [ 16h 59m ] +INFO: Idle: 3512, Running: 4, Completed: 12504 [ 16h 59m ] +INFO: Idle: 3511, Running: 4, Completed: 12505 [ 17h 0m ] +INFO: Idle: 3510, Running: 4, Completed: 12506 [ 17h 0m ] +INFO: Idle: 3508, Running: 4, Completed: 12508 [ 17h 0m ] +INFO: Idle: 3505, Running: 4, Completed: 12511 [ 17h 0m ] +INFO: Idle: 3504, Running: 4, Completed: 12512 [ 17h 0m ] +INFO: Idle: 3503, Running: 4, Completed: 12513 [ 17h 0m ] +INFO: Idle: 3500, Running: 4, Completed: 12516 [ 17h 1m ] +INFO: Idle: 3499, Running: 4, Completed: 12517 [ 17h 1m ] +INFO: Idle: 3496, Running: 4, Completed: 12520 [ 17h 1m ] +INFO: Idle: 3495, Running: 4, Completed: 12521 [ 17h 1m ] +INFO: Idle: 3492, Running: 4, Completed: 12524 [ 17h 1m ] +INFO: Idle: 3490, Running: 4, Completed: 12526 [ 17h 1m ] +INFO: Idle: 3489, Running: 4, Completed: 12527 [ 17h 1m ] +INFO: Idle: 3488, Running: 4, Completed: 12528 [ 17h 2m ] +INFO: Idle: 3486, Running: 4, Completed: 12530 [ 17h 2m ] +INFO: Idle: 3485, Running: 4, Completed: 12531 [ 17h 2m ] +INFO: Idle: 3484, Running: 4, Completed: 12532 [ 17h 2m ] +INFO: Idle: 3483, Running: 4, Completed: 12533 [ 17h 2m ] +INFO: Idle: 3481, Running: 4, Completed: 12535 [ 17h 2m ] +INFO: Idle: 3480, Running: 4, Completed: 12536 [ 17h 2m ] +INFO: Idle: 3479, Running: 4, Completed: 12537 [ 17h 2m ] +INFO: Idle: 3477, Running: 4, Completed: 12539 [ 17h 2m ] +INFO: Idle: 3475, Running: 4, Completed: 12541 [ 17h 3m ] +INFO: Idle: 3473, Running: 4, Completed: 12543 [ 17h 3m ] +INFO: Idle: 3472, Running: 4, Completed: 12544 [ 17h 3m ] +INFO: Idle: 3471, Running: 4, Completed: 12545 [ 17h 3m ] +INFO: Idle: 3468, Running: 4, Completed: 12548 [ 17h 3m ] +INFO: Idle: 3465, Running: 4, Completed: 12551 [ 17h 3m ] +INFO: Idle: 3464, Running: 4, Completed: 12552 [ 17h 4m ] +INFO: Idle: 3463, Running: 4, Completed: 12553 [ 17h 4m ] +INFO: Idle: 3461, Running: 4, Completed: 12555 [ 17h 4m ] +INFO: Idle: 3460, Running: 4, Completed: 12556 [ 17h 4m ] +INFO: Idle: 3459, Running: 4, Completed: 12557 [ 17h 4m ] +INFO: Idle: 3457, Running: 4, Completed: 12559 [ 17h 4m ] +INFO: Idle: 3456, Running: 4, Completed: 12560 [ 17h 4m ] +INFO: Idle: 3454, Running: 4, Completed: 12562 [ 17h 4m ] +INFO: Idle: 3452, Running: 4, Completed: 12564 [ 17h 5m ] +INFO: Idle: 3450, Running: 4, Completed: 12566 [ 17h 5m ] +INFO: Idle: 3448, Running: 4, Completed: 12568 [ 17h 5m ] +INFO: Idle: 3446, Running: 4, Completed: 12570 [ 17h 5m ] +INFO: Idle: 3444, Running: 4, Completed: 12572 [ 17h 5m ] +INFO: Idle: 3442, Running: 4, Completed: 12574 [ 17h 5m ] +INFO: Idle: 3441, Running: 4, Completed: 12575 [ 17h 5m ] +INFO: Idle: 3440, Running: 4, Completed: 12576 [ 17h 6m ] +INFO: Idle: 3437, Running: 4, Completed: 12579 [ 17h 6m ] +INFO: Idle: 3436, Running: 4, Completed: 12580 [ 17h 6m ] +INFO: Idle: 3434, Running: 4, Completed: 12582 [ 17h 6m ] +INFO: Idle: 3432, Running: 4, Completed: 12584 [ 17h 6m ] +INFO: Idle: 3430, Running: 4, Completed: 12586 [ 17h 6m ] +INFO: Idle: 3428, Running: 4, Completed: 12588 [ 17h 6m ] +INFO: Idle: 3425, Running: 4, Completed: 12591 [ 17h 7m ] +INFO: Idle: 3424, Running: 4, Completed: 12592 [ 17h 7m ] +INFO: Idle: 3422, Running: 4, Completed: 12594 [ 17h 7m ] +INFO: Idle: 3421, Running: 4, Completed: 12595 [ 17h 7m ] +INFO: Idle: 3419, Running: 4, Completed: 12597 [ 17h 7m ] +INFO: Idle: 3417, Running: 4, Completed: 12599 [ 17h 7m ] +INFO: Idle: 3415, Running: 4, Completed: 12601 [ 17h 7m ] +INFO: Idle: 3411, Running: 4, Completed: 12605 [ 17h 8m ] +INFO: Idle: 3410, Running: 4, Completed: 12606 [ 17h 8m ] +INFO: Idle: 3408, Running: 4, Completed: 12608 [ 17h 8m ] +INFO: Idle: 3407, Running: 4, Completed: 12609 [ 17h 8m ] +INFO: Idle: 3405, Running: 4, Completed: 12611 [ 17h 8m ] +INFO: Idle: 3404, Running: 4, Completed: 12612 [ 17h 8m ] +INFO: Idle: 3402, Running: 4, Completed: 12614 [ 17h 8m ] +INFO: Idle: 3401, Running: 4, Completed: 12615 [ 17h 8m ] +INFO: Idle: 3400, Running: 4, Completed: 12616 [ 17h 8m ] +INFO: Idle: 3398, Running: 4, Completed: 12618 [ 17h 9m ] +INFO: Idle: 3397, Running: 4, Completed: 12619 [ 17h 9m ] +INFO: Idle: 3395, Running: 4, Completed: 12621 [ 17h 9m ] +INFO: Idle: 3394, Running: 4, Completed: 12622 [ 17h 9m ] +INFO: Idle: 3393, Running: 4, Completed: 12623 [ 17h 9m ] +INFO: Idle: 3390, Running: 4, Completed: 12626 [ 17h 9m ] +INFO: Idle: 3389, Running: 4, Completed: 12627 [ 17h 9m ] +INFO: Idle: 3388, Running: 4, Completed: 12628 [ 17h 9m ] +INFO: Idle: 3385, Running: 4, Completed: 12631 [ 17h 10m ] +INFO: Idle: 3384, Running: 4, Completed: 12632 [ 17h 10m ] +INFO: Idle: 3382, Running: 4, Completed: 12634 [ 17h 10m ] +INFO: Idle: 3381, Running: 4, Completed: 12635 [ 17h 10m ] +INFO: Idle: 3380, Running: 4, Completed: 12636 [ 17h 10m ] +INFO: Idle: 3379, Running: 4, Completed: 12637 [ 17h 10m ] +INFO: Idle: 3377, Running: 4, Completed: 12639 [ 17h 10m ] +INFO: Idle: 3375, Running: 4, Completed: 12641 [ 17h 10m ] +INFO: Idle: 3373, Running: 4, Completed: 12643 [ 17h 10m ] +INFO: Idle: 3372, Running: 4, Completed: 12644 [ 17h 11m ] +INFO: Idle: 3370, Running: 4, Completed: 12646 [ 17h 11m ] +INFO: Idle: 3369, Running: 4, Completed: 12647 [ 17h 11m ] +INFO: Idle: 3368, Running: 4, Completed: 12648 [ 17h 11m ] +INFO: Idle: 3366, Running: 4, Completed: 12650 [ 17h 11m ] +INFO: Idle: 3365, Running: 4, Completed: 12651 [ 17h 11m ] +INFO: Idle: 3364, Running: 4, Completed: 12652 [ 17h 11m ] +INFO: Idle: 3363, Running: 4, Completed: 12653 [ 17h 11m ] +INFO: Idle: 3361, Running: 4, Completed: 12655 [ 17h 12m ] +INFO: Idle: 3360, Running: 4, Completed: 12656 [ 17h 12m ] +INFO: Idle: 3358, Running: 4, Completed: 12658 [ 17h 12m ] +INFO: Idle: 3357, Running: 4, Completed: 12659 [ 17h 12m ] +INFO: Idle: 3356, Running: 4, Completed: 12660 [ 17h 12m ] +INFO: Idle: 3355, Running: 4, Completed: 12661 [ 17h 12m ] +INFO: Idle: 3352, Running: 4, Completed: 12664 [ 17h 13m ] +INFO: Idle: 3350, Running: 4, Completed: 12666 [ 17h 13m ] +INFO: Idle: 3349, Running: 4, Completed: 12667 [ 17h 13m ] +INFO: Idle: 3348, Running: 4, Completed: 12668 [ 17h 13m ] +INFO: Idle: 3345, Running: 4, Completed: 12671 [ 17h 13m ] +INFO: Idle: 3344, Running: 4, Completed: 12672 [ 17h 13m ] +INFO: Idle: 3341, Running: 4, Completed: 12675 [ 17h 13m ] +INFO: Idle: 3340, Running: 4, Completed: 12676 [ 17h 14m ] +INFO: Idle: 3337, Running: 4, Completed: 12679 [ 17h 14m ] +INFO: Idle: 3336, Running: 4, Completed: 12680 [ 17h 14m ] +INFO: Idle: 3335, Running: 4, Completed: 12681 [ 17h 14m ] +INFO: Idle: 3332, Running: 4, Completed: 12684 [ 17h 14m ] +INFO: Idle: 3331, Running: 4, Completed: 12685 [ 17h 14m ] +INFO: Idle: 3329, Running: 4, Completed: 12687 [ 17h 15m ] +INFO: Idle: 3328, Running: 4, Completed: 12688 [ 17h 15m ] +INFO: Idle: 3327, Running: 4, Completed: 12689 [ 17h 15m ] +INFO: Idle: 3325, Running: 4, Completed: 12691 [ 17h 15m ] +INFO: Idle: 3324, Running: 4, Completed: 12692 [ 17h 15m ] +INFO: Idle: 3323, Running: 4, Completed: 12693 [ 17h 15m ] +INFO: Idle: 3321, Running: 4, Completed: 12695 [ 17h 15m ] +INFO: Idle: 3319, Running: 4, Completed: 12697 [ 17h 15m ] +INFO: Idle: 3318, Running: 4, Completed: 12698 [ 17h 15m ] +INFO: Idle: 3316, Running: 4, Completed: 12700 [ 17h 16m ] +INFO: Idle: 3315, Running: 4, Completed: 12701 [ 17h 16m ] +INFO: Idle: 3312, Running: 4, Completed: 12704 [ 17h 16m ] +INFO: Idle: 3311, Running: 4, Completed: 12705 [ 17h 16m ] +INFO: Idle: 3310, Running: 4, Completed: 12706 [ 17h 16m ] +INFO: Idle: 3308, Running: 4, Completed: 12708 [ 17h 16m ] +INFO: Idle: 3307, Running: 4, Completed: 12709 [ 17h 16m ] +INFO: Idle: 3306, Running: 4, Completed: 12710 [ 17h 16m ] +INFO: Idle: 3304, Running: 4, Completed: 12712 [ 17h 16m ] +INFO: Idle: 3303, Running: 4, Completed: 12713 [ 17h 17m ] +INFO: Idle: 3301, Running: 4, Completed: 12715 [ 17h 17m ] +INFO: Idle: 3299, Running: 4, Completed: 12717 [ 17h 17m ] +INFO: Idle: 3296, Running: 4, Completed: 12720 [ 17h 17m ] +INFO: Idle: 3295, Running: 4, Completed: 12721 [ 17h 17m ] +INFO: Idle: 3293, Running: 4, Completed: 12723 [ 17h 17m ] +INFO: Idle: 3291, Running: 4, Completed: 12725 [ 17h 17m ] +INFO: Idle: 3289, Running: 4, Completed: 12727 [ 17h 17m ] +INFO: Idle: 3287, Running: 4, Completed: 12729 [ 17h 18m ] +INFO: Idle: 3284, Running: 4, Completed: 12732 [ 17h 18m ] +INFO: Idle: 3282, Running: 4, Completed: 12734 [ 17h 18m ] +INFO: Idle: 3280, Running: 4, Completed: 12736 [ 17h 18m ] +INFO: Idle: 3278, Running: 4, Completed: 12738 [ 17h 18m ] +INFO: Idle: 3277, Running: 4, Completed: 12739 [ 17h 18m ] +INFO: Idle: 3275, Running: 4, Completed: 12741 [ 17h 19m ] +INFO: Idle: 3273, Running: 4, Completed: 12743 [ 17h 19m ] +INFO: Idle: 3272, Running: 4, Completed: 12744 [ 17h 19m ] +INFO: Idle: 3269, Running: 4, Completed: 12747 [ 17h 19m ] +INFO: Idle: 3268, Running: 4, Completed: 12748 [ 17h 19m ] +INFO: Idle: 3266, Running: 4, Completed: 12750 [ 17h 19m ] +INFO: Idle: 3265, Running: 4, Completed: 12751 [ 17h 19m ] +INFO: Idle: 3262, Running: 4, Completed: 12754 [ 17h 19m ] +INFO: Idle: 3261, Running: 4, Completed: 12755 [ 17h 20m ] +INFO: Idle: 3259, Running: 4, Completed: 12757 [ 17h 20m ] +INFO: Idle: 3258, Running: 4, Completed: 12758 [ 17h 20m ] +INFO: Idle: 3257, Running: 4, Completed: 12759 [ 17h 20m ] +INFO: Idle: 3256, Running: 4, Completed: 12760 [ 17h 20m ] +INFO: Idle: 3255, Running: 4, Completed: 12761 [ 17h 20m ] +INFO: Idle: 3253, Running: 4, Completed: 12763 [ 17h 20m ] +INFO: Idle: 3252, Running: 4, Completed: 12764 [ 17h 20m ] +INFO: Idle: 3250, Running: 4, Completed: 12766 [ 17h 20m ] +INFO: Idle: 3248, Running: 4, Completed: 12768 [ 17h 21m ] +INFO: Idle: 3246, Running: 4, Completed: 12770 [ 17h 21m ] +INFO: Idle: 3244, Running: 4, Completed: 12772 [ 17h 21m ] +INFO: Idle: 3243, Running: 4, Completed: 12773 [ 17h 21m ] +INFO: Idle: 3242, Running: 4, Completed: 12774 [ 17h 21m ] +INFO: Idle: 3241, Running: 4, Completed: 12775 [ 17h 21m ] +INFO: Idle: 3240, Running: 4, Completed: 12776 [ 17h 21m ] +INFO: Idle: 3239, Running: 4, Completed: 12777 [ 17h 21m ] +INFO: Idle: 3238, Running: 4, Completed: 12778 [ 17h 21m ] +INFO: Idle: 3237, Running: 4, Completed: 12779 [ 17h 21m ] +INFO: Idle: 3235, Running: 4, Completed: 12781 [ 17h 22m ] +INFO: Idle: 3233, Running: 4, Completed: 12783 [ 17h 22m ] +INFO: Idle: 3231, Running: 4, Completed: 12785 [ 17h 22m ] +INFO: Idle: 3227, Running: 4, Completed: 12789 [ 17h 22m ] +INFO: Idle: 3225, Running: 4, Completed: 12791 [ 17h 22m ] +INFO: Idle: 3223, Running: 4, Completed: 12793 [ 17h 22m ] +INFO: Idle: 3221, Running: 4, Completed: 12795 [ 17h 23m ] +INFO: Idle: 3219, Running: 4, Completed: 12797 [ 17h 23m ] +INFO: Idle: 3217, Running: 4, Completed: 12799 [ 17h 23m ] +INFO: Idle: 3216, Running: 4, Completed: 12800 [ 17h 23m ] +INFO: Idle: 3214, Running: 4, Completed: 12802 [ 17h 23m ] +INFO: Idle: 3212, Running: 4, Completed: 12804 [ 17h 23m ] +INFO: Idle: 3210, Running: 4, Completed: 12806 [ 17h 23m ] +INFO: Idle: 3209, Running: 4, Completed: 12807 [ 17h 23m ] +INFO: Idle: 3207, Running: 4, Completed: 12809 [ 17h 24m ] +INFO: Idle: 3205, Running: 4, Completed: 12811 [ 17h 24m ] +INFO: Idle: 3204, Running: 4, Completed: 12812 [ 17h 24m ] +INFO: Idle: 3201, Running: 4, Completed: 12815 [ 17h 24m ] +INFO: Idle: 3200, Running: 4, Completed: 12816 [ 17h 24m ] +INFO: Idle: 3198, Running: 4, Completed: 12818 [ 17h 24m ] +INFO: Idle: 3197, Running: 4, Completed: 12819 [ 17h 24m ] +INFO: Idle: 3196, Running: 4, Completed: 12820 [ 17h 25m ] +INFO: Idle: 3194, Running: 4, Completed: 12822 [ 17h 25m ] +INFO: Idle: 3193, Running: 4, Completed: 12823 [ 17h 25m ] +INFO: Idle: 3190, Running: 4, Completed: 12826 [ 17h 25m ] +INFO: Idle: 3189, Running: 4, Completed: 12827 [ 17h 25m ] +INFO: Idle: 3187, Running: 4, Completed: 12829 [ 17h 25m ] +INFO: Idle: 3186, Running: 4, Completed: 12830 [ 17h 25m ] +INFO: Idle: 3185, Running: 4, Completed: 12831 [ 17h 25m ] +INFO: Idle: 3183, Running: 4, Completed: 12833 [ 17h 26m ] +INFO: Idle: 3182, Running: 4, Completed: 12834 [ 17h 26m ] +INFO: Idle: 3181, Running: 4, Completed: 12835 [ 17h 26m ] +INFO: Idle: 3180, Running: 4, Completed: 12836 [ 17h 26m ] +INFO: Idle: 3177, Running: 4, Completed: 12839 [ 17h 26m ] +INFO: Idle: 3174, Running: 4, Completed: 12842 [ 17h 26m ] +INFO: Idle: 3173, Running: 4, Completed: 12843 [ 17h 27m ] +INFO: Idle: 3170, Running: 4, Completed: 12846 [ 17h 27m ] +INFO: Idle: 3169, Running: 4, Completed: 12847 [ 17h 27m ] +INFO: Idle: 3166, Running: 4, Completed: 12850 [ 17h 27m ] +INFO: Idle: 3165, Running: 4, Completed: 12851 [ 17h 27m ] +INFO: Idle: 3161, Running: 4, Completed: 12855 [ 17h 27m ] +INFO: Idle: 3159, Running: 4, Completed: 12857 [ 17h 28m ] +INFO: Idle: 3157, Running: 4, Completed: 12859 [ 17h 28m ] +INFO: Idle: 3155, Running: 4, Completed: 12861 [ 17h 28m ] +INFO: Idle: 3153, Running: 4, Completed: 12863 [ 17h 28m ] +INFO: Idle: 3151, Running: 4, Completed: 12865 [ 17h 28m ] +INFO: Idle: 3149, Running: 4, Completed: 12867 [ 17h 29m ] +INFO: Idle: 3147, Running: 4, Completed: 12869 [ 17h 29m ] +INFO: Idle: 3146, Running: 4, Completed: 12870 [ 17h 29m ] +INFO: Idle: 3145, Running: 4, Completed: 12871 [ 17h 29m ] +INFO: Idle: 3144, Running: 4, Completed: 12872 [ 17h 29m ] +INFO: Idle: 3141, Running: 4, Completed: 12875 [ 17h 29m ] +INFO: Idle: 3138, Running: 4, Completed: 12878 [ 17h 29m ] +INFO: Idle: 3137, Running: 4, Completed: 12879 [ 17h 30m ] +INFO: Idle: 3134, Running: 4, Completed: 12882 [ 17h 30m ] +INFO: Idle: 3133, Running: 4, Completed: 12883 [ 17h 30m ] +INFO: Idle: 3129, Running: 4, Completed: 12887 [ 17h 30m ] +INFO: Idle: 3125, Running: 4, Completed: 12891 [ 17h 31m ] +INFO: Idle: 3122, Running: 4, Completed: 12894 [ 17h 31m ] +INFO: Idle: 3121, Running: 4, Completed: 12895 [ 17h 31m ] +INFO: Idle: 3118, Running: 4, Completed: 12898 [ 17h 31m ] +INFO: Idle: 3117, Running: 4, Completed: 12899 [ 17h 31m ] +INFO: Idle: 3116, Running: 4, Completed: 12900 [ 17h 31m ] +INFO: Idle: 3115, Running: 4, Completed: 12901 [ 17h 31m ] +INFO: Idle: 3113, Running: 4, Completed: 12903 [ 17h 32m ] +INFO: Idle: 3110, Running: 4, Completed: 12906 [ 17h 32m ] +INFO: Idle: 3109, Running: 4, Completed: 12907 [ 17h 32m ] +INFO: Idle: 3106, Running: 4, Completed: 12910 [ 17h 32m ] +INFO: Idle: 3105, Running: 4, Completed: 12911 [ 17h 32m ] +INFO: Idle: 3102, Running: 4, Completed: 12914 [ 17h 32m ] +INFO: Idle: 3101, Running: 4, Completed: 12915 [ 17h 33m ] +INFO: Idle: 3098, Running: 4, Completed: 12918 [ 17h 33m ] +INFO: Idle: 3097, Running: 4, Completed: 12919 [ 17h 33m ] +INFO: Idle: 3094, Running: 4, Completed: 12922 [ 17h 33m ] +INFO: Idle: 3093, Running: 4, Completed: 12923 [ 17h 33m ] +INFO: Idle: 3092, Running: 4, Completed: 12924 [ 17h 33m ] +INFO: Idle: 3090, Running: 4, Completed: 12926 [ 17h 33m ] +INFO: Idle: 3089, Running: 4, Completed: 12927 [ 17h 34m ] +INFO: Idle: 3088, Running: 4, Completed: 12928 [ 17h 34m ] +INFO: Idle: 3085, Running: 4, Completed: 12931 [ 17h 34m ] +INFO: Idle: 3084, Running: 4, Completed: 12932 [ 17h 34m ] +INFO: Idle: 3082, Running: 4, Completed: 12934 [ 17h 34m ] +INFO: Idle: 3081, Running: 4, Completed: 12935 [ 17h 34m ] +INFO: Idle: 3080, Running: 4, Completed: 12936 [ 17h 34m ] +INFO: Idle: 3078, Running: 4, Completed: 12938 [ 17h 34m ] +INFO: Idle: 3077, Running: 4, Completed: 12939 [ 17h 34m ] +INFO: Idle: 3075, Running: 4, Completed: 12941 [ 17h 34m ] +INFO: Idle: 3074, Running: 4, Completed: 12942 [ 17h 35m ] +INFO: Idle: 3073, Running: 4, Completed: 12943 [ 17h 35m ] +INFO: Idle: 3072, Running: 4, Completed: 12944 [ 17h 35m ] +INFO: Idle: 3070, Running: 4, Completed: 12946 [ 17h 35m ] +INFO: Idle: 3069, Running: 4, Completed: 12947 [ 17h 35m ] +INFO: Idle: 3068, Running: 4, Completed: 12948 [ 17h 35m ] +INFO: Idle: 3065, Running: 4, Completed: 12951 [ 17h 35m ] +INFO: Idle: 3064, Running: 4, Completed: 12952 [ 17h 35m ] +INFO: Idle: 3062, Running: 4, Completed: 12954 [ 17h 36m ] +INFO: Idle: 3061, Running: 4, Completed: 12955 [ 17h 36m ] +INFO: Idle: 3060, Running: 4, Completed: 12956 [ 17h 36m ] +INFO: Idle: 3058, Running: 4, Completed: 12958 [ 17h 36m ] +INFO: Idle: 3057, Running: 4, Completed: 12959 [ 17h 36m ] +INFO: Idle: 3054, Running: 4, Completed: 12962 [ 17h 36m ] +INFO: Idle: 3053, Running: 4, Completed: 12963 [ 17h 37m ] +INFO: Idle: 3051, Running: 4, Completed: 12965 [ 17h 37m ] +INFO: Idle: 3050, Running: 4, Completed: 12966 [ 17h 37m ] +INFO: Idle: 3049, Running: 4, Completed: 12967 [ 17h 37m ] +INFO: Idle: 3045, Running: 4, Completed: 12971 [ 17h 37m ] +INFO: Idle: 3042, Running: 4, Completed: 12974 [ 17h 37m ] +INFO: Idle: 3041, Running: 4, Completed: 12975 [ 17h 38m ] +INFO: Idle: 3038, Running: 4, Completed: 12978 [ 17h 38m ] +INFO: Idle: 3037, Running: 4, Completed: 12979 [ 17h 38m ] +INFO: Idle: 3036, Running: 4, Completed: 12980 [ 17h 38m ] +INFO: Idle: 3033, Running: 4, Completed: 12983 [ 17h 38m ] +INFO: Idle: 3030, Running: 4, Completed: 12986 [ 17h 38m ] +INFO: Idle: 3029, Running: 4, Completed: 12987 [ 17h 39m ] +INFO: Idle: 3026, Running: 4, Completed: 12990 [ 17h 39m ] +INFO: Idle: 3025, Running: 4, Completed: 12991 [ 17h 39m ] +INFO: Idle: 3022, Running: 4, Completed: 12994 [ 17h 39m ] +INFO: Idle: 3021, Running: 4, Completed: 12995 [ 17h 39m ] +INFO: Idle: 3017, Running: 4, Completed: 12999 [ 17h 40m ] +INFO: Idle: 3013, Running: 4, Completed: 13003 [ 17h 40m ] +INFO: Idle: 3010, Running: 4, Completed: 13006 [ 17h 40m ] +INFO: Idle: 3009, Running: 4, Completed: 13007 [ 17h 40m ] +INFO: Idle: 3006, Running: 4, Completed: 13010 [ 17h 40m ] +INFO: Idle: 3005, Running: 4, Completed: 13011 [ 17h 40m ] +INFO: Idle: 3002, Running: 4, Completed: 13014 [ 17h 41m ] +INFO: Idle: 3001, Running: 4, Completed: 13015 [ 17h 41m ] +INFO: Idle: 2999, Running: 4, Completed: 13017 [ 17h 41m ] +INFO: Idle: 2997, Running: 4, Completed: 13019 [ 17h 41m ] +INFO: Idle: 2995, Running: 4, Completed: 13021 [ 17h 41m ] +INFO: Idle: 2993, Running: 4, Completed: 13023 [ 17h 41m ] +INFO: Idle: 2991, Running: 4, Completed: 13025 [ 17h 42m ] +INFO: Idle: 2990, Running: 4, Completed: 13026 [ 17h 42m ] +INFO: Idle: 2989, Running: 4, Completed: 13027 [ 17h 42m ] +INFO: Idle: 2988, Running: 4, Completed: 13028 [ 17h 42m ] +INFO: Idle: 2986, Running: 4, Completed: 13030 [ 17h 42m ] +INFO: Idle: 2985, Running: 4, Completed: 13031 [ 17h 42m ] +INFO: Idle: 2982, Running: 4, Completed: 13034 [ 17h 42m ] +INFO: Idle: 2981, Running: 4, Completed: 13035 [ 17h 43m ] +INFO: Idle: 2980, Running: 4, Completed: 13036 [ 17h 43m ] +INFO: Idle: 2979, Running: 4, Completed: 13037 [ 17h 43m ] +INFO: Idle: 2978, Running: 4, Completed: 13038 [ 17h 43m ] +INFO: Idle: 2977, Running: 4, Completed: 13039 [ 17h 43m ] +INFO: Idle: 2973, Running: 4, Completed: 13043 [ 17h 43m ] +INFO: Idle: 2970, Running: 4, Completed: 13046 [ 17h 43m ] +INFO: Idle: 2969, Running: 4, Completed: 13047 [ 17h 44m ] +INFO: Idle: 2966, Running: 4, Completed: 13050 [ 17h 44m ] +INFO: Idle: 2965, Running: 4, Completed: 13051 [ 17h 44m ] +INFO: Idle: 2964, Running: 4, Completed: 13052 [ 17h 44m ] +INFO: Idle: 2961, Running: 4, Completed: 13055 [ 17h 44m ] +INFO: Idle: 2959, Running: 4, Completed: 13057 [ 17h 44m ] +INFO: Idle: 2957, Running: 4, Completed: 13059 [ 17h 45m ] +INFO: Idle: 2954, Running: 4, Completed: 13062 [ 17h 45m ] +INFO: Idle: 2953, Running: 4, Completed: 13063 [ 17h 45m ] +INFO: Idle: 2950, Running: 4, Completed: 13066 [ 17h 45m ] +INFO: Idle: 2949, Running: 4, Completed: 13067 [ 17h 45m ] +INFO: Idle: 2945, Running: 4, Completed: 13071 [ 17h 46m ] +INFO: Idle: 2943, Running: 4, Completed: 13073 [ 17h 46m ] +INFO: Idle: 2942, Running: 4, Completed: 13074 [ 17h 46m ] +INFO: Idle: 2941, Running: 4, Completed: 13075 [ 17h 46m ] +INFO: Idle: 2938, Running: 4, Completed: 13078 [ 17h 46m ] +INFO: Idle: 2937, Running: 4, Completed: 13079 [ 17h 46m ] +INFO: Idle: 2934, Running: 4, Completed: 13082 [ 17h 47m ] +INFO: Idle: 2933, Running: 4, Completed: 13083 [ 17h 47m ] +INFO: Idle: 2930, Running: 4, Completed: 13086 [ 17h 47m ] +INFO: Idle: 2929, Running: 4, Completed: 13087 [ 17h 47m ] +INFO: Idle: 2928, Running: 4, Completed: 13088 [ 17h 47m ] +INFO: Idle: 2925, Running: 4, Completed: 13091 [ 17h 47m ] +INFO: Idle: 2924, Running: 4, Completed: 13092 [ 17h 48m ] +INFO: Idle: 2921, Running: 4, Completed: 13095 [ 17h 48m ] +INFO: Idle: 2920, Running: 4, Completed: 13096 [ 17h 48m ] +INFO: Idle: 2919, Running: 4, Completed: 13097 [ 17h 48m ] +INFO: Idle: 2917, Running: 4, Completed: 13099 [ 17h 48m ] +INFO: Idle: 2916, Running: 4, Completed: 13100 [ 17h 48m ] +INFO: Idle: 2914, Running: 4, Completed: 13102 [ 17h 48m ] +INFO: Idle: 2912, Running: 4, Completed: 13104 [ 17h 49m ] +INFO: Idle: 2911, Running: 4, Completed: 13105 [ 17h 49m ] +INFO: Idle: 2910, Running: 4, Completed: 13106 [ 17h 49m ] +INFO: Idle: 2909, Running: 4, Completed: 13107 [ 17h 49m ] +INFO: Idle: 2908, Running: 4, Completed: 13108 [ 17h 49m ] +INFO: Idle: 2905, Running: 4, Completed: 13111 [ 17h 49m ] +INFO: Idle: 2904, Running: 4, Completed: 13112 [ 17h 49m ] +INFO: Idle: 2903, Running: 4, Completed: 13113 [ 17h 49m ] +INFO: Idle: 2900, Running: 4, Completed: 13116 [ 17h 50m ] +INFO: Idle: 2899, Running: 4, Completed: 13117 [ 17h 50m ] +INFO: Idle: 2897, Running: 4, Completed: 13119 [ 17h 50m ] +INFO: Idle: 2896, Running: 4, Completed: 13120 [ 17h 50m ] +INFO: Idle: 2895, Running: 4, Completed: 13121 [ 17h 50m ] +INFO: Idle: 2893, Running: 4, Completed: 13123 [ 17h 50m ] +INFO: Idle: 2892, Running: 4, Completed: 13124 [ 17h 50m ] +INFO: Idle: 2890, Running: 4, Completed: 13126 [ 17h 50m ] +INFO: Idle: 2888, Running: 4, Completed: 13128 [ 17h 51m ] +INFO: Idle: 2885, Running: 4, Completed: 13131 [ 17h 51m ] +INFO: Idle: 2884, Running: 4, Completed: 13132 [ 17h 51m ] +INFO: Idle: 2883, Running: 4, Completed: 13133 [ 17h 51m ] +INFO: Idle: 2880, Running: 4, Completed: 13136 [ 17h 51m ] +INFO: Idle: 2879, Running: 4, Completed: 13137 [ 17h 52m ] +INFO: Idle: 2878, Running: 4, Completed: 13138 [ 17h 52m ] +INFO: Idle: 2876, Running: 4, Completed: 13140 [ 17h 52m ] +INFO: Idle: 2874, Running: 4, Completed: 13142 [ 17h 52m ] +INFO: Idle: 2872, Running: 4, Completed: 13144 [ 17h 52m ] +INFO: Idle: 2870, Running: 4, Completed: 13146 [ 17h 52m ] +INFO: Idle: 2868, Running: 4, Completed: 13148 [ 17h 52m ] +INFO: Idle: 2867, Running: 4, Completed: 13149 [ 17h 52m ] +INFO: Idle: 2864, Running: 4, Completed: 13152 [ 17h 53m ] +INFO: Idle: 2861, Running: 4, Completed: 13155 [ 17h 53m ] +INFO: Idle: 2860, Running: 4, Completed: 13156 [ 17h 53m ] +INFO: Idle: 2859, Running: 4, Completed: 13157 [ 17h 53m ] +INFO: Idle: 2858, Running: 4, Completed: 13158 [ 17h 53m ] +INFO: Idle: 2856, Running: 4, Completed: 13160 [ 17h 53m ] +INFO: Idle: 2855, Running: 4, Completed: 13161 [ 17h 54m ] +INFO: Idle: 2853, Running: 4, Completed: 13163 [ 17h 54m ] +INFO: Idle: 2851, Running: 4, Completed: 13165 [ 17h 54m ] +INFO: Idle: 2849, Running: 4, Completed: 13167 [ 17h 54m ] +INFO: Idle: 2848, Running: 4, Completed: 13168 [ 17h 54m ] +INFO: Idle: 2847, Running: 4, Completed: 13169 [ 17h 54m ] +INFO: Idle: 2845, Running: 4, Completed: 13171 [ 17h 54m ] +INFO: Idle: 2844, Running: 4, Completed: 13172 [ 17h 55m ] +INFO: Idle: 2843, Running: 4, Completed: 13173 [ 17h 55m ] +INFO: Idle: 2842, Running: 4, Completed: 13174 [ 17h 55m ] +INFO: Idle: 2840, Running: 4, Completed: 13176 [ 17h 55m ] +INFO: Idle: 2839, Running: 4, Completed: 13177 [ 17h 55m ] +INFO: Idle: 2838, Running: 4, Completed: 13178 [ 17h 55m ] +INFO: Idle: 2836, Running: 4, Completed: 13180 [ 17h 55m ] +INFO: Idle: 2835, Running: 4, Completed: 13181 [ 17h 55m ] +INFO: Idle: 2834, Running: 4, Completed: 13182 [ 17h 56m ] +INFO: Idle: 2833, Running: 4, Completed: 13183 [ 17h 56m ] +INFO: Idle: 2831, Running: 4, Completed: 13185 [ 17h 56m ] +INFO: Idle: 2830, Running: 4, Completed: 13186 [ 17h 56m ] +INFO: Idle: 2829, Running: 4, Completed: 13187 [ 17h 56m ] +INFO: Idle: 2827, Running: 4, Completed: 13189 [ 17h 56m ] +INFO: Idle: 2826, Running: 4, Completed: 13190 [ 17h 56m ] +INFO: Idle: 2825, Running: 4, Completed: 13191 [ 17h 56m ] +INFO: Idle: 2824, Running: 4, Completed: 13192 [ 17h 56m ] +INFO: Idle: 2823, Running: 4, Completed: 13193 [ 17h 56m ] +INFO: Idle: 2822, Running: 4, Completed: 13194 [ 17h 56m ] +INFO: Idle: 2820, Running: 4, Completed: 13196 [ 17h 57m ] +INFO: Idle: 2819, Running: 4, Completed: 13197 [ 17h 57m ] +INFO: Idle: 2818, Running: 4, Completed: 13198 [ 17h 57m ] +INFO: Idle: 2817, Running: 4, Completed: 13199 [ 17h 57m ] +INFO: Idle: 2815, Running: 4, Completed: 13201 [ 17h 57m ] +INFO: Idle: 2814, Running: 4, Completed: 13202 [ 17h 57m ] +INFO: Idle: 2813, Running: 4, Completed: 13203 [ 17h 57m ] +INFO: Idle: 2811, Running: 4, Completed: 13205 [ 17h 57m ] +INFO: Idle: 2810, Running: 4, Completed: 13206 [ 17h 58m ] +INFO: Idle: 2809, Running: 4, Completed: 13207 [ 17h 58m ] +INFO: Idle: 2807, Running: 4, Completed: 13209 [ 17h 58m ] +INFO: Idle: 2806, Running: 4, Completed: 13210 [ 17h 58m ] +INFO: Idle: 2804, Running: 4, Completed: 13212 [ 17h 58m ] +INFO: Idle: 2802, Running: 4, Completed: 13214 [ 17h 58m ] +INFO: Idle: 2799, Running: 4, Completed: 13217 [ 17h 58m ] +INFO: Idle: 2798, Running: 4, Completed: 13218 [ 17h 59m ] +INFO: Idle: 2797, Running: 4, Completed: 13219 [ 17h 59m ] +INFO: Idle: 2794, Running: 4, Completed: 13222 [ 17h 59m ] +INFO: Idle: 2793, Running: 4, Completed: 13223 [ 17h 59m ] +INFO: Idle: 2791, Running: 4, Completed: 13225 [ 17h 59m ] +INFO: Idle: 2790, Running: 4, Completed: 13226 [ 17h 59m ] +INFO: Idle: 2789, Running: 4, Completed: 13227 [ 17h 59m ] +INFO: Idle: 2788, Running: 4, Completed: 13228 [ 18h 0m ] +INFO: Idle: 2787, Running: 4, Completed: 13229 [ 18h 0m ] +INFO: Idle: 2786, Running: 4, Completed: 13230 [ 18h 0m ] +INFO: Idle: 2785, Running: 4, Completed: 13231 [ 18h 0m ] +INFO: Idle: 2782, Running: 4, Completed: 13234 [ 18h 0m ] +INFO: Idle: 2781, Running: 4, Completed: 13235 [ 18h 0m ] +INFO: Idle: 2778, Running: 4, Completed: 13238 [ 18h 0m ] +INFO: Idle: 2776, Running: 4, Completed: 13240 [ 18h 0m ] +INFO: Idle: 2775, Running: 4, Completed: 13241 [ 18h 0m ] +INFO: Idle: 2774, Running: 4, Completed: 13242 [ 18h 1m ] +INFO: Idle: 2773, Running: 4, Completed: 13243 [ 18h 1m ] +INFO: Idle: 2772, Running: 4, Completed: 13244 [ 18h 1m ] +INFO: Idle: 2771, Running: 4, Completed: 13245 [ 18h 1m ] +INFO: Idle: 2769, Running: 4, Completed: 13247 [ 18h 1m ] +INFO: Idle: 2768, Running: 4, Completed: 13248 [ 18h 1m ] +INFO: Idle: 2767, Running: 4, Completed: 13249 [ 18h 1m ] +INFO: Idle: 2766, Running: 4, Completed: 13250 [ 18h 1m ] +INFO: Idle: 2765, Running: 4, Completed: 13251 [ 18h 2m ] +INFO: Idle: 2764, Running: 4, Completed: 13252 [ 18h 2m ] +INFO: Idle: 2763, Running: 4, Completed: 13253 [ 18h 2m ] +INFO: Idle: 2762, Running: 4, Completed: 13254 [ 18h 2m ] +INFO: Idle: 2761, Running: 4, Completed: 13255 [ 18h 2m ] +INFO: Idle: 2758, Running: 4, Completed: 13258 [ 18h 2m ] +INFO: Idle: 2757, Running: 4, Completed: 13259 [ 18h 2m ] +INFO: Idle: 2756, Running: 4, Completed: 13260 [ 18h 2m ] +INFO: Idle: 2754, Running: 4, Completed: 13262 [ 18h 3m ] +INFO: Idle: 2753, Running: 4, Completed: 13263 [ 18h 3m ] +INFO: Idle: 2751, Running: 4, Completed: 13265 [ 18h 3m ] +INFO: Idle: 2750, Running: 4, Completed: 13266 [ 18h 3m ] +INFO: Idle: 2749, Running: 4, Completed: 13267 [ 18h 3m ] +INFO: Idle: 2748, Running: 4, Completed: 13268 [ 18h 3m ] +INFO: Idle: 2745, Running: 4, Completed: 13271 [ 18h 3m ] +INFO: Idle: 2744, Running: 4, Completed: 13272 [ 18h 3m ] +INFO: Idle: 2742, Running: 4, Completed: 13274 [ 18h 4m ] +INFO: Idle: 2741, Running: 4, Completed: 13275 [ 18h 4m ] +INFO: Idle: 2740, Running: 4, Completed: 13276 [ 18h 4m ] +INFO: Idle: 2737, Running: 4, Completed: 13279 [ 18h 4m ] +INFO: Idle: 2734, Running: 4, Completed: 13282 [ 18h 4m ] +INFO: Idle: 2733, Running: 4, Completed: 13283 [ 18h 4m ] +INFO: Idle: 2732, Running: 4, Completed: 13284 [ 18h 4m ] +INFO: Idle: 2729, Running: 4, Completed: 13287 [ 18h 5m ] +INFO: Idle: 2728, Running: 4, Completed: 13288 [ 18h 5m ] +INFO: Idle: 2727, Running: 4, Completed: 13289 [ 18h 5m ] +INFO: Idle: 2725, Running: 4, Completed: 13291 [ 18h 5m ] +INFO: Idle: 2724, Running: 4, Completed: 13292 [ 18h 5m ] +INFO: Idle: 2721, Running: 4, Completed: 13295 [ 18h 5m ] +INFO: Idle: 2719, Running: 4, Completed: 13297 [ 18h 5m ] +INFO: Idle: 2718, Running: 4, Completed: 13298 [ 18h 6m ] +INFO: Idle: 2717, Running: 4, Completed: 13299 [ 18h 6m ] +INFO: Idle: 2715, Running: 4, Completed: 13301 [ 18h 6m ] +INFO: Idle: 2714, Running: 4, Completed: 13302 [ 18h 6m ] +INFO: Idle: 2713, Running: 4, Completed: 13303 [ 18h 6m ] +INFO: Idle: 2711, Running: 4, Completed: 13305 [ 18h 6m ] +INFO: Idle: 2710, Running: 4, Completed: 13306 [ 18h 6m ] +INFO: Idle: 2709, Running: 4, Completed: 13307 [ 18h 6m ] +INFO: Idle: 2708, Running: 4, Completed: 13308 [ 18h 6m ] +INFO: Idle: 2706, Running: 4, Completed: 13310 [ 18h 7m ] +INFO: Idle: 2705, Running: 4, Completed: 13311 [ 18h 7m ] +INFO: Idle: 2703, Running: 4, Completed: 13313 [ 18h 7m ] +INFO: Idle: 2702, Running: 4, Completed: 13314 [ 18h 7m ] +INFO: Idle: 2701, Running: 4, Completed: 13315 [ 18h 7m ] +INFO: Idle: 2699, Running: 4, Completed: 13317 [ 18h 7m ] +INFO: Idle: 2698, Running: 4, Completed: 13318 [ 18h 7m ] +INFO: Idle: 2697, Running: 4, Completed: 13319 [ 18h 8m ] +INFO: Idle: 2696, Running: 4, Completed: 13320 [ 18h 8m ] +INFO: Idle: 2694, Running: 4, Completed: 13322 [ 18h 8m ] +INFO: Idle: 2693, Running: 4, Completed: 13323 [ 18h 8m ] +INFO: Idle: 2692, Running: 4, Completed: 13324 [ 18h 8m ] +INFO: Idle: 2690, Running: 4, Completed: 13326 [ 18h 8m ] +INFO: Idle: 2689, Running: 4, Completed: 13327 [ 18h 8m ] +INFO: Idle: 2688, Running: 4, Completed: 13328 [ 18h 8m ] +INFO: Idle: 2685, Running: 4, Completed: 13331 [ 18h 8m ] +INFO: Idle: 2682, Running: 4, Completed: 13334 [ 18h 9m ] +INFO: Idle: 2681, Running: 4, Completed: 13335 [ 18h 9m ] +INFO: Idle: 2680, Running: 4, Completed: 13336 [ 18h 9m ] +INFO: Idle: 2678, Running: 4, Completed: 13338 [ 18h 9m ] +INFO: Idle: 2677, Running: 4, Completed: 13339 [ 18h 9m ] +INFO: Idle: 2676, Running: 4, Completed: 13340 [ 18h 9m ] +INFO: Idle: 2674, Running: 4, Completed: 13342 [ 18h 9m ] +INFO: Idle: 2673, Running: 4, Completed: 13343 [ 18h 9m ] +INFO: Idle: 2672, Running: 4, Completed: 13344 [ 18h 9m ] +INFO: Idle: 2669, Running: 4, Completed: 13347 [ 18h 10m ] +INFO: Idle: 2667, Running: 4, Completed: 13349 [ 18h 10m ] +INFO: Idle: 2666, Running: 4, Completed: 13350 [ 18h 10m ] +INFO: Idle: 2665, Running: 4, Completed: 13351 [ 18h 10m ] +INFO: Idle: 2662, Running: 4, Completed: 13354 [ 18h 10m ] +INFO: Idle: 2661, Running: 4, Completed: 13355 [ 18h 11m ] +INFO: Idle: 2660, Running: 4, Completed: 13356 [ 18h 11m ] +INFO: Idle: 2658, Running: 4, Completed: 13358 [ 18h 11m ] +INFO: Idle: 2657, Running: 4, Completed: 13359 [ 18h 11m ] +INFO: Idle: 2656, Running: 4, Completed: 13360 [ 18h 11m ] +INFO: Idle: 2654, Running: 4, Completed: 13362 [ 18h 11m ] +INFO: Idle: 2653, Running: 4, Completed: 13363 [ 18h 11m ] +INFO: Idle: 2651, Running: 4, Completed: 13365 [ 18h 11m ] +INFO: Idle: 2649, Running: 4, Completed: 13367 [ 18h 11m ] +INFO: Idle: 2648, Running: 4, Completed: 13368 [ 18h 12m ] +INFO: Idle: 2647, Running: 4, Completed: 13369 [ 18h 12m ] +INFO: Idle: 2645, Running: 4, Completed: 13371 [ 18h 12m ] +INFO: Idle: 2644, Running: 4, Completed: 13372 [ 18h 12m ] +INFO: Idle: 2643, Running: 4, Completed: 13373 [ 18h 12m ] +INFO: Idle: 2641, Running: 4, Completed: 13375 [ 18h 12m ] +INFO: Idle: 2640, Running: 4, Completed: 13376 [ 18h 12m ] +INFO: Idle: 2639, Running: 4, Completed: 13377 [ 18h 12m ] +INFO: Idle: 2637, Running: 4, Completed: 13379 [ 18h 12m ] +INFO: Idle: 2635, Running: 4, Completed: 13381 [ 18h 12m ] +INFO: Idle: 2633, Running: 4, Completed: 13383 [ 18h 13m ] +INFO: Idle: 2631, Running: 4, Completed: 13385 [ 18h 13m ] +INFO: Idle: 2629, Running: 4, Completed: 13387 [ 18h 13m ] +INFO: Idle: 2626, Running: 4, Completed: 13390 [ 18h 13m ] +INFO: Idle: 2625, Running: 4, Completed: 13391 [ 18h 14m ] +INFO: Idle: 2622, Running: 4, Completed: 13394 [ 18h 14m ] +INFO: Idle: 2621, Running: 4, Completed: 13395 [ 18h 14m ] +INFO: Idle: 2619, Running: 4, Completed: 13397 [ 18h 14m ] +INFO: Idle: 2618, Running: 4, Completed: 13398 [ 18h 14m ] +INFO: Idle: 2617, Running: 4, Completed: 13399 [ 18h 14m ] +INFO: Idle: 2616, Running: 4, Completed: 13400 [ 18h 14m ] +INFO: Idle: 2614, Running: 4, Completed: 13402 [ 18h 14m ] +INFO: Idle: 2613, Running: 4, Completed: 13403 [ 18h 14m ] +INFO: Idle: 2612, Running: 4, Completed: 13404 [ 18h 15m ] +INFO: Idle: 2610, Running: 4, Completed: 13406 [ 18h 15m ] +INFO: Idle: 2609, Running: 4, Completed: 13407 [ 18h 15m ] +INFO: Idle: 2608, Running: 4, Completed: 13408 [ 18h 15m ] +INFO: Idle: 2606, Running: 4, Completed: 13410 [ 18h 15m ] +INFO: Idle: 2605, Running: 4, Completed: 13411 [ 18h 15m ] +INFO: Idle: 2604, Running: 4, Completed: 13412 [ 18h 15m ] +INFO: Idle: 2603, Running: 4, Completed: 13413 [ 18h 15m ] +INFO: Idle: 2600, Running: 4, Completed: 13416 [ 18h 15m ] +INFO: Idle: 2599, Running: 4, Completed: 13417 [ 18h 15m ] +INFO: Idle: 2597, Running: 4, Completed: 13419 [ 18h 16m ] +INFO: Idle: 2596, Running: 4, Completed: 13420 [ 18h 16m ] +INFO: Idle: 2595, Running: 4, Completed: 13421 [ 18h 16m ] +INFO: Idle: 2594, Running: 4, Completed: 13422 [ 18h 16m ] +INFO: Idle: 2592, Running: 4, Completed: 13424 [ 18h 16m ] +INFO: Idle: 2591, Running: 4, Completed: 13425 [ 18h 16m ] +INFO: Idle: 2590, Running: 4, Completed: 13426 [ 18h 16m ] +INFO: Idle: 2588, Running: 4, Completed: 13428 [ 18h 16m ] +INFO: Idle: 2587, Running: 4, Completed: 13429 [ 18h 16m ] +INFO: Idle: 2586, Running: 4, Completed: 13430 [ 18h 17m ] +INFO: Idle: 2585, Running: 4, Completed: 13431 [ 18h 17m ] +INFO: Idle: 2583, Running: 4, Completed: 13433 [ 18h 17m ] +INFO: Idle: 2581, Running: 4, Completed: 13435 [ 18h 17m ] +INFO: Idle: 2579, Running: 4, Completed: 13437 [ 18h 17m ] +INFO: Idle: 2577, Running: 4, Completed: 13439 [ 18h 17m ] +INFO: Idle: 2575, Running: 4, Completed: 13441 [ 18h 17m ] +INFO: Idle: 2573, Running: 4, Completed: 13443 [ 18h 18m ] +INFO: Idle: 2571, Running: 4, Completed: 13445 [ 18h 18m ] +INFO: Idle: 2569, Running: 4, Completed: 13447 [ 18h 18m ] +INFO: Idle: 2567, Running: 4, Completed: 13449 [ 18h 18m ] +INFO: Idle: 2565, Running: 4, Completed: 13451 [ 18h 18m ] +INFO: Idle: 2562, Running: 4, Completed: 13454 [ 18h 18m ] +INFO: Idle: 2561, Running: 4, Completed: 13455 [ 18h 19m ] +INFO: Idle: 2559, Running: 4, Completed: 13457 [ 18h 19m ] +INFO: Idle: 2557, Running: 4, Completed: 13459 [ 18h 19m ] +INFO: Idle: 2555, Running: 4, Completed: 13461 [ 18h 19m ] +INFO: Idle: 2553, Running: 4, Completed: 13463 [ 18h 20m ] +INFO: Idle: 2551, Running: 4, Completed: 13465 [ 18h 20m ] +INFO: Idle: 2550, Running: 4, Completed: 13466 [ 18h 20m ] +INFO: Idle: 2549, Running: 4, Completed: 13467 [ 18h 20m ] +INFO: Idle: 2546, Running: 4, Completed: 13470 [ 18h 20m ] +INFO: Idle: 2545, Running: 4, Completed: 13471 [ 18h 20m ] +INFO: Idle: 2544, Running: 4, Completed: 13472 [ 18h 20m ] +INFO: Idle: 2542, Running: 4, Completed: 13474 [ 18h 20m ] +INFO: Idle: 2541, Running: 4, Completed: 13475 [ 18h 20m ] +INFO: Idle: 2540, Running: 4, Completed: 13476 [ 18h 21m ] +INFO: Idle: 2537, Running: 4, Completed: 13479 [ 18h 21m ] +INFO: Idle: 2536, Running: 4, Completed: 13480 [ 18h 21m ] +INFO: Idle: 2534, Running: 4, Completed: 13482 [ 18h 21m ] +INFO: Idle: 2533, Running: 4, Completed: 13483 [ 18h 21m ] +INFO: Idle: 2532, Running: 4, Completed: 13484 [ 18h 21m ] +INFO: Idle: 2531, Running: 4, Completed: 13485 [ 18h 21m ] +INFO: Idle: 2529, Running: 4, Completed: 13487 [ 18h 21m ] +INFO: Idle: 2528, Running: 4, Completed: 13488 [ 18h 21m ] +INFO: Idle: 2527, Running: 4, Completed: 13489 [ 18h 21m ] +INFO: Idle: 2525, Running: 4, Completed: 13491 [ 18h 22m ] +INFO: Idle: 2524, Running: 4, Completed: 13492 [ 18h 22m ] +INFO: Idle: 2523, Running: 4, Completed: 13493 [ 18h 22m ] +INFO: Idle: 2522, Running: 4, Completed: 13494 [ 18h 22m ] +INFO: Idle: 2520, Running: 4, Completed: 13496 [ 18h 22m ] +INFO: Idle: 2519, Running: 4, Completed: 13497 [ 18h 22m ] +INFO: Idle: 2518, Running: 4, Completed: 13498 [ 18h 22m ] +INFO: Idle: 2516, Running: 4, Completed: 13500 [ 18h 22m ] +INFO: Idle: 2515, Running: 4, Completed: 13501 [ 18h 22m ] +INFO: Idle: 2514, Running: 4, Completed: 13502 [ 18h 23m ] +INFO: Idle: 2513, Running: 4, Completed: 13503 [ 18h 23m ] +INFO: Idle: 2511, Running: 4, Completed: 13505 [ 18h 23m ] +INFO: Idle: 2509, Running: 4, Completed: 13507 [ 18h 23m ] +INFO: Idle: 2507, Running: 4, Completed: 13509 [ 18h 23m ] +INFO: Idle: 2506, Running: 4, Completed: 13510 [ 18h 23m ] +INFO: Idle: 2505, Running: 4, Completed: 13511 [ 18h 23m ] +INFO: Idle: 2503, Running: 4, Completed: 13513 [ 18h 23m ] +INFO: Idle: 2501, Running: 4, Completed: 13515 [ 18h 24m ] +INFO: Idle: 2499, Running: 4, Completed: 13517 [ 18h 24m ] +INFO: Idle: 2497, Running: 4, Completed: 13519 [ 18h 24m ] +INFO: Idle: 2495, Running: 4, Completed: 13521 [ 18h 24m ] +INFO: Idle: 2493, Running: 4, Completed: 13523 [ 18h 24m ] +INFO: Idle: 2491, Running: 4, Completed: 13525 [ 18h 24m ] +INFO: Idle: 2489, Running: 4, Completed: 13527 [ 18h 25m ] +INFO: Idle: 2487, Running: 4, Completed: 13529 [ 18h 25m ] +INFO: Idle: 2486, Running: 4, Completed: 13530 [ 18h 25m ] +INFO: Idle: 2485, Running: 4, Completed: 13531 [ 18h 25m ] +INFO: Idle: 2483, Running: 4, Completed: 13533 [ 18h 25m ] +INFO: Idle: 2481, Running: 4, Completed: 13535 [ 18h 25m ] +INFO: Idle: 2479, Running: 4, Completed: 13537 [ 18h 26m ] +INFO: Idle: 2477, Running: 4, Completed: 13539 [ 18h 26m ] +INFO: Idle: 2475, Running: 4, Completed: 13541 [ 18h 26m ] +INFO: Idle: 2474, Running: 4, Completed: 13542 [ 18h 26m ] +INFO: Idle: 2473, Running: 4, Completed: 13543 [ 18h 26m ] +INFO: Idle: 2471, Running: 4, Completed: 13545 [ 18h 26m ] +INFO: Idle: 2469, Running: 4, Completed: 13547 [ 18h 26m ] +INFO: Idle: 2468, Running: 4, Completed: 13548 [ 18h 27m ] +INFO: Idle: 2465, Running: 4, Completed: 13551 [ 18h 27m ] +INFO: Idle: 2464, Running: 4, Completed: 13552 [ 18h 27m ] +INFO: Idle: 2463, Running: 4, Completed: 13553 [ 18h 27m ] +INFO: Idle: 2461, Running: 4, Completed: 13555 [ 18h 27m ] +INFO: Idle: 2460, Running: 4, Completed: 13556 [ 18h 27m ] +INFO: Idle: 2459, Running: 4, Completed: 13557 [ 18h 27m ] +INFO: Idle: 2456, Running: 4, Completed: 13560 [ 18h 27m ] +INFO: Idle: 2455, Running: 4, Completed: 13561 [ 18h 27m ] +INFO: Idle: 2453, Running: 4, Completed: 13563 [ 18h 28m ] +INFO: Idle: 2452, Running: 4, Completed: 13564 [ 18h 28m ] +INFO: Idle: 2449, Running: 4, Completed: 13567 [ 18h 28m ] +INFO: Idle: 2448, Running: 4, Completed: 13568 [ 18h 28m ] +INFO: Idle: 2447, Running: 4, Completed: 13569 [ 18h 28m ] +INFO: Idle: 2445, Running: 4, Completed: 13571 [ 18h 28m ] +INFO: Idle: 2444, Running: 4, Completed: 13572 [ 18h 28m ] +INFO: Idle: 2441, Running: 4, Completed: 13575 [ 18h 28m ] +INFO: Idle: 2440, Running: 4, Completed: 13576 [ 18h 29m ] +INFO: Idle: 2437, Running: 4, Completed: 13579 [ 18h 29m ] +INFO: Idle: 2436, Running: 4, Completed: 13580 [ 18h 29m ] +INFO: Idle: 2433, Running: 4, Completed: 13583 [ 18h 29m ] +INFO: Idle: 2432, Running: 4, Completed: 13584 [ 18h 29m ] +INFO: Idle: 2431, Running: 4, Completed: 13585 [ 18h 29m ] +INFO: Idle: 2429, Running: 4, Completed: 13587 [ 18h 30m ] +INFO: Idle: 2428, Running: 4, Completed: 13588 [ 18h 30m ] +INFO: Idle: 2427, Running: 4, Completed: 13589 [ 18h 30m ] +INFO: Idle: 2425, Running: 4, Completed: 13591 [ 18h 30m ] +INFO: Idle: 2424, Running: 4, Completed: 13592 [ 18h 30m ] +INFO: Idle: 2423, Running: 4, Completed: 13593 [ 18h 30m ] +INFO: Idle: 2422, Running: 4, Completed: 13594 [ 18h 30m ] +INFO: Idle: 2420, Running: 4, Completed: 13596 [ 18h 30m ] +INFO: Idle: 2419, Running: 4, Completed: 13597 [ 18h 30m ] +INFO: Idle: 2416, Running: 4, Completed: 13600 [ 18h 31m ] +INFO: Idle: 2415, Running: 4, Completed: 13601 [ 18h 31m ] +INFO: Idle: 2413, Running: 4, Completed: 13603 [ 18h 31m ] +INFO: Idle: 2412, Running: 4, Completed: 13604 [ 18h 31m ] +INFO: Idle: 2411, Running: 4, Completed: 13605 [ 18h 31m ] +INFO: Idle: 2409, Running: 4, Completed: 13607 [ 18h 31m ] +INFO: Idle: 2408, Running: 4, Completed: 13608 [ 18h 31m ] +INFO: Idle: 2405, Running: 4, Completed: 13611 [ 18h 31m ] +INFO: Idle: 2404, Running: 4, Completed: 13612 [ 18h 32m ] +INFO: Idle: 2403, Running: 4, Completed: 13613 [ 18h 32m ] +INFO: Idle: 2401, Running: 4, Completed: 13615 [ 18h 32m ] +INFO: Idle: 2400, Running: 4, Completed: 13616 [ 18h 32m ] +INFO: Idle: 2399, Running: 4, Completed: 13617 [ 18h 32m ] +INFO: Idle: 2398, Running: 4, Completed: 13618 [ 18h 32m ] +INFO: Idle: 2396, Running: 4, Completed: 13620 [ 18h 32m ] +INFO: Idle: 2394, Running: 4, Completed: 13622 [ 18h 32m ] +INFO: Idle: 2390, Running: 4, Completed: 13626 [ 18h 33m ] +INFO: Idle: 2387, Running: 4, Completed: 13629 [ 18h 33m ] +INFO: Idle: 2386, Running: 4, Completed: 13630 [ 18h 33m ] +INFO: Idle: 2383, Running: 4, Completed: 13633 [ 18h 33m ] +INFO: Idle: 2382, Running: 4, Completed: 13634 [ 18h 33m ] +INFO: Idle: 2381, Running: 4, Completed: 13635 [ 18h 33m ] +INFO: Idle: 2378, Running: 4, Completed: 13638 [ 18h 34m ] +INFO: Idle: 2377, Running: 4, Completed: 13639 [ 18h 34m ] +INFO: Idle: 2374, Running: 4, Completed: 13642 [ 18h 34m ] +INFO: Idle: 2373, Running: 4, Completed: 13643 [ 18h 34m ] +INFO: Idle: 2371, Running: 4, Completed: 13645 [ 18h 34m ] +INFO: Idle: 2370, Running: 4, Completed: 13646 [ 18h 34m ] +INFO: Idle: 2369, Running: 4, Completed: 13647 [ 18h 34m ] +INFO: Idle: 2367, Running: 4, Completed: 13649 [ 18h 35m ] +INFO: Idle: 2366, Running: 4, Completed: 13650 [ 18h 35m ] +INFO: Idle: 2365, Running: 4, Completed: 13651 [ 18h 35m ] +INFO: Idle: 2362, Running: 4, Completed: 13654 [ 18h 35m ] +INFO: Idle: 2361, Running: 4, Completed: 13655 [ 18h 35m ] +INFO: Idle: 2359, Running: 4, Completed: 13657 [ 18h 35m ] +INFO: Idle: 2358, Running: 4, Completed: 13658 [ 18h 35m ] +INFO: Idle: 2355, Running: 4, Completed: 13661 [ 18h 35m ] +INFO: Idle: 2354, Running: 4, Completed: 13662 [ 18h 36m ] +INFO: Idle: 2353, Running: 4, Completed: 13663 [ 18h 36m ] +INFO: Idle: 2350, Running: 4, Completed: 13666 [ 18h 36m ] +INFO: Idle: 2349, Running: 4, Completed: 13667 [ 18h 36m ] +INFO: Idle: 2348, Running: 4, Completed: 13668 [ 18h 36m ] +INFO: Idle: 2345, Running: 4, Completed: 13671 [ 18h 36m ] +INFO: Idle: 2344, Running: 4, Completed: 13672 [ 18h 36m ] +INFO: Idle: 2340, Running: 4, Completed: 13676 [ 18h 37m ] +INFO: Idle: 2338, Running: 4, Completed: 13678 [ 18h 37m ] +INFO: Idle: 2336, Running: 4, Completed: 13680 [ 18h 37m ] +INFO: Idle: 2334, Running: 4, Completed: 13682 [ 18h 37m ] +INFO: Idle: 2331, Running: 4, Completed: 13685 [ 18h 37m ] +INFO: Idle: 2330, Running: 4, Completed: 13686 [ 18h 37m ] +INFO: Idle: 2327, Running: 4, Completed: 13689 [ 18h 37m ] +INFO: Idle: 2326, Running: 4, Completed: 13690 [ 18h 38m ] +INFO: Idle: 2325, Running: 4, Completed: 13691 [ 18h 38m ] +INFO: Idle: 2322, Running: 4, Completed: 13694 [ 18h 38m ] +INFO: Idle: 2318, Running: 4, Completed: 13698 [ 18h 38m ] +INFO: Idle: 2317, Running: 4, Completed: 13699 [ 18h 38m ] +INFO: Idle: 2315, Running: 4, Completed: 13701 [ 18h 38m ] +INFO: Idle: 2314, Running: 4, Completed: 13702 [ 18h 39m ] +INFO: Idle: 2313, Running: 4, Completed: 13703 [ 18h 39m ] +INFO: Idle: 2312, Running: 4, Completed: 13704 [ 18h 39m ] +INFO: Idle: 2309, Running: 4, Completed: 13707 [ 18h 39m ] +INFO: Idle: 2308, Running: 4, Completed: 13708 [ 18h 39m ] +INFO: Idle: 2306, Running: 4, Completed: 13710 [ 18h 39m ] +INFO: Idle: 2305, Running: 4, Completed: 13711 [ 18h 39m ] +INFO: Idle: 2304, Running: 4, Completed: 13712 [ 18h 39m ] +INFO: Idle: 2303, Running: 4, Completed: 13713 [ 18h 39m ] +INFO: Idle: 2301, Running: 4, Completed: 13715 [ 18h 40m ] +INFO: Idle: 2300, Running: 4, Completed: 13716 [ 18h 40m ] +INFO: Idle: 2299, Running: 4, Completed: 13717 [ 18h 40m ] +INFO: Idle: 2297, Running: 4, Completed: 13719 [ 18h 40m ] +INFO: Idle: 2296, Running: 4, Completed: 13720 [ 18h 40m ] +INFO: Idle: 2293, Running: 4, Completed: 13723 [ 18h 40m ] +INFO: Idle: 2290, Running: 4, Completed: 13726 [ 18h 41m ] +INFO: Idle: 2289, Running: 4, Completed: 13727 [ 18h 41m ] +INFO: Idle: 2288, Running: 4, Completed: 13728 [ 18h 41m ] +INFO: Idle: 2286, Running: 4, Completed: 13730 [ 18h 41m ] +INFO: Idle: 2285, Running: 4, Completed: 13731 [ 18h 41m ] +INFO: Idle: 2284, Running: 4, Completed: 13732 [ 18h 41m ] +INFO: Idle: 2282, Running: 4, Completed: 13734 [ 18h 41m ] +INFO: Idle: 2281, Running: 4, Completed: 13735 [ 18h 42m ] +INFO: Idle: 2280, Running: 4, Completed: 13736 [ 18h 42m ] +INFO: Idle: 2278, Running: 4, Completed: 13738 [ 18h 42m ] +INFO: Idle: 2277, Running: 4, Completed: 13739 [ 18h 42m ] +INFO: Idle: 2275, Running: 4, Completed: 13741 [ 18h 42m ] +INFO: Idle: 2274, Running: 4, Completed: 13742 [ 18h 42m ] +INFO: Idle: 2273, Running: 4, Completed: 13743 [ 18h 42m ] +INFO: Idle: 2271, Running: 4, Completed: 13745 [ 18h 42m ] +INFO: Idle: 2269, Running: 4, Completed: 13747 [ 18h 43m ] +INFO: Idle: 2268, Running: 4, Completed: 13748 [ 18h 43m ] +INFO: Idle: 2265, Running: 4, Completed: 13751 [ 18h 43m ] +INFO: Idle: 2264, Running: 4, Completed: 13752 [ 18h 43m ] +INFO: Idle: 2263, Running: 4, Completed: 13753 [ 18h 43m ] +INFO: Idle: 2261, Running: 4, Completed: 13755 [ 18h 43m ] +INFO: Idle: 2260, Running: 4, Completed: 13756 [ 18h 43m ] +INFO: Idle: 2259, Running: 4, Completed: 13757 [ 18h 43m ] +INFO: Idle: 2256, Running: 4, Completed: 13760 [ 18h 43m ] +INFO: Idle: 2255, Running: 4, Completed: 13761 [ 18h 44m ] +INFO: Idle: 2253, Running: 4, Completed: 13763 [ 18h 44m ] +INFO: Idle: 2252, Running: 4, Completed: 13764 [ 18h 44m ] +INFO: Idle: 2251, Running: 4, Completed: 13765 [ 18h 44m ] +INFO: Idle: 2250, Running: 4, Completed: 13766 [ 18h 44m ] +INFO: Idle: 2248, Running: 4, Completed: 13768 [ 18h 44m ] +INFO: Idle: 2247, Running: 4, Completed: 13769 [ 18h 44m ] +INFO: Idle: 2246, Running: 4, Completed: 13770 [ 18h 44m ] +INFO: Idle: 2244, Running: 4, Completed: 13772 [ 18h 45m ] +INFO: Idle: 2242, Running: 4, Completed: 13774 [ 18h 45m ] +INFO: Idle: 2241, Running: 4, Completed: 13775 [ 18h 45m ] +INFO: Idle: 2239, Running: 4, Completed: 13777 [ 18h 45m ] +INFO: Idle: 2237, Running: 4, Completed: 13779 [ 18h 45m ] +INFO: Idle: 2236, Running: 4, Completed: 13780 [ 18h 45m ] +INFO: Idle: 2235, Running: 4, Completed: 13781 [ 18h 45m ] +INFO: Idle: 2233, Running: 4, Completed: 13783 [ 18h 45m ] +INFO: Idle: 2231, Running: 4, Completed: 13785 [ 18h 46m ] +INFO: Idle: 2229, Running: 4, Completed: 13787 [ 18h 46m ] +INFO: Idle: 2227, Running: 4, Completed: 13789 [ 18h 46m ] +INFO: Idle: 2225, Running: 4, Completed: 13791 [ 18h 46m ] +INFO: Idle: 2223, Running: 4, Completed: 13793 [ 18h 46m ] +INFO: Idle: 2221, Running: 4, Completed: 13795 [ 18h 46m ] +INFO: Idle: 2219, Running: 4, Completed: 13797 [ 18h 47m ] +INFO: Idle: 2217, Running: 4, Completed: 13799 [ 18h 47m ] +INFO: Idle: 2215, Running: 4, Completed: 13801 [ 18h 47m ] +INFO: Idle: 2213, Running: 4, Completed: 13803 [ 18h 47m ] +INFO: Idle: 2211, Running: 4, Completed: 13805 [ 18h 47m ] +INFO: Idle: 2210, Running: 4, Completed: 13806 [ 18h 47m ] +INFO: Idle: 2209, Running: 4, Completed: 13807 [ 18h 48m ] +INFO: Idle: 2207, Running: 4, Completed: 13809 [ 18h 48m ] +INFO: Idle: 2206, Running: 4, Completed: 13810 [ 18h 48m ] +INFO: Idle: 2205, Running: 4, Completed: 13811 [ 18h 48m ] +INFO: Idle: 2202, Running: 4, Completed: 13814 [ 18h 48m ] +INFO: Idle: 2201, Running: 4, Completed: 13815 [ 18h 48m ] +INFO: Idle: 2200, Running: 4, Completed: 13816 [ 18h 48m ] +INFO: Idle: 2197, Running: 4, Completed: 13819 [ 18h 49m ] +INFO: Idle: 2196, Running: 4, Completed: 13820 [ 18h 49m ] +INFO: Idle: 2193, Running: 4, Completed: 13823 [ 18h 49m ] +INFO: Idle: 2192, Running: 4, Completed: 13824 [ 18h 49m ] +INFO: Idle: 2190, Running: 4, Completed: 13826 [ 18h 49m ] +INFO: Idle: 2189, Running: 4, Completed: 13827 [ 18h 49m ] +INFO: Idle: 2188, Running: 4, Completed: 13828 [ 18h 49m ] +INFO: Idle: 2187, Running: 4, Completed: 13829 [ 18h 49m ] +INFO: Idle: 2184, Running: 4, Completed: 13832 [ 18h 50m ] +INFO: Idle: 2183, Running: 4, Completed: 13833 [ 18h 50m ] +INFO: Idle: 2181, Running: 4, Completed: 13835 [ 18h 50m ] +INFO: Idle: 2180, Running: 4, Completed: 13836 [ 18h 50m ] +INFO: Idle: 2179, Running: 4, Completed: 13837 [ 18h 50m ] +INFO: Idle: 2177, Running: 4, Completed: 13839 [ 18h 50m ] +INFO: Idle: 2176, Running: 4, Completed: 13840 [ 18h 50m ] +INFO: Idle: 2175, Running: 4, Completed: 13841 [ 18h 50m ] +INFO: Idle: 2173, Running: 4, Completed: 13843 [ 18h 50m ] +INFO: Idle: 2172, Running: 4, Completed: 13844 [ 18h 51m ] +INFO: Idle: 2170, Running: 4, Completed: 13846 [ 18h 51m ] +INFO: Idle: 2168, Running: 4, Completed: 13848 [ 18h 51m ] +INFO: Idle: 2165, Running: 4, Completed: 13851 [ 18h 51m ] +INFO: Idle: 2164, Running: 4, Completed: 13852 [ 18h 51m ] +INFO: Idle: 2161, Running: 4, Completed: 13855 [ 18h 51m ] +INFO: Idle: 2160, Running: 4, Completed: 13856 [ 18h 52m ] +INFO: Idle: 2157, Running: 4, Completed: 13859 [ 18h 52m ] +INFO: Idle: 2156, Running: 4, Completed: 13860 [ 18h 52m ] +INFO: Idle: 2154, Running: 4, Completed: 13862 [ 18h 52m ] +INFO: Idle: 2153, Running: 4, Completed: 13863 [ 18h 52m ] +INFO: Idle: 2152, Running: 4, Completed: 13864 [ 18h 52m ] +INFO: Idle: 2150, Running: 4, Completed: 13866 [ 18h 52m ] +INFO: Idle: 2149, Running: 4, Completed: 13867 [ 18h 52m ] +INFO: Idle: 2146, Running: 4, Completed: 13870 [ 18h 53m ] +INFO: Idle: 2145, Running: 4, Completed: 13871 [ 18h 53m ] +INFO: Idle: 2143, Running: 4, Completed: 13873 [ 18h 53m ] +INFO: Idle: 2141, Running: 4, Completed: 13875 [ 18h 53m ] +INFO: Idle: 2137, Running: 4, Completed: 13879 [ 18h 54m ] +INFO: Idle: 2134, Running: 4, Completed: 13882 [ 18h 54m ] +INFO: Idle: 2133, Running: 4, Completed: 13883 [ 18h 54m ] +INFO: Idle: 2130, Running: 4, Completed: 13886 [ 18h 54m ] +INFO: Idle: 2129, Running: 4, Completed: 13887 [ 18h 54m ] +INFO: Idle: 2128, Running: 4, Completed: 13888 [ 18h 54m ] +INFO: Idle: 2125, Running: 4, Completed: 13891 [ 18h 55m ] +INFO: Idle: 2124, Running: 4, Completed: 13892 [ 18h 55m ] +INFO: Idle: 2121, Running: 4, Completed: 13895 [ 18h 55m ] +INFO: Idle: 2120, Running: 4, Completed: 13896 [ 18h 55m ] +INFO: Idle: 2118, Running: 4, Completed: 13898 [ 18h 55m ] +INFO: Idle: 2117, Running: 4, Completed: 13899 [ 18h 55m ] +INFO: Idle: 2116, Running: 4, Completed: 13900 [ 18h 55m ] +INFO: Idle: 2115, Running: 4, Completed: 13901 [ 18h 55m ] +INFO: Idle: 2113, Running: 4, Completed: 13903 [ 18h 55m ] +INFO: Idle: 2112, Running: 4, Completed: 13904 [ 18h 55m ] +INFO: Idle: 2111, Running: 4, Completed: 13905 [ 18h 56m ] +INFO: Idle: 2109, Running: 4, Completed: 13907 [ 18h 56m ] +INFO: Idle: 2108, Running: 4, Completed: 13908 [ 18h 56m ] +INFO: Idle: 2106, Running: 4, Completed: 13910 [ 18h 56m ] +INFO: Idle: 2104, Running: 4, Completed: 13912 [ 18h 56m ] +INFO: Idle: 2103, Running: 4, Completed: 13913 [ 18h 56m ] +INFO: Idle: 2102, Running: 4, Completed: 13914 [ 18h 56m ] +INFO: Idle: 2100, Running: 4, Completed: 13916 [ 18h 56m ] +INFO: Idle: 2099, Running: 4, Completed: 13917 [ 18h 57m ] +INFO: Idle: 2098, Running: 4, Completed: 13918 [ 18h 57m ] +INFO: Idle: 2097, Running: 4, Completed: 13919 [ 18h 57m ] +INFO: Idle: 2095, Running: 4, Completed: 13921 [ 18h 57m ] +INFO: Idle: 2093, Running: 4, Completed: 13923 [ 18h 57m ] +INFO: Idle: 2091, Running: 4, Completed: 13925 [ 18h 57m ] +INFO: Idle: 2089, Running: 4, Completed: 13927 [ 18h 57m ] +INFO: Idle: 2087, Running: 4, Completed: 13929 [ 18h 58m ] +INFO: Idle: 2085, Running: 4, Completed: 13931 [ 18h 58m ] +INFO: Idle: 2083, Running: 4, Completed: 13933 [ 18h 58m ] +INFO: Idle: 2081, Running: 4, Completed: 13935 [ 18h 58m ] +INFO: Idle: 2079, Running: 4, Completed: 13937 [ 18h 58m ] +INFO: Idle: 2077, Running: 4, Completed: 13939 [ 18h 58m ] +INFO: Idle: 2074, Running: 4, Completed: 13942 [ 18h 59m ] +INFO: Idle: 2073, Running: 4, Completed: 13943 [ 18h 59m ] +INFO: Idle: 2072, Running: 4, Completed: 13944 [ 18h 59m ] +INFO: Idle: 2071, Running: 4, Completed: 13945 [ 18h 59m ] +INFO: Idle: 2069, Running: 4, Completed: 13947 [ 18h 59m ] +INFO: Idle: 2067, Running: 4, Completed: 13949 [ 18h 59m ] +INFO: Idle: 2065, Running: 4, Completed: 13951 [ 19h 0m ] +INFO: Idle: 2063, Running: 4, Completed: 13953 [ 19h 0m ] +INFO: Idle: 2062, Running: 4, Completed: 13954 [ 19h 0m ] +INFO: Idle: 2061, Running: 4, Completed: 13955 [ 19h 0m ] +INFO: Idle: 2059, Running: 4, Completed: 13957 [ 19h 0m ] +INFO: Idle: 2058, Running: 4, Completed: 13958 [ 19h 0m ] +INFO: Idle: 2057, Running: 4, Completed: 13959 [ 19h 0m ] +INFO: Idle: 2055, Running: 4, Completed: 13961 [ 19h 0m ] +INFO: Idle: 2053, Running: 4, Completed: 13963 [ 19h 1m ] +INFO: Idle: 2052, Running: 4, Completed: 13964 [ 19h 1m ] +INFO: Idle: 2050, Running: 4, Completed: 13966 [ 19h 1m ] +INFO: Idle: 2049, Running: 4, Completed: 13967 [ 19h 1m ] +INFO: Idle: 2048, Running: 4, Completed: 13968 [ 19h 1m ] +INFO: Idle: 2047, Running: 4, Completed: 13969 [ 19h 1m ] +INFO: Idle: 2045, Running: 4, Completed: 13971 [ 19h 1m ] +INFO: Idle: 2044, Running: 4, Completed: 13972 [ 19h 1m ] +INFO: Idle: 2043, Running: 4, Completed: 13973 [ 19h 1m ] +INFO: Idle: 2040, Running: 4, Completed: 13976 [ 19h 1m ] +INFO: Idle: 2039, Running: 4, Completed: 13977 [ 19h 2m ] +INFO: Idle: 2037, Running: 4, Completed: 13979 [ 19h 2m ] +INFO: Idle: 2036, Running: 4, Completed: 13980 [ 19h 2m ] +INFO: Idle: 2035, Running: 4, Completed: 13981 [ 19h 2m ] +INFO: Idle: 2034, Running: 4, Completed: 13982 [ 19h 2m ] +INFO: Idle: 2032, Running: 4, Completed: 13984 [ 19h 2m ] +INFO: Idle: 2031, Running: 4, Completed: 13985 [ 19h 2m ] +INFO: Idle: 2030, Running: 4, Completed: 13986 [ 19h 2m ] +INFO: Idle: 2028, Running: 4, Completed: 13988 [ 19h 2m ] +INFO: Idle: 2027, Running: 4, Completed: 13989 [ 19h 2m ] +INFO: Idle: 2026, Running: 4, Completed: 13990 [ 19h 3m ] +INFO: Idle: 2025, Running: 4, Completed: 13991 [ 19h 3m ] +INFO: Idle: 2023, Running: 4, Completed: 13993 [ 19h 3m ] +INFO: Idle: 2021, Running: 4, Completed: 13995 [ 19h 3m ] +INFO: Idle: 2019, Running: 4, Completed: 13997 [ 19h 3m ] +INFO: Idle: 2017, Running: 4, Completed: 13999 [ 19h 3m ] +INFO: Idle: 2015, Running: 4, Completed: 14001 [ 19h 3m ] +INFO: Idle: 2013, Running: 4, Completed: 14003 [ 19h 4m ] +INFO: Idle: 2011, Running: 4, Completed: 14005 [ 19h 4m ] +INFO: Idle: 2009, Running: 4, Completed: 14007 [ 19h 4m ] +INFO: Idle: 2007, Running: 4, Completed: 14009 [ 19h 4m ] +INFO: Idle: 2005, Running: 4, Completed: 14011 [ 19h 4m ] +INFO: Idle: 2002, Running: 4, Completed: 14014 [ 19h 5m ] +INFO: Idle: 2001, Running: 4, Completed: 14015 [ 19h 5m ] +INFO: Idle: 2000, Running: 4, Completed: 14016 [ 19h 5m ] +INFO: Idle: 1999, Running: 4, Completed: 14017 [ 19h 5m ] +INFO: Idle: 1997, Running: 4, Completed: 14019 [ 19h 5m ] +INFO: Idle: 1995, Running: 4, Completed: 14021 [ 19h 5m ] +INFO: Idle: 1993, Running: 4, Completed: 14023 [ 19h 6m ] +INFO: Idle: 1991, Running: 4, Completed: 14025 [ 19h 6m ] +INFO: Idle: 1990, Running: 4, Completed: 14026 [ 19h 6m ] +INFO: Idle: 1989, Running: 4, Completed: 14027 [ 19h 6m ] +INFO: Idle: 1987, Running: 4, Completed: 14029 [ 19h 6m ] +INFO: Idle: 1986, Running: 4, Completed: 14030 [ 19h 6m ] +INFO: Idle: 1985, Running: 4, Completed: 14031 [ 19h 6m ] +INFO: Idle: 1984, Running: 4, Completed: 14032 [ 19h 6m ] +INFO: Idle: 1982, Running: 4, Completed: 14034 [ 19h 6m ] +INFO: Idle: 1981, Running: 4, Completed: 14035 [ 19h 7m ] +INFO: Idle: 1980, Running: 4, Completed: 14036 [ 19h 7m ] +INFO: Idle: 1979, Running: 4, Completed: 14037 [ 19h 7m ] +INFO: Idle: 1977, Running: 4, Completed: 14039 [ 19h 7m ] +INFO: Idle: 1976, Running: 4, Completed: 14040 [ 19h 7m ] +INFO: Idle: 1975, Running: 4, Completed: 14041 [ 19h 7m ] +INFO: Idle: 1973, Running: 4, Completed: 14043 [ 19h 7m ] +INFO: Idle: 1972, Running: 4, Completed: 14044 [ 19h 7m ] +INFO: Idle: 1971, Running: 4, Completed: 14045 [ 19h 7m ] +INFO: Idle: 1969, Running: 4, Completed: 14047 [ 19h 7m ] +INFO: Idle: 1968, Running: 4, Completed: 14048 [ 19h 7m ] +INFO: Idle: 1967, Running: 4, Completed: 14049 [ 19h 8m ] +INFO: Idle: 1965, Running: 4, Completed: 14051 [ 19h 8m ] +INFO: Idle: 1964, Running: 4, Completed: 14052 [ 19h 8m ] +INFO: Idle: 1962, Running: 4, Completed: 14054 [ 19h 8m ] +INFO: Idle: 1961, Running: 4, Completed: 14055 [ 19h 8m ] +INFO: Idle: 1960, Running: 4, Completed: 14056 [ 19h 8m ] +INFO: Idle: 1958, Running: 4, Completed: 14058 [ 19h 8m ] +INFO: Idle: 1957, Running: 4, Completed: 14059 [ 19h 8m ] +INFO: Idle: 1955, Running: 4, Completed: 14061 [ 19h 9m ] +INFO: Idle: 1953, Running: 4, Completed: 14063 [ 19h 9m ] +INFO: Idle: 1950, Running: 4, Completed: 14066 [ 19h 9m ] +INFO: Idle: 1949, Running: 4, Completed: 14067 [ 19h 9m ] +INFO: Idle: 1945, Running: 4, Completed: 14071 [ 19h 9m ] +INFO: Idle: 1942, Running: 4, Completed: 14074 [ 19h 9m ] +INFO: Idle: 1941, Running: 4, Completed: 14075 [ 19h 10m ] +INFO: Idle: 1937, Running: 4, Completed: 14079 [ 19h 10m ] +INFO: Idle: 1933, Running: 4, Completed: 14083 [ 19h 10m ] +INFO: Idle: 1931, Running: 4, Completed: 14085 [ 19h 10m ] +INFO: Idle: 1930, Running: 4, Completed: 14086 [ 19h 11m ] +INFO: Idle: 1929, Running: 4, Completed: 14087 [ 19h 11m ] +INFO: Idle: 1927, Running: 4, Completed: 14089 [ 19h 11m ] +INFO: Idle: 1926, Running: 4, Completed: 14090 [ 19h 11m ] +INFO: Idle: 1925, Running: 4, Completed: 14091 [ 19h 11m ] +INFO: Idle: 1923, Running: 4, Completed: 14093 [ 19h 11m ] +INFO: Idle: 1922, Running: 4, Completed: 14094 [ 19h 11m ] +INFO: Idle: 1920, Running: 4, Completed: 14096 [ 19h 11m ] +INFO: Idle: 1919, Running: 4, Completed: 14097 [ 19h 12m ] +INFO: Idle: 1917, Running: 4, Completed: 14099 [ 19h 12m ] +INFO: Idle: 1916, Running: 4, Completed: 14100 [ 19h 12m ] +INFO: Idle: 1915, Running: 4, Completed: 14101 [ 19h 12m ] +INFO: Idle: 1914, Running: 4, Completed: 14102 [ 19h 12m ] +INFO: Idle: 1913, Running: 4, Completed: 14103 [ 19h 12m ] +INFO: Idle: 1911, Running: 4, Completed: 14105 [ 19h 12m ] +INFO: Idle: 1910, Running: 4, Completed: 14106 [ 19h 12m ] +INFO: Idle: 1908, Running: 4, Completed: 14108 [ 19h 12m ] +INFO: Idle: 1907, Running: 4, Completed: 14109 [ 19h 13m ] +INFO: Idle: 1906, Running: 4, Completed: 14110 [ 19h 13m ] +INFO: Idle: 1905, Running: 4, Completed: 14111 [ 19h 13m ] +INFO: Idle: 1903, Running: 4, Completed: 14113 [ 19h 13m ] +INFO: Idle: 1902, Running: 4, Completed: 14114 [ 19h 13m ] +INFO: Idle: 1899, Running: 4, Completed: 14117 [ 19h 13m ] +INFO: Idle: 1898, Running: 4, Completed: 14118 [ 19h 13m ] +INFO: Idle: 1896, Running: 4, Completed: 14120 [ 19h 13m ] +INFO: Idle: 1895, Running: 4, Completed: 14121 [ 19h 14m ] +INFO: Idle: 1894, Running: 4, Completed: 14122 [ 19h 14m ] +INFO: Idle: 1892, Running: 4, Completed: 14124 [ 19h 14m ] +INFO: Idle: 1891, Running: 4, Completed: 14125 [ 19h 14m ] +INFO: Idle: 1890, Running: 4, Completed: 14126 [ 19h 14m ] +INFO: Idle: 1888, Running: 4, Completed: 14128 [ 19h 14m ] +INFO: Idle: 1887, Running: 4, Completed: 14129 [ 19h 14m ] +INFO: Idle: 1886, Running: 4, Completed: 14130 [ 19h 14m ] +INFO: Idle: 1884, Running: 4, Completed: 14132 [ 19h 14m ] +INFO: Idle: 1883, Running: 4, Completed: 14133 [ 19h 14m ] +INFO: Idle: 1879, Running: 4, Completed: 14137 [ 19h 15m ] +INFO: Idle: 1877, Running: 4, Completed: 14139 [ 19h 15m ] +INFO: Idle: 1876, Running: 4, Completed: 14140 [ 19h 15m ] +INFO: Idle: 1875, Running: 4, Completed: 14141 [ 19h 15m ] +INFO: Idle: 1874, Running: 4, Completed: 14142 [ 19h 15m ] +INFO: Idle: 1873, Running: 4, Completed: 14143 [ 19h 15m ] +INFO: Idle: 1871, Running: 4, Completed: 14145 [ 19h 15m ] +INFO: Idle: 1870, Running: 4, Completed: 14146 [ 19h 15m ] +INFO: Idle: 1866, Running: 4, Completed: 14150 [ 19h 16m ] +INFO: Idle: 1864, Running: 4, Completed: 14152 [ 19h 16m ] +INFO: Idle: 1862, Running: 4, Completed: 14154 [ 19h 16m ] +INFO: Idle: 1860, Running: 4, Completed: 14156 [ 19h 16m ] +INFO: Idle: 1858, Running: 4, Completed: 14158 [ 19h 16m ] +INFO: Idle: 1856, Running: 4, Completed: 14160 [ 19h 16m ] +INFO: Idle: 1854, Running: 4, Completed: 14162 [ 19h 17m ] +INFO: Idle: 1852, Running: 4, Completed: 14164 [ 19h 17m ] +INFO: Idle: 1850, Running: 4, Completed: 14166 [ 19h 17m ] +INFO: Idle: 1846, Running: 4, Completed: 14170 [ 19h 17m ] +INFO: Idle: 1842, Running: 4, Completed: 14174 [ 19h 18m ] +INFO: Idle: 1838, Running: 4, Completed: 14178 [ 19h 18m ] +INFO: Idle: 1834, Running: 4, Completed: 14182 [ 19h 18m ] +INFO: Idle: 1832, Running: 4, Completed: 14184 [ 19h 18m ] +INFO: Idle: 1830, Running: 4, Completed: 14186 [ 19h 18m ] +INFO: Idle: 1828, Running: 4, Completed: 14188 [ 19h 19m ] +INFO: Idle: 1827, Running: 4, Completed: 14189 [ 19h 19m ] +INFO: Idle: 1826, Running: 4, Completed: 14190 [ 19h 19m ] +INFO: Idle: 1825, Running: 4, Completed: 14191 [ 19h 19m ] +INFO: Idle: 1824, Running: 4, Completed: 14192 [ 19h 19m ] +INFO: Idle: 1823, Running: 4, Completed: 14193 [ 19h 19m ] +INFO: Idle: 1821, Running: 4, Completed: 14195 [ 19h 19m ] +INFO: Idle: 1819, Running: 4, Completed: 14197 [ 19h 19m ] +INFO: Idle: 1817, Running: 4, Completed: 14199 [ 19h 20m ] +INFO: Idle: 1815, Running: 4, Completed: 14201 [ 19h 20m ] +INFO: Idle: 1814, Running: 4, Completed: 14202 [ 19h 20m ] +INFO: Idle: 1811, Running: 4, Completed: 14205 [ 19h 20m ] +INFO: Idle: 1810, Running: 4, Completed: 14206 [ 19h 20m ] +INFO: Idle: 1808, Running: 4, Completed: 14208 [ 19h 20m ] +INFO: Idle: 1806, Running: 4, Completed: 14210 [ 19h 20m ] +INFO: Idle: 1805, Running: 4, Completed: 14211 [ 19h 20m ] +INFO: Idle: 1804, Running: 4, Completed: 14212 [ 19h 21m ] +INFO: Idle: 1802, Running: 4, Completed: 14214 [ 19h 21m ] +INFO: Idle: 1800, Running: 4, Completed: 14216 [ 19h 21m ] +INFO: Idle: 1798, Running: 4, Completed: 14218 [ 19h 21m ] +INFO: Idle: 1796, Running: 4, Completed: 14220 [ 19h 21m ] +INFO: Idle: 1794, Running: 4, Completed: 14222 [ 19h 21m ] +INFO: Idle: 1793, Running: 4, Completed: 14223 [ 19h 21m ] +INFO: Idle: 1792, Running: 4, Completed: 14224 [ 19h 22m ] +INFO: Idle: 1790, Running: 4, Completed: 14226 [ 19h 22m ] +INFO: Idle: 1788, Running: 4, Completed: 14228 [ 19h 22m ] +INFO: Idle: 1787, Running: 4, Completed: 14229 [ 19h 22m ] +INFO: Idle: 1785, Running: 4, Completed: 14231 [ 19h 22m ] +INFO: Idle: 1784, Running: 4, Completed: 14232 [ 19h 22m ] +INFO: Idle: 1783, Running: 4, Completed: 14233 [ 19h 22m ] +INFO: Idle: 1782, Running: 4, Completed: 14234 [ 19h 22m ] +INFO: Idle: 1780, Running: 4, Completed: 14236 [ 19h 23m ] +INFO: Idle: 1779, Running: 4, Completed: 14237 [ 19h 23m ] +INFO: Idle: 1777, Running: 4, Completed: 14239 [ 19h 23m ] +INFO: Idle: 1776, Running: 4, Completed: 14240 [ 19h 23m ] +INFO: Idle: 1773, Running: 4, Completed: 14243 [ 19h 23m ] +INFO: Idle: 1772, Running: 4, Completed: 14244 [ 19h 23m ] +INFO: Idle: 1771, Running: 4, Completed: 14245 [ 19h 23m ] +INFO: Idle: 1769, Running: 4, Completed: 14247 [ 19h 23m ] +INFO: Idle: 1768, Running: 4, Completed: 14248 [ 19h 24m ] +INFO: Idle: 1765, Running: 4, Completed: 14251 [ 19h 24m ] +INFO: Idle: 1764, Running: 4, Completed: 14252 [ 19h 24m ] +INFO: Idle: 1763, Running: 4, Completed: 14253 [ 19h 24m ] +INFO: Idle: 1760, Running: 4, Completed: 14256 [ 19h 24m ] +INFO: Idle: 1759, Running: 4, Completed: 14257 [ 19h 24m ] +INFO: Idle: 1758, Running: 4, Completed: 14258 [ 19h 24m ] +INFO: Idle: 1757, Running: 4, Completed: 14259 [ 19h 24m ] +INFO: Idle: 1755, Running: 4, Completed: 14261 [ 19h 24m ] +INFO: Idle: 1754, Running: 4, Completed: 14262 [ 19h 25m ] +INFO: Idle: 1750, Running: 4, Completed: 14266 [ 19h 25m ] +INFO: Idle: 1747, Running: 4, Completed: 14269 [ 19h 25m ] +INFO: Idle: 1746, Running: 4, Completed: 14270 [ 19h 25m ] +INFO: Idle: 1743, Running: 4, Completed: 14273 [ 19h 25m ] +INFO: Idle: 1742, Running: 4, Completed: 14274 [ 19h 26m ] +INFO: Idle: 1741, Running: 4, Completed: 14275 [ 19h 26m ] +INFO: Idle: 1739, Running: 4, Completed: 14277 [ 19h 26m ] +INFO: Idle: 1738, Running: 4, Completed: 14278 [ 19h 26m ] +INFO: Idle: 1737, Running: 4, Completed: 14279 [ 19h 26m ] +INFO: Idle: 1734, Running: 4, Completed: 14282 [ 19h 26m ] +INFO: Idle: 1733, Running: 4, Completed: 14283 [ 19h 26m ] +INFO: Idle: 1732, Running: 4, Completed: 14284 [ 19h 27m ] +INFO: Idle: 1730, Running: 4, Completed: 14286 [ 19h 27m ] +INFO: Idle: 1729, Running: 4, Completed: 14287 [ 19h 27m ] +INFO: Idle: 1727, Running: 4, Completed: 14289 [ 19h 27m ] +INFO: Idle: 1726, Running: 4, Completed: 14290 [ 19h 27m ] +INFO: Idle: 1724, Running: 4, Completed: 14292 [ 19h 27m ] +INFO: Idle: 1722, Running: 4, Completed: 14294 [ 19h 27m ] +INFO: Idle: 1721, Running: 4, Completed: 14295 [ 19h 27m ] +INFO: Idle: 1719, Running: 4, Completed: 14297 [ 19h 27m ] +INFO: Idle: 1718, Running: 4, Completed: 14298 [ 19h 28m ] +INFO: Idle: 1715, Running: 4, Completed: 14301 [ 19h 28m ] +INFO: Idle: 1714, Running: 4, Completed: 14302 [ 19h 28m ] +INFO: Idle: 1713, Running: 4, Completed: 14303 [ 19h 28m ] +INFO: Idle: 1711, Running: 4, Completed: 14305 [ 19h 28m ] +INFO: Idle: 1710, Running: 4, Completed: 14306 [ 19h 28m ] +INFO: Idle: 1709, Running: 4, Completed: 14307 [ 19h 28m ] +INFO: Idle: 1708, Running: 4, Completed: 14308 [ 19h 28m ] +INFO: Idle: 1706, Running: 4, Completed: 14310 [ 19h 28m ] +INFO: Idle: 1704, Running: 4, Completed: 14312 [ 19h 29m ] +INFO: Idle: 1700, Running: 4, Completed: 14316 [ 19h 29m ] +INFO: Idle: 1699, Running: 4, Completed: 14317 [ 19h 29m ] +INFO: Idle: 1696, Running: 4, Completed: 14320 [ 19h 29m ] +INFO: Idle: 1692, Running: 4, Completed: 14324 [ 19h 30m ] +INFO: Idle: 1691, Running: 4, Completed: 14325 [ 19h 30m ] +INFO: Idle: 1688, Running: 4, Completed: 14328 [ 19h 30m ] +INFO: Idle: 1687, Running: 4, Completed: 14329 [ 19h 30m ] +INFO: Idle: 1686, Running: 4, Completed: 14330 [ 19h 30m ] +INFO: Idle: 1683, Running: 4, Completed: 14333 [ 19h 30m ] +INFO: Idle: 1682, Running: 4, Completed: 14334 [ 19h 30m ] +INFO: Idle: 1680, Running: 4, Completed: 14336 [ 19h 31m ] +INFO: Idle: 1678, Running: 4, Completed: 14338 [ 19h 31m ] +INFO: Idle: 1677, Running: 4, Completed: 14339 [ 19h 31m ] +INFO: Idle: 1675, Running: 4, Completed: 14341 [ 19h 31m ] +INFO: Idle: 1674, Running: 4, Completed: 14342 [ 19h 31m ] +INFO: Idle: 1672, Running: 4, Completed: 14344 [ 19h 31m ] +INFO: Idle: 1671, Running: 4, Completed: 14345 [ 19h 31m ] +INFO: Idle: 1670, Running: 4, Completed: 14346 [ 19h 31m ] +INFO: Idle: 1668, Running: 4, Completed: 14348 [ 19h 31m ] +INFO: Idle: 1667, Running: 4, Completed: 14349 [ 19h 32m ] +INFO: Idle: 1665, Running: 4, Completed: 14351 [ 19h 32m ] +INFO: Idle: 1664, Running: 4, Completed: 14352 [ 19h 32m ] +INFO: Idle: 1663, Running: 4, Completed: 14353 [ 19h 32m ] +INFO: Idle: 1662, Running: 4, Completed: 14354 [ 19h 32m ] +INFO: Idle: 1660, Running: 4, Completed: 14356 [ 19h 32m ] +INFO: Idle: 1659, Running: 4, Completed: 14357 [ 19h 32m ] +INFO: Idle: 1658, Running: 4, Completed: 14358 [ 19h 33m ] +INFO: Idle: 1656, Running: 4, Completed: 14360 [ 19h 33m ] +INFO: Idle: 1655, Running: 4, Completed: 14361 [ 19h 33m ] +INFO: Idle: 1654, Running: 4, Completed: 14362 [ 19h 33m ] +INFO: Idle: 1652, Running: 4, Completed: 14364 [ 19h 33m ] +INFO: Idle: 1651, Running: 4, Completed: 14365 [ 19h 33m ] +INFO: Idle: 1649, Running: 4, Completed: 14367 [ 19h 33m ] +INFO: Idle: 1648, Running: 4, Completed: 14368 [ 19h 33m ] +INFO: Idle: 1647, Running: 4, Completed: 14369 [ 19h 33m ] +INFO: Idle: 1645, Running: 4, Completed: 14371 [ 19h 33m ] +INFO: Idle: 1643, Running: 4, Completed: 14373 [ 19h 34m ] +INFO: Idle: 1642, Running: 4, Completed: 14374 [ 19h 34m ] +INFO: Idle: 1639, Running: 4, Completed: 14377 [ 19h 34m ] +INFO: Idle: 1638, Running: 4, Completed: 14378 [ 19h 34m ] +INFO: Idle: 1636, Running: 4, Completed: 14380 [ 19h 34m ] +INFO: Idle: 1635, Running: 4, Completed: 14381 [ 19h 34m ] +INFO: Idle: 1634, Running: 4, Completed: 14382 [ 19h 34m ] +INFO: Idle: 1633, Running: 4, Completed: 14383 [ 19h 35m ] +INFO: Idle: 1630, Running: 4, Completed: 14386 [ 19h 35m ] +INFO: Idle: 1629, Running: 4, Completed: 14387 [ 19h 35m ] +INFO: Idle: 1627, Running: 4, Completed: 14389 [ 19h 35m ] +INFO: Idle: 1626, Running: 4, Completed: 14390 [ 19h 35m ] +INFO: Idle: 1623, Running: 4, Completed: 14393 [ 19h 35m ] +INFO: Idle: 1622, Running: 4, Completed: 14394 [ 19h 35m ] +INFO: Idle: 1619, Running: 4, Completed: 14397 [ 19h 36m ] +INFO: Idle: 1618, Running: 4, Completed: 14398 [ 19h 36m ] +INFO: Idle: 1615, Running: 4, Completed: 14401 [ 19h 36m ] +INFO: Idle: 1614, Running: 4, Completed: 14402 [ 19h 36m ] +INFO: Idle: 1613, Running: 4, Completed: 14403 [ 19h 36m ] +INFO: Idle: 1611, Running: 4, Completed: 14405 [ 19h 36m ] +INFO: Idle: 1608, Running: 4, Completed: 14408 [ 19h 36m ] +INFO: Idle: 1607, Running: 4, Completed: 14409 [ 19h 36m ] +INFO: Idle: 1603, Running: 4, Completed: 14413 [ 19h 37m ] +INFO: Idle: 1599, Running: 4, Completed: 14417 [ 19h 37m ] +INFO: Idle: 1596, Running: 4, Completed: 14420 [ 19h 37m ] +INFO: Idle: 1595, Running: 4, Completed: 14421 [ 19h 38m ] +INFO: Idle: 1592, Running: 4, Completed: 14424 [ 19h 38m ] +INFO: Idle: 1591, Running: 4, Completed: 14425 [ 19h 38m ] +INFO: Idle: 1589, Running: 4, Completed: 14427 [ 19h 38m ] +INFO: Idle: 1588, Running: 4, Completed: 14428 [ 19h 38m ] +INFO: Idle: 1587, Running: 4, Completed: 14429 [ 19h 38m ] +INFO: Idle: 1586, Running: 4, Completed: 14430 [ 19h 38m ] +INFO: Idle: 1585, Running: 4, Completed: 14431 [ 19h 38m ] +INFO: Idle: 1584, Running: 4, Completed: 14432 [ 19h 38m ] +INFO: Idle: 1582, Running: 4, Completed: 14434 [ 19h 38m ] +INFO: Idle: 1580, Running: 4, Completed: 14436 [ 19h 39m ] +INFO: Idle: 1577, Running: 4, Completed: 14439 [ 19h 39m ] +INFO: Idle: 1576, Running: 4, Completed: 14440 [ 19h 39m ] +INFO: Idle: 1575, Running: 4, Completed: 14441 [ 19h 39m ] +INFO: Idle: 1572, Running: 4, Completed: 14444 [ 19h 39m ] +INFO: Idle: 1571, Running: 4, Completed: 14445 [ 19h 39m ] +INFO: Idle: 1570, Running: 4, Completed: 14446 [ 19h 39m ] +INFO: Idle: 1569, Running: 4, Completed: 14447 [ 19h 40m ] +INFO: Idle: 1568, Running: 4, Completed: 14448 [ 19h 40m ] +INFO: Idle: 1567, Running: 4, Completed: 14449 [ 19h 40m ] +INFO: Idle: 1566, Running: 4, Completed: 14450 [ 19h 40m ] +INFO: Idle: 1565, Running: 4, Completed: 14451 [ 19h 40m ] +INFO: Idle: 1564, Running: 4, Completed: 14452 [ 19h 40m ] +INFO: Idle: 1562, Running: 4, Completed: 14454 [ 19h 40m ] +INFO: Idle: 1560, Running: 4, Completed: 14456 [ 19h 40m ] +INFO: Idle: 1558, Running: 4, Completed: 14458 [ 19h 40m ] +INFO: Idle: 1556, Running: 4, Completed: 14460 [ 19h 41m ] +INFO: Idle: 1554, Running: 4, Completed: 14462 [ 19h 41m ] +INFO: Idle: 1552, Running: 4, Completed: 14464 [ 19h 41m ] +INFO: Idle: 1550, Running: 4, Completed: 14466 [ 19h 41m ] +INFO: Idle: 1547, Running: 4, Completed: 14469 [ 19h 41m ] +INFO: Idle: 1546, Running: 4, Completed: 14470 [ 19h 41m ] +INFO: Idle: 1543, Running: 4, Completed: 14473 [ 19h 42m ] +INFO: Idle: 1542, Running: 4, Completed: 14474 [ 19h 42m ] +INFO: Idle: 1539, Running: 4, Completed: 14477 [ 19h 42m ] +INFO: Idle: 1538, Running: 4, Completed: 14478 [ 19h 42m ] +INFO: Idle: 1536, Running: 4, Completed: 14480 [ 19h 42m ] +INFO: Idle: 1535, Running: 4, Completed: 14481 [ 19h 42m ] +INFO: Idle: 1534, Running: 4, Completed: 14482 [ 19h 42m ] +INFO: Idle: 1533, Running: 4, Completed: 14483 [ 19h 42m ] +INFO: Idle: 1531, Running: 4, Completed: 14485 [ 19h 43m ] +INFO: Idle: 1528, Running: 4, Completed: 14488 [ 19h 43m ] +INFO: Idle: 1527, Running: 4, Completed: 14489 [ 19h 43m ] +INFO: Idle: 1524, Running: 4, Completed: 14492 [ 19h 43m ] +INFO: Idle: 1523, Running: 4, Completed: 14493 [ 19h 43m ] +INFO: Idle: 1522, Running: 4, Completed: 14494 [ 19h 43m ] +INFO: Idle: 1519, Running: 4, Completed: 14497 [ 19h 43m ] +INFO: Idle: 1518, Running: 4, Completed: 14498 [ 19h 44m ] +INFO: Idle: 1517, Running: 4, Completed: 14499 [ 19h 44m ] +INFO: Idle: 1515, Running: 4, Completed: 14501 [ 19h 44m ] +INFO: Idle: 1514, Running: 4, Completed: 14502 [ 19h 44m ] +INFO: Idle: 1513, Running: 4, Completed: 14503 [ 19h 44m ] +INFO: Idle: 1511, Running: 4, Completed: 14505 [ 19h 44m ] +INFO: Idle: 1509, Running: 4, Completed: 14507 [ 19h 44m ] +INFO: Idle: 1507, Running: 4, Completed: 14509 [ 19h 45m ] +INFO: Idle: 1505, Running: 4, Completed: 14511 [ 19h 45m ] +INFO: Idle: 1503, Running: 4, Completed: 14513 [ 19h 45m ] +INFO: Idle: 1501, Running: 4, Completed: 14515 [ 19h 45m ] +INFO: Idle: 1499, Running: 4, Completed: 14517 [ 19h 45m ] +INFO: Idle: 1497, Running: 4, Completed: 14519 [ 19h 45m ] +INFO: Idle: 1494, Running: 4, Completed: 14522 [ 19h 45m ] +INFO: Idle: 1493, Running: 4, Completed: 14523 [ 19h 46m ] +INFO: Idle: 1490, Running: 4, Completed: 14526 [ 19h 46m ] +INFO: Idle: 1489, Running: 4, Completed: 14527 [ 19h 46m ] +INFO: Idle: 1487, Running: 4, Completed: 14529 [ 19h 46m ] +INFO: Idle: 1485, Running: 4, Completed: 14531 [ 19h 46m ] +INFO: Idle: 1483, Running: 4, Completed: 14533 [ 19h 46m ] +INFO: Idle: 1482, Running: 4, Completed: 14534 [ 19h 46m ] +INFO: Idle: 1481, Running: 4, Completed: 14535 [ 19h 46m ] +INFO: Idle: 1480, Running: 4, Completed: 14536 [ 19h 47m ] +INFO: Idle: 1479, Running: 4, Completed: 14537 [ 19h 47m ] +INFO: Idle: 1478, Running: 4, Completed: 14538 [ 19h 47m ] +INFO: Idle: 1476, Running: 4, Completed: 14540 [ 19h 47m ] +INFO: Idle: 1474, Running: 4, Completed: 14542 [ 19h 47m ] +INFO: Idle: 1472, Running: 4, Completed: 14544 [ 19h 47m ] +INFO: Idle: 1470, Running: 4, Completed: 14546 [ 19h 47m ] +INFO: Idle: 1469, Running: 4, Completed: 14547 [ 19h 48m ] +INFO: Idle: 1468, Running: 4, Completed: 14548 [ 19h 48m ] +INFO: Idle: 1466, Running: 4, Completed: 14550 [ 19h 48m ] +INFO: Idle: 1465, Running: 4, Completed: 14551 [ 19h 48m ] +INFO: Idle: 1464, Running: 4, Completed: 14552 [ 19h 48m ] +INFO: Idle: 1461, Running: 4, Completed: 14555 [ 19h 48m ] +INFO: Idle: 1460, Running: 4, Completed: 14556 [ 19h 48m ] +INFO: Idle: 1457, Running: 4, Completed: 14559 [ 19h 49m ] +INFO: Idle: 1456, Running: 4, Completed: 14560 [ 19h 49m ] +INFO: Idle: 1455, Running: 4, Completed: 14561 [ 19h 49m ] +INFO: Idle: 1453, Running: 4, Completed: 14563 [ 19h 49m ] +INFO: Idle: 1452, Running: 4, Completed: 14564 [ 19h 49m ] +INFO: Idle: 1451, Running: 4, Completed: 14565 [ 19h 49m ] +INFO: Idle: 1448, Running: 4, Completed: 14568 [ 19h 50m ] +INFO: Idle: 1446, Running: 4, Completed: 14570 [ 19h 50m ] +INFO: Idle: 1444, Running: 4, Completed: 14572 [ 19h 50m ] +INFO: Idle: 1442, Running: 4, Completed: 14574 [ 19h 50m ] +INFO: Idle: 1440, Running: 4, Completed: 14576 [ 19h 50m ] +INFO: Idle: 1439, Running: 4, Completed: 14577 [ 19h 50m ] +INFO: Idle: 1437, Running: 4, Completed: 14579 [ 19h 50m ] +INFO: Idle: 1436, Running: 4, Completed: 14580 [ 19h 50m ] +INFO: Idle: 1435, Running: 4, Completed: 14581 [ 19h 50m ] +INFO: Idle: 1433, Running: 4, Completed: 14583 [ 19h 51m ] +INFO: Idle: 1432, Running: 4, Completed: 14584 [ 19h 51m ] +INFO: Idle: 1430, Running: 4, Completed: 14586 [ 19h 51m ] +INFO: Idle: 1428, Running: 4, Completed: 14588 [ 19h 51m ] +INFO: Idle: 1426, Running: 4, Completed: 14590 [ 19h 51m ] +INFO: Idle: 1424, Running: 4, Completed: 14592 [ 19h 51m ] +INFO: Idle: 1421, Running: 4, Completed: 14595 [ 19h 51m ] +INFO: Idle: 1420, Running: 4, Completed: 14596 [ 19h 52m ] +INFO: Idle: 1418, Running: 4, Completed: 14598 [ 19h 52m ] +INFO: Idle: 1416, Running: 4, Completed: 14600 [ 19h 52m ] +INFO: Idle: 1412, Running: 4, Completed: 14604 [ 19h 52m ] +INFO: Idle: 1409, Running: 4, Completed: 14607 [ 19h 52m ] +INFO: Idle: 1408, Running: 4, Completed: 14608 [ 19h 53m ] +INFO: Idle: 1405, Running: 4, Completed: 14611 [ 19h 53m ] +INFO: Idle: 1404, Running: 4, Completed: 14612 [ 19h 53m ] +INFO: Idle: 1403, Running: 4, Completed: 14613 [ 19h 53m ] +INFO: Idle: 1400, Running: 4, Completed: 14616 [ 19h 53m ] +INFO: Idle: 1398, Running: 4, Completed: 14618 [ 19h 53m ] +INFO: Idle: 1396, Running: 4, Completed: 14620 [ 19h 54m ] +INFO: Idle: 1394, Running: 4, Completed: 14622 [ 19h 54m ] +INFO: Idle: 1392, Running: 4, Completed: 14624 [ 19h 54m ] +INFO: Idle: 1391, Running: 4, Completed: 14625 [ 19h 54m ] +INFO: Idle: 1389, Running: 4, Completed: 14627 [ 19h 54m ] +INFO: Idle: 1388, Running: 4, Completed: 14628 [ 19h 54m ] +INFO: Idle: 1387, Running: 4, Completed: 14629 [ 19h 54m ] +INFO: Idle: 1385, Running: 4, Completed: 14631 [ 19h 54m ] +INFO: Idle: 1384, Running: 4, Completed: 14632 [ 19h 54m ] +INFO: Idle: 1383, Running: 4, Completed: 14633 [ 19h 55m ] +INFO: Idle: 1381, Running: 4, Completed: 14635 [ 19h 55m ] +INFO: Idle: 1380, Running: 4, Completed: 14636 [ 19h 55m ] +INFO: Idle: 1378, Running: 4, Completed: 14638 [ 19h 55m ] +INFO: Idle: 1376, Running: 4, Completed: 14640 [ 19h 55m ] +INFO: Idle: 1373, Running: 4, Completed: 14643 [ 19h 55m ] +INFO: Idle: 1372, Running: 4, Completed: 14644 [ 19h 55m ] +INFO: Idle: 1369, Running: 4, Completed: 14647 [ 19h 55m ] +INFO: Idle: 1368, Running: 4, Completed: 14648 [ 19h 56m ] +INFO: Idle: 1365, Running: 4, Completed: 14651 [ 19h 56m ] +INFO: Idle: 1364, Running: 4, Completed: 14652 [ 19h 56m ] +INFO: Idle: 1363, Running: 4, Completed: 14653 [ 19h 56m ] +INFO: Idle: 1362, Running: 4, Completed: 14654 [ 19h 56m ] +INFO: Idle: 1360, Running: 4, Completed: 14656 [ 19h 56m ] +INFO: Idle: 1357, Running: 4, Completed: 14659 [ 19h 56m ] +INFO: Idle: 1356, Running: 4, Completed: 14660 [ 19h 57m ] +INFO: Idle: 1353, Running: 4, Completed: 14663 [ 19h 57m ] +INFO: Idle: 1352, Running: 4, Completed: 14664 [ 19h 57m ] +INFO: Idle: 1349, Running: 4, Completed: 14667 [ 19h 57m ] +INFO: Idle: 1348, Running: 4, Completed: 14668 [ 19h 57m ] +INFO: Idle: 1346, Running: 4, Completed: 14670 [ 19h 57m ] +INFO: Idle: 1345, Running: 4, Completed: 14671 [ 19h 57m ] +INFO: Idle: 1344, Running: 4, Completed: 14672 [ 19h 58m ] +INFO: Idle: 1341, Running: 4, Completed: 14675 [ 19h 58m ] +INFO: Idle: 1340, Running: 4, Completed: 14676 [ 19h 58m ] +INFO: Idle: 1339, Running: 4, Completed: 14677 [ 19h 58m ] +INFO: Idle: 1336, Running: 4, Completed: 14680 [ 19h 58m ] +INFO: Idle: 1334, Running: 4, Completed: 14682 [ 19h 58m ] +INFO: Idle: 1332, Running: 4, Completed: 14684 [ 19h 59m ] +INFO: Idle: 1331, Running: 4, Completed: 14685 [ 19h 59m ] +INFO: Idle: 1330, Running: 4, Completed: 14686 [ 19h 59m ] +INFO: Idle: 1328, Running: 4, Completed: 14688 [ 19h 59m ] +INFO: Idle: 1327, Running: 4, Completed: 14689 [ 19h 59m ] +INFO: Idle: 1326, Running: 4, Completed: 14690 [ 19h 59m ] +INFO: Idle: 1325, Running: 4, Completed: 14691 [ 19h 59m ] +INFO: Idle: 1323, Running: 4, Completed: 14693 [ 19h 59m ] +INFO: Idle: 1322, Running: 4, Completed: 14694 [ 19h 59m ] +INFO: Idle: 1320, Running: 4, Completed: 14696 [ 20h 0m ] +INFO: Idle: 1319, Running: 4, Completed: 14697 [ 20h 0m ] +INFO: Idle: 1318, Running: 4, Completed: 14698 [ 20h 0m ] +INFO: Idle: 1316, Running: 4, Completed: 14700 [ 20h 0m ] +INFO: Idle: 1315, Running: 4, Completed: 14701 [ 20h 0m ] +INFO: Idle: 1314, Running: 4, Completed: 14702 [ 20h 0m ] +INFO: Idle: 1313, Running: 4, Completed: 14703 [ 20h 0m ] +INFO: Idle: 1311, Running: 4, Completed: 14705 [ 20h 0m ] +INFO: Idle: 1310, Running: 4, Completed: 14706 [ 20h 0m ] +INFO: Idle: 1309, Running: 4, Completed: 14707 [ 20h 0m ] +INFO: Idle: 1306, Running: 4, Completed: 14710 [ 20h 1m ] +INFO: Idle: 1303, Running: 4, Completed: 14713 [ 20h 1m ] +INFO: Idle: 1302, Running: 4, Completed: 14714 [ 20h 1m ] +INFO: Idle: 1299, Running: 4, Completed: 14717 [ 20h 1m ] +INFO: Idle: 1298, Running: 4, Completed: 14718 [ 20h 1m ] +INFO: Idle: 1295, Running: 4, Completed: 14721 [ 20h 2m ] +INFO: Idle: 1294, Running: 4, Completed: 14722 [ 20h 2m ] +INFO: Idle: 1290, Running: 4, Completed: 14726 [ 20h 2m ] +INFO: Idle: 1287, Running: 4, Completed: 14729 [ 20h 2m ] +INFO: Idle: 1286, Running: 4, Completed: 14730 [ 20h 2m ] +INFO: Idle: 1283, Running: 4, Completed: 14733 [ 20h 3m ] +INFO: Idle: 1282, Running: 4, Completed: 14734 [ 20h 3m ] +INFO: Idle: 1281, Running: 4, Completed: 14735 [ 20h 3m ] +INFO: Idle: 1280, Running: 4, Completed: 14736 [ 20h 3m ] +INFO: Idle: 1278, Running: 4, Completed: 14738 [ 20h 3m ] +INFO: Idle: 1275, Running: 4, Completed: 14741 [ 20h 3m ] +INFO: Idle: 1274, Running: 4, Completed: 14742 [ 20h 3m ] +INFO: Idle: 1271, Running: 4, Completed: 14745 [ 20h 4m ] +INFO: Idle: 1270, Running: 4, Completed: 14746 [ 20h 4m ] +INFO: Idle: 1267, Running: 4, Completed: 14749 [ 20h 4m ] +INFO: Idle: 1266, Running: 4, Completed: 14750 [ 20h 4m ] +INFO: Idle: 1264, Running: 4, Completed: 14752 [ 20h 4m ] +INFO: Idle: 1263, Running: 4, Completed: 14753 [ 20h 4m ] +INFO: Idle: 1262, Running: 4, Completed: 14754 [ 20h 4m ] +INFO: Idle: 1259, Running: 4, Completed: 14757 [ 20h 4m ] +INFO: Idle: 1258, Running: 4, Completed: 14758 [ 20h 5m ] +INFO: Idle: 1257, Running: 4, Completed: 14759 [ 20h 5m ] +INFO: Idle: 1254, Running: 4, Completed: 14762 [ 20h 5m ] +INFO: Idle: 1252, Running: 4, Completed: 14764 [ 20h 5m ] +INFO: Idle: 1250, Running: 4, Completed: 14766 [ 20h 5m ] +INFO: Idle: 1248, Running: 4, Completed: 14768 [ 20h 5m ] +INFO: Idle: 1246, Running: 4, Completed: 14770 [ 20h 5m ] +INFO: Idle: 1245, Running: 4, Completed: 14771 [ 20h 6m ] +INFO: Idle: 1244, Running: 4, Completed: 14772 [ 20h 6m ] +INFO: Idle: 1243, Running: 4, Completed: 14773 [ 20h 6m ] +INFO: Idle: 1241, Running: 4, Completed: 14775 [ 20h 6m ] +INFO: Idle: 1240, Running: 4, Completed: 14776 [ 20h 6m ] +INFO: Idle: 1239, Running: 4, Completed: 14777 [ 20h 6m ] +INFO: Idle: 1237, Running: 4, Completed: 14779 [ 20h 6m ] +INFO: Idle: 1236, Running: 4, Completed: 14780 [ 20h 6m ] +INFO: Idle: 1235, Running: 4, Completed: 14781 [ 20h 7m ] +INFO: Idle: 1233, Running: 4, Completed: 14783 [ 20h 7m ] +INFO: Idle: 1232, Running: 4, Completed: 14784 [ 20h 7m ] +INFO: Idle: 1231, Running: 4, Completed: 14785 [ 20h 7m ] +INFO: Idle: 1229, Running: 4, Completed: 14787 [ 20h 7m ] +INFO: Idle: 1228, Running: 4, Completed: 14788 [ 20h 7m ] +INFO: Idle: 1225, Running: 4, Completed: 14791 [ 20h 7m ] +INFO: Idle: 1224, Running: 4, Completed: 14792 [ 20h 8m ] +INFO: Idle: 1221, Running: 4, Completed: 14795 [ 20h 8m ] +INFO: Idle: 1220, Running: 4, Completed: 14796 [ 20h 8m ] +INFO: Idle: 1217, Running: 4, Completed: 14799 [ 20h 8m ] +INFO: Idle: 1216, Running: 4, Completed: 14800 [ 20h 8m ] +INFO: Idle: 1213, Running: 4, Completed: 14803 [ 20h 8m ] +INFO: Idle: 1212, Running: 4, Completed: 14804 [ 20h 8m ] +INFO: Idle: 1208, Running: 4, Completed: 14808 [ 20h 9m ] +INFO: Idle: 1204, Running: 4, Completed: 14812 [ 20h 9m ] +INFO: Idle: 1200, Running: 4, Completed: 14816 [ 20h 10m ] +INFO: Idle: 1197, Running: 4, Completed: 14819 [ 20h 10m ] +INFO: Idle: 1196, Running: 4, Completed: 14820 [ 20h 10m ] +INFO: Idle: 1192, Running: 4, Completed: 14824 [ 20h 10m ] +INFO: Idle: 1190, Running: 4, Completed: 14826 [ 20h 10m ] +INFO: Idle: 1188, Running: 4, Completed: 14828 [ 20h 10m ] +INFO: Idle: 1186, Running: 4, Completed: 14830 [ 20h 11m ] +INFO: Idle: 1184, Running: 4, Completed: 14832 [ 20h 11m ] +INFO: Idle: 1181, Running: 4, Completed: 14835 [ 20h 11m ] +INFO: Idle: 1180, Running: 4, Completed: 14836 [ 20h 11m ] +INFO: Idle: 1179, Running: 4, Completed: 14837 [ 20h 11m ] +INFO: Idle: 1177, Running: 4, Completed: 14839 [ 20h 11m ] +INFO: Idle: 1176, Running: 4, Completed: 14840 [ 20h 11m ] +INFO: Idle: 1174, Running: 4, Completed: 14842 [ 20h 12m ] +INFO: Idle: 1173, Running: 4, Completed: 14843 [ 20h 12m ] +INFO: Idle: 1172, Running: 4, Completed: 14844 [ 20h 12m ] +INFO: Idle: 1171, Running: 4, Completed: 14845 [ 20h 12m ] +INFO: Idle: 1169, Running: 4, Completed: 14847 [ 20h 12m ] +INFO: Idle: 1168, Running: 4, Completed: 14848 [ 20h 12m ] +INFO: Idle: 1165, Running: 4, Completed: 14851 [ 20h 12m ] +INFO: Idle: 1164, Running: 4, Completed: 14852 [ 20h 13m ] +INFO: Idle: 1161, Running: 4, Completed: 14855 [ 20h 13m ] +INFO: Idle: 1160, Running: 4, Completed: 14856 [ 20h 13m ] +INFO: Idle: 1157, Running: 4, Completed: 14859 [ 20h 13m ] +INFO: Idle: 1156, Running: 4, Completed: 14860 [ 20h 13m ] +INFO: Idle: 1155, Running: 4, Completed: 14861 [ 20h 13m ] +INFO: Idle: 1152, Running: 4, Completed: 14864 [ 20h 14m ] +INFO: Idle: 1150, Running: 4, Completed: 14866 [ 20h 14m ] +INFO: Idle: 1148, Running: 4, Completed: 14868 [ 20h 14m ] +INFO: Idle: 1146, Running: 4, Completed: 14870 [ 20h 14m ] +INFO: Idle: 1144, Running: 4, Completed: 14872 [ 20h 14m ] +INFO: Idle: 1143, Running: 4, Completed: 14873 [ 20h 14m ] +INFO: Idle: 1141, Running: 4, Completed: 14875 [ 20h 14m ] +INFO: Idle: 1140, Running: 4, Completed: 14876 [ 20h 15m ] +INFO: Idle: 1139, Running: 4, Completed: 14877 [ 20h 15m ] +INFO: Idle: 1137, Running: 4, Completed: 14879 [ 20h 15m ] +INFO: Idle: 1136, Running: 4, Completed: 14880 [ 20h 15m ] +INFO: Idle: 1135, Running: 4, Completed: 14881 [ 20h 15m ] +INFO: Idle: 1132, Running: 4, Completed: 14884 [ 20h 15m ] +INFO: Idle: 1130, Running: 4, Completed: 14886 [ 20h 15m ] +INFO: Idle: 1128, Running: 4, Completed: 14888 [ 20h 15m ] +INFO: Idle: 1126, Running: 4, Completed: 14890 [ 20h 15m ] +INFO: Idle: 1124, Running: 4, Completed: 14892 [ 20h 16m ] +INFO: Idle: 1122, Running: 4, Completed: 14894 [ 20h 16m ] +INFO: Idle: 1121, Running: 4, Completed: 14895 [ 20h 16m ] +INFO: Idle: 1120, Running: 4, Completed: 14896 [ 20h 16m ] +INFO: Idle: 1117, Running: 4, Completed: 14899 [ 20h 16m ] +INFO: Idle: 1116, Running: 4, Completed: 14900 [ 20h 16m ] +INFO: Idle: 1115, Running: 4, Completed: 14901 [ 20h 16m ] +INFO: Idle: 1114, Running: 4, Completed: 14902 [ 20h 16m ] +INFO: Idle: 1112, Running: 4, Completed: 14904 [ 20h 17m ] +INFO: Idle: 1109, Running: 4, Completed: 14907 [ 20h 17m ] +INFO: Idle: 1108, Running: 4, Completed: 14908 [ 20h 17m ] +INFO: Idle: 1105, Running: 4, Completed: 14911 [ 20h 17m ] +INFO: Idle: 1104, Running: 4, Completed: 14912 [ 20h 17m ] +INFO: Idle: 1101, Running: 4, Completed: 14915 [ 20h 17m ] +INFO: Idle: 1100, Running: 4, Completed: 14916 [ 20h 18m ] +INFO: Idle: 1098, Running: 4, Completed: 14918 [ 20h 18m ] +INFO: Idle: 1097, Running: 4, Completed: 14919 [ 20h 18m ] +INFO: Idle: 1096, Running: 4, Completed: 14920 [ 20h 18m ] +INFO: Idle: 1093, Running: 4, Completed: 14923 [ 20h 18m ] +INFO: Idle: 1092, Running: 4, Completed: 14924 [ 20h 18m ] +INFO: Idle: 1091, Running: 4, Completed: 14925 [ 20h 18m ] +INFO: Idle: 1088, Running: 4, Completed: 14928 [ 20h 18m ] +INFO: Idle: 1087, Running: 4, Completed: 14929 [ 20h 19m ] +INFO: Idle: 1086, Running: 4, Completed: 14930 [ 20h 19m ] +INFO: Idle: 1084, Running: 4, Completed: 14932 [ 20h 19m ] +INFO: Idle: 1083, Running: 4, Completed: 14933 [ 20h 19m ] +INFO: Idle: 1082, Running: 4, Completed: 14934 [ 20h 19m ] +INFO: Idle: 1080, Running: 4, Completed: 14936 [ 20h 19m ] +INFO: Idle: 1079, Running: 4, Completed: 14937 [ 20h 19m ] +INFO: Idle: 1078, Running: 4, Completed: 14938 [ 20h 19m ] +INFO: Idle: 1077, Running: 4, Completed: 14939 [ 20h 19m ] +INFO: Idle: 1075, Running: 4, Completed: 14941 [ 20h 20m ] +INFO: Idle: 1074, Running: 4, Completed: 14942 [ 20h 20m ] +INFO: Idle: 1073, Running: 4, Completed: 14943 [ 20h 20m ] +INFO: Idle: 1071, Running: 4, Completed: 14945 [ 20h 20m ] +INFO: Idle: 1070, Running: 4, Completed: 14946 [ 20h 20m ] +INFO: Idle: 1069, Running: 4, Completed: 14947 [ 20h 20m ] +INFO: Idle: 1067, Running: 4, Completed: 14949 [ 20h 20m ] +INFO: Idle: 1066, Running: 4, Completed: 14950 [ 20h 20m ] +INFO: Idle: 1065, Running: 4, Completed: 14951 [ 20h 20m ] +INFO: Idle: 1063, Running: 4, Completed: 14953 [ 20h 21m ] +INFO: Idle: 1062, Running: 4, Completed: 14954 [ 20h 21m ] +INFO: Idle: 1059, Running: 4, Completed: 14957 [ 20h 21m ] +INFO: Idle: 1058, Running: 4, Completed: 14958 [ 20h 21m ] +INFO: Idle: 1055, Running: 4, Completed: 14961 [ 20h 21m ] +INFO: Idle: 1054, Running: 4, Completed: 14962 [ 20h 21m ] +INFO: Idle: 1051, Running: 4, Completed: 14965 [ 20h 22m ] +INFO: Idle: 1050, Running: 4, Completed: 14966 [ 20h 22m ] +INFO: Idle: 1047, Running: 4, Completed: 14969 [ 20h 22m ] +INFO: Idle: 1046, Running: 4, Completed: 14970 [ 20h 22m ] +INFO: Idle: 1042, Running: 4, Completed: 14974 [ 20h 22m ] +INFO: Idle: 1039, Running: 4, Completed: 14977 [ 20h 23m ] +INFO: Idle: 1038, Running: 4, Completed: 14978 [ 20h 23m ] +INFO: Idle: 1035, Running: 4, Completed: 14981 [ 20h 23m ] +INFO: Idle: 1034, Running: 4, Completed: 14982 [ 20h 23m ] +INFO: Idle: 1033, Running: 4, Completed: 14983 [ 20h 23m ] +INFO: Idle: 1032, Running: 4, Completed: 14984 [ 20h 23m ] +INFO: Idle: 1030, Running: 4, Completed: 14986 [ 20h 23m ] +INFO: Idle: 1027, Running: 4, Completed: 14989 [ 20h 24m ] +INFO: Idle: 1026, Running: 4, Completed: 14990 [ 20h 24m ] +INFO: Idle: 1023, Running: 4, Completed: 14993 [ 20h 24m ] +INFO: Idle: 1022, Running: 4, Completed: 14994 [ 20h 24m ] +INFO: Idle: 1019, Running: 4, Completed: 14997 [ 20h 24m ] +INFO: Idle: 1018, Running: 4, Completed: 14998 [ 20h 24m ] +INFO: Idle: 1016, Running: 4, Completed: 15000 [ 20h 24m ] +INFO: Idle: 1015, Running: 4, Completed: 15001 [ 20h 25m ] +INFO: Idle: 1014, Running: 4, Completed: 15002 [ 20h 25m ] +INFO: Idle: 1011, Running: 4, Completed: 15005 [ 20h 25m ] +INFO: Idle: 1010, Running: 4, Completed: 15006 [ 20h 25m ] +INFO: Idle: 1009, Running: 4, Completed: 15007 [ 20h 25m ] +INFO: Idle: 1007, Running: 4, Completed: 15009 [ 20h 25m ] +INFO: Idle: 1006, Running: 4, Completed: 15010 [ 20h 25m ] +INFO: Idle: 1004, Running: 4, Completed: 15012 [ 20h 25m ] +INFO: Idle: 1002, Running: 4, Completed: 15014 [ 20h 26m ] +INFO: Idle: 1000, Running: 4, Completed: 15016 [ 20h 26m ] +INFO: Idle: 998, Running: 4, Completed: 15018 [ 20h 26m ] +INFO: Idle: 997, Running: 4, Completed: 15019 [ 20h 26m ] +INFO: Idle: 996, Running: 4, Completed: 15020 [ 20h 26m ] +INFO: Idle: 995, Running: 4, Completed: 15021 [ 20h 26m ] +INFO: Idle: 993, Running: 4, Completed: 15023 [ 20h 26m ] +INFO: Idle: 992, Running: 4, Completed: 15024 [ 20h 27m ] +INFO: Idle: 991, Running: 4, Completed: 15025 [ 20h 27m ] +INFO: Idle: 989, Running: 4, Completed: 15027 [ 20h 27m ] +INFO: Idle: 988, Running: 4, Completed: 15028 [ 20h 27m ] +INFO: Idle: 987, Running: 4, Completed: 15029 [ 20h 27m ] +INFO: Idle: 985, Running: 4, Completed: 15031 [ 20h 27m ] +INFO: Idle: 984, Running: 4, Completed: 15032 [ 20h 27m ] +INFO: Idle: 983, Running: 4, Completed: 15033 [ 20h 27m ] +INFO: Idle: 981, Running: 4, Completed: 15035 [ 20h 27m ] +INFO: Idle: 980, Running: 4, Completed: 15036 [ 20h 27m ] +INFO: Idle: 978, Running: 4, Completed: 15038 [ 20h 28m ] +INFO: Idle: 976, Running: 4, Completed: 15040 [ 20h 28m ] +INFO: Idle: 973, Running: 4, Completed: 15043 [ 20h 28m ] +INFO: Idle: 972, Running: 4, Completed: 15044 [ 20h 28m ] +INFO: Idle: 969, Running: 4, Completed: 15047 [ 20h 28m ] +INFO: Idle: 968, Running: 4, Completed: 15048 [ 20h 29m ] +INFO: Idle: 965, Running: 4, Completed: 15051 [ 20h 29m ] +INFO: Idle: 964, Running: 4, Completed: 15052 [ 20h 29m ] +INFO: Idle: 960, Running: 4, Completed: 15056 [ 20h 29m ] +INFO: Idle: 958, Running: 4, Completed: 15058 [ 20h 29m ] +INFO: Idle: 956, Running: 4, Completed: 15060 [ 20h 29m ] +INFO: Idle: 952, Running: 4, Completed: 15064 [ 20h 30m ] +INFO: Idle: 949, Running: 4, Completed: 15067 [ 20h 30m ] +INFO: Idle: 948, Running: 4, Completed: 15068 [ 20h 30m ] +INFO: Idle: 945, Running: 4, Completed: 15071 [ 20h 30m ] +INFO: Idle: 944, Running: 4, Completed: 15072 [ 20h 30m ] +INFO: Idle: 943, Running: 4, Completed: 15073 [ 20h 30m ] +INFO: Idle: 940, Running: 4, Completed: 15076 [ 20h 31m ] +INFO: Idle: 939, Running: 4, Completed: 15077 [ 20h 31m ] +INFO: Idle: 936, Running: 4, Completed: 15080 [ 20h 31m ] +INFO: Idle: 935, Running: 4, Completed: 15081 [ 20h 31m ] +INFO: Idle: 933, Running: 4, Completed: 15083 [ 20h 31m ] +INFO: Idle: 932, Running: 4, Completed: 15084 [ 20h 31m ] +INFO: Idle: 931, Running: 4, Completed: 15085 [ 20h 31m ] +INFO: Idle: 930, Running: 4, Completed: 15086 [ 20h 32m ] +INFO: Idle: 928, Running: 4, Completed: 15088 [ 20h 32m ] +INFO: Idle: 927, Running: 4, Completed: 15089 [ 20h 32m ] +INFO: Idle: 926, Running: 4, Completed: 15090 [ 20h 32m ] +INFO: Idle: 924, Running: 4, Completed: 15092 [ 20h 32m ] +INFO: Idle: 922, Running: 4, Completed: 15094 [ 20h 32m ] +INFO: Idle: 921, Running: 4, Completed: 15095 [ 20h 32m ] +INFO: Idle: 920, Running: 4, Completed: 15096 [ 20h 32m ] +INFO: Idle: 917, Running: 4, Completed: 15099 [ 20h 32m ] +INFO: Idle: 916, Running: 4, Completed: 15100 [ 20h 33m ] +INFO: Idle: 913, Running: 4, Completed: 15103 [ 20h 33m ] +INFO: Idle: 912, Running: 4, Completed: 15104 [ 20h 33m ] +INFO: Idle: 909, Running: 4, Completed: 15107 [ 20h 33m ] +INFO: Idle: 908, Running: 4, Completed: 15108 [ 20h 33m ] +INFO: Idle: 907, Running: 4, Completed: 15109 [ 20h 33m ] +INFO: Idle: 905, Running: 4, Completed: 15111 [ 20h 33m ] +INFO: Idle: 904, Running: 4, Completed: 15112 [ 20h 34m ] +INFO: Idle: 903, Running: 4, Completed: 15113 [ 20h 34m ] +INFO: Idle: 900, Running: 4, Completed: 15116 [ 20h 34m ] +INFO: Idle: 899, Running: 4, Completed: 15117 [ 20h 34m ] +INFO: Idle: 898, Running: 4, Completed: 15118 [ 20h 34m ] +INFO: Idle: 896, Running: 4, Completed: 15120 [ 20h 34m ] +INFO: Idle: 895, Running: 4, Completed: 15121 [ 20h 34m ] +INFO: Idle: 894, Running: 4, Completed: 15122 [ 20h 34m ] +INFO: Idle: 892, Running: 4, Completed: 15124 [ 20h 35m ] +INFO: Idle: 891, Running: 4, Completed: 15125 [ 20h 35m ] +INFO: Idle: 890, Running: 4, Completed: 15126 [ 20h 35m ] +INFO: Idle: 888, Running: 4, Completed: 15128 [ 20h 35m ] +INFO: Idle: 887, Running: 4, Completed: 15129 [ 20h 35m ] +INFO: Idle: 886, Running: 4, Completed: 15130 [ 20h 35m ] +INFO: Idle: 885, Running: 4, Completed: 15131 [ 20h 35m ] +INFO: Idle: 884, Running: 4, Completed: 15132 [ 20h 35m ] +INFO: Idle: 882, Running: 4, Completed: 15134 [ 20h 35m ] +INFO: Idle: 881, Running: 4, Completed: 15135 [ 20h 35m ] +INFO: Idle: 879, Running: 4, Completed: 15137 [ 20h 36m ] +INFO: Idle: 878, Running: 4, Completed: 15138 [ 20h 36m ] +INFO: Idle: 877, Running: 4, Completed: 15139 [ 20h 36m ] +INFO: Idle: 876, Running: 4, Completed: 15140 [ 20h 36m ] +INFO: Idle: 875, Running: 4, Completed: 15141 [ 20h 36m ] +INFO: Idle: 872, Running: 4, Completed: 15144 [ 20h 36m ] +INFO: Idle: 871, Running: 4, Completed: 15145 [ 20h 36m ] +INFO: Idle: 870, Running: 4, Completed: 15146 [ 20h 36m ] +INFO: Idle: 868, Running: 4, Completed: 15148 [ 20h 36m ] +INFO: Idle: 867, Running: 4, Completed: 15149 [ 20h 36m ] +INFO: Idle: 866, Running: 4, Completed: 15150 [ 20h 37m ] +INFO: Idle: 864, Running: 4, Completed: 15152 [ 20h 37m ] +INFO: Idle: 863, Running: 4, Completed: 15153 [ 20h 37m ] +INFO: Idle: 862, Running: 4, Completed: 15154 [ 20h 37m ] +INFO: Idle: 860, Running: 4, Completed: 15156 [ 20h 37m ] +INFO: Idle: 859, Running: 4, Completed: 15157 [ 20h 37m ] +INFO: Idle: 857, Running: 4, Completed: 15159 [ 20h 37m ] +INFO: Idle: 856, Running: 4, Completed: 15160 [ 20h 38m ] +INFO: Idle: 853, Running: 4, Completed: 15163 [ 20h 38m ] +INFO: Idle: 852, Running: 4, Completed: 15164 [ 20h 38m ] +INFO: Idle: 848, Running: 4, Completed: 15168 [ 20h 38m ] +INFO: Idle: 847, Running: 4, Completed: 15169 [ 20h 38m ] +INFO: Idle: 846, Running: 4, Completed: 15170 [ 20h 38m ] +INFO: Idle: 844, Running: 4, Completed: 15172 [ 20h 38m ] +INFO: Idle: 843, Running: 4, Completed: 15173 [ 20h 38m ] +INFO: Idle: 842, Running: 4, Completed: 15174 [ 20h 38m ] +INFO: Idle: 839, Running: 4, Completed: 15177 [ 20h 39m ] +INFO: Idle: 838, Running: 4, Completed: 15178 [ 20h 39m ] +INFO: Idle: 835, Running: 4, Completed: 15181 [ 20h 39m ] +INFO: Idle: 834, Running: 4, Completed: 15182 [ 20h 39m ] +INFO: Idle: 833, Running: 4, Completed: 15183 [ 20h 39m ] +INFO: Idle: 831, Running: 4, Completed: 15185 [ 20h 39m ] +INFO: Idle: 830, Running: 4, Completed: 15186 [ 20h 39m ] +INFO: Idle: 829, Running: 4, Completed: 15187 [ 20h 39m ] +INFO: Idle: 826, Running: 4, Completed: 15190 [ 20h 40m ] +INFO: Idle: 825, Running: 4, Completed: 15191 [ 20h 40m ] +INFO: Idle: 824, Running: 4, Completed: 15192 [ 20h 40m ] +INFO: Idle: 821, Running: 4, Completed: 15195 [ 20h 40m ] +INFO: Idle: 819, Running: 4, Completed: 15197 [ 20h 40m ] +INFO: Idle: 816, Running: 4, Completed: 15200 [ 20h 40m ] +INFO: Idle: 815, Running: 4, Completed: 15201 [ 20h 40m ] +INFO: Idle: 812, Running: 4, Completed: 15204 [ 20h 40m ] +INFO: Idle: 811, Running: 4, Completed: 15205 [ 20h 41m ] +INFO: Idle: 810, Running: 4, Completed: 15206 [ 20h 41m ] +INFO: Idle: 809, Running: 4, Completed: 15207 [ 20h 41m ] +INFO: Idle: 807, Running: 4, Completed: 15209 [ 20h 41m ] +INFO: Idle: 806, Running: 4, Completed: 15210 [ 20h 41m ] +INFO: Idle: 803, Running: 4, Completed: 15213 [ 20h 41m ] +INFO: Idle: 802, Running: 4, Completed: 15214 [ 20h 42m ] +INFO: Idle: 799, Running: 4, Completed: 15217 [ 20h 42m ] +INFO: Idle: 798, Running: 4, Completed: 15218 [ 20h 42m ] +INFO: Idle: 796, Running: 4, Completed: 15220 [ 20h 42m ] +INFO: Idle: 795, Running: 4, Completed: 15221 [ 20h 42m ] +INFO: Idle: 794, Running: 4, Completed: 15222 [ 20h 42m ] +INFO: Idle: 793, Running: 4, Completed: 15223 [ 20h 42m ] +INFO: Idle: 792, Running: 4, Completed: 15224 [ 20h 42m ] +INFO: Idle: 790, Running: 4, Completed: 15226 [ 20h 43m ] +INFO: Idle: 789, Running: 4, Completed: 15227 [ 20h 43m ] +INFO: Idle: 786, Running: 4, Completed: 15230 [ 20h 43m ] +INFO: Idle: 785, Running: 4, Completed: 15231 [ 20h 43m ] +INFO: Idle: 784, Running: 4, Completed: 15232 [ 20h 43m ] +INFO: Idle: 782, Running: 4, Completed: 15234 [ 20h 43m ] +INFO: Idle: 781, Running: 4, Completed: 15235 [ 20h 43m ] +INFO: Idle: 780, Running: 4, Completed: 15236 [ 20h 43m ] +INFO: Idle: 777, Running: 4, Completed: 15239 [ 20h 43m ] +INFO: Idle: 776, Running: 4, Completed: 15240 [ 20h 44m ] +INFO: Idle: 774, Running: 4, Completed: 15242 [ 20h 44m ] +INFO: Idle: 773, Running: 4, Completed: 15243 [ 20h 44m ] +INFO: Idle: 771, Running: 4, Completed: 15245 [ 20h 44m ] +INFO: Idle: 770, Running: 4, Completed: 15246 [ 20h 44m ] +INFO: Idle: 766, Running: 4, Completed: 15250 [ 20h 45m ] +INFO: Idle: 764, Running: 4, Completed: 15252 [ 20h 45m ] +INFO: Idle: 762, Running: 4, Completed: 15254 [ 20h 45m ] +INFO: Idle: 760, Running: 4, Completed: 15256 [ 20h 45m ] +INFO: Idle: 759, Running: 4, Completed: 15257 [ 20h 45m ] +INFO: Idle: 758, Running: 4, Completed: 15258 [ 20h 45m ] +INFO: Idle: 756, Running: 4, Completed: 15260 [ 20h 45m ] +INFO: Idle: 754, Running: 4, Completed: 15262 [ 20h 45m ] +INFO: Idle: 753, Running: 4, Completed: 15263 [ 20h 46m ] +INFO: Idle: 750, Running: 4, Completed: 15266 [ 20h 46m ] +INFO: Idle: 749, Running: 4, Completed: 15267 [ 20h 46m ] +INFO: Idle: 748, Running: 4, Completed: 15268 [ 20h 46m ] +INFO: Idle: 746, Running: 4, Completed: 15270 [ 20h 46m ] +INFO: Idle: 745, Running: 4, Completed: 15271 [ 20h 46m ] +INFO: Idle: 744, Running: 4, Completed: 15272 [ 20h 46m ] +INFO: Idle: 741, Running: 4, Completed: 15275 [ 20h 46m ] +INFO: Idle: 740, Running: 4, Completed: 15276 [ 20h 46m ] +INFO: Idle: 738, Running: 4, Completed: 15278 [ 20h 47m ] +INFO: Idle: 737, Running: 4, Completed: 15279 [ 20h 47m ] +INFO: Idle: 736, Running: 4, Completed: 15280 [ 20h 47m ] +INFO: Idle: 735, Running: 4, Completed: 15281 [ 20h 47m ] +INFO: Idle: 733, Running: 4, Completed: 15283 [ 20h 47m ] +INFO: Idle: 732, Running: 4, Completed: 15284 [ 20h 47m ] +INFO: Idle: 731, Running: 4, Completed: 15285 [ 20h 47m ] +INFO: Idle: 729, Running: 4, Completed: 15287 [ 20h 47m ] +INFO: Idle: 727, Running: 4, Completed: 15289 [ 20h 47m ] +INFO: Idle: 726, Running: 4, Completed: 15290 [ 20h 48m ] +INFO: Idle: 724, Running: 4, Completed: 15292 [ 20h 48m ] +INFO: Idle: 722, Running: 4, Completed: 15294 [ 20h 48m ] +INFO: Idle: 720, Running: 4, Completed: 15296 [ 20h 48m ] +INFO: Idle: 718, Running: 4, Completed: 15298 [ 20h 48m ] +INFO: Idle: 716, Running: 4, Completed: 15300 [ 20h 48m ] +INFO: Idle: 714, Running: 4, Completed: 15302 [ 20h 49m ] +INFO: Idle: 712, Running: 4, Completed: 15304 [ 20h 49m ] +INFO: Idle: 710, Running: 4, Completed: 15306 [ 20h 49m ] +INFO: Idle: 708, Running: 4, Completed: 15308 [ 20h 49m ] +INFO: Idle: 706, Running: 4, Completed: 15310 [ 20h 49m ] +INFO: Idle: 703, Running: 4, Completed: 15313 [ 20h 49m ] +INFO: Idle: 702, Running: 4, Completed: 15314 [ 20h 50m ] +INFO: Idle: 700, Running: 4, Completed: 15316 [ 20h 50m ] +INFO: Idle: 698, Running: 4, Completed: 15318 [ 20h 50m ] +INFO: Idle: 697, Running: 4, Completed: 15319 [ 20h 50m ] +INFO: Idle: 694, Running: 4, Completed: 15322 [ 20h 50m ] +INFO: Idle: 693, Running: 4, Completed: 15323 [ 20h 50m ] +INFO: Idle: 690, Running: 4, Completed: 15326 [ 20h 50m ] +INFO: Idle: 689, Running: 4, Completed: 15327 [ 20h 51m ] +INFO: Idle: 686, Running: 4, Completed: 15330 [ 20h 51m ] +INFO: Idle: 685, Running: 4, Completed: 15331 [ 20h 51m ] +INFO: Idle: 682, Running: 4, Completed: 15334 [ 20h 51m ] +INFO: Idle: 681, Running: 4, Completed: 15335 [ 20h 51m ] +INFO: Idle: 678, Running: 4, Completed: 15338 [ 20h 51m ] +INFO: Idle: 675, Running: 4, Completed: 15341 [ 20h 52m ] +INFO: Idle: 674, Running: 4, Completed: 15342 [ 20h 52m ] +INFO: Idle: 673, Running: 4, Completed: 15343 [ 20h 52m ] +INFO: Idle: 671, Running: 4, Completed: 15345 [ 20h 52m ] +INFO: Idle: 670, Running: 4, Completed: 15346 [ 20h 52m ] +INFO: Idle: 667, Running: 4, Completed: 15349 [ 20h 52m ] +INFO: Idle: 666, Running: 4, Completed: 15350 [ 20h 53m ] +INFO: Idle: 665, Running: 4, Completed: 15351 [ 20h 53m ] +INFO: Idle: 664, Running: 4, Completed: 15352 [ 20h 53m ] +INFO: Idle: 662, Running: 4, Completed: 15354 [ 20h 53m ] +INFO: Idle: 658, Running: 4, Completed: 15358 [ 20h 53m ] +INFO: Idle: 655, Running: 4, Completed: 15361 [ 20h 53m ] +INFO: Idle: 654, Running: 4, Completed: 15362 [ 20h 54m ] +INFO: Idle: 651, Running: 4, Completed: 15365 [ 20h 54m ] +INFO: Idle: 650, Running: 4, Completed: 15366 [ 20h 54m ] +INFO: Idle: 649, Running: 4, Completed: 15367 [ 20h 54m ] +INFO: Idle: 646, Running: 4, Completed: 15370 [ 20h 54m ] +INFO: Idle: 645, Running: 4, Completed: 15371 [ 20h 54m ] +INFO: Idle: 642, Running: 4, Completed: 15374 [ 20h 55m ] +INFO: Idle: 641, Running: 4, Completed: 15375 [ 20h 55m ] +INFO: Idle: 639, Running: 4, Completed: 15377 [ 20h 55m ] +INFO: Idle: 638, Running: 4, Completed: 15378 [ 20h 55m ] +INFO: Idle: 637, Running: 4, Completed: 15379 [ 20h 55m ] +INFO: Idle: 636, Running: 4, Completed: 15380 [ 20h 55m ] +INFO: Idle: 634, Running: 4, Completed: 15382 [ 20h 55m ] +INFO: Idle: 633, Running: 4, Completed: 15383 [ 20h 55m ] +INFO: Idle: 632, Running: 4, Completed: 15384 [ 20h 55m ] +INFO: Idle: 630, Running: 4, Completed: 15386 [ 20h 55m ] +INFO: Idle: 629, Running: 4, Completed: 15387 [ 20h 56m ] +INFO: Idle: 628, Running: 4, Completed: 15388 [ 20h 56m ] +INFO: Idle: 626, Running: 4, Completed: 15390 [ 20h 56m ] +INFO: Idle: 625, Running: 4, Completed: 15391 [ 20h 56m ] +INFO: Idle: 622, Running: 4, Completed: 15394 [ 20h 56m ] +INFO: Idle: 621, Running: 4, Completed: 15395 [ 20h 56m ] +INFO: Idle: 618, Running: 4, Completed: 15398 [ 20h 56m ] +INFO: Idle: 617, Running: 4, Completed: 15399 [ 20h 56m ] +INFO: Idle: 614, Running: 4, Completed: 15402 [ 20h 57m ] +INFO: Idle: 613, Running: 4, Completed: 15403 [ 20h 57m ] +INFO: Idle: 610, Running: 4, Completed: 15406 [ 20h 57m ] +INFO: Idle: 609, Running: 4, Completed: 15407 [ 20h 57m ] +INFO: Idle: 606, Running: 4, Completed: 15410 [ 20h 57m ] +INFO: Idle: 605, Running: 4, Completed: 15411 [ 20h 57m ] +INFO: Idle: 603, Running: 4, Completed: 15413 [ 20h 57m ] +INFO: Idle: 602, Running: 4, Completed: 15414 [ 20h 58m ] +INFO: Idle: 600, Running: 4, Completed: 15416 [ 20h 58m ] +INFO: Idle: 598, Running: 4, Completed: 15418 [ 20h 58m ] +INFO: Idle: 594, Running: 4, Completed: 15422 [ 20h 58m ] +INFO: Idle: 592, Running: 4, Completed: 15424 [ 20h 59m ] +INFO: Idle: 590, Running: 4, Completed: 15426 [ 20h 59m ] +INFO: Idle: 586, Running: 4, Completed: 15430 [ 20h 59m ] +INFO: Idle: 583, Running: 4, Completed: 15433 [ 20h 59m ] +INFO: Idle: 582, Running: 4, Completed: 15434 [ 20h 59m ] +INFO: Idle: 579, Running: 4, Completed: 15437 [ 21h 0m ] +INFO: Idle: 578, Running: 4, Completed: 15438 [ 21h 0m ] +INFO: Idle: 577, Running: 4, Completed: 15439 [ 21h 0m ] +INFO: Idle: 574, Running: 4, Completed: 15442 [ 21h 0m ] +INFO: Idle: 573, Running: 4, Completed: 15443 [ 21h 0m ] +INFO: Idle: 570, Running: 4, Completed: 15446 [ 21h 0m ] +INFO: Idle: 569, Running: 4, Completed: 15447 [ 21h 1m ] +INFO: Idle: 567, Running: 4, Completed: 15449 [ 21h 1m ] +INFO: Idle: 566, Running: 4, Completed: 15450 [ 21h 1m ] +INFO: Idle: 565, Running: 4, Completed: 15451 [ 21h 1m ] +INFO: Idle: 563, Running: 4, Completed: 15453 [ 21h 1m ] +INFO: Idle: 561, Running: 4, Completed: 15455 [ 21h 1m ] +INFO: Idle: 560, Running: 4, Completed: 15456 [ 21h 1m ] +INFO: Idle: 558, Running: 4, Completed: 15458 [ 21h 1m ] +INFO: Idle: 556, Running: 4, Completed: 15460 [ 21h 1m ] +INFO: Idle: 555, Running: 4, Completed: 15461 [ 21h 2m ] +INFO: Idle: 554, Running: 4, Completed: 15462 [ 21h 2m ] +INFO: Idle: 552, Running: 4, Completed: 15464 [ 21h 2m ] +INFO: Idle: 550, Running: 4, Completed: 15466 [ 21h 2m ] +INFO: Idle: 549, Running: 4, Completed: 15467 [ 21h 2m ] +INFO: Idle: 546, Running: 4, Completed: 15470 [ 21h 2m ] +INFO: Idle: 545, Running: 4, Completed: 15471 [ 21h 2m ] +INFO: Idle: 542, Running: 4, Completed: 15474 [ 21h 3m ] +INFO: Idle: 539, Running: 4, Completed: 15477 [ 21h 3m ] +INFO: Idle: 538, Running: 4, Completed: 15478 [ 21h 3m ] +INFO: Idle: 535, Running: 4, Completed: 15481 [ 21h 3m ] +INFO: Idle: 534, Running: 4, Completed: 15482 [ 21h 3m ] +INFO: Idle: 532, Running: 4, Completed: 15484 [ 21h 3m ] +INFO: Idle: 531, Running: 4, Completed: 15485 [ 21h 4m ] +INFO: Idle: 530, Running: 4, Completed: 15486 [ 21h 4m ] +INFO: Idle: 529, Running: 4, Completed: 15487 [ 21h 4m ] +INFO: Idle: 527, Running: 4, Completed: 15489 [ 21h 4m ] +INFO: Idle: 526, Running: 4, Completed: 15490 [ 21h 4m ] +INFO: Idle: 523, Running: 4, Completed: 15493 [ 21h 4m ] +INFO: Idle: 522, Running: 4, Completed: 15494 [ 21h 4m ] +INFO: Idle: 521, Running: 4, Completed: 15495 [ 21h 4m ] +INFO: Idle: 518, Running: 4, Completed: 15498 [ 21h 5m ] +INFO: Idle: 515, Running: 4, Completed: 15501 [ 21h 5m ] +INFO: Idle: 514, Running: 4, Completed: 15502 [ 21h 5m ] +INFO: Idle: 513, Running: 4, Completed: 15503 [ 21h 5m ] +INFO: Idle: 511, Running: 4, Completed: 15505 [ 21h 5m ] +INFO: Idle: 510, Running: 4, Completed: 15506 [ 21h 5m ] +INFO: Idle: 509, Running: 4, Completed: 15507 [ 21h 5m ] +INFO: Idle: 507, Running: 4, Completed: 15509 [ 21h 5m ] +INFO: Idle: 506, Running: 4, Completed: 15510 [ 21h 6m ] +INFO: Idle: 504, Running: 4, Completed: 15512 [ 21h 6m ] +INFO: Idle: 502, Running: 4, Completed: 15514 [ 21h 6m ] +INFO: Idle: 500, Running: 4, Completed: 15516 [ 21h 6m ] +INFO: Idle: 498, Running: 4, Completed: 15518 [ 21h 6m ] +INFO: Idle: 496, Running: 4, Completed: 15520 [ 21h 6m ] +INFO: Idle: 495, Running: 4, Completed: 15521 [ 21h 6m ] +INFO: Idle: 494, Running: 4, Completed: 15522 [ 21h 7m ] +INFO: Idle: 492, Running: 4, Completed: 15524 [ 21h 7m ] +INFO: Idle: 491, Running: 4, Completed: 15525 [ 21h 7m ] +INFO: Idle: 490, Running: 4, Completed: 15526 [ 21h 7m ] +INFO: Idle: 487, Running: 4, Completed: 15529 [ 21h 7m ] +INFO: Idle: 486, Running: 4, Completed: 15530 [ 21h 7m ] +INFO: Idle: 485, Running: 4, Completed: 15531 [ 21h 7m ] +INFO: Idle: 482, Running: 4, Completed: 15534 [ 21h 7m ] +INFO: Idle: 480, Running: 4, Completed: 15536 [ 21h 7m ] +INFO: Idle: 478, Running: 4, Completed: 15538 [ 21h 8m ] +INFO: Idle: 475, Running: 4, Completed: 15541 [ 21h 8m ] +INFO: Idle: 474, Running: 4, Completed: 15542 [ 21h 8m ] +INFO: Idle: 471, Running: 4, Completed: 15545 [ 21h 8m ] +INFO: Idle: 470, Running: 4, Completed: 15546 [ 21h 8m ] +INFO: Idle: 469, Running: 4, Completed: 15547 [ 21h 8m ] +INFO: Idle: 467, Running: 4, Completed: 15549 [ 21h 8m ] +INFO: Idle: 465, Running: 4, Completed: 15551 [ 21h 9m ] +INFO: Idle: 461, Running: 4, Completed: 15555 [ 21h 9m ] +INFO: Idle: 458, Running: 4, Completed: 15558 [ 21h 9m ] +INFO: Idle: 457, Running: 4, Completed: 15559 [ 21h 9m ] +INFO: Idle: 455, Running: 4, Completed: 15561 [ 21h 9m ] +INFO: Idle: 454, Running: 4, Completed: 15562 [ 21h 9m ] +INFO: Idle: 452, Running: 4, Completed: 15564 [ 21h 9m ] +INFO: Idle: 450, Running: 4, Completed: 15566 [ 21h 10m ] +INFO: Idle: 448, Running: 4, Completed: 15568 [ 21h 10m ] +INFO: Idle: 447, Running: 4, Completed: 15569 [ 21h 10m ] +INFO: Idle: 445, Running: 4, Completed: 15571 [ 21h 10m ] +INFO: Idle: 444, Running: 4, Completed: 15572 [ 21h 10m ] +INFO: Idle: 443, Running: 4, Completed: 15573 [ 21h 10m ] +INFO: Idle: 439, Running: 4, Completed: 15577 [ 21h 10m ] +INFO: Idle: 438, Running: 4, Completed: 15578 [ 21h 11m ] +INFO: Idle: 435, Running: 4, Completed: 15581 [ 21h 11m ] +INFO: Idle: 434, Running: 4, Completed: 15582 [ 21h 11m ] +INFO: Idle: 432, Running: 4, Completed: 15584 [ 21h 11m ] +INFO: Idle: 431, Running: 4, Completed: 15585 [ 21h 11m ] +INFO: Idle: 430, Running: 4, Completed: 15586 [ 21h 11m ] +INFO: Idle: 428, Running: 4, Completed: 15588 [ 21h 11m ] +INFO: Idle: 427, Running: 4, Completed: 15589 [ 21h 11m ] +INFO: Idle: 425, Running: 4, Completed: 15591 [ 21h 11m ] +INFO: Idle: 423, Running: 4, Completed: 15593 [ 21h 12m ] +INFO: Idle: 422, Running: 4, Completed: 15594 [ 21h 12m ] +INFO: Idle: 420, Running: 4, Completed: 15596 [ 21h 12m ] +INFO: Idle: 419, Running: 4, Completed: 15597 [ 21h 12m ] +INFO: Idle: 418, Running: 4, Completed: 15598 [ 21h 12m ] +INFO: Idle: 415, Running: 4, Completed: 15601 [ 21h 12m ] +INFO: Idle: 414, Running: 4, Completed: 15602 [ 21h 12m ] +INFO: Idle: 413, Running: 4, Completed: 15603 [ 21h 12m ] +INFO: Idle: 412, Running: 4, Completed: 15604 [ 21h 13m ] +INFO: Idle: 410, Running: 4, Completed: 15606 [ 21h 13m ] +INFO: Idle: 407, Running: 4, Completed: 15609 [ 21h 13m ] +INFO: Idle: 406, Running: 4, Completed: 15610 [ 21h 13m ] +INFO: Idle: 404, Running: 4, Completed: 15612 [ 21h 13m ] +INFO: Idle: 402, Running: 4, Completed: 15614 [ 21h 14m ] +INFO: Idle: 400, Running: 4, Completed: 15616 [ 21h 14m ] +INFO: Idle: 399, Running: 4, Completed: 15617 [ 21h 14m ] +INFO: Idle: 398, Running: 4, Completed: 15618 [ 21h 14m ] +INFO: Idle: 394, Running: 4, Completed: 15622 [ 21h 14m ] +INFO: Idle: 391, Running: 4, Completed: 15625 [ 21h 14m ] +INFO: Idle: 390, Running: 4, Completed: 15626 [ 21h 15m ] +INFO: Idle: 387, Running: 4, Completed: 15629 [ 21h 15m ] +INFO: Idle: 386, Running: 4, Completed: 15630 [ 21h 15m ] +INFO: Idle: 385, Running: 4, Completed: 15631 [ 21h 15m ] +INFO: Idle: 383, Running: 4, Completed: 15633 [ 21h 15m ] +INFO: Idle: 382, Running: 4, Completed: 15634 [ 21h 15m ] +INFO: Idle: 381, Running: 4, Completed: 15635 [ 21h 15m ] +INFO: Idle: 378, Running: 4, Completed: 15638 [ 21h 16m ] +INFO: Idle: 377, Running: 4, Completed: 15639 [ 21h 16m ] +INFO: Idle: 376, Running: 4, Completed: 15640 [ 21h 16m ] +INFO: Idle: 374, Running: 4, Completed: 15642 [ 21h 16m ] +INFO: Idle: 373, Running: 4, Completed: 15643 [ 21h 16m ] +INFO: Idle: 372, Running: 4, Completed: 15644 [ 21h 16m ] +INFO: Idle: 369, Running: 4, Completed: 15647 [ 21h 16m ] +INFO: Idle: 368, Running: 4, Completed: 15648 [ 21h 16m ] +INFO: Idle: 366, Running: 4, Completed: 15650 [ 21h 16m ] +INFO: Idle: 365, Running: 4, Completed: 15651 [ 21h 16m ] +INFO: Idle: 362, Running: 4, Completed: 15654 [ 21h 17m ] +INFO: Idle: 361, Running: 4, Completed: 15655 [ 21h 17m ] +INFO: Idle: 360, Running: 4, Completed: 15656 [ 21h 17m ] +INFO: Idle: 359, Running: 4, Completed: 15657 [ 21h 17m ] +INFO: Idle: 357, Running: 4, Completed: 15659 [ 21h 17m ] +INFO: Idle: 356, Running: 4, Completed: 15660 [ 21h 17m ] +INFO: Idle: 353, Running: 4, Completed: 15663 [ 21h 17m ] +INFO: Idle: 352, Running: 4, Completed: 15664 [ 21h 17m ] +INFO: Idle: 349, Running: 4, Completed: 15667 [ 21h 18m ] +INFO: Idle: 346, Running: 4, Completed: 15670 [ 21h 18m ] +INFO: Idle: 345, Running: 4, Completed: 15671 [ 21h 18m ] +INFO: Idle: 343, Running: 4, Completed: 15673 [ 21h 18m ] +INFO: Idle: 341, Running: 4, Completed: 15675 [ 21h 18m ] +INFO: Idle: 337, Running: 4, Completed: 15679 [ 21h 18m ] +INFO: Idle: 335, Running: 4, Completed: 15681 [ 21h 19m ] +INFO: Idle: 334, Running: 4, Completed: 15682 [ 21h 19m ] +INFO: Idle: 333, Running: 4, Completed: 15683 [ 21h 19m ] +INFO: Idle: 332, Running: 4, Completed: 15684 [ 21h 19m ] +INFO: Idle: 330, Running: 4, Completed: 15686 [ 21h 19m ] +INFO: Idle: 328, Running: 4, Completed: 15688 [ 21h 19m ] +INFO: Idle: 327, Running: 4, Completed: 15689 [ 21h 19m ] +INFO: Idle: 325, Running: 4, Completed: 15691 [ 21h 19m ] +INFO: Idle: 323, Running: 4, Completed: 15693 [ 21h 19m ] +INFO: Idle: 321, Running: 4, Completed: 15695 [ 21h 20m ] +INFO: Idle: 319, Running: 4, Completed: 15697 [ 21h 20m ] +INFO: Idle: 316, Running: 4, Completed: 15700 [ 21h 20m ] +INFO: Idle: 315, Running: 4, Completed: 15701 [ 21h 20m ] +INFO: Idle: 311, Running: 4, Completed: 15705 [ 21h 20m ] +INFO: Idle: 309, Running: 4, Completed: 15707 [ 21h 20m ] +INFO: Idle: 308, Running: 4, Completed: 15708 [ 21h 21m ] +INFO: Idle: 307, Running: 4, Completed: 15709 [ 21h 21m ] +INFO: Idle: 306, Running: 4, Completed: 15710 [ 21h 21m ] +INFO: Idle: 305, Running: 4, Completed: 15711 [ 21h 21m ] +INFO: Idle: 303, Running: 4, Completed: 15713 [ 21h 21m ] +INFO: Idle: 302, Running: 4, Completed: 15714 [ 21h 21m ] +INFO: Idle: 298, Running: 4, Completed: 15718 [ 21h 21m ] +INFO: Idle: 296, Running: 4, Completed: 15720 [ 21h 21m ] +INFO: Idle: 294, Running: 4, Completed: 15722 [ 21h 22m ] +INFO: Idle: 293, Running: 4, Completed: 15723 [ 21h 22m ] +INFO: Idle: 291, Running: 4, Completed: 15725 [ 21h 22m ] +INFO: Idle: 289, Running: 4, Completed: 15727 [ 21h 22m ] +INFO: Idle: 288, Running: 4, Completed: 15728 [ 21h 22m ] +INFO: Idle: 285, Running: 4, Completed: 15731 [ 21h 22m ] +INFO: Idle: 282, Running: 4, Completed: 15734 [ 21h 22m ] +INFO: Idle: 281, Running: 4, Completed: 15735 [ 21h 22m ] +INFO: Idle: 277, Running: 4, Completed: 15739 [ 21h 23m ] +INFO: Idle: 275, Running: 4, Completed: 15741 [ 21h 23m ] +INFO: Idle: 273, Running: 4, Completed: 15743 [ 21h 23m ] +INFO: Idle: 271, Running: 4, Completed: 15745 [ 21h 23m ] +INFO: Idle: 269, Running: 4, Completed: 15747 [ 21h 23m ] +INFO: Idle: 267, Running: 4, Completed: 15749 [ 21h 23m ] +INFO: Idle: 266, Running: 4, Completed: 15750 [ 21h 24m ] +INFO: Idle: 264, Running: 4, Completed: 15752 [ 21h 24m ] +INFO: Idle: 262, Running: 4, Completed: 15754 [ 21h 24m ] +INFO: Idle: 260, Running: 4, Completed: 15756 [ 21h 24m ] +INFO: Idle: 259, Running: 4, Completed: 15757 [ 21h 24m ] +INFO: Idle: 257, Running: 4, Completed: 15759 [ 21h 24m ] +INFO: Idle: 255, Running: 4, Completed: 15761 [ 21h 24m ] +INFO: Idle: 253, Running: 4, Completed: 15763 [ 21h 24m ] +INFO: Idle: 251, Running: 4, Completed: 15765 [ 21h 25m ] +INFO: Idle: 249, Running: 4, Completed: 15767 [ 21h 25m ] +INFO: Idle: 247, Running: 4, Completed: 15769 [ 21h 25m ] +INFO: Idle: 246, Running: 4, Completed: 15770 [ 21h 25m ] +INFO: Idle: 244, Running: 4, Completed: 15772 [ 21h 25m ] +INFO: Idle: 243, Running: 4, Completed: 15773 [ 21h 25m ] +INFO: Idle: 239, Running: 4, Completed: 15777 [ 21h 26m ] +INFO: Idle: 236, Running: 4, Completed: 15780 [ 21h 26m ] +INFO: Idle: 235, Running: 4, Completed: 15781 [ 21h 26m ] +INFO: Idle: 232, Running: 4, Completed: 15784 [ 21h 26m ] +INFO: Idle: 231, Running: 4, Completed: 15785 [ 21h 26m ] +INFO: Idle: 230, Running: 4, Completed: 15786 [ 21h 26m ] +INFO: Idle: 227, Running: 4, Completed: 15789 [ 21h 27m ] +INFO: Idle: 224, Running: 4, Completed: 15792 [ 21h 27m ] +INFO: Idle: 223, Running: 4, Completed: 15793 [ 21h 27m ] +INFO: Idle: 220, Running: 4, Completed: 15796 [ 21h 27m ] +INFO: Idle: 219, Running: 4, Completed: 15797 [ 21h 27m ] +INFO: Idle: 216, Running: 4, Completed: 15800 [ 21h 27m ] +INFO: Idle: 215, Running: 4, Completed: 15801 [ 21h 28m ] +INFO: Idle: 211, Running: 4, Completed: 15805 [ 21h 28m ] +INFO: Idle: 209, Running: 4, Completed: 15807 [ 21h 28m ] +INFO: Idle: 207, Running: 4, Completed: 15809 [ 21h 28m ] +INFO: Idle: 203, Running: 4, Completed: 15813 [ 21h 29m ] +INFO: Idle: 200, Running: 4, Completed: 15816 [ 21h 29m ] +INFO: Idle: 199, Running: 4, Completed: 15817 [ 21h 29m ] +INFO: Idle: 196, Running: 4, Completed: 15820 [ 21h 29m ] +INFO: Idle: 195, Running: 4, Completed: 15821 [ 21h 29m ] +INFO: Idle: 194, Running: 4, Completed: 15822 [ 21h 29m ] +INFO: Idle: 191, Running: 4, Completed: 15825 [ 21h 30m ] +INFO: Idle: 189, Running: 4, Completed: 15827 [ 21h 30m ] +INFO: Idle: 187, Running: 4, Completed: 15829 [ 21h 30m ] +INFO: Idle: 184, Running: 4, Completed: 15832 [ 21h 30m ] +INFO: Idle: 183, Running: 4, Completed: 15833 [ 21h 30m ] +INFO: Idle: 179, Running: 4, Completed: 15837 [ 21h 30m ] +INFO: Idle: 175, Running: 4, Completed: 15841 [ 21h 31m ] +INFO: Idle: 172, Running: 4, Completed: 15844 [ 21h 31m ] +INFO: Idle: 171, Running: 4, Completed: 15845 [ 21h 31m ] +INFO: Idle: 168, Running: 4, Completed: 15848 [ 21h 31m ] +INFO: Idle: 167, Running: 4, Completed: 15849 [ 21h 31m ] +INFO: Idle: 166, Running: 4, Completed: 15850 [ 21h 31m ] +INFO: Idle: 165, Running: 4, Completed: 15851 [ 21h 32m ] +INFO: Idle: 163, Running: 4, Completed: 15853 [ 21h 32m ] +INFO: Idle: 160, Running: 4, Completed: 15856 [ 21h 32m ] +INFO: Idle: 159, Running: 4, Completed: 15857 [ 21h 32m ] +INFO: Idle: 156, Running: 4, Completed: 15860 [ 21h 32m ] +INFO: Idle: 155, Running: 4, Completed: 15861 [ 21h 32m ] +INFO: Idle: 152, Running: 4, Completed: 15864 [ 21h 32m ] +INFO: Idle: 151, Running: 4, Completed: 15865 [ 21h 33m ] +INFO: Idle: 148, Running: 4, Completed: 15868 [ 21h 33m ] +INFO: Idle: 147, Running: 4, Completed: 15869 [ 21h 33m ] +INFO: Idle: 144, Running: 4, Completed: 15872 [ 21h 33m ] +INFO: Idle: 143, Running: 4, Completed: 15873 [ 21h 33m ] +INFO: Idle: 142, Running: 4, Completed: 15874 [ 21h 33m ] +INFO: Idle: 139, Running: 4, Completed: 15877 [ 21h 34m ] +INFO: Idle: 138, Running: 4, Completed: 15878 [ 21h 34m ] +INFO: Idle: 135, Running: 4, Completed: 15881 [ 21h 34m ] +INFO: Idle: 134, Running: 4, Completed: 15882 [ 21h 34m ] +INFO: Idle: 132, Running: 4, Completed: 15884 [ 21h 34m ] +INFO: Idle: 131, Running: 4, Completed: 15885 [ 21h 34m ] +INFO: Idle: 130, Running: 4, Completed: 15886 [ 21h 34m ] +INFO: Idle: 128, Running: 4, Completed: 15888 [ 21h 35m ] +INFO: Idle: 127, Running: 4, Completed: 15889 [ 21h 35m ] +INFO: Idle: 125, Running: 4, Completed: 15891 [ 21h 35m ] +INFO: Idle: 124, Running: 4, Completed: 15892 [ 21h 35m ] +INFO: Idle: 123, Running: 4, Completed: 15893 [ 21h 35m ] +INFO: Idle: 122, Running: 4, Completed: 15894 [ 21h 35m ] +INFO: Idle: 121, Running: 4, Completed: 15895 [ 21h 35m ] +INFO: Idle: 119, Running: 4, Completed: 15897 [ 21h 35m ] +INFO: Idle: 117, Running: 4, Completed: 15899 [ 21h 36m ] +INFO: Idle: 115, Running: 4, Completed: 15901 [ 21h 36m ] +INFO: Idle: 113, Running: 4, Completed: 15903 [ 21h 36m ] +INFO: Idle: 112, Running: 4, Completed: 15904 [ 21h 36m ] +INFO: Idle: 111, Running: 4, Completed: 15905 [ 21h 36m ] +INFO: Idle: 110, Running: 4, Completed: 15906 [ 21h 36m ] +INFO: Idle: 109, Running: 4, Completed: 15907 [ 21h 36m ] +INFO: Idle: 107, Running: 4, Completed: 15909 [ 21h 36m ] +INFO: Idle: 103, Running: 4, Completed: 15913 [ 21h 37m ] +INFO: Idle: 101, Running: 4, Completed: 15915 [ 21h 37m ] +INFO: Idle: 99, Running: 4, Completed: 15917 [ 21h 37m ] +INFO: Idle: 96, Running: 4, Completed: 15920 [ 21h 37m ] +INFO: Idle: 95, Running: 4, Completed: 15921 [ 21h 38m ] +INFO: Idle: 93, Running: 4, Completed: 15923 [ 21h 38m ] +INFO: Idle: 92, Running: 4, Completed: 15924 [ 21h 38m ] +INFO: Idle: 91, Running: 4, Completed: 15925 [ 21h 38m ] +INFO: Idle: 89, Running: 4, Completed: 15927 [ 21h 38m ] +INFO: Idle: 88, Running: 4, Completed: 15928 [ 21h 38m ] +INFO: Idle: 87, Running: 4, Completed: 15929 [ 21h 38m ] +INFO: Idle: 84, Running: 4, Completed: 15932 [ 21h 38m ] +INFO: Idle: 83, Running: 4, Completed: 15933 [ 21h 39m ] +INFO: Idle: 80, Running: 4, Completed: 15936 [ 21h 39m ] +INFO: Idle: 79, Running: 4, Completed: 15937 [ 21h 39m ] +INFO: Idle: 76, Running: 4, Completed: 15940 [ 21h 39m ] +INFO: Idle: 75, Running: 4, Completed: 15941 [ 21h 39m ] +INFO: Idle: 71, Running: 4, Completed: 15945 [ 21h 39m ] +INFO: Idle: 67, Running: 4, Completed: 15949 [ 21h 40m ] +INFO: Idle: 63, Running: 4, Completed: 15953 [ 21h 40m ] +INFO: Idle: 60, Running: 4, Completed: 15956 [ 21h 40m ] +INFO: Idle: 59, Running: 4, Completed: 15957 [ 21h 40m ] +INFO: Idle: 56, Running: 4, Completed: 15960 [ 21h 40m ] +INFO: Idle: 55, Running: 4, Completed: 15961 [ 21h 41m ] +INFO: Idle: 52, Running: 4, Completed: 15964 [ 21h 41m ] +INFO: Idle: 51, Running: 4, Completed: 15965 [ 21h 41m ] +INFO: Idle: 50, Running: 4, Completed: 15966 [ 21h 41m ] +INFO: Idle: 48, Running: 4, Completed: 15968 [ 21h 41m ] +INFO: Idle: 47, Running: 4, Completed: 15969 [ 21h 41m ] +INFO: Idle: 45, Running: 4, Completed: 15971 [ 21h 41m ] +INFO: Idle: 43, Running: 4, Completed: 15973 [ 21h 42m ] +INFO: Idle: 41, Running: 4, Completed: 15975 [ 21h 42m ] +INFO: Idle: 40, Running: 4, Completed: 15976 [ 21h 42m ] +INFO: Idle: 39, Running: 4, Completed: 15977 [ 21h 42m ] +INFO: Idle: 38, Running: 4, Completed: 15978 [ 21h 42m ] +INFO: Idle: 36, Running: 4, Completed: 15980 [ 21h 42m ] +INFO: Idle: 35, Running: 4, Completed: 15981 [ 21h 42m ] +INFO: Idle: 32, Running: 4, Completed: 15984 [ 21h 42m ] +INFO: Idle: 31, Running: 4, Completed: 15985 [ 21h 43m ] +INFO: Idle: 30, Running: 4, Completed: 15986 [ 21h 43m ] +INFO: Idle: 29, Running: 4, Completed: 15987 [ 21h 43m ] +INFO: Idle: 27, Running: 4, Completed: 15989 [ 21h 43m ] +INFO: Idle: 23, Running: 4, Completed: 15993 [ 21h 43m ] +INFO: Idle: 20, Running: 4, Completed: 15996 [ 21h 43m ] +INFO: Idle: 19, Running: 4, Completed: 15997 [ 21h 44m ] +INFO: Idle: 16, Running: 4, Completed: 16000 [ 21h 44m ] +INFO: Idle: 15, Running: 4, Completed: 16001 [ 21h 44m ] +INFO: Idle: 14, Running: 4, Completed: 16002 [ 21h 44m ] +INFO: Idle: 11, Running: 4, Completed: 16005 [ 21h 44m ] +INFO: Idle: 8, Running: 4, Completed: 16008 [ 21h 44m ] +INFO: Idle: 7, Running: 4, Completed: 16009 [ 21h 45m ] +INFO: Idle: 4, Running: 4, Completed: 16012 [ 21h 45m ] +INFO: Idle: 3, Running: 4, Completed: 16013 [ 21h 45m ] +INFO: Idle: 0, Running: 4, Completed: 16016 [ 21h 45m ] +INFO: Idle: 0, Running: 3, Completed: 16017 [ 21h 45m ] +INFO: Idle: 0, Running: 0, Completed: 16020 [ 21h 45m ] +sum of cpu time of last step: 3 days,23h23m17s + === Results Summary for run: run_01 tag: tag_1 === + + Cross-section : 7748 +- 52.04 pb + Nb of events : 0 + +INFO: End survey +combine_events +INFO: Combining Events +combination of events done in 17.617920875549316 s +store_events +INFO: Storing parton level results +INFO: End Parton +decay_events -from_cards +create_gridpack +INFO: Creating gridpack +P0_gg_epemgguux +P0_gg_epemggddx +P0_gg_taptamgguux +P0_gg_taptamggddx +P0_gu_epemgggu +P0_gd_epemgggd +P0_gux_epemgggux +P0_gdx_epemgggdx +P0_gu_taptamgggu +P0_gd_taptamgggd +P0_gux_taptamgggux +P0_gdx_taptamgggdx +P0_uux_epemgggg +P0_ddx_epemgggg +P0_uux_taptamgggg +P0_ddx_taptamgggg +P0_gg_epemuuuxux +P0_gg_epemdddxdx +P0_gg_epemuduxdx +P0_gg_epemucuxcx +P0_gg_epemdsdxsx +P0_gg_taptamuuuxux +P0_gg_taptamdddxdx +P0_gg_taptamuduxdx +P0_gg_taptamucuxcx +P0_gg_taptamdsdxsx +P0_gu_epemguuux +P0_gd_epemgdddx +P0_gux_epemguuxux +P0_gdx_epemgddxdx +P0_gu_epemguddx +P0_gu_epemguccx +P0_gd_epemgudux +P0_gd_epemgdssx +P0_gux_epemgduxdx +P0_gux_epemgcuxcx +P0_gdx_epemguuxdx +P0_gdx_epemgsdxsx +P0_gu_taptamguuux +P0_gd_taptamgdddx +P0_gux_taptamguuxux +P0_gdx_taptamgddxdx +P0_gu_taptamguddx +P0_gu_taptamguccx +P0_gd_taptamgudux +P0_gd_taptamgdssx +P0_gux_taptamgduxdx +P0_gux_taptamgcuxcx +P0_gdx_taptamguuxdx +P0_gdx_taptamgsdxsx +P0_uu_epemgguu +P0_uux_epemgguux +P0_dd_epemggdd +P0_ddx_epemggddx +P0_uxux_epemgguxux +P0_dxdx_epemggdxdx +P0_ud_epemggud +P0_uc_epemgguc +P0_uux_epemggddx +P0_uux_epemggccx +P0_udx_epemggudx +P0_ucx_epemggucx +P0_dc_epemggdc +P0_ds_epemggds +P0_dux_epemggdux +P0_ddx_epemgguux +P0_ddx_epemggssx +P0_dsx_epemggdsx +P0_uxdx_epemgguxdx +P0_uxcx_epemgguxcx +P0_dxcx_epemggdxcx +P0_dxsx_epemggdxsx +P0_uu_taptamgguu +P0_uux_taptamgguux +P0_dd_taptamggdd +P0_ddx_taptamggddx +P0_uxux_taptamgguxux +P0_dxdx_taptamggdxdx +P0_ud_taptamggud +P0_uc_taptamgguc +P0_uux_taptamggddx +P0_uux_taptamggccx +P0_udx_taptamggudx +P0_ucx_taptamggucx +P0_dc_taptamggdc +P0_ds_taptamggds +P0_dux_taptamggdux +P0_ddx_taptamgguux +P0_ddx_taptamggssx +P0_dsx_taptamggdsx +P0_uxdx_taptamgguxdx +P0_uxcx_taptamgguxcx +P0_dxcx_taptamggdxcx +P0_dxsx_taptamggdxsx +P0_uu_epemuuuux +P0_uux_epemuuuxux +P0_dd_epemddddx +P0_ddx_epemdddxdx +P0_uxux_epemuuxuxux +P0_dxdx_epemddxdxdx +P0_uu_epemuuddx +P0_uu_epemuuccx +P0_ud_epemuudux +P0_ud_epemudddx +P0_uc_epemuucux +P0_uc_epemucccx +P0_uux_epemuduxdx +P0_uux_epemucuxcx +P0_uux_epemdddxdx +P0_uux_epemcccxcx +P0_udx_epemuuuxdx +P0_udx_epemuddxdx +P0_ucx_epemuuuxcx +P0_ucx_epemuccxcx +P0_dd_epemuddux +P0_dd_epemddssx +P0_dc_epemddcdx +P0_dc_epemdcccx +P0_ds_epemddsdx +P0_ds_epemdsssx +P0_dux_epemuduxux +P0_dux_epemdduxdx +P0_ddx_epemuuuxux +P0_ddx_epemuduxdx +P0_ddx_epemdcdxcx +P0_ddx_epemdsdxsx +P0_ddx_epemsssxsx +P0_dsx_epemdddxsx +P0_dsx_epemdssxsx +P0_ccx_epemucuxcx +P0_ccx_epemdcdxcx +P0_ssx_epemdsdxsx +P0_uxux_epemduxuxdx +P0_uxux_epemcuxuxcx +P0_uxdx_epemuuxuxdx +P0_uxdx_epemduxdxdx +P0_uxcx_epemuuxuxcx +P0_uxcx_epemcuxcxcx +P0_dxdx_epemuuxdxdx +P0_dxdx_epemsdxdxsx +P0_dxcx_epemddxdxcx +P0_dxcx_epemcdxcxcx +P0_dxsx_epemddxdxsx +P0_dxsx_epemsdxsxsx +P0_ud_epemudccx +P0_ud_epemudssx +P0_uc_epemudcdx +P0_uux_epemdcdxcx +P0_uux_epemdsdxsx +P0_uux_epemcscxsx +P0_udx_epemucdxcx +P0_udx_epemusdxsx +P0_ucx_epemuddxcx +P0_dc_epemudcux +P0_dc_epemdcssx +P0_ds_epemudsux +P0_ds_epemdsbbx +P0_dux_epemdcuxcx +P0_dux_epemdsuxsx +P0_ddx_epemucuxcx +P0_ddx_epemusuxsx +P0_ddx_epemsbsxbx +P0_dsx_epemuduxsx +P0_dsx_epemdbsxbx +P0_ccx_epemuduxdx +P0_ccx_epemdsdxsx +P0_ssx_epemuduxdx +P0_ssx_epemucuxcx +P0_ssx_epemdcdxcx +P0_ssx_epemdbdxbx +P0_bbx_epemdsdxsx +P0_uxdx_epemcuxdxcx +P0_uxdx_epemsuxdxsx +P0_uxcx_epemduxdxcx +P0_dxcx_epemuuxdxcx +P0_dxcx_epemsdxcxsx +P0_dxsx_epemuuxdxsx +P0_dxsx_epembdxsxbx +P0_uu_taptamuuuux +P0_uux_taptamuuuxux +P0_dd_taptamddddx +P0_ddx_taptamdddxdx +P0_uxux_taptamuuxuxux +P0_dxdx_taptamddxdxdx +P0_uu_taptamuuddx +P0_uu_taptamuuccx +P0_ud_taptamuudux +P0_ud_taptamudddx +P0_uc_taptamuucux +P0_uc_taptamucccx +P0_uux_taptamuduxdx +P0_uux_taptamucuxcx +P0_uux_taptamdddxdx +P0_uux_taptamcccxcx +P0_udx_taptamuuuxdx +P0_udx_taptamuddxdx +P0_ucx_taptamuuuxcx +P0_ucx_taptamuccxcx +P0_dd_taptamuddux +P0_dd_taptamddssx +P0_dc_taptamddcdx +P0_dc_taptamdcccx +P0_ds_taptamddsdx +P0_ds_taptamdsssx +P0_dux_taptamuduxux +P0_dux_taptamdduxdx +P0_ddx_taptamuuuxux +P0_ddx_taptamuduxdx +P0_ddx_taptamdcdxcx +P0_ddx_taptamdsdxsx +P0_ddx_taptamsssxsx +P0_dsx_taptamdddxsx +P0_dsx_taptamdssxsx +P0_ccx_taptamucuxcx +P0_ccx_taptamdcdxcx +P0_ssx_taptamdsdxsx +P0_uxux_taptamduxuxdx +P0_uxux_taptamcuxuxcx +P0_uxdx_taptamuuxuxdx +P0_uxdx_taptamduxdxdx +P0_uxcx_taptamuuxuxcx +P0_uxcx_taptamcuxcxcx +P0_dxdx_taptamuuxdxdx +P0_dxdx_taptamsdxdxsx +P0_dxcx_taptamddxdxcx +P0_dxcx_taptamcdxcxcx +P0_dxsx_taptamddxdxsx +P0_dxsx_taptamsdxsxsx +P0_ud_taptamudccx +P0_ud_taptamudssx +P0_uc_taptamudcdx +P0_uux_taptamdcdxcx +P0_uux_taptamdsdxsx +P0_uux_taptamcscxsx +P0_udx_taptamucdxcx +P0_udx_taptamusdxsx +P0_ucx_taptamuddxcx +P0_dc_taptamudcux +P0_dc_taptamdcssx +P0_ds_taptamudsux +P0_ds_taptamdsbbx +P0_dux_taptamdcuxcx +P0_dux_taptamdsuxsx +P0_ddx_taptamucuxcx +P0_ddx_taptamusuxsx +P0_ddx_taptamsbsxbx +P0_dsx_taptamuduxsx +P0_dsx_taptamdbsxbx +P0_ccx_taptamuduxdx +P0_ccx_taptamdsdxsx +P0_ssx_taptamuduxdx +P0_ssx_taptamucuxcx +P0_ssx_taptamdcdxcx +P0_ssx_taptamdbdxbx +P0_bbx_taptamdsdxsx +P0_uxdx_taptamcuxdxcx +P0_uxdx_taptamsuxdxsx +P0_uxcx_taptamduxdxcx +P0_dxcx_taptamuuxdxcx +P0_dxcx_taptamsdxcxsx +P0_dxsx_taptamuuxdxsx +P0_dxsx_taptambdxsxbx +Cleaning SubProcesses...................................................................................................................................................................................................................................................................... +INFO: gridpack created +quit +INFO: storing files of previous run +INFO: Done +INFO: +338283.01user 6236.99system 24:42:06elapsed 387%CPU (0avgtext+0avgdata 725876maxresident)k +91242848inputs+137969472outputs (2451major+1254319692minor)pagefaults 0swaps +END: Sat Sep 14 01:48:52 AM CEST 2024 +ELAPSED: 88928 seconds diff --git a/epochX/cudacpp/tlau/lauX.sh b/epochX/cudacpp/tlau/lauX.sh index 3a86b9221a..3129dc6c48 100755 --- a/epochX/cudacpp/tlau/lauX.sh +++ b/epochX/cudacpp/tlau/lauX.sh @@ -10,38 +10,101 @@ scrdir=$(cd $(dirname $0); pwd) function usage() { - echo "Usage: $0 - " + echo "Usage: $0 - [-nomakeclean] [-rndoff ] [-togridpack|-fromgridpack]" echo " (supported values: fortran, cuda, hip, cppnone, cppsse4, cppavx2, cpp512y, cpp512z)" + echo " (*special* value: ALL processes all available backends)" echo "Example: $0 -cppavx2 gg_tt.mad" exit 1 } bckend= proc= +grid= +nomakeclean= +rndoff=0 while [ "$1" != "" ]; do - if [ "$1" == "-fortran" ] || [ "$1" == "-cuda" ] || [ "$1" == "-hip" ] || [ "$1" == "-cppnone" ] || [ "$1" == "-cppsse4" ] || [ "$1" == "-cppavx2" ] || [ "$1" == "-cpp512y" ] || [ "$1" == "-cpp512z" ]; then + if [ "$1" == "-fortran" ] || [ "$1" == "-cuda" ] || [ "$1" == "-hip" ] || [ "$1" == "-cppnone" ] || [ "$1" == "-cppsse4" ] || [ "$1" == "-cppavx2" ] || [ "$1" == "-cpp512y" ] || [ "$1" == "-cpp512z" ] || [ "$1" == "-ALL" ]; then if [ "${bckend}" == "" ]; then bckend=${1/-/}; else echo "ERROR! Backend already set"; usage; fi + elif [ "$1" == "-togridpack" ] && [ "${grid}" == "" ]; then + grid=${1} + elif [ "$1" == "-fromgridpack" ] && [ "${grid}" == "" ]; then + grid=${1} + elif [ "$1" == "-nomakeclean" ]; then + nomakeclean=$1 + elif [ "$1" == "-rndoff" ]; then + rndoff=$2 + shift + elif [ "${1#-}" != "${1}" ]; then + echo "ERROR! Invalid option '$1'" + usage elif [ "${proc}" == "" ]; then proc=$1 else - echo "ERROR! Invalid option '$1': process directory already set to '${proc}'" + echo "ERROR! Invalid input '$1': process directory already set to '${proc}'" usage fi shift done if [ "${bckend}" == "" ]; then echo "ERROR! No backend was specified"; usage; fi -if [ "$proc" == "" ]; then echo "ERROR! No process directory was specified"; usage; fi +if [ "${proc}" == "" ]; then echo "ERROR! No process directory was specified"; usage; fi +if [ "${grid}" != "" ] && [ "${rndoff}" != "0" ]; then echo "ERROR! ${grid} and -rndoff are not compatible"; exit 1; fi + +if [ "${bckend}" == "ALL" ]; then + if [ "${grid}" == "-togridpack" ]; then echo "ERROR! ${grid} and -ALL are not compatible"; exit 1; fi # temporary? + for b in fortran cuda hip cppnone cppsse4 cppavx2 cpp512y cpp512z; do + $0 -${b} ${nomakeclean} ${proc} -rndoff ${rndoff} ${grid} + nomakeclean=-nomakeclean # respect user input only on the first test, then keep the builds + done + exit 0 # successful termination on each loop (and skip the rest of this file) +fi +gridpackdir=${scrdir}/gridpacks/${proc} +###echo gridpackdir=${gridpackdir}; exit 1 suff=.mad if [ "${proc}" == "${proc%${suff}}" ]; then echo "ERROR! Process directory does not end in '${suff}'"; usage; fi proc=${proc%${suff}} +resultsdir=${scrdir}/logs_${proc//_}_${bckend} +if [ "${grid}" == "-togridpack" ]; then + resultsdir=${gridpackdir} +elif [ "${grid}" == "-fromgridpack" ]; then + resultsdir=${gridpackdir/gridpacks/fromgridpacks}/${bckend} + rm -rf ${resultsdir}; mkdir -p ${resultsdir} +fi -cd $(dirname $0)/.. -echo "Execute $(basename $0) for process ${proc} and backend ${bckend} in directory $(pwd)" -procdir=$(pwd)/${proc}${suff} -if [ ! -d ${procdir} ]; then echo "ERROR! Process directory '${procdir}' does not exist"; usage; fi -cd ${procdir} -resultsdir=${scrdir}/logs_${proc//_}_${bckend/} +if [ "${rndoff}" == "x10" ]; then + for i in $(seq 0 9); do + $0 -${bckend} ${nomakeclean} ${proc} ${grid} -rndoff ${i} + nomakeclean=-nomakeclean # respect user input only on the first test, then keep the builds + done + more ${resultsdir}/*txt | \egrep '(Cross-section :)' + exit 0 # successful termination on each loop (and skip the rest of this file) +elif [ ${rndoff} -lt 0 ]; then + echo "ERROR! Invalid rndoff=${rndoff}" + exit 1 +fi + +outfile=output.txt +if [ "${rndoff}" != "0" ]; then outfile=output${rndoff}.txt; fi + +function exit0() +{ + echo "" + echo "********************************************************************************" + echo "" + exit 0 +} + +if [ "${bckend}" == "cuda" ]; then + if ! nvidia-smi -L > /dev/null 2>&1; then + echo "WARNING! No NVidia GPU was found: skip backend ${bckend}" + exit0 + fi +elif [ "${bckend}" == "hip" ]; then + if ! rocm-smi -i > /dev/null 2>&1; then + echo "WARNING! No AMD GPU was found: skip backend ${bckend}" + exit0 + fi +fi function getnevt() { @@ -55,14 +118,109 @@ function getnevt() nevt=100 else echo "WARNING! Unknown process ${proc}" > /dev/stderr - nevt=100 + nevt=500 fi echo $nevt } +# Get the random seed +(( rndseed = 21 + ${rndoff} )) + +# Get the number of unweighted events to generate +nevt=$(getnevt) + +function showcpugpu() +{ + unames=$(uname -s) + unamep=$(uname -p) + if nvidia-smi -L > /dev/null 2>&1; then + gpuTxt="$(nvidia-smi -L | wc -l)x $(nvidia-smi -L | awk '{print $3,$4}' | sort -u)" + elif rocm-smi -i > /dev/null 2>&1; then + gpuTxt="$(rocm-smi --showproductname | grep 'Card series' | awk '{print $5,$6,$7}')" + else + gpuTxt=none + fi + if [ "${unames}" == "Darwin" ]; then + cpuTxt=$(sysctl -h machdep.cpu.brand_string) + cpuTxt=${cpuTxt/machdep.cpu.brand_string: } + elif [ "${unamep}" == "ppc64le" ]; then + cpuTxt=$(cat /proc/cpuinfo | grep ^machine | awk '{print substr($0,index($0,"Power"))", "}')$(cat /proc/cpuinfo | grep ^cpu | head -1 | awk '{print substr($0,index($0,"POWER"))}') + else + cpuTxt=$(cat /proc/cpuinfo | grep '^model name' |& head -1 | awk '{i0=index($0,"Intel"); if (i0==0) i0=index($0,"AMD"); i1=index($0," @"); if (i1>0) {print substr($0,i0,i1-i0)} else {print substr($0,i0)}}') + fi + cpuTxt="${cpuTxt} (nproc=$(nproc))" + echo -e "On $HOSTNAME [CPU: $cpuTxt] [GPU: $gpuTxt]:" +} + +# --- OPTION 2: RUN FROM GRIDPACK --- +if [ "${grid}" == "-fromgridpack" ]; then + echo "Execute $(basename $0) for process ${proc} and backend ${bckend} from gridpack directory $(pwd)" + if [ ! -d ${gridpackdir} ]; then echo "ERROR! Gridpack directory '${gridpackdir}' does not exist"; usage; fi + # Untar the gridpack (NB: this is faster than doing an rsync of an untarred backup!) + set -x + cd ${gridpackdir} + rm -rf madevent run.sh events.lhe* + tar -xzf run_01_gridpack.tar.gz + set +x + # Configure gridpack patches + dir=madevent/bin/internal + pushd $dir >& /dev/null + echo "INFO: configure gridpack patches in ${dir}" + mv madevent_interface.py madevent_interface.py.BKP + mv gen_ximprove.py gen_ximprove.py.BKP + mv cluster.py cluster.py.BKP + \cp ../../../../MG5aMC_patches/madevent_interface.py . + \cp ../../../../MG5aMC_patches/gen_ximprove.py . + \cp ../../../../MG5aMC_patches/cluster.py . + popd >& /dev/null + # Configure the appropriate backend + for dir in madevent/SubProcesses/P*; do + pushd $dir >& /dev/null + echo "INFO: redefine madevent symlink for backend ${bckend} in ${dir}" + if [ "${bckend}" == "fortran" ]; then + exe=$(\ls madevent_fortran) + else + exe=$(\ls build.*${bckend/cpp/}*/madevent_*) + fi + if [ "$(echo ${exe} | wc -w)" == "0" ]; then + echo "ERROR! No madevent executable found for backend $bckend"; exit 1 + elif [ "$(echo ${exe} | wc -w)" != "1" ]; then + echo "ERROR! Too many madevent executables found for backend $bckend:"; echo "$exe"; exit 1 + fi + \rm -f madevent + ln -sf $exe madevent + echo "----> madevent symlink will now point to $exe" + popd >& /dev/null + done + # Run the test for the appropriate backend + START=$(date +%s) + echo "START: $(date)" |& tee ${resultsdir}/${outfile} + showcpugpu |& tee -a ${resultsdir}/${outfile} + if [ -v CUDACPP_RUNTIME_DISABLEFPE ]; then echo CUDACPP_RUNTIME_DISABLEFPE is set |& tee -a ${resultsdir}/${outfile}; else echo CUDACPP_RUNTIME_DISABLEFPE is not set |& tee -a ${resultsdir}/${outfile}; fi # temporary? (debug FPEs in CMS DY #942) + ls -l madevent/SubProcesses/P*/madevent |& tee -a ${resultsdir}/${outfile} + ./run.sh ${nevt} ${rndseed} |& tee -a ${resultsdir}/${outfile} + mv events* ${resultsdir} + echo "END: $(date)" |& tee -a ${resultsdir}/${outfile} + END=$(date +%s) + echo "ELAPSED: $((END-START)) seconds" |& tee -a ${resultsdir}/${outfile} + exit0 +fi + +# --- OPTION 1: RUN FROM SOURCE CODE (AND OPTIONALLY CREATE A GRIDPACK) --- +cd $(dirname $0)/.. +echo "Execute $(basename $0) for process ${proc} and backend ${bckend} in directory $(pwd)" +procdir=$(pwd)/${proc}${suff} +if [ ! -d ${procdir} ]; then echo "ERROR! Process directory '${procdir}' does not exist"; usage; fi +cd ${procdir} + function lauX_makeclean() { - for d in SubProcesses/P*; do cd $d; make cleanall; cd -; break; done + if [ "${nomakeclean}" == "" ]; then + echo "INFO: clean all builds" + for d in SubProcesses/P*; do cd $d; make cleanall > /dev/null; cd - > /dev/null; break; done + else + echo "WARNING! Keep all builds (-nomakeclean option was specified)" + fi } function lauX_cleanup() @@ -71,29 +229,49 @@ function lauX_cleanup() rm -f SubProcesses/results.dat rm -rf Events HTML; mkdir Events HTML; touch Events/.keep HTML/.keep for d in SubProcesses/P*; do cd $d; rm -rf gensym input_app.txt symfact.dat G[0-9]* ajob[0-9]*; cd - > /dev/null; done + if [ "${grid}" == "-togridpack" ]; then + rm -rf bin/TheChopper-pl + rm -rf bin/clean4grid + rm -rf bin/compile + rm -rf bin/gridrun + rm -rf bin/internal/gen_ximprove + rm -rf bin/refine4grid + rm -rf bin/replace.pl + rm -rf bin/run.sh + fi } # Clean builds before launch -lauX_makeclean >& /dev/null +lauX_makeclean + +# Back up config before launch +cp SubProcesses/randinit SubProcesses/randinit.BKP # save the initial file +cp Cards/run_card.dat Cards/run_card.dat.BKP # save the initial file +cp Cards/grid_card.dat Cards/grid_card.dat.BKP # save the initial file +cp Source/run_card.inc Source/run_card.inc.BKP # save the initial file +cp bin/internal/gen_ximprove.py bin_internal_gen_ximprove.py.BKP # save the initial file +cp bin/internal/madevent_interface.py bin_internal_madevent_interface.py.BKP # save the initial file +cp Source/make_opts Source/make_opts.BKP # save the initial file +cp Source/param_card.inc Source/param_card.inc.BKP # save the initial file # Clean config before launch -rm -rf ${resultsdir}; mkdir ${resultsdir} +# (NB: "just in case" actions below should normally keep the defaults of generated code in the repo?) +# (NB: but some small differences have been observed, e.g. "False = gridpack" vs "False = gridpack") +if [ "${rndoff}" == "0" ]; then rm -rf ${resultsdir}; mkdir ${resultsdir}; fi lauX_cleanup rm -f SubProcesses/ME5_debug -echo "r=21" > SubProcesses/randinit # just in case a previous test was not cleaned up -cp SubProcesses/randinit SubProcesses/randinit.BKP # save the initial file +echo "r=21" > SubProcesses/randinit # just in case sed -i "s/.* = nevents/ 10000 = nevents/" Cards/run_card.dat # just in case sed -i "s/.* = cudacpp_backend/ cpp = cudacpp_backend/" Cards/run_card.dat # just in case -cp Cards/run_card.dat Cards/run_card.dat.BKP # save the initial file +sed -i "s/.* = cudacpp_bldall/ False = cudacpp_bldall/" Cards/run_card.dat # just in case +sed -i "s/.* = gridpack/ False = gridpack/" Cards/run_card.dat # just in case sed -i "s/ NEVENTS = .*/ NEVENTS = 10000/" Source/run_card.inc # just in case -cp Source/run_card.inc Source/run_card.inc.BKP # save the initial file sed -i "s/8192 1 1/%(event)s %(maxiter)s %(miniter)s/" bin/internal/gen_ximprove.py # just in case -cp bin/internal/gen_ximprove.py bin/internal/gen_ximprove.py.BKP # save the initial file sed -i "s/'int', 8192,'Number of points/'int', 1000,'Number of points/" bin/internal/madevent_interface.py # just in case sed -i "s/'int', 1, 'Number of iterations'/'int', 5, 'Number of iterations'/" bin/internal/madevent_interface.py # just in case -cp bin/internal/madevent_interface.py bin/internal/madevent_interface.py.BKP # save the initial file -cp Source/make_opts Source/make_opts.BKP # save the initial file -cp Source/param_card.inc Source/param_card.inc.BKP # save the initial file + +# Set the random seed +echo "r=${rndseed}" > SubProcesses/randinit # just in case a previous test was not cleaned up # Set the number of events and iterations in the survey step sed -i "s/'int', 1000,'Number of points/'int', 8192,'Number of points/" bin/internal/madevent_interface.py @@ -103,44 +281,57 @@ sed -i "s/'int', 5, 'Number of iterations'/'int', 1, 'Number of iterations'/" bi sed -i "s/%(event)s %(maxiter)s %(miniter)s/8192 1 1/" bin/internal/gen_ximprove.py # Set the number of unweighted events in run_card.dat -nevt=$(getnevt) sed -i "s/ 10000 = nevents/ ${nevt} = nevents/" Cards/run_card.dat # Set the backend in run_card.dat sed -i "s/ cpp = cudacpp_backend/${bckend} = cudacpp_backend/" Cards/run_card.dat +# Set gridpack mode in run_card.dat +if [ "${grid}" == "-togridpack" ]; then sed -i "s/.* = gridpack/ True = gridpack/" Cards/run_card.dat; fi + +# Configure bldall in run_card.dat +sed -i "s/.* = cudacpp_bldall/ True = cudacpp_bldall/" Cards/run_card.dat + # Launch (generate_events) # (BUG #683: generate_events does not return an error code even if it fails) ###set -x # verbose START=$(date +%s) -echo "START: $(date)" |& tee ${resultsdir}/output.txt -MG5AMC_CARD_PATH=$(pwd)/Cards time ./bin/generate_events -f |& tee -a ${resultsdir}/output.txt -echo "END: $(date)" |& tee -a ${resultsdir}/output.txt +echo "START: $(date)" |& tee ${resultsdir}/${outfile} +showcpugpu |& tee -a ${resultsdir}/${outfile} +if [ -v CUDACPP_RUNTIME_DISABLEFPE ]; then echo CUDACPP_RUNTIME_DISABLEFPE is set |& tee -a ${resultsdir}/${outfile}; else echo CUDACPP_RUNTIME_DISABLEFPE is not set |& tee -a ${resultsdir}/${outfile}; fi # temporary? (debug FPEs in CMS DY #942) +MG5AMC_CARD_PATH=$(pwd)/Cards time ./bin/generate_events -f |& tee -a ${resultsdir}/${outfile} +echo "END: $(date)" |& tee -a ${resultsdir}/${outfile} END=$(date +%s) -echo "ELAPSED: $((END-START)) seconds" |& tee -a ${resultsdir}/output.txt +echo "ELAPSED: $((END-START)) seconds" |& tee -a ${resultsdir}/${outfile} ###set +x # not verbose -# Process and keep results -\rm HTML/results.pkl -mv Events ${resultsdir}; mv HTML ${resultsdir} -gunzip ${resultsdir}/Events/run_01/unweighted_events.lhe.gz +# Copy output gridpack to tlau/gridpacks directory +if [ "${grid}" == "-togridpack" ]; then + mv run_01_gridpack.tar.gz ${gridpackdir}/run_01_gridpack.tar.gz + echo "Gridpack created: ${gridpackdir}/run_01_gridpack.tar.gz" +fi -# FIXME! No need to keep events in git, there is no lhe file comparison yet anyway (20-DEC-2023) -\rm ${resultsdir}/Events/run_01/unweighted_events.lhe -\rm ${resultsdir}/Events/run_01/run_01_tag_1_banner.txt -touch ${resultsdir}/Events/run_01/.keep +# Process and keep results (only for the default rndoff) +if [ "${grid}" != "-togridpack" ] && [ "${rndoff}" == "0" ]; then + \rm HTML/results.pkl + mv Events ${resultsdir}; mv HTML ${resultsdir} + gunzip ${resultsdir}/Events/run_01/unweighted_events.lhe.gz + # FIXME! No need to keep events in git, there is no lhe file comparison yet anyway (20-DEC-2023) + \rm ${resultsdir}/Events/run_01/unweighted_events.lhe + \rm ${resultsdir}/Events/run_01/run_01_tag_1_banner.txt + touch ${resultsdir}/Events/run_01/.keep +fi # Clean config after launch lauX_cleanup mv SubProcesses/randinit.BKP SubProcesses/randinit # restore the initial file mv Cards/run_card.dat.BKP Cards/run_card.dat # restore the initial file +mv Cards/grid_card.dat.BKP Cards/grid_card.dat # restore the initial file mv Source/run_card.inc.BKP Source/run_card.inc # restore the initial file -mv bin/internal/gen_ximprove.py.BKP bin/internal/gen_ximprove.py # restore the initial file -mv bin/internal/madevent_interface.py.BKP bin/internal/madevent_interface.py # restore the initial file +mv bin_internal_gen_ximprove.py.BKP bin/internal/gen_ximprove.py # restore the initial file +mv bin_internal_madevent_interface.py.BKP bin/internal/madevent_interface.py # restore the initial file mv Source/make_opts.BKP Source/make_opts # restore the initial file mv Source/param_card.inc.BKP Source/param_card.inc # restore the initial file -# Add an 80-character separator -echo "" -echo "********************************************************************************" -echo "" +# Add an 80-character separator and exit +exit0 diff --git a/epochX/cudacpp/tlau/logs_ggtt_cpp512y/HTML/run_01/results.html b/epochX/cudacpp/tlau/logs_ggtt_cpp512y/HTML/run_01/results.html index 531abb52e8..33486e3031 100644 --- a/epochX/cudacpp/tlau/logs_ggtt_cpp512y/HTML/run_01/results.html +++ b/epochX/cudacpp/tlau/logs_ggtt_cpp512y/HTML/run_01/results.html @@ -21,7 +21,7 @@
-      s= 440.4 ± 0.445 (pb)

+      s= 440.12 ± 0.547 (pb)

@@ -32,10 +32,10 @@ - - - - + + + + @@ -45,7 +45,7 @@

-      s= 440.4 ± 0.445 (pb)

+      s= 440.12 ± 0.547 (pb)

Graph Cross-Section
/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx 440.4 0.445 638.976 21258.0 440.1 0.547 557.056 23147.0 0
@@ -56,11 +56,11 @@ - - - - - + + + + + diff --git a/epochX/cudacpp/tlau/logs_ggtt_cpp512y/output.txt b/epochX/cudacpp/tlau/logs_ggtt_cpp512y/output.txt index 4d95da7367..282a4308a8 100644 --- a/epochX/cudacpp/tlau/logs_ggtt_cpp512y/output.txt +++ b/epochX/cudacpp/tlau/logs_ggtt_cpp512y/output.txt @@ -1,4 +1,4 @@ -START: Wed Dec 20 12:28:02 AM CET 2023 +START: Mon Aug 5 04:12:46 PM CEST 2024 ************************************************************ * * * W E L C O M E to * @@ -11,7 +11,7 @@ START: Wed Dec 20 12:28:02 AM CET 2023 * * * * * * * * * * * * -* VERSION 3.5.2_lo_vect * +* VERSION 3.5.3_lo_vect * * * * The MadGraph5_aMC@NLO Development Team - Find us at * * https://server06.fynu.ucl.ac.be/projects/madgraph * @@ -34,34 +34,36 @@ INFO: Running Survey Creating Jobs Working on SubProcesses INFO: P1_gg_ttx -INFO: Building madevent in madevent_interface.py with 'cpp512y' matrix elements -INFO: Idle: 1, Running: 0, Completed: 0 [ current time: 00h28 ] -INFO: Idle: 0, Running: 0, Completed: 1 [ 1.8s ] -INFO: Idle: 0, Running: 0, Completed: 1 [ 1.8s ] +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cpp512y' matrix elements +INFO: Idle: 1, Running: 0, Completed: 0 [ current time: 16h12 ] +INFO: Idle: 0, Running: 0, Completed: 1 [ 2.7s ] +INFO: Idle: 0, Running: 0, Completed: 1 [ 2.7s ] INFO: End survey refine 10000 Creating Jobs INFO: Refine results to 10000 INFO: Generating 10000.0 unweighted events. -sum of cpu time of last step: 2 seconds -INFO: Effective Luminosity 27.040181710021088 pb^-1 -INFO: need to improve 2 channels -- Current estimate of cross-section: 443.78400000000005 +- 4.146007168348844 +sum of cpu time of last step: 3 seconds +INFO: Effective Luminosity 27.13133601481371 pb^-1 +INFO: need to improve 1 channels +- Current estimate of cross-section: 442.293 +- 4.035736223404597 P1_gg_ttx +Pre-building madevent in madevent_interface.py with ALL matrix elements Building madevent in madevent_interface.py with 'cpp512y' matrix elements -INFO: Idle: 9, Running: 4, Completed: 0 [ current time: 00h28 ] -INFO: Idle: 4, Running: 4, Completed: 5 [ 2.6s ] -INFO: Idle: 0, Running: 0, Completed: 13 [ 5.2s ] +INFO: Idle: 7, Running: 4, Completed: 0 [ current time: 16h12 ] +INFO: Idle: 2, Running: 4, Completed: 5 [ 2.9s ] +INFO: Idle: 0, Running: 0, Completed: 11 [ 4.3s ] INFO: Combining runs -sum of cpu time of last step: 22 seconds +sum of cpu time of last step: 21 seconds INFO: finish refine refine 10000 --treshold=0.9 No need for second refine due to stability of cross-section INFO: Combining Events -combination of events done in 2.0271809101104736 s +combination of events done in 2.16316819190979 s === Results Summary for run: run_01 tag: tag_1 === - Cross-section : 440.4 +- 0.4447 pb + Cross-section : 440.1 +- 0.547 pb Nb of events : 10000 INFO: No version of lhapdf. Can not run systematics computation @@ -74,7 +76,7 @@ INFO: storing files of previous run INFO: Done quit INFO: -38.15user 4.10system 0:17.72elapsed 238%CPU (0avgtext+0avgdata 95444maxresident)k -0inputs+422576outputs (31major+852227minor)pagefaults 0swaps -END: Wed Dec 20 12:28:20 AM CET 2023 -ELAPSED: 18 seconds +24.32user 2.02system 0:12.74elapsed 206%CPU (0avgtext+0avgdata 95356maxresident)k +0inputs+403112outputs (2major+357321minor)pagefaults 0swaps +END: Mon Aug 5 04:12:59 PM CEST 2024 +ELAPSED: 13 seconds diff --git a/epochX/cudacpp/tlau/logs_ggtt_cpp512z/Events/.keep b/epochX/cudacpp/tlau/logs_ggtt_cpp512z/Events/.keep new file mode 100644 index 0000000000..e69de29bb2 diff --git a/epochX/cudacpp/tlau/logs_ggtt_cpp512z/HTML/.keep b/epochX/cudacpp/tlau/logs_ggtt_cpp512z/HTML/.keep new file mode 100644 index 0000000000..e69de29bb2 diff --git a/epochX/cudacpp/tlau/logs_ggtt_cpp512z/output.txt b/epochX/cudacpp/tlau/logs_ggtt_cpp512z/output.txt new file mode 100644 index 0000000000..fa39815a0c --- /dev/null +++ b/epochX/cudacpp/tlau/logs_ggtt_cpp512z/output.txt @@ -0,0 +1,82 @@ +START: Mon Aug 5 04:12:59 PM CEST 2024 +************************************************************ +* * +* W E L C O M E to * +* M A D G R A P H 5 _ a M C @ N L O * +* M A D E V E N T * +* * +* * * * +* * * * * * +* * * * * 5 * * * * * +* * * * * * +* * * * +* * +* VERSION 3.5.3_lo_vect * +* * +* The MadGraph5_aMC@NLO Development Team - Find us at * +* https://server06.fynu.ucl.ac.be/projects/madgraph * +* * +* Type 'help' for in-line help. * +* * +************************************************************ +INFO: load configuration from /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/Cards/me5_configuration.txt +INFO: load configuration from /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/Cards/me5_configuration.txt +Using default text editor "vi". Set another one in ./input/mg5_configuration.txt +Using default eps viewer "evince". Set another one in ./input/mg5_configuration.txt +Using default web browser "firefox". Set another one in ./input/mg5_configuration.txt +generate_events -f +Generating 10000 events with run name run_01 +survey run_01 +INFO: compile directory +compile Source Directory +Using random number seed offset = 21 +INFO: Running Survey +Creating Jobs +Working on SubProcesses +INFO: P1_gg_ttx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cpp512z' matrix elements +INFO: Idle: 1, Running: 0, Completed: 0 [ current time: 16h13 ] +INFO: Idle: 0, Running: 0, Completed: 1 [ 2.8s ] +INFO: Idle: 0, Running: 0, Completed: 1 [ 2.8s ] +INFO: End survey +refine 10000 +Creating Jobs +INFO: Refine results to 10000 +INFO: Generating 10000.0 unweighted events. +sum of cpu time of last step: 3 seconds +INFO: Effective Luminosity 27.13133601481371 pb^-1 +INFO: need to improve 1 channels +- Current estimate of cross-section: 442.293 +- 4.035736223404597 + P1_gg_ttx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cpp512z' matrix elements +INFO: Idle: 7, Running: 4, Completed: 0 [ current time: 16h13 ] +INFO: Idle: 2, Running: 4, Completed: 5 [ 3.1s ] +INFO: Idle: 0, Running: 0, Completed: 11 [ 4.6s ] +INFO: Combining runs +sum of cpu time of last step: 22 seconds +INFO: finish refine +refine 10000 --treshold=0.9 +No need for second refine due to stability of cross-section +INFO: Combining Events +combination of events done in 2.1579713821411133 s + === Results Summary for run: run_01 tag: tag_1 === + + Cross-section : 440.1 +- 0.547 pb + Nb of events : 10000 + +INFO: No version of lhapdf. Can not run systematics computation +store_events +INFO: Storing parton level results +INFO: End Parton +reweight -from_cards +decay_events -from_cards +INFO: storing files of previous run +INFO: Done +quit +INFO: +25.13user 2.04system 0:12.40elapsed 219%CPU (0avgtext+0avgdata 95372maxresident)k +0inputs+403112outputs (0major+356548minor)pagefaults 0swaps +END: Mon Aug 5 04:13:11 PM CEST 2024 +ELAPSED: 12 seconds diff --git a/epochX/cudacpp/tlau/logs_ggtt_cppavx2/Events/.keep b/epochX/cudacpp/tlau/logs_ggtt_cppavx2/Events/.keep new file mode 100644 index 0000000000..e69de29bb2 diff --git a/epochX/cudacpp/tlau/logs_ggtt_cppavx2/HTML/.keep b/epochX/cudacpp/tlau/logs_ggtt_cppavx2/HTML/.keep new file mode 100644 index 0000000000..e69de29bb2 diff --git a/epochX/cudacpp/tlau/logs_ggtt_cppavx2/output.txt b/epochX/cudacpp/tlau/logs_ggtt_cppavx2/output.txt new file mode 100644 index 0000000000..c49bc2d310 --- /dev/null +++ b/epochX/cudacpp/tlau/logs_ggtt_cppavx2/output.txt @@ -0,0 +1,82 @@ +START: Mon Aug 5 04:14:49 PM CEST 2024 +************************************************************ +* * +* W E L C O M E to * +* M A D G R A P H 5 _ a M C @ N L O * +* M A D E V E N T * +* * +* * * * +* * * * * * +* * * * * 5 * * * * * +* * * * * * +* * * * +* * +* VERSION 3.5.3_lo_vect * +* * +* The MadGraph5_aMC@NLO Development Team - Find us at * +* https://server06.fynu.ucl.ac.be/projects/madgraph * +* * +* Type 'help' for in-line help. * +* * +************************************************************ +INFO: load configuration from /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/Cards/me5_configuration.txt +INFO: load configuration from /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/Cards/me5_configuration.txt +Using default text editor "vi". Set another one in ./input/mg5_configuration.txt +Using default eps viewer "evince". Set another one in ./input/mg5_configuration.txt +Using default web browser "firefox". Set another one in ./input/mg5_configuration.txt +generate_events -f +Generating 10000 events with run name run_01 +survey run_01 +INFO: compile directory +compile Source Directory +Using random number seed offset = 21 +INFO: Running Survey +Creating Jobs +Working on SubProcesses +INFO: P1_gg_ttx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cppavx2' matrix elements +INFO: Idle: 1, Running: 0, Completed: 0 [ current time: 16h14 ] +INFO: Idle: 0, Running: 0, Completed: 1 [ 2.7s ] +INFO: Idle: 0, Running: 0, Completed: 1 [ 2.7s ] +INFO: End survey +refine 10000 +Creating Jobs +INFO: Refine results to 10000 +INFO: Generating 10000.0 unweighted events. +sum of cpu time of last step: 3 seconds +INFO: Effective Luminosity 27.13133601481371 pb^-1 +INFO: need to improve 1 channels +- Current estimate of cross-section: 442.293 +- 4.035736223404597 + P1_gg_ttx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cppavx2' matrix elements +INFO: Idle: 7, Running: 4, Completed: 0 [ current time: 16h14 ] +INFO: Idle: 2, Running: 4, Completed: 5 [ 2.9s ] +INFO: Idle: 0, Running: 0, Completed: 11 [ 4.4s ] +INFO: Combining runs +sum of cpu time of last step: 21 seconds +INFO: finish refine +refine 10000 --treshold=0.9 +No need for second refine due to stability of cross-section +INFO: Combining Events +combination of events done in 2.1775732040405273 s + === Results Summary for run: run_01 tag: tag_1 === + + Cross-section : 440.1 +- 0.547 pb + Nb of events : 10000 + +INFO: No version of lhapdf. Can not run systematics computation +store_events +INFO: Storing parton level results +INFO: End Parton +reweight -from_cards +decay_events -from_cards +INFO: storing files of previous run +INFO: Done +quit +INFO: +24.57user 1.95system 0:12.11elapsed 218%CPU (0avgtext+0avgdata 95364maxresident)k +0inputs+403112outputs (2major+353650minor)pagefaults 0swaps +END: Mon Aug 5 04:15:01 PM CEST 2024 +ELAPSED: 12 seconds diff --git a/epochX/cudacpp/tlau/logs_ggtt_cppnone/Events/.keep b/epochX/cudacpp/tlau/logs_ggtt_cppnone/Events/.keep new file mode 100644 index 0000000000..e69de29bb2 diff --git a/epochX/cudacpp/tlau/logs_ggtt_cppnone/HTML/.keep b/epochX/cudacpp/tlau/logs_ggtt_cppnone/HTML/.keep new file mode 100644 index 0000000000..e69de29bb2 diff --git a/epochX/cudacpp/tlau/logs_ggtt_cppnone/output.txt b/epochX/cudacpp/tlau/logs_ggtt_cppnone/output.txt new file mode 100644 index 0000000000..64d4e50a12 --- /dev/null +++ b/epochX/cudacpp/tlau/logs_ggtt_cppnone/output.txt @@ -0,0 +1,82 @@ +START: Mon Aug 5 04:12:08 PM CEST 2024 +************************************************************ +* * +* W E L C O M E to * +* M A D G R A P H 5 _ a M C @ N L O * +* M A D E V E N T * +* * +* * * * +* * * * * * +* * * * * 5 * * * * * +* * * * * * +* * * * +* * +* VERSION 3.5.3_lo_vect * +* * +* The MadGraph5_aMC@NLO Development Team - Find us at * +* https://server06.fynu.ucl.ac.be/projects/madgraph * +* * +* Type 'help' for in-line help. * +* * +************************************************************ +INFO: load configuration from /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/Cards/me5_configuration.txt +INFO: load configuration from /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/Cards/me5_configuration.txt +Using default text editor "vi". Set another one in ./input/mg5_configuration.txt +Using default eps viewer "evince". Set another one in ./input/mg5_configuration.txt +Using default web browser "firefox". Set another one in ./input/mg5_configuration.txt +generate_events -f +Generating 10000 events with run name run_01 +survey run_01 +INFO: compile directory +compile Source Directory +Using random number seed offset = 21 +INFO: Running Survey +Creating Jobs +Working on SubProcesses +INFO: P1_gg_ttx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cppnone' matrix elements +INFO: Idle: 1, Running: 0, Completed: 0 [ current time: 16h12 ] +INFO: Idle: 0, Running: 0, Completed: 1 [ 2.9s ] +INFO: Idle: 0, Running: 0, Completed: 1 [ 2.9s ] +INFO: End survey +refine 10000 +Creating Jobs +INFO: Refine results to 10000 +INFO: Generating 10000.0 unweighted events. +sum of cpu time of last step: 3 seconds +INFO: Effective Luminosity 27.13133601481371 pb^-1 +INFO: need to improve 1 channels +- Current estimate of cross-section: 442.293 +- 4.035736223404597 + P1_gg_ttx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cppnone' matrix elements +INFO: Idle: 7, Running: 4, Completed: 0 [ current time: 16h12 ] +INFO: Idle: 2, Running: 4, Completed: 5 [ 3.4s ] +INFO: Idle: 0, Running: 0, Completed: 11 [ 5.2s ] +INFO: Combining runs +sum of cpu time of last step: 24 seconds +INFO: finish refine +refine 10000 --treshold=0.9 +No need for second refine due to stability of cross-section +INFO: Combining Events +combination of events done in 2.260270595550537 s + === Results Summary for run: run_01 tag: tag_1 === + + Cross-section : 440.1 +- 0.547 pb + Nb of events : 10000 + +INFO: No version of lhapdf. Can not run systematics computation +store_events +INFO: Storing parton level results +INFO: End Parton +reweight -from_cards +decay_events -from_cards +INFO: storing files of previous run +INFO: Done +quit +INFO: +27.52user 2.03system 0:13.19elapsed 224%CPU (0avgtext+0avgdata 95336maxresident)k +0inputs+403112outputs (2major+358627minor)pagefaults 0swaps +END: Mon Aug 5 04:12:21 PM CEST 2024 +ELAPSED: 13 seconds diff --git a/epochX/cudacpp/tlau/logs_ggtt_cppsse4/Events/.keep b/epochX/cudacpp/tlau/logs_ggtt_cppsse4/Events/.keep new file mode 100644 index 0000000000..e69de29bb2 diff --git a/epochX/cudacpp/tlau/logs_ggtt_cppsse4/HTML/.keep b/epochX/cudacpp/tlau/logs_ggtt_cppsse4/HTML/.keep new file mode 100644 index 0000000000..e69de29bb2 diff --git a/epochX/cudacpp/tlau/logs_ggtt_cppsse4/output.txt b/epochX/cudacpp/tlau/logs_ggtt_cppsse4/output.txt new file mode 100644 index 0000000000..274c5df6c9 --- /dev/null +++ b/epochX/cudacpp/tlau/logs_ggtt_cppsse4/output.txt @@ -0,0 +1,82 @@ +START: Mon Aug 5 04:12:21 PM CEST 2024 +************************************************************ +* * +* W E L C O M E to * +* M A D G R A P H 5 _ a M C @ N L O * +* M A D E V E N T * +* * +* * * * +* * * * * * +* * * * * 5 * * * * * +* * * * * * +* * * * +* * +* VERSION 3.5.3_lo_vect * +* * +* The MadGraph5_aMC@NLO Development Team - Find us at * +* https://server06.fynu.ucl.ac.be/projects/madgraph * +* * +* Type 'help' for in-line help. * +* * +************************************************************ +INFO: load configuration from /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/Cards/me5_configuration.txt +INFO: load configuration from /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/Cards/me5_configuration.txt +Using default text editor "vi". Set another one in ./input/mg5_configuration.txt +Using default eps viewer "evince". Set another one in ./input/mg5_configuration.txt +Using default web browser "firefox". Set another one in ./input/mg5_configuration.txt +generate_events -f +Generating 10000 events with run name run_01 +survey run_01 +INFO: compile directory +compile Source Directory +Using random number seed offset = 21 +INFO: Running Survey +Creating Jobs +Working on SubProcesses +INFO: P1_gg_ttx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cppsse4' matrix elements +INFO: Idle: 1, Running: 0, Completed: 0 [ current time: 16h12 ] +INFO: Idle: 0, Running: 0, Completed: 1 [ 2.8s ] +INFO: Idle: 0, Running: 0, Completed: 1 [ 2.8s ] +INFO: End survey +refine 10000 +Creating Jobs +INFO: Refine results to 10000 +INFO: Generating 10000.0 unweighted events. +sum of cpu time of last step: 3 seconds +INFO: Effective Luminosity 27.13133601481371 pb^-1 +INFO: need to improve 1 channels +- Current estimate of cross-section: 442.293 +- 4.035736223404597 + P1_gg_ttx +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cppsse4' matrix elements +INFO: Idle: 7, Running: 4, Completed: 0 [ current time: 16h12 ] +INFO: Idle: 2, Running: 4, Completed: 5 [ 3s ] +INFO: Idle: 0, Running: 0, Completed: 11 [ 4.6s ] +INFO: Combining runs +sum of cpu time of last step: 22 seconds +INFO: finish refine +refine 10000 --treshold=0.9 +No need for second refine due to stability of cross-section +INFO: Combining Events +combination of events done in 2.187422752380371 s + === Results Summary for run: run_01 tag: tag_1 === + + Cross-section : 440.1 +- 0.547 pb + Nb of events : 10000 + +INFO: No version of lhapdf. Can not run systematics computation +store_events +INFO: Storing parton level results +INFO: End Parton +reweight -from_cards +decay_events -from_cards +INFO: storing files of previous run +INFO: Done +quit +INFO: +25.40user 2.02system 0:12.42elapsed 220%CPU (0avgtext+0avgdata 95340maxresident)k +0inputs+403112outputs (1major+355669minor)pagefaults 0swaps +END: Mon Aug 5 04:12:33 PM CEST 2024 +ELAPSED: 12 seconds diff --git a/epochX/cudacpp/tlau/logs_ggtt_cuda/HTML/run_01/results.html b/epochX/cudacpp/tlau/logs_ggtt_cuda/HTML/run_01/results.html index 531abb52e8..33486e3031 100644 --- a/epochX/cudacpp/tlau/logs_ggtt_cuda/HTML/run_01/results.html +++ b/epochX/cudacpp/tlau/logs_ggtt_cuda/HTML/run_01/results.html @@ -21,7 +21,7 @@
-      s= 440.4 ± 0.445 (pb)

+      s= 440.12 ± 0.547 (pb)

Graph Cross-Section
G1 47.1 0.222 98.304 3292.0 69.9 46.82 0.388 16.384 5181.0 111
G2
@@ -32,10 +32,10 @@ - - - - + + + + @@ -45,7 +45,7 @@

-      s= 440.4 ± 0.445 (pb)

+      s= 440.12 ± 0.547 (pb)

Graph Cross-Section
/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx 440.4 0.445 638.976 21258.0 440.1 0.547 557.056 23147.0 0
@@ -56,11 +56,11 @@ - - - - - + + + + + diff --git a/epochX/cudacpp/tlau/logs_ggtt_cuda/output.txt b/epochX/cudacpp/tlau/logs_ggtt_cuda/output.txt index c818e6b32f..8436fea8a0 100644 --- a/epochX/cudacpp/tlau/logs_ggtt_cuda/output.txt +++ b/epochX/cudacpp/tlau/logs_ggtt_cuda/output.txt @@ -1,4 +1,4 @@ -START: Wed Dec 20 12:27:20 AM CET 2023 +START: Mon Aug 5 04:11:53 PM CEST 2024 ************************************************************ * * * W E L C O M E to * @@ -11,7 +11,7 @@ START: Wed Dec 20 12:27:20 AM CET 2023 * * * * * * * * * * * * -* VERSION 3.5.2_lo_vect * +* VERSION 3.5.3_lo_vect * * * * The MadGraph5_aMC@NLO Development Team - Find us at * * https://server06.fynu.ucl.ac.be/projects/madgraph * @@ -34,34 +34,36 @@ INFO: Running Survey Creating Jobs Working on SubProcesses INFO: P1_gg_ttx -INFO: Building madevent in madevent_interface.py with 'cuda' matrix elements -INFO: Idle: 1, Running: 0, Completed: 0 [ current time: 00h27 ] -INFO: Idle: 0, Running: 0, Completed: 1 [ 3.1s ] -INFO: Idle: 0, Running: 0, Completed: 1 [ 3.1s ] +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'cuda' matrix elements +INFO: Idle: 1, Running: 0, Completed: 0 [ current time: 16h11 ] +INFO: Idle: 0, Running: 0, Completed: 1 [ 4s ] +INFO: Idle: 0, Running: 0, Completed: 1 [ 4s ] INFO: End survey refine 10000 Creating Jobs INFO: Refine results to 10000 INFO: Generating 10000.0 unweighted events. sum of cpu time of last step: 3 seconds -INFO: Effective Luminosity 27.040181710021088 pb^-1 -INFO: need to improve 2 channels -- Current estimate of cross-section: 443.78400000000005 +- 4.146007168348844 +INFO: Effective Luminosity 27.13133601481371 pb^-1 +INFO: need to improve 1 channels +- Current estimate of cross-section: 442.293 +- 4.035736223404597 P1_gg_ttx +Pre-building madevent in madevent_interface.py with ALL matrix elements Building madevent in madevent_interface.py with 'cuda' matrix elements -INFO: Idle: 9, Running: 4, Completed: 0 [ current time: 00h27 ] -INFO: Idle: 4, Running: 4, Completed: 5 [ 3.1s ] -INFO: Idle: 0, Running: 0, Completed: 13 [ 6s ] +INFO: Idle: 7, Running: 4, Completed: 0 [ current time: 16h12 ] +INFO: Idle: 2, Running: 4, Completed: 5 [ 3.5s ] +INFO: Idle: 0, Running: 0, Completed: 11 [ 5.4s ] INFO: Combining runs -sum of cpu time of last step: 23 seconds +sum of cpu time of last step: 22 seconds INFO: finish refine refine 10000 --treshold=0.9 No need for second refine due to stability of cross-section INFO: Combining Events -combination of events done in 2.0331735610961914 s +combination of events done in 2.185755729675293 s === Results Summary for run: run_01 tag: tag_1 === - Cross-section : 440.4 +- 0.4447 pb + Cross-section : 440.1 +- 0.547 pb Nb of events : 10000 INFO: No version of lhapdf. Can not run systematics computation @@ -74,7 +76,7 @@ INFO: storing files of previous run INFO: Done quit INFO: -37.92user 8.02system 0:20.53elapsed 223%CPU (0avgtext+0avgdata 232700maxresident)k -0inputs+436576outputs (6major+1045210minor)pagefaults 0swaps -END: Wed Dec 20 12:27:41 AM CET 2023 -ELAPSED: 21 seconds +23.46user 5.51system 0:14.32elapsed 202%CPU (0avgtext+0avgdata 233464maxresident)k +0inputs+403112outputs (3major+435834minor)pagefaults 0swaps +END: Mon Aug 5 04:12:07 PM CEST 2024 +ELAPSED: 14 seconds diff --git a/epochX/cudacpp/tlau/logs_ggtt_fortran/HTML/run_01/results.html b/epochX/cudacpp/tlau/logs_ggtt_fortran/HTML/run_01/results.html index 531abb52e8..33486e3031 100644 --- a/epochX/cudacpp/tlau/logs_ggtt_fortran/HTML/run_01/results.html +++ b/epochX/cudacpp/tlau/logs_ggtt_fortran/HTML/run_01/results.html @@ -21,7 +21,7 @@
-      s= 440.4 ± 0.445 (pb)

+      s= 440.12 ± 0.547 (pb)

Graph Cross-Section
G1 47.1 0.222 98.304 3292.0 69.9 46.82 0.388 16.384 5181.0 111
G2
@@ -32,10 +32,10 @@ - - - - + + + + @@ -45,7 +45,7 @@

-      s= 440.4 ± 0.445 (pb)

+      s= 440.12 ± 0.547 (pb)

Graph Cross-Section
/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx 440.4 0.445 638.976 21258.0 440.1 0.547 557.056 23147.0 0
@@ -56,11 +56,11 @@ - - - - - + + + + + diff --git a/epochX/cudacpp/tlau/logs_ggtt_fortran/output.txt b/epochX/cudacpp/tlau/logs_ggtt_fortran/output.txt index 97c9a738cf..b573d5825e 100644 --- a/epochX/cudacpp/tlau/logs_ggtt_fortran/output.txt +++ b/epochX/cudacpp/tlau/logs_ggtt_fortran/output.txt @@ -1,4 +1,4 @@ -START: Wed Dec 20 12:27:42 AM CET 2023 +START: Mon Aug 5 04:11:33 PM CEST 2024 ************************************************************ * * * W E L C O M E to * @@ -11,7 +11,7 @@ START: Wed Dec 20 12:27:42 AM CET 2023 * * * * * * * * * * * * -* VERSION 3.5.2_lo_vect * +* VERSION 3.5.3_lo_vect * * * * The MadGraph5_aMC@NLO Development Team - Find us at * * https://server06.fynu.ucl.ac.be/projects/madgraph * @@ -34,34 +34,36 @@ INFO: Running Survey Creating Jobs Working on SubProcesses INFO: P1_gg_ttx -INFO: Building madevent in madevent_interface.py with 'fortran' matrix elements -INFO: Idle: 1, Running: 0, Completed: 0 [ current time: 00h27 ] -INFO: Idle: 0, Running: 0, Completed: 1 [ 1.8s ] -INFO: Idle: 0, Running: 0, Completed: 1 [ 1.8s ] +Pre-building madevent in madevent_interface.py with ALL matrix elements +Building madevent in madevent_interface.py with 'fortran' matrix elements +INFO: Idle: 1, Running: 0, Completed: 0 [ current time: 16h11 ] +INFO: Idle: 0, Running: 0, Completed: 1 [ 2.8s ] +INFO: Idle: 0, Running: 0, Completed: 1 [ 2.8s ] INFO: End survey refine 10000 Creating Jobs INFO: Refine results to 10000 INFO: Generating 10000.0 unweighted events. -sum of cpu time of last step: 2 seconds -INFO: Effective Luminosity 27.040181710021088 pb^-1 -INFO: need to improve 2 channels -- Current estimate of cross-section: 443.78400000000005 +- 4.146007168348844 +sum of cpu time of last step: 3 seconds +INFO: Effective Luminosity 27.13133601481371 pb^-1 +INFO: need to improve 1 channels +- Current estimate of cross-section: 442.293 +- 4.035736223404597 P1_gg_ttx +Pre-building madevent in madevent_interface.py with ALL matrix elements Building madevent in madevent_interface.py with 'fortran' matrix elements -INFO: Idle: 9, Running: 4, Completed: 0 [ current time: 00h27 ] -INFO: Idle: 4, Running: 4, Completed: 5 [ 2.9s ] -INFO: Idle: 0, Running: 0, Completed: 13 [ 5.8s ] +INFO: Idle: 7, Running: 4, Completed: 0 [ current time: 16h11 ] +INFO: Idle: 2, Running: 4, Completed: 5 [ 3.3s ] +INFO: Idle: 0, Running: 0, Completed: 11 [ 4.9s ] INFO: Combining runs -sum of cpu time of last step: 24 seconds +sum of cpu time of last step: 23 seconds INFO: finish refine refine 10000 --treshold=0.9 No need for second refine due to stability of cross-section INFO: Combining Events -combination of events done in 2.059004545211792 s +combination of events done in 2.1742801666259766 s === Results Summary for run: run_01 tag: tag_1 === - Cross-section : 440.4 +- 0.4447 pb + Cross-section : 440.1 +- 0.547 pb Nb of events : 10000 INFO: No version of lhapdf. Can not run systematics computation @@ -74,7 +76,7 @@ INFO: storing files of previous run INFO: Done quit INFO: -39.24user 3.11system 0:18.01elapsed 235%CPU (0avgtext+0avgdata 91800maxresident)k -0inputs+415632outputs (11major+651311minor)pagefaults 0swaps -END: Wed Dec 20 12:28:00 AM CET 2023 -ELAPSED: 18 seconds +43.42user 6.81system 0:20.04elapsed 250%CPU (0avgtext+0avgdata 91768maxresident)k +0inputs+507016outputs (10major+1344940minor)pagefaults 0swaps +END: Mon Aug 5 04:11:53 PM CEST 2024 +ELAPSED: 20 seconds diff --git a/epochX/cudacpp/tmad/logs_eemumu_mad/log_eemumu_mad_d_inl0_hrd0.txt b/epochX/cudacpp/tmad/logs_eemumu_mad/log_eemumu_mad_d_inl0_hrd0.txt index c9c9460105..f6e6a6c583 100644 --- a/epochX/cudacpp/tmad/logs_eemumu_mad/log_eemumu_mad_d_inl0_hrd0.txt +++ b/epochX/cudacpp/tmad/logs_eemumu_mad/log_eemumu_mad_d_inl0_hrd0.txt @@ -2,8 +2,8 @@ Working directory (build): /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/e make USEBUILDDIR=1 BACKEND=cuda -make USEBUILDDIR=1 BACKEND=cppnone +make USEBUILDDIR=1 BACKEND=cppnone make USEBUILDDIR=1 BACKEND=cppsse4 @@ -13,8 +13,8 @@ make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum' make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum' make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum' - make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum' + make USEBUILDDIR=1 BACKEND=cpp512z make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum' make[1]: Nothing to be done for 'all'. @@ -32,7 +32,7 @@ make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ OMP_NUM_THREADS= -DATE: 2024-10-06_10:04:14 +DATE: 2024-10-06_14:20:07 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: Working directory (run): /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum @@ -58,9 +58,9 @@ Executing ' ./madevent_fortran < /tmp/avalassi/input_eemumu_x1_fortran > /tmp/av [XSECTION] ChannelId = 1 [XSECTION] Cross section = 0.09243 [9.2432789448173985E-002] fbridge_mode=0 [UNWEIGHT] Wrote 3837 events (found 8192 events) - [COUNTERS] PROGRAM TOTAL : 0.7444s - [COUNTERS] Fortran Overhead ( 0 ) : 0.7368s - [COUNTERS] Fortran MEs ( 1 ) : 0.0075s for 8192 events => throughput is 1.09E+06 events/s + [COUNTERS] PROGRAM TOTAL : 0.7496s + [COUNTERS] Fortran Overhead ( 0 ) : 0.7419s + [COUNTERS] Fortran MEs ( 1 ) : 0.0076s for 8192 events => throughput is 1.07E+06 events/s *** (1) EXECUTE MADEVENT_FORTRAN x1 (create events.lhe) *** -------------------- @@ -83,9 +83,34 @@ Executing ' ./madevent_fortran < /tmp/avalassi/input_eemumu_x1_fortran > /tmp/av [XSECTION] ChannelId = 1 [XSECTION] Cross section = 0.09243 [9.2432789448173985E-002] fbridge_mode=0 [UNWEIGHT] Wrote 1589 events (found 1593 events) - [COUNTERS] PROGRAM TOTAL : 0.2176s - [COUNTERS] Fortran Overhead ( 0 ) : 0.2101s - [COUNTERS] Fortran MEs ( 1 ) : 0.0075s for 8192 events => throughput is 1.09E+06 events/s + [COUNTERS] PROGRAM TOTAL : 0.2181s + [COUNTERS] Fortran Overhead ( 0 ) : 0.2108s + [COUNTERS] Fortran MEs ( 1 ) : 0.0073s for 8192 events => throughput is 1.12E+06 events/s + +*** (1) EXECUTE MADEVENT_FORTRAN x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./madevent_fortran < /tmp/avalassi/input_eemumu_x10_fortran > /tmp/avalassi/output_eemumu_x10_fortran' + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 4/16 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 0.09171 [9.1711103909519892E-002] fbridge_mode=0 + [UNWEIGHT] Wrote 1655 events (found 1660 events) + [COUNTERS] PROGRAM TOTAL : 0.7241s + [COUNTERS] Fortran Overhead ( 0 ) : 0.6494s + [COUNTERS] Fortran MEs ( 1 ) : 0.0747s for 81920 events => throughput is 1.10E+06 events/s *** (2-none) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -109,9 +134,9 @@ DEBUG: MEK processed 8192 events across 2 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 0.09243 [9.2432789448173944E-002] fbridge_mode=1 [UNWEIGHT] Wrote 1589 events (found 1593 events) - [COUNTERS] PROGRAM TOTAL : 0.2197s - [COUNTERS] Fortran Overhead ( 0 ) : 0.2123s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0071s for 8192 events => throughput is 1.15E+06 events/s + [COUNTERS] PROGRAM TOTAL : 0.2216s + [COUNTERS] Fortran Overhead ( 0 ) : 0.2141s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0072s for 8192 events => throughput is 1.13E+06 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0003s *** (2-none) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -122,15 +147,50 @@ OK! xsec from fortran (9.2432789448173985E-002) and cpp (9.2432789448173944E-002 OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-none) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.none_d_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_eemumu_x10_cudacpp > /tmp/avalassi/output_eemumu_x10_cudacpp' +DEBUG: MEK processed 81920 events across 2 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 4/16 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 0.09171 [9.1711103909519906E-002] fbridge_mode=1 + [UNWEIGHT] Wrote 1655 events (found 1660 events) + [COUNTERS] PROGRAM TOTAL : 0.7216s + [COUNTERS] Fortran Overhead ( 0 ) : 0.6491s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0722s for 81920 events => throughput is 1.13E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0003s + +*** (2-none) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (9.1711103909519892E-002) and cpp (9.1711103909519906E-002) differ by less than 3E-14 (2.220446049250313e-16) + +*** (2-none) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+BRDHST/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.158620e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.155506e+06 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.163690e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.172380e+06 ) sec^-1 *** (2-sse4) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -154,9 +214,9 @@ DEBUG: MEK processed 8192 events across 2 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 0.09243 [9.2432789448173944E-002] fbridge_mode=1 [UNWEIGHT] Wrote 1589 events (found 1593 events) - [COUNTERS] PROGRAM TOTAL : 0.2221s - [COUNTERS] Fortran Overhead ( 0 ) : 0.2173s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0045s for 8192 events => throughput is 1.81E+06 events/s + [COUNTERS] PROGRAM TOTAL : 0.2181s + [COUNTERS] Fortran Overhead ( 0 ) : 0.2134s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0045s for 8192 events => throughput is 1.84E+06 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0003s *** (2-sse4) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -167,15 +227,50 @@ OK! xsec from fortran (9.2432789448173985E-002) and cpp (9.2432789448173944E-002 OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-sse4) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.sse4_d_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_eemumu_x10_cudacpp > /tmp/avalassi/output_eemumu_x10_cudacpp' +DEBUG: MEK processed 81920 events across 2 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 4/16 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 0.09171 [9.1711103909519906E-002] fbridge_mode=1 + [UNWEIGHT] Wrote 1655 events (found 1660 events) + [COUNTERS] PROGRAM TOTAL : 0.7042s + [COUNTERS] Fortran Overhead ( 0 ) : 0.6608s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0431s for 81920 events => throughput is 1.90E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0003s + +*** (2-sse4) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (9.1711103909519892E-002) and cpp (9.1711103909519906E-002) differ by less than 3E-14 (2.220446049250313e-16) + +*** (2-sse4) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+BRDHST/sse4+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.887925e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.927547e+06 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.991506e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.997228e+06 ) sec^-1 *** (2-avx2) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -199,8 +294,8 @@ DEBUG: MEK processed 8192 events across 2 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 0.09243 [9.2432789448173971E-002] fbridge_mode=1 [UNWEIGHT] Wrote 1589 events (found 1593 events) - [COUNTERS] PROGRAM TOTAL : 0.2160s - [COUNTERS] Fortran Overhead ( 0 ) : 0.2124s + [COUNTERS] PROGRAM TOTAL : 0.2167s + [COUNTERS] Fortran Overhead ( 0 ) : 0.2132s [COUNTERS] CudaCpp MEs ( 2 ) : 0.0033s for 8192 events => throughput is 2.50E+06 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0003s @@ -212,15 +307,50 @@ OK! xsec from fortran (9.2432789448173985E-002) and cpp (9.2432789448173971E-002 OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-avx2) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.avx2_d_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_eemumu_x10_cudacpp > /tmp/avalassi/output_eemumu_x10_cudacpp' +DEBUG: MEK processed 81920 events across 2 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 4/16 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 0.09171 [9.1711103909519906E-002] fbridge_mode=1 + [UNWEIGHT] Wrote 1655 events (found 1660 events) + [COUNTERS] PROGRAM TOTAL : 0.6875s + [COUNTERS] Fortran Overhead ( 0 ) : 0.6550s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0322s for 81920 events => throughput is 2.54E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0003s + +*** (2-avx2) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (9.1711103909519892E-002) and cpp (9.1711103909519906E-002) differ by less than 3E-14 (2.220446049250313e-16) + +*** (2-avx2) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+BRDHST/avx2+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 2.590914e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.600529e+06 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 2.667984e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.671423e+06 ) sec^-1 *** (2-512y) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -244,9 +374,9 @@ DEBUG: MEK processed 8192 events across 2 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 0.09243 [9.2432789448173971E-002] fbridge_mode=1 [UNWEIGHT] Wrote 1589 events (found 1593 events) - [COUNTERS] PROGRAM TOTAL : 0.2167s - [COUNTERS] Fortran Overhead ( 0 ) : 0.2131s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0033s for 8192 events => throughput is 2.46E+06 events/s + [COUNTERS] PROGRAM TOTAL : 0.2157s + [COUNTERS] Fortran Overhead ( 0 ) : 0.2122s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0033s for 8192 events => throughput is 2.52E+06 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0003s *** (2-512y) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -257,15 +387,50 @@ OK! xsec from fortran (9.2432789448173985E-002) and cpp (9.2432789448173971E-002 OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-512y) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.512y_d_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_eemumu_x10_cudacpp > /tmp/avalassi/output_eemumu_x10_cudacpp' +DEBUG: MEK processed 81920 events across 2 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 4/16 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 0.09171 [9.1711103909519906E-002] fbridge_mode=1 + [UNWEIGHT] Wrote 1655 events (found 1660 events) + [COUNTERS] PROGRAM TOTAL : 0.6769s + [COUNTERS] Fortran Overhead ( 0 ) : 0.6447s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0319s for 81920 events => throughput is 2.57E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0003s + +*** (2-512y) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (9.1711103909519892E-002) and cpp (9.1711103909519906E-002) differ by less than 3E-14 (2.220446049250313e-16) + +*** (2-512y) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+BRDHST/512y+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 2.636316e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.668766e+06 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 2.730901e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.733435e+06 ) sec^-1 *** (2-512z) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -289,9 +454,9 @@ DEBUG: MEK processed 8192 events across 2 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 0.09243 [9.2432789448173971E-002] fbridge_mode=1 [UNWEIGHT] Wrote 1589 events (found 1593 events) - [COUNTERS] PROGRAM TOTAL : 0.2188s - [COUNTERS] Fortran Overhead ( 0 ) : 0.2145s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0041s for 8192 events => throughput is 2.01E+06 events/s + [COUNTERS] PROGRAM TOTAL : 0.2217s + [COUNTERS] Fortran Overhead ( 0 ) : 0.2173s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0041s for 8192 events => throughput is 2.02E+06 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0003s *** (2-512z) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -302,15 +467,50 @@ OK! xsec from fortran (9.2432789448173985E-002) and cpp (9.2432789448173971E-002 OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-512z) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.512z_d_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_eemumu_x10_cudacpp > /tmp/avalassi/output_eemumu_x10_cudacpp' +DEBUG: MEK processed 81920 events across 2 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 4/16 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 0.09171 [9.1711103909519892E-002] fbridge_mode=1 + [UNWEIGHT] Wrote 1655 events (found 1660 events) + [COUNTERS] PROGRAM TOTAL : 0.7054s + [COUNTERS] Fortran Overhead ( 0 ) : 0.6643s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0408s for 81920 events => throughput is 2.01E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0003s + +*** (2-512z) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (9.1711103909519892E-002) and cpp (9.1711103909519892E-002) differ by less than 3E-14 (0.0) + +*** (2-512z) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+BRDHST/512z+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 2.085135e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.024845e+06 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 2.218811e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.208238e+06 ) sec^-1 *** (3-cuda) EXECUTE MADEVENT_CUDA x1 (create events.lhe) *** -------------------- @@ -334,9 +534,9 @@ DEBUG: MEK processed 8192 events across 2 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 0.09243 [9.2432789448173971E-002] fbridge_mode=1 [UNWEIGHT] Wrote 1589 events (found 1593 events) - [COUNTERS] PROGRAM TOTAL : 0.6526s - [COUNTERS] Fortran Overhead ( 0 ) : 0.6492s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0028s for 8192 events => throughput is 2.88E+06 events/s + [COUNTERS] PROGRAM TOTAL : 0.6581s + [COUNTERS] Fortran Overhead ( 0 ) : 0.6547s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0028s for 8192 events => throughput is 2.90E+06 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0006s *** (3-cuda) Compare MADEVENT_CUDA x1 xsec to MADEVENT_FORTRAN xsec *** @@ -347,45 +547,80 @@ OK! xsec from fortran (9.2432789448173985E-002) and cuda (9.2432789448173971E-00 OK! events.lhe.cuda.1 and events.lhe.ref.1 are identical +*** (3-cuda) EXECUTE MADEVENT_CUDA x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.cuda_d_inl0_hrd0/madevent_cuda < /tmp/avalassi/input_eemumu_x10_cudacpp > /tmp/avalassi/output_eemumu_x10_cudacpp' +DEBUG: MEK processed 81920 events across 2 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 4/16 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 0.09171 [9.1711103909519892E-002] fbridge_mode=1 + [UNWEIGHT] Wrote 1655 events (found 1660 events) + [COUNTERS] PROGRAM TOTAL : 1.1162s + [COUNTERS] Fortran Overhead ( 0 ) : 1.1080s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0076s for 81920 events => throughput is 1.07E+07 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0006s + +*** (3-cuda) Compare MADEVENT_CUDA x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (9.1711103909519892E-002) and cuda (9.1711103909519892E-002) differ by less than 3E-14 (0.0) + +*** (3-cuda) Compare MADEVENT_CUDA x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cuda.10 and events.lhe.ref.10 are identical + *** EXECUTE GCHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SM_EPEM_MUPMUM_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.299210e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.367801e+06 ) sec^-1 *** EXECUTE GCHECK(8192) -p 256 32 1 *** Process = SIGMA_SM_EPEM_MUPMUM_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.632885e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.659981e+06 ) sec^-1 *** EXECUTE GCHECK(MAX) -p 16384 32 1 --bridge *** Process = SIGMA_SM_EPEM_MUPMUM_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.507229e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.524192e+07 ) sec^-1 *** EXECUTE GCHECK(MAX) -p 16384 32 1 *** Process = SIGMA_SM_EPEM_MUPMUM_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.868548e+08 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.897997e+08 ) sec^-1 *** EXECUTE GCHECK(MAX128THR) -p 4096 128 1 --bridge *** Process = SIGMA_SM_EPEM_MUPMUM_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.543060e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.576606e+07 ) sec^-1 *** EXECUTE GCHECK(MAX128THR) -p 4096 128 1 *** Process = SIGMA_SM_EPEM_MUPMUM_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.911449e+08 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.909861e+08 ) sec^-1 *** EXECUTE GCHECK(MAX8THR) -p 65536 8 1 --bridge *** Process = SIGMA_SM_EPEM_MUPMUM_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.533062e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.538380e+07 ) sec^-1 *** EXECUTE GCHECK(MAX8THR) -p 65536 8 1 *** Process = SIGMA_SM_EPEM_MUPMUM_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.164979e+08 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.189972e+08 ) sec^-1 *** (3-hip) WARNING! SKIP MADEVENT_HIP (hip is not supported on this node) *** diff --git a/epochX/cudacpp/tmad/logs_eemumu_mad/log_eemumu_mad_f_inl0_hrd0.txt b/epochX/cudacpp/tmad/logs_eemumu_mad/log_eemumu_mad_f_inl0_hrd0.txt index 13ceac3a87..64db1657be 100644 --- a/epochX/cudacpp/tmad/logs_eemumu_mad/log_eemumu_mad_f_inl0_hrd0.txt +++ b/epochX/cudacpp/tmad/logs_eemumu_mad/log_eemumu_mad_f_inl0_hrd0.txt @@ -1,22 +1,22 @@ Working directory (build): /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum + make USEBUILDDIR=1 BACKEND=cuda make USEBUILDDIR=1 BACKEND=cppnone make USEBUILDDIR=1 BACKEND=cppsse4 - - make USEBUILDDIR=1 BACKEND=cppavx2 + make USEBUILDDIR=1 BACKEND=cpp512y make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum' make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum' make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum' make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum' +make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum' make USEBUILDDIR=1 BACKEND=cpp512z make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum' -make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum' make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum' make[1]: Nothing to be done for 'all'. @@ -32,7 +32,7 @@ make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ OMP_NUM_THREADS= -DATE: 2024-10-06_10:04:39 +DATE: 2024-10-06_14:20:46 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: Working directory (run): /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum @@ -58,9 +58,9 @@ Executing ' ./madevent_fortran < /tmp/avalassi/input_eemumu_x1_fortran > /tmp/av [XSECTION] ChannelId = 1 [XSECTION] Cross section = 0.09243 [9.2432789448173985E-002] fbridge_mode=0 [UNWEIGHT] Wrote 3837 events (found 8192 events) - [COUNTERS] PROGRAM TOTAL : 0.7443s - [COUNTERS] Fortran Overhead ( 0 ) : 0.7370s - [COUNTERS] Fortran MEs ( 1 ) : 0.0073s for 8192 events => throughput is 1.12E+06 events/s + [COUNTERS] PROGRAM TOTAL : 0.7392s + [COUNTERS] Fortran Overhead ( 0 ) : 0.7318s + [COUNTERS] Fortran MEs ( 1 ) : 0.0075s for 8192 events => throughput is 1.10E+06 events/s *** (1) EXECUTE MADEVENT_FORTRAN x1 (create events.lhe) *** -------------------- @@ -83,9 +83,34 @@ Executing ' ./madevent_fortran < /tmp/avalassi/input_eemumu_x1_fortran > /tmp/av [XSECTION] ChannelId = 1 [XSECTION] Cross section = 0.09243 [9.2432789448173985E-002] fbridge_mode=0 [UNWEIGHT] Wrote 1589 events (found 1593 events) - [COUNTERS] PROGRAM TOTAL : 0.2183s - [COUNTERS] Fortran Overhead ( 0 ) : 0.2108s - [COUNTERS] Fortran MEs ( 1 ) : 0.0075s for 8192 events => throughput is 1.09E+06 events/s + [COUNTERS] PROGRAM TOTAL : 0.2195s + [COUNTERS] Fortran Overhead ( 0 ) : 0.2118s + [COUNTERS] Fortran MEs ( 1 ) : 0.0078s for 8192 events => throughput is 1.05E+06 events/s + +*** (1) EXECUTE MADEVENT_FORTRAN x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./madevent_fortran < /tmp/avalassi/input_eemumu_x10_fortran > /tmp/avalassi/output_eemumu_x10_fortran' + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 4/16 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 0.09171 [9.1711103909519892E-002] fbridge_mode=0 + [UNWEIGHT] Wrote 1655 events (found 1660 events) + [COUNTERS] PROGRAM TOTAL : 0.7225s + [COUNTERS] Fortran Overhead ( 0 ) : 0.6478s + [COUNTERS] Fortran MEs ( 1 ) : 0.0748s for 81920 events => throughput is 1.10E+06 events/s *** (2-none) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -109,9 +134,9 @@ DEBUG: MEK processed 8192 events across 2 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 0.09243 [9.2432777382586498E-002] fbridge_mode=1 [UNWEIGHT] Wrote 1589 events (found 1593 events) - [COUNTERS] PROGRAM TOTAL : 0.2266s - [COUNTERS] Fortran Overhead ( 0 ) : 0.2197s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0067s for 8192 events => throughput is 1.21E+06 events/s + [COUNTERS] PROGRAM TOTAL : 0.2184s + [COUNTERS] Fortran Overhead ( 0 ) : 0.2114s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0068s for 8192 events => throughput is 1.21E+06 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0002s *** (2-none) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -122,15 +147,50 @@ OK! xsec from fortran (9.2432789448173985E-002) and cpp (9.2432777382586498E-002 OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-none) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.none_f_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_eemumu_x10_cudacpp > /tmp/avalassi/output_eemumu_x10_cudacpp' +DEBUG: MEK processed 81920 events across 2 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 4/16 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 0.09171 [9.1711091925143637E-002] fbridge_mode=1 + [UNWEIGHT] Wrote 1655 events (found 1660 events) + [COUNTERS] PROGRAM TOTAL : 0.7150s + [COUNTERS] Fortran Overhead ( 0 ) : 0.6465s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0683s for 81920 events => throughput is 1.20E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0002s + +*** (2-none) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (9.1711103909519892E-002) and cpp (9.1711091925143637E-002) differ by less than 4E-4 (1.3067530257870885e-07) + +*** (2-none) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+BRDHST/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.221258e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.206383e+06 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.225429e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.228334e+06 ) sec^-1 *** (2-sse4) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -154,9 +214,9 @@ DEBUG: MEK processed 8192 events across 2 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 0.09243 [9.2432774839452045E-002] fbridge_mode=1 [UNWEIGHT] Wrote 1589 events (found 1593 events) - [COUNTERS] PROGRAM TOTAL : 0.2220s - [COUNTERS] Fortran Overhead ( 0 ) : 0.2190s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0028s for 8192 events => throughput is 2.89E+06 events/s + [COUNTERS] PROGRAM TOTAL : 0.2158s + [COUNTERS] Fortran Overhead ( 0 ) : 0.2129s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0027s for 8192 events => throughput is 2.99E+06 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0002s *** (2-sse4) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -167,15 +227,50 @@ OK! xsec from fortran (9.2432789448173985E-002) and cpp (9.2432774839452045E-002 OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-sse4) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.sse4_f_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_eemumu_x10_cudacpp > /tmp/avalassi/output_eemumu_x10_cudacpp' +DEBUG: MEK processed 81920 events across 2 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 4/16 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 0.09171 [9.1711089416628339E-002] fbridge_mode=1 + [UNWEIGHT] Wrote 1655 events (found 1660 events) + [COUNTERS] PROGRAM TOTAL : 0.6780s + [COUNTERS] Fortran Overhead ( 0 ) : 0.6508s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0271s for 81920 events => throughput is 3.03E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0002s + +*** (2-sse4) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (9.1711103909519892E-002) and cpp (9.1711089416628339E-002) differ by less than 4E-4 (1.5802766439865223e-07) + +*** (2-sse4) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+BRDHST/sse4+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.137547e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.131873e+06 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.221144e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.232105e+06 ) sec^-1 *** (2-avx2) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -199,9 +294,9 @@ DEBUG: MEK processed 8192 events across 2 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 0.09243 [9.2432774915924193E-002] fbridge_mode=1 [UNWEIGHT] Wrote 1589 events (found 1593 events) - [COUNTERS] PROGRAM TOTAL : 0.2228s - [COUNTERS] Fortran Overhead ( 0 ) : 0.2200s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0026s for 8192 events => throughput is 3.13E+06 events/s + [COUNTERS] PROGRAM TOTAL : 0.2163s + [COUNTERS] Fortran Overhead ( 0 ) : 0.2136s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0025s for 8192 events => throughput is 3.25E+06 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0002s *** (2-avx2) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -212,15 +307,50 @@ OK! xsec from fortran (9.2432789448173985E-002) and cpp (9.2432774915924193E-002 OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-avx2) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.avx2_f_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_eemumu_x10_cudacpp > /tmp/avalassi/output_eemumu_x10_cudacpp' +DEBUG: MEK processed 81920 events across 2 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 4/16 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 0.09171 [9.1711089453554426E-002] fbridge_mode=1 + [UNWEIGHT] Wrote 1655 events (found 1660 events) + [COUNTERS] PROGRAM TOTAL : 0.6753s + [COUNTERS] Fortran Overhead ( 0 ) : 0.6499s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0252s for 81920 events => throughput is 3.25E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0002s + +*** (2-avx2) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (9.1711103909519892E-002) and cpp (9.1711089453554426E-002) differ by less than 4E-4 (1.5762502958427405e-07) + +*** (2-avx2) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+BRDHST/avx2+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.328121e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.398192e+06 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.556846e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.627173e+06 ) sec^-1 *** (2-512y) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -244,9 +374,9 @@ DEBUG: MEK processed 8192 events across 2 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 0.09243 [9.2432774915924193E-002] fbridge_mode=1 [UNWEIGHT] Wrote 1589 events (found 1593 events) - [COUNTERS] PROGRAM TOTAL : 0.2241s - [COUNTERS] Fortran Overhead ( 0 ) : 0.2212s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0027s for 8192 events => throughput is 3.09E+06 events/s + [COUNTERS] PROGRAM TOTAL : 0.2221s + [COUNTERS] Fortran Overhead ( 0 ) : 0.2194s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0025s for 8192 events => throughput is 3.24E+06 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0002s *** (2-512y) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -257,15 +387,50 @@ OK! xsec from fortran (9.2432789448173985E-002) and cpp (9.2432774915924193E-002 OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-512y) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.512y_f_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_eemumu_x10_cudacpp > /tmp/avalassi/output_eemumu_x10_cudacpp' +DEBUG: MEK processed 81920 events across 2 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 4/16 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 0.09171 [9.1711089453554426E-002] fbridge_mode=1 + [UNWEIGHT] Wrote 1655 events (found 1660 events) + [COUNTERS] PROGRAM TOTAL : 0.6772s + [COUNTERS] Fortran Overhead ( 0 ) : 0.6532s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0238s for 81920 events => throughput is 3.44E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0002s + +*** (2-512y) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (9.1711103909519892E-002) and cpp (9.1711089453554426E-002) differ by less than 4E-4 (1.5762502958427405e-07) + +*** (2-512y) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+BRDHST/512y+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.452418e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.434955e+06 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.604389e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.680336e+06 ) sec^-1 *** (2-512z) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -289,10 +454,10 @@ DEBUG: MEK processed 8192 events across 2 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 0.09243 [9.2432778556608516E-002] fbridge_mode=1 [UNWEIGHT] Wrote 1589 events (found 1593 events) - [COUNTERS] PROGRAM TOTAL : 0.2173s - [COUNTERS] Fortran Overhead ( 0 ) : 0.2144s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0027s for 8192 events => throughput is 3.08E+06 events/s - [COUNTERS] CudaCpp HEL ( 3 ) : 0.0002s + [COUNTERS] PROGRAM TOTAL : 0.2178s + [COUNTERS] Fortran Overhead ( 0 ) : 0.2149s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0027s for 8192 events => throughput is 3.03E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0003s *** (2-512z) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -302,15 +467,50 @@ OK! xsec from fortran (9.2432789448173985E-002) and cpp (9.2432778556608516E-002 OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-512z) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.512z_f_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_eemumu_x10_cudacpp > /tmp/avalassi/output_eemumu_x10_cudacpp' +DEBUG: MEK processed 81920 events across 2 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 4/16 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 0.09171 [9.1711093118690828E-002] fbridge_mode=1 + [UNWEIGHT] Wrote 1655 events (found 1660 events) + [COUNTERS] PROGRAM TOTAL : 0.6863s + [COUNTERS] Fortran Overhead ( 0 ) : 0.6605s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0256s for 81920 events => throughput is 3.20E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0002s + +*** (2-512z) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (9.1711103909519892E-002) and cpp (9.1711093118690828E-002) differ by less than 4E-4 (1.1766109664357316e-07) + +*** (2-512z) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+BRDHST/512z+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.402847e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.375073e+06 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.641263e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.617423e+06 ) sec^-1 *** (3-cuda) EXECUTE MADEVENT_CUDA x1 (create events.lhe) *** -------------------- @@ -334,9 +534,9 @@ DEBUG: MEK processed 8192 events across 2 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 0.09243 [9.2432780016531851E-002] fbridge_mode=1 [UNWEIGHT] Wrote 1589 events (found 1593 events) - [COUNTERS] PROGRAM TOTAL : 0.6500s - [COUNTERS] Fortran Overhead ( 0 ) : 0.6467s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0028s for 8192 events => throughput is 2.92E+06 events/s + [COUNTERS] PROGRAM TOTAL : 0.6551s + [COUNTERS] Fortran Overhead ( 0 ) : 0.6518s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0028s for 8192 events => throughput is 2.91E+06 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0005s *** (3-cuda) Compare MADEVENT_CUDA x1 xsec to MADEVENT_FORTRAN xsec *** @@ -347,45 +547,80 @@ OK! xsec from fortran (9.2432789448173985E-002) and cuda (9.2432780016531851E-00 OK! events.lhe.cuda.1 and events.lhe.ref.1 are identical +*** (3-cuda) EXECUTE MADEVENT_CUDA x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.cuda_f_inl0_hrd0/madevent_cuda < /tmp/avalassi/input_eemumu_x10_cudacpp > /tmp/avalassi/output_eemumu_x10_cudacpp' +DEBUG: MEK processed 81920 events across 2 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 4/16 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 0.09171 [9.1711094767039689E-002] fbridge_mode=1 + [UNWEIGHT] Wrote 1655 events (found 1660 events) + [COUNTERS] PROGRAM TOTAL : 1.0867s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0792s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0069s for 81920 events => throughput is 1.18E+07 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0005s + +*** (3-cuda) Compare MADEVENT_CUDA x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (9.1711103909519892E-002) and cuda (9.1711094767039689E-002) differ by less than 4E-4 (9.968782199720749e-08) + +*** (3-cuda) Compare MADEVENT_CUDA x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cuda.10 and events.lhe.ref.10 are identical + *** EXECUTE GCHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SM_EPEM_MUPMUM_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.451436e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.393792e+06 ) sec^-1 *** EXECUTE GCHECK(8192) -p 256 32 1 *** Process = SIGMA_SM_EPEM_MUPMUM_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.688055e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.688058e+06 ) sec^-1 *** EXECUTE GCHECK(MAX) -p 16384 32 1 --bridge *** Process = SIGMA_SM_EPEM_MUPMUM_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 6.014252e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 5.851874e+07 ) sec^-1 *** EXECUTE GCHECK(MAX) -p 16384 32 1 *** Process = SIGMA_SM_EPEM_MUPMUM_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 2.229387e+08 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.268525e+08 ) sec^-1 *** EXECUTE GCHECK(MAX128THR) -p 4096 128 1 --bridge *** Process = SIGMA_SM_EPEM_MUPMUM_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 5.787718e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 5.807265e+07 ) sec^-1 *** EXECUTE GCHECK(MAX128THR) -p 4096 128 1 *** Process = SIGMA_SM_EPEM_MUPMUM_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 2.220221e+08 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.259927e+08 ) sec^-1 *** EXECUTE GCHECK(MAX8THR) -p 65536 8 1 --bridge *** Process = SIGMA_SM_EPEM_MUPMUM_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 5.380548e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 5.477067e+07 ) sec^-1 *** EXECUTE GCHECK(MAX8THR) -p 65536 8 1 *** Process = SIGMA_SM_EPEM_MUPMUM_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.826286e+08 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.855612e+08 ) sec^-1 *** (3-hip) WARNING! SKIP MADEVENT_HIP (hip is not supported on this node) *** diff --git a/epochX/cudacpp/tmad/logs_eemumu_mad/log_eemumu_mad_m_inl0_hrd0.txt b/epochX/cudacpp/tmad/logs_eemumu_mad/log_eemumu_mad_m_inl0_hrd0.txt index 093bec81e5..2668e88aa0 100644 --- a/epochX/cudacpp/tmad/logs_eemumu_mad/log_eemumu_mad_m_inl0_hrd0.txt +++ b/epochX/cudacpp/tmad/logs_eemumu_mad/log_eemumu_mad_m_inl0_hrd0.txt @@ -1,22 +1,22 @@ Working directory (build): /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum -make USEBUILDDIR=1 BACKEND=cuda +make USEBUILDDIR=1 BACKEND=cuda make USEBUILDDIR=1 BACKEND=cppnone make USEBUILDDIR=1 BACKEND=cppsse4 make USEBUILDDIR=1 BACKEND=cppavx2 -make USEBUILDDIR=1 BACKEND=cpp512y -make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum' make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum' +make USEBUILDDIR=1 BACKEND=cpp512y make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum' make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum' make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum' make USEBUILDDIR=1 BACKEND=cpp512z make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum' +make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum' make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum' make[1]: Nothing to be done for 'all'. @@ -32,7 +32,7 @@ make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ OMP_NUM_THREADS= -DATE: 2024-10-06_10:04:27 +DATE: 2024-10-06_14:20:27 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: Working directory (run): /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum @@ -58,9 +58,9 @@ Executing ' ./madevent_fortran < /tmp/avalassi/input_eemumu_x1_fortran > /tmp/av [XSECTION] ChannelId = 1 [XSECTION] Cross section = 0.09243 [9.2432789448173985E-002] fbridge_mode=0 [UNWEIGHT] Wrote 3837 events (found 8192 events) - [COUNTERS] PROGRAM TOTAL : 0.7605s - [COUNTERS] Fortran Overhead ( 0 ) : 0.7527s - [COUNTERS] Fortran MEs ( 1 ) : 0.0078s for 8192 events => throughput is 1.06E+06 events/s + [COUNTERS] PROGRAM TOTAL : 0.7468s + [COUNTERS] Fortran Overhead ( 0 ) : 0.7392s + [COUNTERS] Fortran MEs ( 1 ) : 0.0076s for 8192 events => throughput is 1.08E+06 events/s *** (1) EXECUTE MADEVENT_FORTRAN x1 (create events.lhe) *** -------------------- @@ -83,9 +83,34 @@ Executing ' ./madevent_fortran < /tmp/avalassi/input_eemumu_x1_fortran > /tmp/av [XSECTION] ChannelId = 1 [XSECTION] Cross section = 0.09243 [9.2432789448173985E-002] fbridge_mode=0 [UNWEIGHT] Wrote 1589 events (found 1593 events) - [COUNTERS] PROGRAM TOTAL : 0.2221s - [COUNTERS] Fortran Overhead ( 0 ) : 0.2144s - [COUNTERS] Fortran MEs ( 1 ) : 0.0077s for 8192 events => throughput is 1.06E+06 events/s + [COUNTERS] PROGRAM TOTAL : 0.2203s + [COUNTERS] Fortran Overhead ( 0 ) : 0.2128s + [COUNTERS] Fortran MEs ( 1 ) : 0.0075s for 8192 events => throughput is 1.09E+06 events/s + +*** (1) EXECUTE MADEVENT_FORTRAN x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./madevent_fortran < /tmp/avalassi/input_eemumu_x10_fortran > /tmp/avalassi/output_eemumu_x10_fortran' + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 4/16 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 0.09171 [9.1711103909519892E-002] fbridge_mode=0 + [UNWEIGHT] Wrote 1655 events (found 1660 events) + [COUNTERS] PROGRAM TOTAL : 0.7161s + [COUNTERS] Fortran Overhead ( 0 ) : 0.6414s + [COUNTERS] Fortran MEs ( 1 ) : 0.0748s for 81920 events => throughput is 1.10E+06 events/s *** (2-none) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -109,9 +134,9 @@ DEBUG: MEK processed 8192 events across 2 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 0.09243 [9.2432789444986618E-002] fbridge_mode=1 [UNWEIGHT] Wrote 1589 events (found 1593 events) - [COUNTERS] PROGRAM TOTAL : 0.2212s - [COUNTERS] Fortran Overhead ( 0 ) : 0.2136s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0073s for 8192 events => throughput is 1.12E+06 events/s + [COUNTERS] PROGRAM TOTAL : 0.2236s + [COUNTERS] Fortran Overhead ( 0 ) : 0.2159s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0074s for 8192 events => throughput is 1.11E+06 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0003s *** (2-none) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -122,15 +147,50 @@ OK! xsec from fortran (9.2432789448173985E-002) and cpp (9.2432789444986618E-002 OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-none) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.none_m_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_eemumu_x10_cudacpp > /tmp/avalassi/output_eemumu_x10_cudacpp' +DEBUG: MEK processed 81920 events across 2 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 4/16 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 0.09171 [9.1711103904317928E-002] fbridge_mode=1 + [UNWEIGHT] Wrote 1655 events (found 1660 events) + [COUNTERS] PROGRAM TOTAL : 0.7270s + [COUNTERS] Fortran Overhead ( 0 ) : 0.6522s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0745s for 81920 events => throughput is 1.10E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0003s + +*** (2-none) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (9.1711103909519892E-002) and cpp (9.1711103904317928E-002) differ by less than 2E-4 (5.6721183305796785e-11) + +*** (2-none) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+BRDHST/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.133245e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.127918e+06 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.115304e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.146074e+06 ) sec^-1 *** (2-sse4) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -154,9 +214,9 @@ DEBUG: MEK processed 8192 events across 2 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 0.09243 [9.2432789444986618E-002] fbridge_mode=1 [UNWEIGHT] Wrote 1589 events (found 1593 events) - [COUNTERS] PROGRAM TOTAL : 0.2168s - [COUNTERS] Fortran Overhead ( 0 ) : 0.2123s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0042s for 8192 events => throughput is 1.94E+06 events/s + [COUNTERS] PROGRAM TOTAL : 0.2163s + [COUNTERS] Fortran Overhead ( 0 ) : 0.2118s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0043s for 8192 events => throughput is 1.91E+06 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0003s *** (2-sse4) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -167,15 +227,50 @@ OK! xsec from fortran (9.2432789448173985E-002) and cpp (9.2432789444986618E-002 OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-sse4) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.sse4_m_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_eemumu_x10_cudacpp > /tmp/avalassi/output_eemumu_x10_cudacpp' +DEBUG: MEK processed 81920 events across 2 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 4/16 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 0.09171 [9.1711103904317928E-002] fbridge_mode=1 + [UNWEIGHT] Wrote 1655 events (found 1660 events) + [COUNTERS] PROGRAM TOTAL : 0.6921s + [COUNTERS] Fortran Overhead ( 0 ) : 0.6497s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0421s for 81920 events => throughput is 1.95E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0003s + +*** (2-sse4) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (9.1711103909519892E-002) and cpp (9.1711103904317928E-002) differ by less than 2E-4 (5.6721183305796785e-11) + +*** (2-sse4) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+BRDHST/sse4+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.993139e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.942921e+06 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 2.058944e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.086391e+06 ) sec^-1 *** (2-avx2) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -199,9 +294,9 @@ DEBUG: MEK processed 8192 events across 2 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 0.09243 [9.2432789444494415E-002] fbridge_mode=1 [UNWEIGHT] Wrote 1589 events (found 1593 events) - [COUNTERS] PROGRAM TOTAL : 0.2174s - [COUNTERS] Fortran Overhead ( 0 ) : 0.2138s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0034s for 8192 events => throughput is 2.42E+06 events/s + [COUNTERS] PROGRAM TOTAL : 0.2166s + [COUNTERS] Fortran Overhead ( 0 ) : 0.2129s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0034s for 8192 events => throughput is 2.41E+06 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0003s *** (2-avx2) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -212,15 +307,50 @@ OK! xsec from fortran (9.2432789448173985E-002) and cpp (9.2432789444494415E-002 OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-avx2) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.avx2_m_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_eemumu_x10_cudacpp > /tmp/avalassi/output_eemumu_x10_cudacpp' +DEBUG: MEK processed 81920 events across 2 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 4/16 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 0.09171 [9.1711103899063451E-002] fbridge_mode=1 + [UNWEIGHT] Wrote 1655 events (found 1660 events) + [COUNTERS] PROGRAM TOTAL : 0.6886s + [COUNTERS] Fortran Overhead ( 0 ) : 0.6554s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0329s for 81920 events => throughput is 2.49E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0003s + +*** (2-avx2) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (9.1711103909519892E-002) and cpp (9.1711103899063451E-002) differ by less than 2E-4 (1.1401501964769523e-10) + +*** (2-avx2) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+BRDHST/avx2+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 2.549665e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.496979e+06 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 2.708708e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.649957e+06 ) sec^-1 *** (2-512y) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -244,9 +374,9 @@ DEBUG: MEK processed 8192 events across 2 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 0.09243 [9.2432789444494415E-002] fbridge_mode=1 [UNWEIGHT] Wrote 1589 events (found 1593 events) - [COUNTERS] PROGRAM TOTAL : 0.2179s - [COUNTERS] Fortran Overhead ( 0 ) : 0.2143s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0033s for 8192 events => throughput is 2.49E+06 events/s + [COUNTERS] PROGRAM TOTAL : 0.2180s + [COUNTERS] Fortran Overhead ( 0 ) : 0.2145s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0032s for 8192 events => throughput is 2.54E+06 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0003s *** (2-512y) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -257,15 +387,50 @@ OK! xsec from fortran (9.2432789448173985E-002) and cpp (9.2432789444494415E-002 OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-512y) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.512y_m_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_eemumu_x10_cudacpp > /tmp/avalassi/output_eemumu_x10_cudacpp' +DEBUG: MEK processed 81920 events across 2 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 4/16 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 0.09171 [9.1711103899063451E-002] fbridge_mode=1 + [UNWEIGHT] Wrote 1655 events (found 1660 events) + [COUNTERS] PROGRAM TOTAL : 0.6837s + [COUNTERS] Fortran Overhead ( 0 ) : 0.6513s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0321s for 81920 events => throughput is 2.55E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0003s + +*** (2-512y) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (9.1711103909519892E-002) and cpp (9.1711103899063451E-002) differ by less than 2E-4 (1.1401501964769523e-10) + +*** (2-512y) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+BRDHST/512y+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 2.606715e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.660018e+06 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 2.748967e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.795960e+06 ) sec^-1 *** (2-512z) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -289,9 +454,9 @@ DEBUG: MEK processed 8192 events across 2 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 0.09243 [9.2432789444494415E-002] fbridge_mode=1 [UNWEIGHT] Wrote 1589 events (found 1593 events) - [COUNTERS] PROGRAM TOTAL : 0.2165s - [COUNTERS] Fortran Overhead ( 0 ) : 0.2123s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0039s for 8192 events => throughput is 2.10E+06 events/s + [COUNTERS] PROGRAM TOTAL : 0.2177s + [COUNTERS] Fortran Overhead ( 0 ) : 0.2135s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0039s for 8192 events => throughput is 2.08E+06 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0003s *** (2-512z) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -302,15 +467,50 @@ OK! xsec from fortran (9.2432789448173985E-002) and cpp (9.2432789444494415E-002 OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-512z) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.512z_m_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_eemumu_x10_cudacpp > /tmp/avalassi/output_eemumu_x10_cudacpp' +DEBUG: MEK processed 81920 events across 2 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 4/16 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 0.09171 [9.1711103899063451E-002] fbridge_mode=1 + [UNWEIGHT] Wrote 1655 events (found 1660 events) + [COUNTERS] PROGRAM TOTAL : 0.7075s + [COUNTERS] Fortran Overhead ( 0 ) : 0.6689s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0383s for 81920 events => throughput is 2.14E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0003s + +*** (2-512z) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (9.1711103909519892E-002) and cpp (9.1711103899063451E-002) differ by less than 2E-4 (1.1401501964769523e-10) + +*** (2-512z) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+BRDHST/512z+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 2.203720e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.206357e+06 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 2.284212e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.274939e+06 ) sec^-1 *** (3-cuda) EXECUTE MADEVENT_CUDA x1 (create events.lhe) *** -------------------- @@ -334,9 +534,9 @@ DEBUG: MEK processed 8192 events across 2 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 0.09243 [9.2432789437826970E-002] fbridge_mode=1 [UNWEIGHT] Wrote 1589 events (found 1593 events) - [COUNTERS] PROGRAM TOTAL : 0.6505s - [COUNTERS] Fortran Overhead ( 0 ) : 0.6470s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0030s for 8192 events => throughput is 2.77E+06 events/s + [COUNTERS] PROGRAM TOTAL : 0.6523s + [COUNTERS] Fortran Overhead ( 0 ) : 0.6489s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0028s for 8192 events => throughput is 2.96E+06 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0006s *** (3-cuda) Compare MADEVENT_CUDA x1 xsec to MADEVENT_FORTRAN xsec *** @@ -347,45 +547,80 @@ OK! xsec from fortran (9.2432789448173985E-002) and cuda (9.2432789437826970E-00 OK! events.lhe.cuda.1 and events.lhe.ref.1 are identical +*** (3-cuda) EXECUTE MADEVENT_CUDA x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.cuda_m_inl0_hrd0/madevent_cuda < /tmp/avalassi/input_eemumu_x10_cudacpp > /tmp/avalassi/output_eemumu_x10_cudacpp' +DEBUG: MEK processed 81920 events across 2 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 4/16 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 0.09171 [9.1711103901050417E-002] fbridge_mode=1 + [UNWEIGHT] Wrote 1655 events (found 1660 events) + [COUNTERS] PROGRAM TOTAL : 1.0910s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0833s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0072s for 81920 events => throughput is 1.14E+07 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0006s + +*** (3-cuda) Compare MADEVENT_CUDA x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (9.1711103909519892E-002) and cuda (9.1711103901050417E-002) differ by less than 2E-4 (9.234946141134515e-11) + +*** (3-cuda) Compare MADEVENT_CUDA x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cuda.10 and events.lhe.ref.10 are identical + *** EXECUTE GCHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SM_EPEM_MUPMUM_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:MIX+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.269035e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.343961e+06 ) sec^-1 *** EXECUTE GCHECK(8192) -p 256 32 1 *** Process = SIGMA_SM_EPEM_MUPMUM_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:MIX+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.550305e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.649233e+06 ) sec^-1 *** EXECUTE GCHECK(MAX) -p 16384 32 1 --bridge *** Process = SIGMA_SM_EPEM_MUPMUM_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:MIX+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.523745e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.585332e+07 ) sec^-1 *** EXECUTE GCHECK(MAX) -p 16384 32 1 *** Process = SIGMA_SM_EPEM_MUPMUM_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:MIX+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.857337e+08 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.881833e+08 ) sec^-1 *** EXECUTE GCHECK(MAX128THR) -p 4096 128 1 --bridge *** Process = SIGMA_SM_EPEM_MUPMUM_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:MIX+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.551254e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.584126e+07 ) sec^-1 *** EXECUTE GCHECK(MAX128THR) -p 4096 128 1 *** Process = SIGMA_SM_EPEM_MUPMUM_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:MIX+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.897534e+08 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.876758e+08 ) sec^-1 *** EXECUTE GCHECK(MAX8THR) -p 65536 8 1 --bridge *** Process = SIGMA_SM_EPEM_MUPMUM_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:MIX+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.503798e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.560231e+07 ) sec^-1 *** EXECUTE GCHECK(MAX8THR) -p 65536 8 1 *** Process = SIGMA_SM_EPEM_MUPMUM_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:MIX+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.184430e+08 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.174540e+08 ) sec^-1 *** (3-hip) WARNING! SKIP MADEVENT_HIP (hip is not supported on this node) *** diff --git a/epochX/cudacpp/tmad/logs_ggtt_mad/log_ggtt_mad_d_inl0_hrd0.txt b/epochX/cudacpp/tmad/logs_ggtt_mad/log_ggtt_mad_d_inl0_hrd0.txt index 794f102690..1b6011dabe 100644 --- a/epochX/cudacpp/tmad/logs_ggtt_mad/log_ggtt_mad_d_inl0_hrd0.txt +++ b/epochX/cudacpp/tmad/logs_ggtt_mad/log_ggtt_mad_d_inl0_hrd0.txt @@ -4,9 +4,9 @@ make USEBUILDDIR=1 BACKEND=cuda make USEBUILDDIR=1 BACKEND=cppnone - make USEBUILDDIR=1 BACKEND=cppsse4 + make USEBUILDDIR=1 BACKEND=cppavx2 make USEBUILDDIR=1 BACKEND=cpp512y make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx' @@ -32,7 +32,7 @@ make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ OMP_NUM_THREADS= -DATE: 2024-10-06_10:04:53 +DATE: 2024-10-06_14:21:05 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: Working directory (run): /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx @@ -58,9 +58,9 @@ Executing ' ./madevent_fortran < /tmp/avalassi/input_ggtt_x1_fortran > /tmp/aval [XSECTION] ChannelId = 1 [XSECTION] Cross section = 47.14 [47.138611968034162] fbridge_mode=0 [UNWEIGHT] Wrote 2613 events (found 5374 events) - [COUNTERS] PROGRAM TOTAL : 0.8494s - [COUNTERS] Fortran Overhead ( 0 ) : 0.8073s - [COUNTERS] Fortran MEs ( 1 ) : 0.0421s for 8192 events => throughput is 1.94E+05 events/s + [COUNTERS] PROGRAM TOTAL : 0.8395s + [COUNTERS] Fortran Overhead ( 0 ) : 0.7969s + [COUNTERS] Fortran MEs ( 1 ) : 0.0426s for 8192 events => throughput is 1.92E+05 events/s *** (1) EXECUTE MADEVENT_FORTRAN x1 (create events.lhe) *** -------------------- @@ -83,10 +83,35 @@ Executing ' ./madevent_fortran < /tmp/avalassi/input_ggtt_x1_fortran > /tmp/aval [XSECTION] ChannelId = 1 [XSECTION] Cross section = 47.14 [47.138611968034162] fbridge_mode=0 [UNWEIGHT] Wrote 1618 events (found 1623 events) - [COUNTERS] PROGRAM TOTAL : 0.4510s - [COUNTERS] Fortran Overhead ( 0 ) : 0.4085s + [COUNTERS] PROGRAM TOTAL : 0.4503s + [COUNTERS] Fortran Overhead ( 0 ) : 0.4078s [COUNTERS] Fortran MEs ( 1 ) : 0.0425s for 8192 events => throughput is 1.93E+05 events/s +*** (1) EXECUTE MADEVENT_FORTRAN x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./madevent_fortran < /tmp/avalassi/input_ggtt_x10_fortran > /tmp/avalassi/output_ggtt_x10_fortran' + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 16/16 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 47.14 [47.144596232268185] fbridge_mode=0 + [UNWEIGHT] Wrote 1613 events (found 1618 events) + [COUNTERS] PROGRAM TOTAL : 2.0110s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5878s + [COUNTERS] Fortran MEs ( 1 ) : 0.4233s for 81920 events => throughput is 1.94E+05 events/s + *** (2-none) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- CUDACPP_RUNTIME_FBRIDGEMODE = (not set) @@ -109,9 +134,9 @@ DEBUG: MEK processed 8192 events across 3 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 47.14 [47.138611968034162] fbridge_mode=1 [UNWEIGHT] Wrote 1618 events (found 1623 events) - [COUNTERS] PROGRAM TOTAL : 0.4555s - [COUNTERS] Fortran Overhead ( 0 ) : 0.4098s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0453s for 8192 events => throughput is 1.81E+05 events/s + [COUNTERS] PROGRAM TOTAL : 0.4526s + [COUNTERS] Fortran Overhead ( 0 ) : 0.4072s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0450s for 8192 events => throughput is 1.82E+05 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0004s *** (2-none) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -122,15 +147,50 @@ OK! xsec from fortran (47.138611968034162) and cpp (47.138611968034162) differ b OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-none) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.none_d_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_ggtt_x10_cudacpp > /tmp/avalassi/output_ggtt_x10_cudacpp' +DEBUG: MEK processed 81920 events across 3 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 16/16 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 47.14 [47.144596232268192] fbridge_mode=1 + [UNWEIGHT] Wrote 1613 events (found 1618 events) + [COUNTERS] PROGRAM TOTAL : 2.0260s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5803s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4452s for 81920 events => throughput is 1.84E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0004s + +*** (2-none) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (47.144596232268185) and cpp (47.144596232268192) differ by less than 3E-14 (2.220446049250313e-16) + +*** (2-none) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+BRDHST/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.856020e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.825193e+05 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.865986e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.862678e+05 ) sec^-1 *** (2-sse4) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -154,10 +214,10 @@ DEBUG: MEK processed 8192 events across 3 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 47.14 [47.138611968034162] fbridge_mode=1 [UNWEIGHT] Wrote 1618 events (found 1623 events) - [COUNTERS] PROGRAM TOTAL : 0.4352s - [COUNTERS] Fortran Overhead ( 0 ) : 0.4103s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0245s for 8192 events => throughput is 3.34E+05 events/s - [COUNTERS] CudaCpp HEL ( 3 ) : 0.0003s + [COUNTERS] PROGRAM TOTAL : 0.4321s + [COUNTERS] Fortran Overhead ( 0 ) : 0.4064s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0254s for 8192 events => throughput is 3.22E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0004s *** (2-sse4) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -167,15 +227,50 @@ OK! xsec from fortran (47.138611968034162) and cpp (47.138611968034162) differ b OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-sse4) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.sse4_d_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_ggtt_x10_cudacpp > /tmp/avalassi/output_ggtt_x10_cudacpp' +DEBUG: MEK processed 81920 events across 3 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 16/16 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 47.14 [47.144596232268192] fbridge_mode=1 + [UNWEIGHT] Wrote 1613 events (found 1618 events) + [COUNTERS] PROGRAM TOTAL : 1.8263s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5750s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2510s for 81920 events => throughput is 3.26E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0004s + +*** (2-sse4) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (47.144596232268185) and cpp (47.144596232268192) differ by less than 3E-14 (2.220446049250313e-16) + +*** (2-sse4) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+BRDHST/sse4+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.314758e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.271165e+05 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.321531e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.297184e+05 ) sec^-1 *** (2-avx2) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -199,10 +294,10 @@ DEBUG: MEK processed 8192 events across 3 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 47.14 [47.138611968034162] fbridge_mode=1 [UNWEIGHT] Wrote 1618 events (found 1623 events) - [COUNTERS] PROGRAM TOTAL : 0.4235s - [COUNTERS] Fortran Overhead ( 0 ) : 0.4077s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0154s for 8192 events => throughput is 5.30E+05 events/s - [COUNTERS] CudaCpp HEL ( 3 ) : 0.0003s + [COUNTERS] PROGRAM TOTAL : 0.4318s + [COUNTERS] Fortran Overhead ( 0 ) : 0.4157s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0157s for 8192 events => throughput is 5.23E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0004s *** (2-avx2) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -212,15 +307,50 @@ OK! xsec from fortran (47.138611968034162) and cpp (47.138611968034162) differ b OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-avx2) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.avx2_d_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_ggtt_x10_cudacpp > /tmp/avalassi/output_ggtt_x10_cudacpp' +DEBUG: MEK processed 81920 events across 3 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 16/16 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 47.14 [47.144596232268192] fbridge_mode=1 + [UNWEIGHT] Wrote 1613 events (found 1618 events) + [COUNTERS] PROGRAM TOTAL : 1.7323s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5756s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1564s for 81920 events => throughput is 5.24E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0003s + +*** (2-avx2) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (47.144596232268185) and cpp (47.144596232268192) differ by less than 3E-14 (2.220446049250313e-16) + +*** (2-avx2) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+BRDHST/avx2+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 5.263509e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 5.036550e+05 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 5.327379e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 5.336415e+05 ) sec^-1 *** (2-512y) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -244,10 +374,10 @@ DEBUG: MEK processed 8192 events across 3 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 47.14 [47.138611968034162] fbridge_mode=1 [UNWEIGHT] Wrote 1618 events (found 1623 events) - [COUNTERS] PROGRAM TOTAL : 0.4237s - [COUNTERS] Fortran Overhead ( 0 ) : 0.4087s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0147s for 8192 events => throughput is 5.58E+05 events/s - [COUNTERS] CudaCpp HEL ( 3 ) : 0.0004s + [COUNTERS] PROGRAM TOTAL : 0.4215s + [COUNTERS] Fortran Overhead ( 0 ) : 0.4072s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0139s for 8192 events => throughput is 5.88E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0003s *** (2-512y) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -257,15 +387,50 @@ OK! xsec from fortran (47.138611968034162) and cpp (47.138611968034162) differ b OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-512y) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.512y_d_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_ggtt_x10_cudacpp > /tmp/avalassi/output_ggtt_x10_cudacpp' +DEBUG: MEK processed 81920 events across 3 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 16/16 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 47.14 [47.144596232268192] fbridge_mode=1 + [UNWEIGHT] Wrote 1613 events (found 1618 events) + [COUNTERS] PROGRAM TOTAL : 1.7384s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5933s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1448s for 81920 events => throughput is 5.66E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0004s + +*** (2-512y) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (47.144596232268185) and cpp (47.144596232268192) differ by less than 3E-14 (2.220446049250313e-16) + +*** (2-512y) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+BRDHST/512y+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 5.648502e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 5.691535e+05 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 5.831851e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 5.836325e+05 ) sec^-1 *** (2-512z) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -289,9 +454,9 @@ DEBUG: MEK processed 8192 events across 3 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 47.14 [47.138611968034169] fbridge_mode=1 [UNWEIGHT] Wrote 1618 events (found 1623 events) - [COUNTERS] PROGRAM TOTAL : 0.4297s - [COUNTERS] Fortran Overhead ( 0 ) : 0.4071s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0222s for 8192 events => throughput is 3.69E+05 events/s + [COUNTERS] PROGRAM TOTAL : 0.4296s + [COUNTERS] Fortran Overhead ( 0 ) : 0.4066s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0226s for 8192 events => throughput is 3.62E+05 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0004s *** (2-512z) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -302,15 +467,50 @@ OK! xsec from fortran (47.138611968034162) and cpp (47.138611968034169) differ b OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-512z) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.512z_d_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_ggtt_x10_cudacpp > /tmp/avalassi/output_ggtt_x10_cudacpp' +DEBUG: MEK processed 81920 events across 3 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 16/16 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 47.14 [47.144596232268192] fbridge_mode=1 + [UNWEIGHT] Wrote 1613 events (found 1618 events) + [COUNTERS] PROGRAM TOTAL : 1.8095s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5822s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2268s for 81920 events => throughput is 3.61E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0004s + +*** (2-512z) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (47.144596232268185) and cpp (47.144596232268192) differ by less than 3E-14 (2.220446049250313e-16) + +*** (2-512z) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+BRDHST/512z+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.526689e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.604004e+05 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.574003e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.623036e+05 ) sec^-1 *** (3-cuda) EXECUTE MADEVENT_CUDA x1 (create events.lhe) *** -------------------- @@ -334,9 +534,9 @@ DEBUG: MEK processed 8192 events across 3 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 47.14 [47.138611968034176] fbridge_mode=1 [UNWEIGHT] Wrote 1618 events (found 1623 events) - [COUNTERS] PROGRAM TOTAL : 0.8534s - [COUNTERS] Fortran Overhead ( 0 ) : 0.8496s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0031s for 8192 events => throughput is 2.68E+06 events/s + [COUNTERS] PROGRAM TOTAL : 0.8540s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8503s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0029s for 8192 events => throughput is 2.85E+06 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0008s *** (3-cuda) Compare MADEVENT_CUDA x1 xsec to MADEVENT_FORTRAN xsec *** @@ -347,45 +547,80 @@ OK! xsec from fortran (47.138611968034162) and cuda (47.138611968034176) differ OK! events.lhe.cuda.1 and events.lhe.ref.1 are identical +*** (3-cuda) EXECUTE MADEVENT_CUDA x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.cuda_d_inl0_hrd0/madevent_cuda < /tmp/avalassi/input_ggtt_x10_cudacpp > /tmp/avalassi/output_ggtt_x10_cudacpp' +DEBUG: MEK processed 81920 events across 3 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 16/16 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 47.14 [47.144596232268178] fbridge_mode=1 + [UNWEIGHT] Wrote 1613 events (found 1618 events) + [COUNTERS] PROGRAM TOTAL : 2.0288s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0193s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0088s for 81920 events => throughput is 9.35E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0008s + +*** (3-cuda) Compare MADEVENT_CUDA x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (47.144596232268185) and cuda (47.144596232268178) differ by less than 3E-14 (1.1102230246251565e-16) + +*** (3-cuda) Compare MADEVENT_CUDA x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cuda.10 and events.lhe.ref.10 are identical + *** EXECUTE GCHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SM_GG_TTX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.103830e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.178023e+06 ) sec^-1 *** EXECUTE GCHECK(8192) -p 256 32 1 *** Process = SIGMA_SM_GG_TTX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.448285e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.430378e+06 ) sec^-1 *** EXECUTE GCHECK(MAX) -p 16384 32 1 --bridge *** Process = SIGMA_SM_GG_TTX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 2.875229e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.911902e+07 ) sec^-1 *** EXECUTE GCHECK(MAX) -p 16384 32 1 *** Process = SIGMA_SM_GG_TTX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 7.627647e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 7.667950e+07 ) sec^-1 *** EXECUTE GCHECK(MAX128THR) -p 4096 128 1 --bridge *** Process = SIGMA_SM_GG_TTX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 2.886865e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.950961e+07 ) sec^-1 *** EXECUTE GCHECK(MAX128THR) -p 4096 128 1 *** Process = SIGMA_SM_GG_TTX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 8.006782e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 7.963133e+07 ) sec^-1 *** EXECUTE GCHECK(MAX8THR) -p 65536 8 1 --bridge *** Process = SIGMA_SM_GG_TTX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 2.862106e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.904002e+07 ) sec^-1 *** EXECUTE GCHECK(MAX8THR) -p 65536 8 1 *** Process = SIGMA_SM_GG_TTX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 2.715892e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.713805e+07 ) sec^-1 *** (3-hip) WARNING! SKIP MADEVENT_HIP (hip is not supported on this node) *** diff --git a/epochX/cudacpp/tmad/logs_ggtt_mad/log_ggtt_mad_f_inl0_hrd0.txt b/epochX/cudacpp/tmad/logs_ggtt_mad/log_ggtt_mad_f_inl0_hrd0.txt index 2bf2a37cc7..6eaf6df17c 100644 --- a/epochX/cudacpp/tmad/logs_ggtt_mad/log_ggtt_mad_f_inl0_hrd0.txt +++ b/epochX/cudacpp/tmad/logs_ggtt_mad/log_ggtt_mad_f_inl0_hrd0.txt @@ -13,8 +13,8 @@ make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx' make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx' make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx' -make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx' +make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx' make USEBUILDDIR=1 BACKEND=cpp512z make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx' make[1]: Nothing to be done for 'all'. @@ -32,7 +32,7 @@ make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ OMP_NUM_THREADS= -DATE: 2024-10-06_10:05:22 +DATE: 2024-10-06_14:22:02 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: Working directory (run): /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx @@ -58,9 +58,9 @@ Executing ' ./madevent_fortran < /tmp/avalassi/input_ggtt_x1_fortran > /tmp/aval [XSECTION] ChannelId = 1 [XSECTION] Cross section = 47.14 [47.138611968034162] fbridge_mode=0 [UNWEIGHT] Wrote 2613 events (found 5374 events) - [COUNTERS] PROGRAM TOTAL : 0.8450s - [COUNTERS] Fortran Overhead ( 0 ) : 0.8027s - [COUNTERS] Fortran MEs ( 1 ) : 0.0423s for 8192 events => throughput is 1.94E+05 events/s + [COUNTERS] PROGRAM TOTAL : 0.8382s + [COUNTERS] Fortran Overhead ( 0 ) : 0.7954s + [COUNTERS] Fortran MEs ( 1 ) : 0.0428s for 8192 events => throughput is 1.91E+05 events/s *** (1) EXECUTE MADEVENT_FORTRAN x1 (create events.lhe) *** -------------------- @@ -84,8 +84,33 @@ Executing ' ./madevent_fortran < /tmp/avalassi/input_ggtt_x1_fortran > /tmp/aval [XSECTION] Cross section = 47.14 [47.138611968034162] fbridge_mode=0 [UNWEIGHT] Wrote 1618 events (found 1623 events) [COUNTERS] PROGRAM TOTAL : 0.4516s - [COUNTERS] Fortran Overhead ( 0 ) : 0.4094s - [COUNTERS] Fortran MEs ( 1 ) : 0.0422s for 8192 events => throughput is 1.94E+05 events/s + [COUNTERS] Fortran Overhead ( 0 ) : 0.4100s + [COUNTERS] Fortran MEs ( 1 ) : 0.0416s for 8192 events => throughput is 1.97E+05 events/s + +*** (1) EXECUTE MADEVENT_FORTRAN x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./madevent_fortran < /tmp/avalassi/input_ggtt_x10_fortran > /tmp/avalassi/output_ggtt_x10_fortran' + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 16/16 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 47.14 [47.144596232268185] fbridge_mode=0 + [UNWEIGHT] Wrote 1613 events (found 1618 events) + [COUNTERS] PROGRAM TOTAL : 1.9980s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5729s + [COUNTERS] Fortran MEs ( 1 ) : 0.4251s for 81920 events => throughput is 1.93E+05 events/s *** (2-none) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -109,9 +134,9 @@ DEBUG: MEK processed 8192 events across 3 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 47.14 [47.138606099989779] fbridge_mode=1 [UNWEIGHT] Wrote 1618 events (found 1623 events) - [COUNTERS] PROGRAM TOTAL : 0.4548s - [COUNTERS] Fortran Overhead ( 0 ) : 0.4118s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0426s for 8192 events => throughput is 1.92E+05 events/s + [COUNTERS] PROGRAM TOTAL : 0.4508s + [COUNTERS] Fortran Overhead ( 0 ) : 0.4081s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0423s for 8192 events => throughput is 1.93E+05 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0003s *** (2-none) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -122,15 +147,50 @@ OK! xsec from fortran (47.138611968034162) and cpp (47.138606099989779) differ b OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-none) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.none_f_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_ggtt_x10_cudacpp > /tmp/avalassi/output_ggtt_x10_cudacpp' +DEBUG: MEK processed 81920 events across 3 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 16/16 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 47.14 [47.144592707001024] fbridge_mode=1 + [UNWEIGHT] Wrote 1613 events (found 1618 events) + [COUNTERS] PROGRAM TOTAL : 1.9904s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5707s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4193s for 81920 events => throughput is 1.95E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0003s + +*** (2-none) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (47.144596232268185) and cpp (47.144592707001024) differ by less than 4E-4 (7.477563590541081e-08) + +*** (2-none) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+BRDHST/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.973574e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.971575e+05 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.981282e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.978955e+05 ) sec^-1 *** (2-sse4) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -154,10 +214,10 @@ DEBUG: MEK processed 8192 events across 3 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 47.14 [47.138602111070696] fbridge_mode=1 [UNWEIGHT] Wrote 1618 events (found 1623 events) - [COUNTERS] PROGRAM TOTAL : 0.4326s - [COUNTERS] Fortran Overhead ( 0 ) : 0.4154s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0170s for 8192 events => throughput is 4.81E+05 events/s - [COUNTERS] CudaCpp HEL ( 3 ) : 0.0002s + [COUNTERS] PROGRAM TOTAL : 0.4226s + [COUNTERS] Fortran Overhead ( 0 ) : 0.4059s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0164s for 8192 events => throughput is 4.98E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0003s *** (2-sse4) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -167,15 +227,50 @@ OK! xsec from fortran (47.138611968034162) and cpp (47.138602111070696) differ b OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-sse4) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.sse4_f_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_ggtt_x10_cudacpp > /tmp/avalassi/output_ggtt_x10_cudacpp' +DEBUG: MEK processed 81920 events across 3 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 16/16 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 47.14 [47.144588828412729] fbridge_mode=1 + [UNWEIGHT] Wrote 1613 events (found 1618 events) + [COUNTERS] PROGRAM TOTAL : 1.7470s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5749s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1719s for 81920 events => throughput is 4.77E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0002s + +*** (2-sse4) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (47.144596232268185) and cpp (47.144588828412729) differ by less than 4E-4 (1.570456860111591e-07) + +*** (2-sse4) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+BRDHST/sse4+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 4.659841e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 4.717994e+05 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 4.743814e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 4.704729e+05 ) sec^-1 *** (2-avx2) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -201,8 +296,8 @@ DEBUG: MEK processed 8192 events across 3 channels { 1 : 8192 } [UNWEIGHT] Wrote 1618 events (found 1623 events) [COUNTERS] PROGRAM TOTAL : 0.4174s [COUNTERS] Fortran Overhead ( 0 ) : 0.4080s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0091s for 8192 events => throughput is 8.97E+05 events/s - [COUNTERS] CudaCpp HEL ( 3 ) : 0.0003s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0091s for 8192 events => throughput is 8.96E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0002s *** (2-avx2) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -212,15 +307,50 @@ OK! xsec from fortran (47.138611968034162) and cpp (47.138602499179925) differ b OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-avx2) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.avx2_f_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_ggtt_x10_cudacpp > /tmp/avalassi/output_ggtt_x10_cudacpp' +DEBUG: MEK processed 81920 events across 3 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 16/16 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 47.14 [47.144586996341530] fbridge_mode=1 + [UNWEIGHT] Wrote 1613 events (found 1618 events) + [COUNTERS] PROGRAM TOTAL : 1.6555s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5656s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0897s for 81920 events => throughput is 9.13E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0002s + +*** (2-avx2) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (47.144596232268185) and cpp (47.144586996341530) differ by less than 4E-4 (1.9590636879396328e-07) + +*** (2-avx2) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+BRDHST/avx2+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 9.079796e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 9.110993e+05 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 9.235810e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 9.307337e+05 ) sec^-1 *** (2-512y) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -244,10 +374,10 @@ DEBUG: MEK processed 8192 events across 3 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 47.14 [47.138602499179925] fbridge_mode=1 [UNWEIGHT] Wrote 1618 events (found 1623 events) - [COUNTERS] PROGRAM TOTAL : 0.4181s - [COUNTERS] Fortran Overhead ( 0 ) : 0.4092s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0086s for 8192 events => throughput is 9.50E+05 events/s - [COUNTERS] CudaCpp HEL ( 3 ) : 0.0003s + [COUNTERS] PROGRAM TOTAL : 0.4137s + [COUNTERS] Fortran Overhead ( 0 ) : 0.4049s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0086s for 8192 events => throughput is 9.54E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0002s *** (2-512y) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -257,15 +387,50 @@ OK! xsec from fortran (47.138611968034162) and cpp (47.138602499179925) differ b OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-512y) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.512y_f_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_ggtt_x10_cudacpp > /tmp/avalassi/output_ggtt_x10_cudacpp' +DEBUG: MEK processed 81920 events across 3 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 16/16 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 47.14 [47.144586996341530] fbridge_mode=1 + [UNWEIGHT] Wrote 1613 events (found 1618 events) + [COUNTERS] PROGRAM TOTAL : 1.6525s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5673s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0849s for 81920 events => throughput is 9.65E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0002s + +*** (2-512y) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (47.144596232268185) and cpp (47.144586996341530) differ by less than 4E-4 (1.9590636879396328e-07) + +*** (2-512y) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+BRDHST/512y+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 8.970038e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 9.842321e+05 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 9.765544e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 9.815945e+05 ) sec^-1 *** (2-512z) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -289,9 +454,9 @@ DEBUG: MEK processed 8192 events across 3 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 47.14 [47.138606840950104] fbridge_mode=1 [UNWEIGHT] Wrote 1618 events (found 1623 events) - [COUNTERS] PROGRAM TOTAL : 0.4258s - [COUNTERS] Fortran Overhead ( 0 ) : 0.4131s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0124s for 8192 events => throughput is 6.60E+05 events/s + [COUNTERS] PROGRAM TOTAL : 0.4207s + [COUNTERS] Fortran Overhead ( 0 ) : 0.4079s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0125s for 8192 events => throughput is 6.53E+05 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0003s *** (2-512z) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -302,15 +467,50 @@ OK! xsec from fortran (47.138611968034162) and cpp (47.138606840950104) differ b OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-512z) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.512z_f_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_ggtt_x10_cudacpp > /tmp/avalassi/output_ggtt_x10_cudacpp' +DEBUG: MEK processed 81920 events across 3 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 16/16 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 47.14 [47.144591429357156] fbridge_mode=1 + [UNWEIGHT] Wrote 1613 events (found 1618 events) + [COUNTERS] PROGRAM TOTAL : 1.6897s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5703s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1191s for 81920 events => throughput is 6.88E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0003s + +*** (2-512z) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (47.144596232268185) and cpp (47.144591429357156) differ by less than 4E-4 (1.0187617272006122e-07) + +*** (2-512z) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+BRDHST/512z+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 6.636236e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 6.775647e+05 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 6.862568e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 6.831691e+05 ) sec^-1 *** (3-cuda) EXECUTE MADEVENT_CUDA x1 (create events.lhe) *** -------------------- @@ -334,9 +534,9 @@ DEBUG: MEK processed 8192 events across 3 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 47.14 [47.138612402172164] fbridge_mode=1 [UNWEIGHT] Wrote 1618 events (found 1623 events) - [COUNTERS] PROGRAM TOTAL : 0.8671s - [COUNTERS] Fortran Overhead ( 0 ) : 0.8634s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0031s for 8192 events => throughput is 2.65E+06 events/s + [COUNTERS] PROGRAM TOTAL : 0.8477s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8442s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0030s for 8192 events => throughput is 2.77E+06 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0006s *** (3-cuda) Compare MADEVENT_CUDA x1 xsec to MADEVENT_FORTRAN xsec *** @@ -347,45 +547,80 @@ OK! xsec from fortran (47.138611968034162) and cuda (47.138612402172164) differ OK! events.lhe.cuda.1 and events.lhe.ref.1 are identical +*** (3-cuda) EXECUTE MADEVENT_CUDA x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.cuda_f_inl0_hrd0/madevent_cuda < /tmp/avalassi/input_ggtt_x10_cudacpp > /tmp/avalassi/output_ggtt_x10_cudacpp' +DEBUG: MEK processed 81920 events across 3 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 16/16 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 47.14 [47.144596666727985] fbridge_mode=1 + [UNWEIGHT] Wrote 1613 events (found 1618 events) + [COUNTERS] PROGRAM TOTAL : 2.0131s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0048s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0077s for 81920 events => throughput is 1.06E+07 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0006s + +*** (3-cuda) Compare MADEVENT_CUDA x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (47.144596232268185) and cuda (47.144596666727985) differ by less than 4E-4 (9.215473939505614e-09) + +*** (3-cuda) Compare MADEVENT_CUDA x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cuda.10 and events.lhe.ref.10 are identical + *** EXECUTE GCHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SM_GG_TTX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.093880e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.314250e+06 ) sec^-1 *** EXECUTE GCHECK(8192) -p 256 32 1 *** Process = SIGMA_SM_GG_TTX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.450343e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.574647e+06 ) sec^-1 *** EXECUTE GCHECK(MAX) -p 16384 32 1 --bridge *** Process = SIGMA_SM_GG_TTX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 5.021092e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 5.188509e+07 ) sec^-1 *** EXECUTE GCHECK(MAX) -p 16384 32 1 *** Process = SIGMA_SM_GG_TTX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.359313e+08 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.392066e+08 ) sec^-1 *** EXECUTE GCHECK(MAX128THR) -p 4096 128 1 --bridge *** Process = SIGMA_SM_GG_TTX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 5.014796e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 5.101548e+07 ) sec^-1 *** EXECUTE GCHECK(MAX128THR) -p 4096 128 1 *** Process = SIGMA_SM_GG_TTX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.375647e+08 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.412594e+08 ) sec^-1 *** EXECUTE GCHECK(MAX8THR) -p 65536 8 1 --bridge *** Process = SIGMA_SM_GG_TTX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 4.628808e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 4.783894e+07 ) sec^-1 *** EXECUTE GCHECK(MAX8THR) -p 65536 8 1 *** Process = SIGMA_SM_GG_TTX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 7.004427e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 7.129554e+07 ) sec^-1 *** (3-hip) WARNING! SKIP MADEVENT_HIP (hip is not supported on this node) *** diff --git a/epochX/cudacpp/tmad/logs_ggtt_mad/log_ggtt_mad_m_inl0_hrd0.txt b/epochX/cudacpp/tmad/logs_ggtt_mad/log_ggtt_mad_m_inl0_hrd0.txt index 2ae843d323..7e0b4ed09b 100644 --- a/epochX/cudacpp/tmad/logs_ggtt_mad/log_ggtt_mad_m_inl0_hrd0.txt +++ b/epochX/cudacpp/tmad/logs_ggtt_mad/log_ggtt_mad_m_inl0_hrd0.txt @@ -6,15 +6,15 @@ make USEBUILDDIR=1 BACKEND=cuda make USEBUILDDIR=1 BACKEND=cppnone make USEBUILDDIR=1 BACKEND=cppsse4 - make USEBUILDDIR=1 BACKEND=cppavx2 + make USEBUILDDIR=1 BACKEND=cpp512y make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx' make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx' make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx' + make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx' make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx' - make USEBUILDDIR=1 BACKEND=cpp512z make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx' make[1]: Nothing to be done for 'all'. @@ -32,7 +32,7 @@ make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ OMP_NUM_THREADS= -DATE: 2024-10-06_10:05:08 +DATE: 2024-10-06_14:21:33 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: Working directory (run): /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx @@ -58,9 +58,9 @@ Executing ' ./madevent_fortran < /tmp/avalassi/input_ggtt_x1_fortran > /tmp/aval [XSECTION] ChannelId = 1 [XSECTION] Cross section = 47.14 [47.138611968034162] fbridge_mode=0 [UNWEIGHT] Wrote 2613 events (found 5374 events) - [COUNTERS] PROGRAM TOTAL : 0.8439s - [COUNTERS] Fortran Overhead ( 0 ) : 0.8015s - [COUNTERS] Fortran MEs ( 1 ) : 0.0424s for 8192 events => throughput is 1.93E+05 events/s + [COUNTERS] PROGRAM TOTAL : 0.8429s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8008s + [COUNTERS] Fortran MEs ( 1 ) : 0.0421s for 8192 events => throughput is 1.95E+05 events/s *** (1) EXECUTE MADEVENT_FORTRAN x1 (create events.lhe) *** -------------------- @@ -83,9 +83,34 @@ Executing ' ./madevent_fortran < /tmp/avalassi/input_ggtt_x1_fortran > /tmp/aval [XSECTION] ChannelId = 1 [XSECTION] Cross section = 47.14 [47.138611968034162] fbridge_mode=0 [UNWEIGHT] Wrote 1618 events (found 1623 events) - [COUNTERS] PROGRAM TOTAL : 0.4498s - [COUNTERS] Fortran Overhead ( 0 ) : 0.4080s - [COUNTERS] Fortran MEs ( 1 ) : 0.0418s for 8192 events => throughput is 1.96E+05 events/s + [COUNTERS] PROGRAM TOTAL : 0.4561s + [COUNTERS] Fortran Overhead ( 0 ) : 0.4126s + [COUNTERS] Fortran MEs ( 1 ) : 0.0436s for 8192 events => throughput is 1.88E+05 events/s + +*** (1) EXECUTE MADEVENT_FORTRAN x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./madevent_fortran < /tmp/avalassi/input_ggtt_x10_fortran > /tmp/avalassi/output_ggtt_x10_fortran' + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 16/16 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 47.14 [47.144596232268185] fbridge_mode=0 + [UNWEIGHT] Wrote 1613 events (found 1618 events) + [COUNTERS] PROGRAM TOTAL : 1.9923s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5731s + [COUNTERS] Fortran MEs ( 1 ) : 0.4193s for 81920 events => throughput is 1.95E+05 events/s *** (2-none) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -109,9 +134,9 @@ DEBUG: MEK processed 8192 events across 3 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 47.14 [47.138613306947967] fbridge_mode=1 [UNWEIGHT] Wrote 1618 events (found 1623 events) - [COUNTERS] PROGRAM TOTAL : 0.4576s - [COUNTERS] Fortran Overhead ( 0 ) : 0.4121s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0451s for 8192 events => throughput is 1.82E+05 events/s + [COUNTERS] PROGRAM TOTAL : 0.4597s + [COUNTERS] Fortran Overhead ( 0 ) : 0.4132s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0461s for 8192 events => throughput is 1.78E+05 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0004s *** (2-none) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -122,15 +147,50 @@ OK! xsec from fortran (47.138611968034162) and cpp (47.138613306947967) differ b OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-none) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.none_m_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_ggtt_x10_cudacpp > /tmp/avalassi/output_ggtt_x10_cudacpp' +DEBUG: MEK processed 81920 events across 3 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 16/16 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 47.14 [47.144597573367548] fbridge_mode=1 + [UNWEIGHT] Wrote 1613 events (found 1618 events) + [COUNTERS] PROGRAM TOTAL : 2.0410s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5866s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4540s for 81920 events => throughput is 1.80E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0004s + +*** (2-none) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (47.144596232268185) and cpp (47.144597573367548) differ by less than 2E-4 (2.8446512922997158e-08) + +*** (2-none) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+BRDHST/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.815647e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.842865e+05 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.845071e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.842403e+05 ) sec^-1 *** (2-sse4) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -154,9 +214,9 @@ DEBUG: MEK processed 8192 events across 3 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 47.14 [47.138613306947953] fbridge_mode=1 [UNWEIGHT] Wrote 1618 events (found 1623 events) - [COUNTERS] PROGRAM TOTAL : 0.4358s - [COUNTERS] Fortran Overhead ( 0 ) : 0.4106s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0248s for 8192 events => throughput is 3.31E+05 events/s + [COUNTERS] PROGRAM TOTAL : 0.4327s + [COUNTERS] Fortran Overhead ( 0 ) : 0.4078s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0246s for 8192 events => throughput is 3.33E+05 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0004s *** (2-sse4) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -167,15 +227,50 @@ OK! xsec from fortran (47.138611968034162) and cpp (47.138613306947953) differ b OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-sse4) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.sse4_m_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_ggtt_x10_cudacpp > /tmp/avalassi/output_ggtt_x10_cudacpp' +DEBUG: MEK processed 81920 events across 3 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 16/16 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 47.14 [47.144597573367555] fbridge_mode=1 + [UNWEIGHT] Wrote 1613 events (found 1618 events) + [COUNTERS] PROGRAM TOTAL : 1.8103s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5636s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2464s for 81920 events => throughput is 3.33E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0004s + +*** (2-sse4) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (47.144596232268185) and cpp (47.144597573367555) differ by less than 2E-4 (2.8446512922997158e-08) + +*** (2-sse4) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+BRDHST/sse4+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.291111e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.348496e+05 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.339005e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.329623e+05 ) sec^-1 *** (2-avx2) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -199,9 +294,9 @@ DEBUG: MEK processed 8192 events across 3 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 47.14 [47.138613350418019] fbridge_mode=1 [UNWEIGHT] Wrote 1618 events (found 1623 events) - [COUNTERS] PROGRAM TOTAL : 0.4251s - [COUNTERS] Fortran Overhead ( 0 ) : 0.4094s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0154s for 8192 events => throughput is 5.31E+05 events/s + [COUNTERS] PROGRAM TOTAL : 0.4243s + [COUNTERS] Fortran Overhead ( 0 ) : 0.4084s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0156s for 8192 events => throughput is 5.27E+05 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0004s *** (2-avx2) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -212,15 +307,50 @@ OK! xsec from fortran (47.138611968034162) and cpp (47.138613350418019) differ b OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-avx2) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.avx2_m_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_ggtt_x10_cudacpp > /tmp/avalassi/output_ggtt_x10_cudacpp' +DEBUG: MEK processed 81920 events across 3 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 16/16 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 47.14 [47.144597608209963] fbridge_mode=1 + [UNWEIGHT] Wrote 1613 events (found 1618 events) + [COUNTERS] PROGRAM TOTAL : 1.7328s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5764s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1561s for 81920 events => throughput is 5.25E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0003s + +*** (2-avx2) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (47.144596232268185) and cpp (47.144597608209963) differ by less than 2E-4 (2.9185567074208052e-08) + +*** (2-avx2) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+BRDHST/avx2+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 5.315398e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 5.305586e+05 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 5.422217e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 5.367891e+05 ) sec^-1 *** (2-512y) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -244,10 +374,10 @@ DEBUG: MEK processed 8192 events across 3 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 47.14 [47.138613350418019] fbridge_mode=1 [UNWEIGHT] Wrote 1618 events (found 1623 events) - [COUNTERS] PROGRAM TOTAL : 0.4227s - [COUNTERS] Fortran Overhead ( 0 ) : 0.4081s + [COUNTERS] PROGRAM TOTAL : 0.4269s + [COUNTERS] Fortran Overhead ( 0 ) : 0.4123s [COUNTERS] CudaCpp MEs ( 2 ) : 0.0143s for 8192 events => throughput is 5.73E+05 events/s - [COUNTERS] CudaCpp HEL ( 3 ) : 0.0003s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0004s *** (2-512y) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -257,15 +387,50 @@ OK! xsec from fortran (47.138611968034162) and cpp (47.138613350418019) differ b OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-512y) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.512y_m_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_ggtt_x10_cudacpp > /tmp/avalassi/output_ggtt_x10_cudacpp' +DEBUG: MEK processed 81920 events across 3 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 16/16 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 47.14 [47.144597608209963] fbridge_mode=1 + [UNWEIGHT] Wrote 1613 events (found 1618 events) + [COUNTERS] PROGRAM TOTAL : 1.7191s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5771s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1416s for 81920 events => throughput is 5.79E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0003s + +*** (2-512y) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (47.144596232268185) and cpp (47.144597608209963) differ by less than 2E-4 (2.9185567074208052e-08) + +*** (2-512y) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+BRDHST/512y+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 5.854463e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 5.829961e+05 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 5.901611e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 5.911938e+05 ) sec^-1 *** (2-512z) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -289,9 +454,9 @@ DEBUG: MEK processed 8192 events across 3 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 47.14 [47.138613350418019] fbridge_mode=1 [UNWEIGHT] Wrote 1618 events (found 1623 events) - [COUNTERS] PROGRAM TOTAL : 0.4322s - [COUNTERS] Fortran Overhead ( 0 ) : 0.4099s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0219s for 8192 events => throughput is 3.74E+05 events/s + [COUNTERS] PROGRAM TOTAL : 0.4306s + [COUNTERS] Fortran Overhead ( 0 ) : 0.4082s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0220s for 8192 events => throughput is 3.73E+05 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0004s *** (2-512z) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -302,15 +467,50 @@ OK! xsec from fortran (47.138611968034162) and cpp (47.138613350418019) differ b OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-512z) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.512z_m_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_ggtt_x10_cudacpp > /tmp/avalassi/output_ggtt_x10_cudacpp' +DEBUG: MEK processed 81920 events across 3 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 16/16 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 47.14 [47.144597608209963] fbridge_mode=1 + [UNWEIGHT] Wrote 1613 events (found 1618 events) + [COUNTERS] PROGRAM TOTAL : 1.7983s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5785s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2194s for 81920 events => throughput is 3.73E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0004s + +*** (2-512z) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (47.144596232268185) and cpp (47.144597608209963) differ by less than 2E-4 (2.9185567074208052e-08) + +*** (2-512z) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+BRDHST/512z+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.724588e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.669588e+05 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.694617e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.694819e+05 ) sec^-1 *** (3-cuda) EXECUTE MADEVENT_CUDA x1 (create events.lhe) *** -------------------- @@ -334,9 +534,9 @@ DEBUG: MEK processed 8192 events across 3 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 47.14 [47.138611963547788] fbridge_mode=1 [UNWEIGHT] Wrote 1618 events (found 1623 events) - [COUNTERS] PROGRAM TOTAL : 0.8506s - [COUNTERS] Fortran Overhead ( 0 ) : 0.8468s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0031s for 8192 events => throughput is 2.66E+06 events/s + [COUNTERS] PROGRAM TOTAL : 0.8518s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8481s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0030s for 8192 events => throughput is 2.72E+06 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0008s *** (3-cuda) Compare MADEVENT_CUDA x1 xsec to MADEVENT_FORTRAN xsec *** @@ -347,45 +547,80 @@ OK! xsec from fortran (47.138611968034162) and cuda (47.138611963547788) differ OK! events.lhe.cuda.1 and events.lhe.ref.1 are identical +*** (3-cuda) EXECUTE MADEVENT_CUDA x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.cuda_m_inl0_hrd0/madevent_cuda < /tmp/avalassi/input_ggtt_x10_cudacpp > /tmp/avalassi/output_ggtt_x10_cudacpp' +DEBUG: MEK processed 81920 events across 3 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 16/16 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 47.14 [47.144596232269095] fbridge_mode=1 + [UNWEIGHT] Wrote 1613 events (found 1618 events) + [COUNTERS] PROGRAM TOTAL : 2.0278s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0184s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0087s for 81920 events => throughput is 9.42E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0007s + +*** (3-cuda) Compare MADEVENT_CUDA x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (47.144596232268185) and cuda (47.144596232269095) differ by less than 2E-4 (1.9317880628477724e-14) + +*** (3-cuda) Compare MADEVENT_CUDA x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cuda.10 and events.lhe.ref.10 are identical + *** EXECUTE GCHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SM_GG_TTX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:MIX+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 2.987528e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.167835e+06 ) sec^-1 *** EXECUTE GCHECK(8192) -p 256 32 1 *** Process = SIGMA_SM_GG_TTX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:MIX+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.325954e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.468388e+06 ) sec^-1 *** EXECUTE GCHECK(MAX) -p 16384 32 1 --bridge *** Process = SIGMA_SM_GG_TTX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:MIX+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 2.868584e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.936649e+07 ) sec^-1 *** EXECUTE GCHECK(MAX) -p 16384 32 1 *** Process = SIGMA_SM_GG_TTX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:MIX+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 7.589038e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 7.568680e+07 ) sec^-1 *** EXECUTE GCHECK(MAX128THR) -p 4096 128 1 --bridge *** Process = SIGMA_SM_GG_TTX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:MIX+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 2.871326e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.910795e+07 ) sec^-1 *** EXECUTE GCHECK(MAX128THR) -p 4096 128 1 *** Process = SIGMA_SM_GG_TTX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:MIX+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 7.949192e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 8.022955e+07 ) sec^-1 *** EXECUTE GCHECK(MAX8THR) -p 65536 8 1 --bridge *** Process = SIGMA_SM_GG_TTX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:MIX+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 2.873573e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.931292e+07 ) sec^-1 *** EXECUTE GCHECK(MAX8THR) -p 65536 8 1 *** Process = SIGMA_SM_GG_TTX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:MIX+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 2.717025e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.749531e+07 ) sec^-1 *** (3-hip) WARNING! SKIP MADEVENT_HIP (hip is not supported on this node) *** diff --git a/epochX/cudacpp/tmad/logs_ggttg_mad/log_ggttg_mad_d_inl0_hrd0.txt b/epochX/cudacpp/tmad/logs_ggttg_mad/log_ggttg_mad_d_inl0_hrd0.txt index 0c7ed732ed..25a2b29c72 100644 --- a/epochX/cudacpp/tmad/logs_ggttg_mad/log_ggttg_mad_d_inl0_hrd0.txt +++ b/epochX/cudacpp/tmad/logs_ggttg_mad/log_ggttg_mad_d_inl0_hrd0.txt @@ -1,20 +1,20 @@ Working directory (build): /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttg.mad/SubProcesses/P1_gg_ttxg + make USEBUILDDIR=1 BACKEND=cuda make USEBUILDDIR=1 BACKEND=cppnone - make USEBUILDDIR=1 BACKEND=cppsse4 - make USEBUILDDIR=1 BACKEND=cppavx2 + make USEBUILDDIR=1 BACKEND=cpp512y make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttg.mad/SubProcesses/P1_gg_ttxg' make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttg.mad/SubProcesses/P1_gg_ttxg' make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttg.mad/SubProcesses/P1_gg_ttxg' make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttg.mad/SubProcesses/P1_gg_ttxg' - make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttg.mad/SubProcesses/P1_gg_ttxg' + make USEBUILDDIR=1 BACKEND=cpp512z make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttg.mad/SubProcesses/P1_gg_ttxg' make[1]: Nothing to be done for 'all'. @@ -32,7 +32,7 @@ make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ OMP_NUM_THREADS= -DATE: 2024-10-06_10:05:36 +DATE: 2024-10-06_14:22:29 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: Working directory (run): /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttg.mad/SubProcesses/P1_gg_ttxg @@ -58,9 +58,9 @@ Executing ' ./madevent_fortran < /tmp/avalassi/input_ggttg_x1_fortran > /tmp/ava [XSECTION] ChannelId = 1 [XSECTION] Cross section = 0.07847 [7.8471485809748567E-002] fbridge_mode=0 [UNWEIGHT] Wrote 387 events (found 1591 events) - [COUNTERS] PROGRAM TOTAL : 0.7416s - [COUNTERS] Fortran Overhead ( 0 ) : 0.4124s - [COUNTERS] Fortran MEs ( 1 ) : 0.3292s for 8192 events => throughput is 2.49E+04 events/s + [COUNTERS] PROGRAM TOTAL : 0.7378s + [COUNTERS] Fortran Overhead ( 0 ) : 0.4073s + [COUNTERS] Fortran MEs ( 1 ) : 0.3305s for 8192 events => throughput is 2.48E+04 events/s *** (1) EXECUTE MADEVENT_FORTRAN x1 (create events.lhe) *** -------------------- @@ -83,9 +83,34 @@ Executing ' ./madevent_fortran < /tmp/avalassi/input_ggttg_x1_fortran > /tmp/ava [XSECTION] ChannelId = 1 [XSECTION] Cross section = 0.07847 [7.8471485809748567E-002] fbridge_mode=0 [UNWEIGHT] Wrote 376 events (found 1358 events) - [COUNTERS] PROGRAM TOTAL : 0.7177s - [COUNTERS] Fortran Overhead ( 0 ) : 0.3873s - [COUNTERS] Fortran MEs ( 1 ) : 0.3304s for 8192 events => throughput is 2.48E+04 events/s + [COUNTERS] PROGRAM TOTAL : 0.7166s + [COUNTERS] Fortran Overhead ( 0 ) : 0.3888s + [COUNTERS] Fortran MEs ( 1 ) : 0.3278s for 8192 events => throughput is 2.50E+04 events/s + +*** (1) EXECUTE MADEVENT_FORTRAN x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./madevent_fortran < /tmp/avalassi/input_ggttg_x10_fortran > /tmp/avalassi/output_ggttg_x10_fortran' + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 32/32 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 0.07997 [7.9971656827279608E-002] fbridge_mode=0 + [UNWEIGHT] Wrote 2267 events (found 2272 events) + [COUNTERS] PROGRAM TOTAL : 5.1880s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8893s + [COUNTERS] Fortran MEs ( 1 ) : 3.2987s for 81920 events => throughput is 2.48E+04 events/s *** (2-none) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -109,9 +134,9 @@ DEBUG: MEK processed 8192 events across 16 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 0.07847 [7.8471485809748553E-002] fbridge_mode=1 [UNWEIGHT] Wrote 376 events (found 1358 events) - [COUNTERS] PROGRAM TOTAL : 0.7353s - [COUNTERS] Fortran Overhead ( 0 ) : 0.3872s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.3470s for 8192 events => throughput is 2.36E+04 events/s + [COUNTERS] PROGRAM TOTAL : 0.7358s + [COUNTERS] Fortran Overhead ( 0 ) : 0.3890s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.3457s for 8192 events => throughput is 2.37E+04 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0011s *** (2-none) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -122,15 +147,50 @@ OK! xsec from fortran (7.8471485809748567E-002) and cpp (7.8471485809748553E-002 OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-none) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.none_d_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_ggttg_x10_cudacpp > /tmp/avalassi/output_ggttg_x10_cudacpp' +DEBUG: MEK processed 81920 events across 16 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 32/32 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 0.07997 [7.9971656827279622E-002] fbridge_mode=1 + [UNWEIGHT] Wrote 2267 events (found 2272 events) + [COUNTERS] PROGRAM TOTAL : 5.3467s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9056s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.4399s for 81920 events => throughput is 2.38E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0011s + +*** (2-none) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (7.9971656827279608E-002) and cpp (7.9971656827279622E-002) differ by less than 3E-14 (2.220446049250313e-16) + +*** (2-none) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SM_GG_TTXG_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+BRDHST/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 2.455924e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.435276e+04 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_SM_GG_TTXG_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 2.454100e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.456830e+04 ) sec^-1 *** (2-sse4) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -154,9 +214,9 @@ DEBUG: MEK processed 8192 events across 16 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 0.07847 [7.8471485809748567E-002] fbridge_mode=1 [UNWEIGHT] Wrote 376 events (found 1358 events) - [COUNTERS] PROGRAM TOTAL : 0.5656s - [COUNTERS] Fortran Overhead ( 0 ) : 0.3857s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.1793s for 8192 events => throughput is 4.57E+04 events/s + [COUNTERS] PROGRAM TOTAL : 0.5862s + [COUNTERS] Fortran Overhead ( 0 ) : 0.3927s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1928s for 8192 events => throughput is 4.25E+04 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0007s *** (2-sse4) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -167,15 +227,50 @@ OK! xsec from fortran (7.8471485809748567E-002) and cpp (7.8471485809748567E-002 OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-sse4) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.sse4_d_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_ggttg_x10_cudacpp > /tmp/avalassi/output_ggttg_x10_cudacpp' +DEBUG: MEK processed 81920 events across 16 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 32/32 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 0.07997 [7.9971656827279650E-002] fbridge_mode=1 + [UNWEIGHT] Wrote 2267 events (found 2272 events) + [COUNTERS] PROGRAM TOTAL : 3.6929s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8822s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.8099s for 81920 events => throughput is 4.53E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0007s + +*** (2-sse4) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (7.9971656827279608E-002) and cpp (7.9971656827279650E-002) differ by less than 3E-14 (4.440892098500626e-16) + +*** (2-sse4) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SM_GG_TTXG_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+BRDHST/sse4+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 4.669927e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 4.544622e+04 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_SM_GG_TTXG_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 4.620836e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 4.673656e+04 ) sec^-1 *** (2-avx2) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -199,9 +294,9 @@ DEBUG: MEK processed 8192 events across 16 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 0.07847 [7.8471485809748595E-002] fbridge_mode=1 [UNWEIGHT] Wrote 376 events (found 1358 events) - [COUNTERS] PROGRAM TOTAL : 0.4792s - [COUNTERS] Fortran Overhead ( 0 ) : 0.3884s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0903s for 8192 events => throughput is 9.07E+04 events/s + [COUNTERS] PROGRAM TOTAL : 0.4745s + [COUNTERS] Fortran Overhead ( 0 ) : 0.3842s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0896s for 8192 events => throughput is 9.14E+04 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0006s *** (2-avx2) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -212,15 +307,50 @@ OK! xsec from fortran (7.8471485809748567E-002) and cpp (7.8471485809748595E-002 OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-avx2) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.avx2_d_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_ggttg_x10_cudacpp > /tmp/avalassi/output_ggttg_x10_cudacpp' +DEBUG: MEK processed 81920 events across 16 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 32/32 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 0.07997 [7.9971656827279622E-002] fbridge_mode=1 + [UNWEIGHT] Wrote 2267 events (found 2272 events) + [COUNTERS] PROGRAM TOTAL : 2.7846s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8884s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8956s for 81920 events => throughput is 9.15E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0006s + +*** (2-avx2) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (7.9971656827279608E-002) and cpp (7.9971656827279622E-002) differ by less than 3E-14 (2.220446049250313e-16) + +*** (2-avx2) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SM_GG_TTXG_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+BRDHST/avx2+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 9.331277e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 9.295208e+04 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_SM_GG_TTXG_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 9.327490e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 9.198330e+04 ) sec^-1 *** (2-512y) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -244,10 +374,10 @@ DEBUG: MEK processed 8192 events across 16 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 0.07847 [7.8471485809748595E-002] fbridge_mode=1 [UNWEIGHT] Wrote 376 events (found 1358 events) - [COUNTERS] PROGRAM TOTAL : 0.4693s - [COUNTERS] Fortran Overhead ( 0 ) : 0.3876s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0812s for 8192 events => throughput is 1.01E+05 events/s - [COUNTERS] CudaCpp HEL ( 3 ) : 0.0006s + [COUNTERS] PROGRAM TOTAL : 0.4624s + [COUNTERS] Fortran Overhead ( 0 ) : 0.3825s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0793s for 8192 events => throughput is 1.03E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0005s *** (2-512y) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -257,15 +387,50 @@ OK! xsec from fortran (7.8471485809748567E-002) and cpp (7.8471485809748595E-002 OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-512y) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.512y_d_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_ggttg_x10_cudacpp > /tmp/avalassi/output_ggttg_x10_cudacpp' +DEBUG: MEK processed 81920 events across 16 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 32/32 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 0.07997 [7.9971656827279622E-002] fbridge_mode=1 + [UNWEIGHT] Wrote 2267 events (found 2272 events) + [COUNTERS] PROGRAM TOTAL : 2.6798s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8775s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8018s for 81920 events => throughput is 1.02E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0005s + +*** (2-512y) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (7.9971656827279608E-002) and cpp (7.9971656827279622E-002) differ by less than 3E-14 (2.220446049250313e-16) + +*** (2-512y) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SM_GG_TTXG_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+BRDHST/512y+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.048553e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.038595e+05 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_SM_GG_TTXG_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.042752e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.048129e+05 ) sec^-1 *** (2-512z) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -289,10 +454,10 @@ DEBUG: MEK processed 8192 events across 16 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 0.07847 [7.8471485809748581E-002] fbridge_mode=1 [UNWEIGHT] Wrote 376 events (found 1358 events) - [COUNTERS] PROGRAM TOTAL : 0.5035s - [COUNTERS] Fortran Overhead ( 0 ) : 0.3879s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.1150s for 8192 events => throughput is 7.13E+04 events/s - [COUNTERS] CudaCpp HEL ( 3 ) : 0.0007s + [COUNTERS] PROGRAM TOTAL : 0.5012s + [COUNTERS] Fortran Overhead ( 0 ) : 0.3878s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1127s for 8192 events => throughput is 7.27E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0006s *** (2-512z) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -302,15 +467,50 @@ OK! xsec from fortran (7.8471485809748567E-002) and cpp (7.8471485809748581E-002 OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-512z) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.512z_d_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_ggttg_x10_cudacpp > /tmp/avalassi/output_ggttg_x10_cudacpp' +DEBUG: MEK processed 81920 events across 16 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 32/32 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 0.07997 [7.9971656827279622E-002] fbridge_mode=1 + [UNWEIGHT] Wrote 2267 events (found 2272 events) + [COUNTERS] PROGRAM TOTAL : 3.0407s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9036s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.1365s for 81920 events => throughput is 7.21E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0007s + +*** (2-512z) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (7.9971656827279608E-002) and cpp (7.9971656827279622E-002) differ by less than 3E-14 (2.220446049250313e-16) + +*** (2-512z) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SM_GG_TTXG_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+BRDHST/512z+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 7.198283e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 7.151172e+04 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_SM_GG_TTXG_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 7.275587e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 7.242422e+04 ) sec^-1 *** (3-cuda) EXECUTE MADEVENT_CUDA x1 (create events.lhe) *** -------------------- @@ -334,9 +534,9 @@ DEBUG: MEK processed 8192 events across 16 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 0.07847 [7.8471485809748553E-002] fbridge_mode=1 [UNWEIGHT] Wrote 376 events (found 1358 events) - [COUNTERS] PROGRAM TOTAL : 0.8395s - [COUNTERS] Fortran Overhead ( 0 ) : 0.8270s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0095s for 8192 events => throughput is 8.61E+05 events/s + [COUNTERS] PROGRAM TOTAL : 0.8368s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8244s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0094s for 8192 events => throughput is 8.76E+05 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0030s *** (3-cuda) Compare MADEVENT_CUDA x1 xsec to MADEVENT_FORTRAN xsec *** @@ -347,45 +547,80 @@ OK! xsec from fortran (7.8471485809748567E-002) and cuda (7.8471485809748553E-00 OK! events.lhe.cuda.1 and events.lhe.ref.1 are identical +*** (3-cuda) EXECUTE MADEVENT_CUDA x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.cuda_d_inl0_hrd0/madevent_cuda < /tmp/avalassi/input_ggttg_x10_cudacpp > /tmp/avalassi/output_ggttg_x10_cudacpp' +DEBUG: MEK processed 81920 events across 16 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 32/32 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 0.07997 [7.9971656827279636E-002] fbridge_mode=1 + [UNWEIGHT] Wrote 2267 events (found 2272 events) + [COUNTERS] PROGRAM TOTAL : 2.3398s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3110s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0258s for 81920 events => throughput is 3.18E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0030s + +*** (3-cuda) Compare MADEVENT_CUDA x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (7.9971656827279608E-002) and cuda (7.9971656827279636E-002) differ by less than 3E-14 (4.440892098500626e-16) + +*** (3-cuda) Compare MADEVENT_CUDA x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cuda.10 and events.lhe.ref.10 are identical + *** EXECUTE GCHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SM_GG_TTXG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.111479e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.135715e+06 ) sec^-1 *** EXECUTE GCHECK(8192) -p 256 32 1 *** Process = SIGMA_SM_GG_TTXG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 2.523607e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.511561e+06 ) sec^-1 *** EXECUTE GCHECK(MAX) -p 16384 32 1 --bridge *** Process = SIGMA_SM_GG_TTXG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 9.454522e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 9.445845e+06 ) sec^-1 *** EXECUTE GCHECK(MAX) -p 16384 32 1 *** Process = SIGMA_SM_GG_TTXG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.167720e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.168069e+07 ) sec^-1 *** EXECUTE GCHECK(MAX128THR) -p 4096 128 1 --bridge *** Process = SIGMA_SM_GG_TTXG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 9.412863e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 9.442124e+06 ) sec^-1 *** EXECUTE GCHECK(MAX128THR) -p 4096 128 1 *** Process = SIGMA_SM_GG_TTXG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.174227e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.175640e+07 ) sec^-1 *** EXECUTE GCHECK(MAX8THR) -p 65536 8 1 --bridge *** Process = SIGMA_SM_GG_TTXG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 9.441638e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 9.463238e+06 ) sec^-1 *** EXECUTE GCHECK(MAX8THR) -p 65536 8 1 *** Process = SIGMA_SM_GG_TTXG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.653840e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.653728e+06 ) sec^-1 *** (3-hip) WARNING! SKIP MADEVENT_HIP (hip is not supported on this node) *** diff --git a/epochX/cudacpp/tmad/logs_ggttg_mad/log_ggttg_mad_f_inl0_hrd0.txt b/epochX/cudacpp/tmad/logs_ggttg_mad/log_ggttg_mad_f_inl0_hrd0.txt index fbc0c57cb4..84ed973c77 100644 --- a/epochX/cudacpp/tmad/logs_ggttg_mad/log_ggttg_mad_f_inl0_hrd0.txt +++ b/epochX/cudacpp/tmad/logs_ggttg_mad/log_ggttg_mad_f_inl0_hrd0.txt @@ -1,13 +1,13 @@ Working directory (build): /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttg.mad/SubProcesses/P1_gg_ttxg - make USEBUILDDIR=1 BACKEND=cuda -make USEBUILDDIR=1 BACKEND=cppnone -make USEBUILDDIR=1 BACKEND=cppsse4 +make USEBUILDDIR=1 BACKEND=cppnone +make USEBUILDDIR=1 BACKEND=cppsse4 make USEBUILDDIR=1 BACKEND=cppavx2 + make USEBUILDDIR=1 BACKEND=cpp512y make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttg.mad/SubProcesses/P1_gg_ttxg' make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttg.mad/SubProcesses/P1_gg_ttxg' @@ -20,9 +20,9 @@ make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttg.mad/SubProcesses/P1_gg_ttxg' make[1]: Nothing to be done for 'all'. -make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttg.mad/SubProcesses/P1_gg_ttxg' make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttg.mad/SubProcesses/P1_gg_ttxg' +make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttg.mad/SubProcesses/P1_gg_ttxg' make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttg.mad/SubProcesses/P1_gg_ttxg' make[1]: Nothing to be done for 'all'. @@ -32,7 +32,7 @@ make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ OMP_NUM_THREADS= -DATE: 2024-10-06_10:06:11 +DATE: 2024-10-06_14:23:57 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: Working directory (run): /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttg.mad/SubProcesses/P1_gg_ttxg @@ -58,9 +58,9 @@ Executing ' ./madevent_fortran < /tmp/avalassi/input_ggttg_x1_fortran > /tmp/ava [XSECTION] ChannelId = 1 [XSECTION] Cross section = 0.07847 [7.8471485809748567E-002] fbridge_mode=0 [UNWEIGHT] Wrote 387 events (found 1591 events) - [COUNTERS] PROGRAM TOTAL : 0.7420s - [COUNTERS] Fortran Overhead ( 0 ) : 0.4102s - [COUNTERS] Fortran MEs ( 1 ) : 0.3318s for 8192 events => throughput is 2.47E+04 events/s + [COUNTERS] PROGRAM TOTAL : 0.7428s + [COUNTERS] Fortran Overhead ( 0 ) : 0.4117s + [COUNTERS] Fortran MEs ( 1 ) : 0.3311s for 8192 events => throughput is 2.47E+04 events/s *** (1) EXECUTE MADEVENT_FORTRAN x1 (create events.lhe) *** -------------------- @@ -83,9 +83,34 @@ Executing ' ./madevent_fortran < /tmp/avalassi/input_ggttg_x1_fortran > /tmp/ava [XSECTION] ChannelId = 1 [XSECTION] Cross section = 0.07847 [7.8471485809748567E-002] fbridge_mode=0 [UNWEIGHT] Wrote 376 events (found 1358 events) - [COUNTERS] PROGRAM TOTAL : 0.7176s - [COUNTERS] Fortran Overhead ( 0 ) : 0.3860s - [COUNTERS] Fortran MEs ( 1 ) : 0.3316s for 8192 events => throughput is 2.47E+04 events/s + [COUNTERS] PROGRAM TOTAL : 0.7169s + [COUNTERS] Fortran Overhead ( 0 ) : 0.3838s + [COUNTERS] Fortran MEs ( 1 ) : 0.3331s for 8192 events => throughput is 2.46E+04 events/s + +*** (1) EXECUTE MADEVENT_FORTRAN x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./madevent_fortran < /tmp/avalassi/input_ggttg_x10_fortran > /tmp/avalassi/output_ggttg_x10_fortran' + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 32/32 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 0.07997 [7.9971656827279608E-002] fbridge_mode=0 + [UNWEIGHT] Wrote 2267 events (found 2272 events) + [COUNTERS] PROGRAM TOTAL : 5.1954s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9036s + [COUNTERS] Fortran MEs ( 1 ) : 3.2918s for 81920 events => throughput is 2.49E+04 events/s *** (2-none) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -109,10 +134,10 @@ DEBUG: MEK processed 8192 events across 16 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 0.07847 [7.8471473453718410E-002] fbridge_mode=1 [UNWEIGHT] Wrote 376 events (found 1358 events) - [COUNTERS] PROGRAM TOTAL : 0.7234s - [COUNTERS] Fortran Overhead ( 0 ) : 0.3899s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.3325s for 8192 events => throughput is 2.46E+04 events/s - [COUNTERS] CudaCpp HEL ( 3 ) : 0.0009s + [COUNTERS] PROGRAM TOTAL : 0.7183s + [COUNTERS] Fortran Overhead ( 0 ) : 0.3862s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.3311s for 8192 events => throughput is 2.47E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0010s *** (2-none) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -122,15 +147,50 @@ OK! xsec from fortran (7.8471485809748567E-002) and cpp (7.8471473453718410E-002 OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-none) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.none_f_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_ggttg_x10_cudacpp > /tmp/avalassi/output_ggttg_x10_cudacpp' +DEBUG: MEK processed 81920 events across 16 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 32/32 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 0.07997 [7.9971643267110940E-002] fbridge_mode=1 + [UNWEIGHT] Wrote 2267 events (found 2272 events) + [COUNTERS] PROGRAM TOTAL : 5.2056s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8933s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.3114s for 81920 events => throughput is 2.47E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0009s + +*** (2-none) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (7.9971656827279608E-002) and cpp (7.9971643267110940E-002) differ by less than 4E-4 (1.69562182517069e-07) + +*** (2-none) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SM_GG_TTXG_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+BRDHST/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 2.535876e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.530315e+04 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_SM_GG_TTXG_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 2.542086e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.544463e+04 ) sec^-1 *** (2-sse4) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -154,10 +214,10 @@ DEBUG: MEK processed 8192 events across 16 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 0.07847 [7.8471459294758378E-002] fbridge_mode=1 [UNWEIGHT] Wrote 376 events (found 1358 events) - [COUNTERS] PROGRAM TOTAL : 0.4904s - [COUNTERS] Fortran Overhead ( 0 ) : 0.3886s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.1014s for 8192 events => throughput is 8.08E+04 events/s - [COUNTERS] CudaCpp HEL ( 3 ) : 0.0005s + [COUNTERS] PROGRAM TOTAL : 0.4900s + [COUNTERS] Fortran Overhead ( 0 ) : 0.3879s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1016s for 8192 events => throughput is 8.06E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0004s *** (2-sse4) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -167,15 +227,50 @@ OK! xsec from fortran (7.8471485809748567E-002) and cpp (7.8471459294758378E-002 OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-sse4) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.sse4_f_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_ggttg_x10_cudacpp > /tmp/avalassi/output_ggttg_x10_cudacpp' +DEBUG: MEK processed 81920 events across 16 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 32/32 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 0.07997 [7.9971629726281482E-002] fbridge_mode=1 + [UNWEIGHT] Wrote 2267 events (found 2272 events) + [COUNTERS] PROGRAM TOTAL : 2.8920s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8821s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.0095s for 81920 events => throughput is 8.11E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0004s + +*** (2-sse4) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (7.9971656827279608E-002) and cpp (7.9971629726281482E-002) differ by less than 4E-4 (3.38882539141494e-07) + +*** (2-sse4) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SM_GG_TTXG_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+BRDHST/sse4+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 8.182689e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 8.217643e+04 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_SM_GG_TTXG_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 8.204950e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 8.144540e+04 ) sec^-1 *** (2-avx2) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -199,9 +294,9 @@ DEBUG: MEK processed 8192 events across 16 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 0.07847 [7.8471459718665412E-002] fbridge_mode=1 [UNWEIGHT] Wrote 376 events (found 1358 events) - [COUNTERS] PROGRAM TOTAL : 0.4358s - [COUNTERS] Fortran Overhead ( 0 ) : 0.3891s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0464s for 8192 events => throughput is 1.77E+05 events/s + [COUNTERS] PROGRAM TOTAL : 0.4320s + [COUNTERS] Fortran Overhead ( 0 ) : 0.3858s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0458s for 8192 events => throughput is 1.79E+05 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0004s *** (2-avx2) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -212,15 +307,50 @@ OK! xsec from fortran (7.8471485809748567E-002) and cpp (7.8471459718665412E-002 OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-avx2) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.avx2_f_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_ggttg_x10_cudacpp > /tmp/avalassi/output_ggttg_x10_cudacpp' +DEBUG: MEK processed 81920 events across 16 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 32/32 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 0.07997 [7.9971629259822388E-002] fbridge_mode=1 + [UNWEIGHT] Wrote 2267 events (found 2272 events) + [COUNTERS] PROGRAM TOTAL : 2.3564s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8947s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4614s for 81920 events => throughput is 1.78E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0004s + +*** (2-avx2) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (7.9971656827279608E-002) and cpp (7.9971629259822388E-002) differ by less than 4E-4 (3.447153443802975e-07) + +*** (2-avx2) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SM_GG_TTXG_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+BRDHST/avx2+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.782969e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.787925e+05 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_SM_GG_TTXG_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.783579e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.807987e+05 ) sec^-1 *** (2-512y) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -244,9 +374,9 @@ DEBUG: MEK processed 8192 events across 16 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 0.07847 [7.8471459718665412E-002] fbridge_mode=1 [UNWEIGHT] Wrote 376 events (found 1358 events) - [COUNTERS] PROGRAM TOTAL : 0.4301s - [COUNTERS] Fortran Overhead ( 0 ) : 0.3871s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0426s for 8192 events => throughput is 1.92E+05 events/s + [COUNTERS] PROGRAM TOTAL : 0.4368s + [COUNTERS] Fortran Overhead ( 0 ) : 0.3943s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0422s for 8192 events => throughput is 1.94E+05 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0004s *** (2-512y) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -257,15 +387,50 @@ OK! xsec from fortran (7.8471485809748567E-002) and cpp (7.8471459718665412E-002 OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-512y) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.512y_f_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_ggttg_x10_cudacpp > /tmp/avalassi/output_ggttg_x10_cudacpp' +DEBUG: MEK processed 81920 events across 16 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 32/32 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 0.07997 [7.9971629259822388E-002] fbridge_mode=1 + [UNWEIGHT] Wrote 2267 events (found 2272 events) + [COUNTERS] PROGRAM TOTAL : 2.3129s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8905s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4220s for 81920 events => throughput is 1.94E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0004s + +*** (2-512y) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (7.9971656827279608E-002) and cpp (7.9971629259822388E-002) differ by less than 4E-4 (3.447153443802975e-07) + +*** (2-512y) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SM_GG_TTXG_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+BRDHST/512y+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.968891e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.980410e+05 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_SM_GG_TTXG_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.969858e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.982304e+05 ) sec^-1 *** (2-512z) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -289,9 +454,9 @@ DEBUG: MEK processed 8192 events across 16 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 0.07847 [7.8471471932611128E-002] fbridge_mode=1 [UNWEIGHT] Wrote 376 events (found 1358 events) - [COUNTERS] PROGRAM TOTAL : 0.4447s - [COUNTERS] Fortran Overhead ( 0 ) : 0.3892s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0551s for 8192 events => throughput is 1.49E+05 events/s + [COUNTERS] PROGRAM TOTAL : 0.4439s + [COUNTERS] Fortran Overhead ( 0 ) : 0.3876s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0559s for 8192 events => throughput is 1.47E+05 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0004s *** (2-512z) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -302,15 +467,50 @@ OK! xsec from fortran (7.8471485809748567E-002) and cpp (7.8471471932611128E-002 OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-512z) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.512z_f_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_ggttg_x10_cudacpp > /tmp/avalassi/output_ggttg_x10_cudacpp' +DEBUG: MEK processed 81920 events across 16 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 32/32 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 0.07997 [7.9971639934306102E-002] fbridge_mode=1 + [UNWEIGHT] Wrote 2267 events (found 2272 events) + [COUNTERS] PROGRAM TOTAL : 2.4346s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8780s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5561s for 81920 events => throughput is 1.47E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0004s + +*** (2-512z) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (7.9971656827279608E-002) and cpp (7.9971639934306102E-002) differ by less than 4E-4 (2.1123700788550082e-07) + +*** (2-512z) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SM_GG_TTXG_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+BRDHST/512z+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.481854e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.480550e+05 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_SM_GG_TTXG_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.468460e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.423424e+05 ) sec^-1 *** (3-cuda) EXECUTE MADEVENT_CUDA x1 (create events.lhe) *** -------------------- @@ -334,9 +534,9 @@ DEBUG: MEK processed 8192 events across 16 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 0.07847 [7.8471475012321185E-002] fbridge_mode=1 [UNWEIGHT] Wrote 376 events (found 1358 events) - [COUNTERS] PROGRAM TOTAL : 0.8373s - [COUNTERS] Fortran Overhead ( 0 ) : 0.8327s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0035s for 8192 events => throughput is 2.36E+06 events/s + [COUNTERS] PROGRAM TOTAL : 0.8281s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8237s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0033s for 8192 events => throughput is 2.45E+06 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0011s *** (3-cuda) Compare MADEVENT_CUDA x1 xsec to MADEVENT_FORTRAN xsec *** @@ -347,45 +547,80 @@ OK! xsec from fortran (7.8471485809748567E-002) and cuda (7.8471475012321185E-00 OK! events.lhe.cuda.1 and events.lhe.ref.1 are identical +*** (3-cuda) EXECUTE MADEVENT_CUDA x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.cuda_f_inl0_hrd0/madevent_cuda < /tmp/avalassi/input_ggttg_x10_cudacpp > /tmp/avalassi/output_ggttg_x10_cudacpp' +DEBUG: MEK processed 81920 events across 16 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 32/32 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 0.07997 [7.9971648932322295E-002] fbridge_mode=1 + [UNWEIGHT] Wrote 2267 events (found 2272 events) + [COUNTERS] PROGRAM TOTAL : 2.3477s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3338s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0127s for 81920 events => throughput is 6.43E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0011s + +*** (3-cuda) Compare MADEVENT_CUDA x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (7.9971656827279608E-002) and cuda (7.9971648932322295E-002) differ by less than 4E-4 (9.872194262072753e-08) + +*** (3-cuda) Compare MADEVENT_CUDA x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cuda.10 and events.lhe.ref.10 are identical + *** EXECUTE GCHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SM_GG_TTXG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 2.717098e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.742648e+06 ) sec^-1 *** EXECUTE GCHECK(8192) -p 256 32 1 *** Process = SIGMA_SM_GG_TTXG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 2.890243e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.017041e+06 ) sec^-1 *** EXECUTE GCHECK(MAX) -p 16384 32 1 --bridge *** Process = SIGMA_SM_GG_TTXG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 2.313606e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.333242e+07 ) sec^-1 *** EXECUTE GCHECK(MAX) -p 16384 32 1 *** Process = SIGMA_SM_GG_TTXG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.232701e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.227055e+07 ) sec^-1 *** EXECUTE GCHECK(MAX128THR) -p 4096 128 1 --bridge *** Process = SIGMA_SM_GG_TTXG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 2.300307e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.332474e+07 ) sec^-1 *** EXECUTE GCHECK(MAX128THR) -p 4096 128 1 *** Process = SIGMA_SM_GG_TTXG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.230438e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.296207e+07 ) sec^-1 *** EXECUTE GCHECK(MAX8THR) -p 65536 8 1 --bridge *** Process = SIGMA_SM_GG_TTXG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 2.193713e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.210483e+07 ) sec^-1 *** EXECUTE GCHECK(MAX8THR) -p 65536 8 1 *** Process = SIGMA_SM_GG_TTXG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.247962e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.247455e+07 ) sec^-1 *** (3-hip) WARNING! SKIP MADEVENT_HIP (hip is not supported on this node) *** diff --git a/epochX/cudacpp/tmad/logs_ggttg_mad/log_ggttg_mad_m_inl0_hrd0.txt b/epochX/cudacpp/tmad/logs_ggttg_mad/log_ggttg_mad_m_inl0_hrd0.txt index 2422d3068f..cd2604800a 100644 --- a/epochX/cudacpp/tmad/logs_ggttg_mad/log_ggttg_mad_m_inl0_hrd0.txt +++ b/epochX/cudacpp/tmad/logs_ggttg_mad/log_ggttg_mad_m_inl0_hrd0.txt @@ -1,22 +1,22 @@ Working directory (build): /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttg.mad/SubProcesses/P1_gg_ttxg - make USEBUILDDIR=1 BACKEND=cuda + make USEBUILDDIR=1 BACKEND=cppnone make USEBUILDDIR=1 BACKEND=cppsse4 -make USEBUILDDIR=1 BACKEND=cppavx2 +make USEBUILDDIR=1 BACKEND=cppavx2 make USEBUILDDIR=1 BACKEND=cpp512y make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttg.mad/SubProcesses/P1_gg_ttxg' make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttg.mad/SubProcesses/P1_gg_ttxg' make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttg.mad/SubProcesses/P1_gg_ttxg' make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttg.mad/SubProcesses/P1_gg_ttxg' +make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttg.mad/SubProcesses/P1_gg_ttxg' make USEBUILDDIR=1 BACKEND=cpp512z make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttg.mad/SubProcesses/P1_gg_ttxg' -make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttg.mad/SubProcesses/P1_gg_ttxg' make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttg.mad/SubProcesses/P1_gg_ttxg' make[1]: Nothing to be done for 'all'. @@ -32,7 +32,7 @@ make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ OMP_NUM_THREADS= -DATE: 2024-10-06_10:05:53 +DATE: 2024-10-06_14:23:13 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: Working directory (run): /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttg.mad/SubProcesses/P1_gg_ttxg @@ -58,9 +58,9 @@ Executing ' ./madevent_fortran < /tmp/avalassi/input_ggttg_x1_fortran > /tmp/ava [XSECTION] ChannelId = 1 [XSECTION] Cross section = 0.07847 [7.8471485809748567E-002] fbridge_mode=0 [UNWEIGHT] Wrote 387 events (found 1591 events) - [COUNTERS] PROGRAM TOTAL : 0.7391s - [COUNTERS] Fortran Overhead ( 0 ) : 0.4096s - [COUNTERS] Fortran MEs ( 1 ) : 0.3295s for 8192 events => throughput is 2.49E+04 events/s + [COUNTERS] PROGRAM TOTAL : 0.7367s + [COUNTERS] Fortran Overhead ( 0 ) : 0.4081s + [COUNTERS] Fortran MEs ( 1 ) : 0.3285s for 8192 events => throughput is 2.49E+04 events/s *** (1) EXECUTE MADEVENT_FORTRAN x1 (create events.lhe) *** -------------------- @@ -83,9 +83,34 @@ Executing ' ./madevent_fortran < /tmp/avalassi/input_ggttg_x1_fortran > /tmp/ava [XSECTION] ChannelId = 1 [XSECTION] Cross section = 0.07847 [7.8471485809748567E-002] fbridge_mode=0 [UNWEIGHT] Wrote 376 events (found 1358 events) - [COUNTERS] PROGRAM TOTAL : 0.7165s - [COUNTERS] Fortran Overhead ( 0 ) : 0.3855s - [COUNTERS] Fortran MEs ( 1 ) : 0.3310s for 8192 events => throughput is 2.47E+04 events/s + [COUNTERS] PROGRAM TOTAL : 0.7174s + [COUNTERS] Fortran Overhead ( 0 ) : 0.3858s + [COUNTERS] Fortran MEs ( 1 ) : 0.3316s for 8192 events => throughput is 2.47E+04 events/s + +*** (1) EXECUTE MADEVENT_FORTRAN x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./madevent_fortran < /tmp/avalassi/input_ggttg_x10_fortran > /tmp/avalassi/output_ggttg_x10_fortran' + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 32/32 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 0.07997 [7.9971656827279608E-002] fbridge_mode=0 + [UNWEIGHT] Wrote 2267 events (found 2272 events) + [COUNTERS] PROGRAM TOTAL : 5.1807s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8818s + [COUNTERS] Fortran MEs ( 1 ) : 3.2989s for 81920 events => throughput is 2.48E+04 events/s *** (2-none) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -109,9 +134,9 @@ DEBUG: MEK processed 8192 events across 16 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 0.07847 [7.8471486590207584E-002] fbridge_mode=1 [UNWEIGHT] Wrote 376 events (found 1358 events) - [COUNTERS] PROGRAM TOTAL : 0.7396s - [COUNTERS] Fortran Overhead ( 0 ) : 0.3874s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.3510s for 8192 events => throughput is 2.33E+04 events/s + [COUNTERS] PROGRAM TOTAL : 0.7386s + [COUNTERS] Fortran Overhead ( 0 ) : 0.3878s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.3497s for 8192 events => throughput is 2.34E+04 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0011s *** (2-none) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -122,15 +147,50 @@ OK! xsec from fortran (7.8471485809748567E-002) and cpp (7.8471486590207584E-002 OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-none) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.none_m_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_ggttg_x10_cudacpp > /tmp/avalassi/output_ggttg_x10_cudacpp' +DEBUG: MEK processed 81920 events across 16 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 32/32 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 0.07997 [7.9971657589635384E-002] fbridge_mode=1 + [UNWEIGHT] Wrote 2267 events (found 2272 events) + [COUNTERS] PROGRAM TOTAL : 5.3882s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8909s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.4963s for 81920 events => throughput is 2.34E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0011s + +*** (2-none) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (7.9971656827279608E-002) and cpp (7.9971657589635384E-002) differ by less than 2E-4 (9.532824529756567e-09) + +*** (2-none) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SM_GG_TTXG_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+BRDHST/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 2.409349e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.433251e+04 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_SM_GG_TTXG_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 2.415956e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.415920e+04 ) sec^-1 *** (2-sse4) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -154,9 +214,9 @@ DEBUG: MEK processed 8192 events across 16 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 0.07847 [7.8471486540430027E-002] fbridge_mode=1 [UNWEIGHT] Wrote 376 events (found 1358 events) - [COUNTERS] PROGRAM TOTAL : 0.5676s - [COUNTERS] Fortran Overhead ( 0 ) : 0.3876s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.1792s for 8192 events => throughput is 4.57E+04 events/s + [COUNTERS] PROGRAM TOTAL : 0.5658s + [COUNTERS] Fortran Overhead ( 0 ) : 0.3855s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1796s for 8192 events => throughput is 4.56E+04 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0007s *** (2-sse4) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -167,15 +227,50 @@ OK! xsec from fortran (7.8471485809748567E-002) and cpp (7.8471486540430027E-002 OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-sse4) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.sse4_m_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_ggttg_x10_cudacpp > /tmp/avalassi/output_ggttg_x10_cudacpp' +DEBUG: MEK processed 81920 events across 16 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 32/32 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 0.07997 [7.9971657589963913E-002] fbridge_mode=1 + [UNWEIGHT] Wrote 2267 events (found 2272 events) + [COUNTERS] PROGRAM TOTAL : 3.6921s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8884s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.8029s for 81920 events => throughput is 4.54E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0008s + +*** (2-sse4) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (7.9971656827279608E-002) and cpp (7.9971657589963913E-002) differ by less than 2E-4 (9.536932576992285e-09) + +*** (2-sse4) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SM_GG_TTXG_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+BRDHST/sse4+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 4.653483e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 4.680303e+04 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_SM_GG_TTXG_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 4.691370e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 4.702275e+04 ) sec^-1 *** (2-avx2) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -199,9 +294,9 @@ DEBUG: MEK processed 8192 events across 16 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 0.07847 [7.8471486395956899E-002] fbridge_mode=1 [UNWEIGHT] Wrote 376 events (found 1358 events) - [COUNTERS] PROGRAM TOTAL : 0.4809s - [COUNTERS] Fortran Overhead ( 0 ) : 0.3907s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0896s for 8192 events => throughput is 9.14E+04 events/s + [COUNTERS] PROGRAM TOTAL : 0.4751s + [COUNTERS] Fortran Overhead ( 0 ) : 0.3855s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0890s for 8192 events => throughput is 9.21E+04 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0006s *** (2-avx2) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -212,15 +307,50 @@ OK! xsec from fortran (7.8471485809748567E-002) and cpp (7.8471486395956899E-002 OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-avx2) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.avx2_m_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_ggttg_x10_cudacpp > /tmp/avalassi/output_ggttg_x10_cudacpp' +DEBUG: MEK processed 81920 events across 16 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 32/32 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 0.07997 [7.9971657432811344E-002] fbridge_mode=1 + [UNWEIGHT] Wrote 2267 events (found 2272 events) + [COUNTERS] PROGRAM TOTAL : 2.7767s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8808s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8953s for 81920 events => throughput is 9.15E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0006s + +*** (2-avx2) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (7.9971656827279608E-002) and cpp (7.9971657432811344E-002) differ by less than 2E-4 (7.571829385710771e-09) + +*** (2-avx2) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SM_GG_TTXG_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+BRDHST/avx2+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 9.402724e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 9.400388e+04 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_SM_GG_TTXG_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 9.391101e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 9.466659e+04 ) sec^-1 *** (2-512y) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -244,9 +374,9 @@ DEBUG: MEK processed 8192 events across 16 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 0.07847 [7.8471486395956899E-002] fbridge_mode=1 [UNWEIGHT] Wrote 376 events (found 1358 events) - [COUNTERS] PROGRAM TOTAL : 0.4660s - [COUNTERS] Fortran Overhead ( 0 ) : 0.3858s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0796s for 8192 events => throughput is 1.03E+05 events/s + [COUNTERS] PROGRAM TOTAL : 0.4648s + [COUNTERS] Fortran Overhead ( 0 ) : 0.3850s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0792s for 8192 events => throughput is 1.03E+05 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0006s *** (2-512y) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -257,15 +387,50 @@ OK! xsec from fortran (7.8471485809748567E-002) and cpp (7.8471486395956899E-002 OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-512y) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.512y_m_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_ggttg_x10_cudacpp > /tmp/avalassi/output_ggttg_x10_cudacpp' +DEBUG: MEK processed 81920 events across 16 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 32/32 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 0.07997 [7.9971657432811344E-002] fbridge_mode=1 + [UNWEIGHT] Wrote 2267 events (found 2272 events) + [COUNTERS] PROGRAM TOTAL : 2.6780s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8831s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.7943s for 81920 events => throughput is 1.03E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0006s + +*** (2-512y) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (7.9971656827279608E-002) and cpp (7.9971657432811344E-002) differ by less than 2E-4 (7.571829385710771e-09) + +*** (2-512y) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SM_GG_TTXG_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+BRDHST/512y+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.055172e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.043975e+05 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_SM_GG_TTXG_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.066925e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.066058e+05 ) sec^-1 *** (2-512z) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -289,10 +454,10 @@ DEBUG: MEK processed 8192 events across 16 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 0.07847 [7.8471486537749241E-002] fbridge_mode=1 [UNWEIGHT] Wrote 376 events (found 1358 events) - [COUNTERS] PROGRAM TOTAL : 0.5026s - [COUNTERS] Fortran Overhead ( 0 ) : 0.3850s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.1170s for 8192 events => throughput is 7.00E+04 events/s - [COUNTERS] CudaCpp HEL ( 3 ) : 0.0006s + [COUNTERS] PROGRAM TOTAL : 0.5067s + [COUNTERS] Fortran Overhead ( 0 ) : 0.3884s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1176s for 8192 events => throughput is 6.97E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0007s *** (2-512z) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -302,15 +467,50 @@ OK! xsec from fortran (7.8471485809748567E-002) and cpp (7.8471486537749241E-002 OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-512z) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.512z_m_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_ggttg_x10_cudacpp > /tmp/avalassi/output_ggttg_x10_cudacpp' +DEBUG: MEK processed 81920 events across 16 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 32/32 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 0.07997 [7.9971657565670345E-002] fbridge_mode=1 + [UNWEIGHT] Wrote 2267 events (found 2272 events) + [COUNTERS] PROGRAM TOTAL : 3.0673s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8937s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.1729s for 81920 events => throughput is 6.98E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0007s + +*** (2-512z) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (7.9971656827279608E-002) and cpp (7.9971657565670345E-002) differ by less than 2E-4 (9.233155351395794e-09) + +*** (2-512z) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SM_GG_TTXG_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+BRDHST/512z+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 7.005425e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 7.067032e+04 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_SM_GG_TTXG_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 7.056979e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 6.995010e+04 ) sec^-1 *** (3-cuda) EXECUTE MADEVENT_CUDA x1 (create events.lhe) *** -------------------- @@ -334,10 +534,10 @@ DEBUG: MEK processed 8192 events across 16 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 0.07847 [7.8471485791426987E-002] fbridge_mode=1 [UNWEIGHT] Wrote 376 events (found 1358 events) - [COUNTERS] PROGRAM TOTAL : 0.8432s - [COUNTERS] Fortran Overhead ( 0 ) : 0.8306s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0095s for 8192 events => throughput is 8.66E+05 events/s - [COUNTERS] CudaCpp HEL ( 3 ) : 0.0031s + [COUNTERS] PROGRAM TOTAL : 0.8346s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8223s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0093s for 8192 events => throughput is 8.78E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0030s *** (3-cuda) Compare MADEVENT_CUDA x1 xsec to MADEVENT_FORTRAN xsec *** @@ -347,45 +547,80 @@ OK! xsec from fortran (7.8471485809748567E-002) and cuda (7.8471485791426987E-00 OK! events.lhe.cuda.1 and events.lhe.ref.1 are identical +*** (3-cuda) EXECUTE MADEVENT_CUDA x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.cuda_m_inl0_hrd0/madevent_cuda < /tmp/avalassi/input_ggttg_x10_cudacpp > /tmp/avalassi/output_ggttg_x10_cudacpp' +DEBUG: MEK processed 81920 events across 16 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 32/32 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 0.07997 [7.9971656830583548E-002] fbridge_mode=1 + [UNWEIGHT] Wrote 2267 events (found 2272 events) + [COUNTERS] PROGRAM TOTAL : 2.3470s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3183s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0257s for 81920 events => throughput is 3.19E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0030s + +*** (3-cuda) Compare MADEVENT_CUDA x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (7.9971656827279608E-002) and cuda (7.9971656830583548E-002) differ by less than 2E-4 (4.131384123695625e-11) + +*** (3-cuda) Compare MADEVENT_CUDA x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cuda.10 and events.lhe.ref.10 are identical + *** EXECUTE GCHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SM_GG_TTXG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:MIX+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.128450e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.134258e+06 ) sec^-1 *** EXECUTE GCHECK(8192) -p 256 32 1 *** Process = SIGMA_SM_GG_TTXG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:MIX+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 2.439893e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.556986e+06 ) sec^-1 *** EXECUTE GCHECK(MAX) -p 16384 32 1 --bridge *** Process = SIGMA_SM_GG_TTXG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:MIX+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 9.421024e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 9.431676e+06 ) sec^-1 *** EXECUTE GCHECK(MAX) -p 16384 32 1 *** Process = SIGMA_SM_GG_TTXG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:MIX+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.153444e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.161360e+07 ) sec^-1 *** EXECUTE GCHECK(MAX128THR) -p 4096 128 1 --bridge *** Process = SIGMA_SM_GG_TTXG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:MIX+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 9.432988e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 9.436607e+06 ) sec^-1 *** EXECUTE GCHECK(MAX128THR) -p 4096 128 1 *** Process = SIGMA_SM_GG_TTXG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:MIX+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.169695e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.166933e+07 ) sec^-1 *** EXECUTE GCHECK(MAX8THR) -p 65536 8 1 --bridge *** Process = SIGMA_SM_GG_TTXG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:MIX+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 9.432146e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 9.456359e+06 ) sec^-1 *** EXECUTE GCHECK(MAX8THR) -p 65536 8 1 *** Process = SIGMA_SM_GG_TTXG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:MIX+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.638179e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.636220e+06 ) sec^-1 *** (3-hip) WARNING! SKIP MADEVENT_HIP (hip is not supported on this node) *** diff --git a/epochX/cudacpp/tmad/logs_ggttgg_mad/log_ggttgg_mad_d_inl0_hrd0.txt b/epochX/cudacpp/tmad/logs_ggttgg_mad/log_ggttgg_mad_d_inl0_hrd0.txt index 5517ab4292..50d9656ae9 100644 --- a/epochX/cudacpp/tmad/logs_ggttgg_mad/log_ggttgg_mad_d_inl0_hrd0.txt +++ b/epochX/cudacpp/tmad/logs_ggttgg_mad/log_ggttgg_mad_d_inl0_hrd0.txt @@ -1,22 +1,22 @@ Working directory (build): /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg + make USEBUILDDIR=1 BACKEND=cuda make USEBUILDDIR=1 BACKEND=cppnone - - make USEBUILDDIR=1 BACKEND=cppsse4 make USEBUILDDIR=1 BACKEND=cppavx2 + make USEBUILDDIR=1 BACKEND=cpp512y make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg' make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg' make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg' make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg' -make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg' make USEBUILDDIR=1 BACKEND=cpp512z make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg' +make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg' make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg' make[1]: Nothing to be done for 'all'. @@ -32,7 +32,7 @@ make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ OMP_NUM_THREADS= -DATE: 2024-10-06_10:06:26 +DATE: 2024-10-06_14:24:36 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: Working directory (run): /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg @@ -58,9 +58,9 @@ Executing ' ./madevent_fortran < /tmp/avalassi/input_ggttgg_x1_fortran > /tmp/av [XSECTION] ChannelId = 112 [XSECTION] Cross section = 0.3314 [0.33144786561240197] fbridge_mode=0 [UNWEIGHT] Wrote 7 events (found 223 events) - [COUNTERS] PROGRAM TOTAL : 4.6353s - [COUNTERS] Fortran Overhead ( 0 ) : 0.3093s - [COUNTERS] Fortran MEs ( 1 ) : 4.3260s for 8192 events => throughput is 1.89E+03 events/s + [COUNTERS] PROGRAM TOTAL : 4.5901s + [COUNTERS] Fortran Overhead ( 0 ) : 0.2951s + [COUNTERS] Fortran MEs ( 1 ) : 4.2950s for 8192 events => throughput is 1.91E+03 events/s *** (1) EXECUTE MADEVENT_FORTRAN x1 (create events.lhe) *** -------------------- @@ -83,9 +83,34 @@ Executing ' ./madevent_fortran < /tmp/avalassi/input_ggttgg_x1_fortran > /tmp/av [XSECTION] ChannelId = 112 [XSECTION] Cross section = 0.3314 [0.33144786561240197] fbridge_mode=0 [UNWEIGHT] Wrote 7 events (found 213 events) - [COUNTERS] PROGRAM TOTAL : 4.5825s - [COUNTERS] Fortran Overhead ( 0 ) : 0.2904s - [COUNTERS] Fortran MEs ( 1 ) : 4.2921s for 8192 events => throughput is 1.91E+03 events/s + [COUNTERS] PROGRAM TOTAL : 4.6088s + [COUNTERS] Fortran Overhead ( 0 ) : 0.2911s + [COUNTERS] Fortran MEs ( 1 ) : 4.3177s for 8192 events => throughput is 1.90E+03 events/s + +*** (1) EXECUTE MADEVENT_FORTRAN x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +104 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./madevent_fortran < /tmp/avalassi/input_ggttgg_x10_fortran > /tmp/avalassi/output_ggttgg_x10_fortran' + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 64/64 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 104 + [XSECTION] ChannelId = 112 + [XSECTION] Cross section = 0.2093 [0.20930257969248323] fbridge_mode=0 + [UNWEIGHT] Wrote 17 events (found 331 events) + [COUNTERS] PROGRAM TOTAL : 45.2399s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0535s + [COUNTERS] Fortran MEs ( 1 ) : 43.1864s for 81920 events => throughput is 1.90E+03 events/s *** (2-none) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -109,10 +134,10 @@ DEBUG: MEK processed 8192 events across 123 channels { 112 : 8192 } [XSECTION] ChannelId = 112 [XSECTION] Cross section = 0.3314 [0.33144786561240192] fbridge_mode=1 [UNWEIGHT] Wrote 7 events (found 213 events) - [COUNTERS] PROGRAM TOTAL : 4.7512s - [COUNTERS] Fortran Overhead ( 0 ) : 0.2946s - [COUNTERS] CudaCpp MEs ( 2 ) : 4.4476s for 8192 events => throughput is 1.84E+03 events/s - [COUNTERS] CudaCpp HEL ( 3 ) : 0.0090s + [COUNTERS] PROGRAM TOTAL : 4.7496s + [COUNTERS] Fortran Overhead ( 0 ) : 0.2951s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.4457s for 8192 events => throughput is 1.84E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0089s *** (2-none) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -122,15 +147,50 @@ OK! xsec from fortran (0.33144786561240197) and cpp (0.33144786561240192) differ OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-none) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +104 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.none_d_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_ggttgg_x10_cudacpp > /tmp/avalassi/output_ggttgg_x10_cudacpp' +DEBUG: MEK processed 81920 events across 123 channels { 112 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 64/64 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 104 + [XSECTION] ChannelId = 112 + [XSECTION] Cross section = 0.2093 [0.20930257969248320] fbridge_mode=1 + [UNWEIGHT] Wrote 17 events (found 331 events) + [COUNTERS] PROGRAM TOTAL : 46.6448s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0552s + [COUNTERS] CudaCpp MEs ( 2 ) : 44.5805s for 81920 events => throughput is 1.84E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0091s + +*** (2-none) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (0.20930257969248323) and cpp (0.20930257969248320) differ by less than 3E-14 (1.1102230246251565e-16) + +*** (2-none) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+BRDHST/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.894558e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.902623e+03 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.891638e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.896636e+03 ) sec^-1 *** (2-sse4) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -154,10 +214,10 @@ DEBUG: MEK processed 8192 events across 123 channels { 112 : 8192 } [XSECTION] ChannelId = 112 [XSECTION] Cross section = 0.3314 [0.33144786561240192] fbridge_mode=1 [UNWEIGHT] Wrote 7 events (found 213 events) - [COUNTERS] PROGRAM TOTAL : 2.6638s - [COUNTERS] Fortran Overhead ( 0 ) : 0.2927s - [COUNTERS] CudaCpp MEs ( 2 ) : 2.3665s for 8192 events => throughput is 3.46E+03 events/s - [COUNTERS] CudaCpp HEL ( 3 ) : 0.0045s + [COUNTERS] PROGRAM TOTAL : 2.6766s + [COUNTERS] Fortran Overhead ( 0 ) : 0.2935s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.3784s for 8192 events => throughput is 3.44E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0047s *** (2-sse4) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -167,15 +227,50 @@ OK! xsec from fortran (0.33144786561240197) and cpp (0.33144786561240192) differ OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-sse4) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +104 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.sse4_d_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_ggttgg_x10_cudacpp > /tmp/avalassi/output_ggttgg_x10_cudacpp' +DEBUG: MEK processed 81920 events across 123 channels { 112 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 64/64 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 104 + [XSECTION] ChannelId = 112 + [XSECTION] Cross section = 0.2093 [0.20930257969248325] fbridge_mode=1 + [UNWEIGHT] Wrote 17 events (found 331 events) + [COUNTERS] PROGRAM TOTAL : 25.8019s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0520s + [COUNTERS] CudaCpp MEs ( 2 ) : 23.7450s for 81920 events => throughput is 3.45E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0049s + +*** (2-sse4) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (0.20930257969248323) and cpp (0.20930257969248325) differ by less than 3E-14 (2.220446049250313e-16) + +*** (2-sse4) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+BRDHST/sse4+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.547129e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.570223e+03 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.542201e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.576065e+03 ) sec^-1 *** (2-avx2) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -199,9 +294,9 @@ DEBUG: MEK processed 8192 events across 123 channels { 112 : 8192 } [XSECTION] ChannelId = 112 [XSECTION] Cross section = 0.3314 [0.33144786561240197] fbridge_mode=1 [UNWEIGHT] Wrote 7 events (found 213 events) - [COUNTERS] PROGRAM TOTAL : 1.3332s - [COUNTERS] Fortran Overhead ( 0 ) : 0.2951s - [COUNTERS] CudaCpp MEs ( 2 ) : 1.0356s for 8192 events => throughput is 7.91E+03 events/s + [COUNTERS] PROGRAM TOTAL : 1.3258s + [COUNTERS] Fortran Overhead ( 0 ) : 0.2938s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.0295s for 8192 events => throughput is 7.96E+03 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0025s *** (2-avx2) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -212,15 +307,50 @@ OK! xsec from fortran (0.33144786561240197) and cpp (0.33144786561240197) differ OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-avx2) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +104 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.avx2_d_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_ggttgg_x10_cudacpp > /tmp/avalassi/output_ggttgg_x10_cudacpp' +DEBUG: MEK processed 81920 events across 123 channels { 112 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 64/64 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 104 + [XSECTION] ChannelId = 112 + [XSECTION] Cross section = 0.2093 [0.20930257969248320] fbridge_mode=1 + [UNWEIGHT] Wrote 17 events (found 331 events) + [COUNTERS] PROGRAM TOTAL : 12.3901s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0437s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.3440s for 81920 events => throughput is 7.92E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0024s + +*** (2-avx2) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (0.20930257969248323) and cpp (0.20930257969248320) differ by less than 3E-14 (1.1102230246251565e-16) + +*** (2-avx2) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+BRDHST/avx2+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 8.118919e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 8.173621e+03 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 8.114943e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 8.175971e+03 ) sec^-1 *** (2-512y) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -244,10 +374,10 @@ DEBUG: MEK processed 8192 events across 123 channels { 112 : 8192 } [XSECTION] ChannelId = 112 [XSECTION] Cross section = 0.3314 [0.33144786561240197] fbridge_mode=1 [UNWEIGHT] Wrote 7 events (found 213 events) - [COUNTERS] PROGRAM TOTAL : 1.2181s - [COUNTERS] Fortran Overhead ( 0 ) : 0.2951s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.9207s for 8192 events => throughput is 8.90E+03 events/s - [COUNTERS] CudaCpp HEL ( 3 ) : 0.0022s + [COUNTERS] PROGRAM TOTAL : 1.2047s + [COUNTERS] Fortran Overhead ( 0 ) : 0.2901s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.9125s for 8192 events => throughput is 8.98E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0021s *** (2-512y) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -257,15 +387,50 @@ OK! xsec from fortran (0.33144786561240197) and cpp (0.33144786561240197) differ OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-512y) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +104 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.512y_d_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_ggttgg_x10_cudacpp > /tmp/avalassi/output_ggttgg_x10_cudacpp' +DEBUG: MEK processed 81920 events across 123 channels { 112 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 64/64 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 104 + [XSECTION] ChannelId = 112 + [XSECTION] Cross section = 0.2093 [0.20930257969248320] fbridge_mode=1 + [UNWEIGHT] Wrote 17 events (found 331 events) + [COUNTERS] PROGRAM TOTAL : 11.1921s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0515s + [COUNTERS] CudaCpp MEs ( 2 ) : 9.1385s for 81920 events => throughput is 8.96E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0022s + +*** (2-512y) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (0.20930257969248323) and cpp (0.20930257969248320) differ by less than 3E-14 (1.1102230246251565e-16) + +*** (2-512y) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+BRDHST/512y+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 9.276674e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 9.288957e+03 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 9.241984e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 9.259020e+03 ) sec^-1 *** (2-512z) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -289,10 +454,10 @@ DEBUG: MEK processed 8192 events across 123 channels { 112 : 8192 } [XSECTION] ChannelId = 112 [XSECTION] Cross section = 0.3314 [0.33144786561240197] fbridge_mode=1 [UNWEIGHT] Wrote 7 events (found 213 events) - [COUNTERS] PROGRAM TOTAL : 1.4646s - [COUNTERS] Fortran Overhead ( 0 ) : 0.2912s - [COUNTERS] CudaCpp MEs ( 2 ) : 1.1708s for 8192 events => throughput is 7.00E+03 events/s - [COUNTERS] CudaCpp HEL ( 3 ) : 0.0026s + [COUNTERS] PROGRAM TOTAL : 1.4589s + [COUNTERS] Fortran Overhead ( 0 ) : 0.2916s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.1645s for 8192 events => throughput is 7.03E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0027s *** (2-512z) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -302,15 +467,50 @@ OK! xsec from fortran (0.33144786561240197) and cpp (0.33144786561240197) differ OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-512z) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +104 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.512z_d_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_ggttgg_x10_cudacpp > /tmp/avalassi/output_ggttgg_x10_cudacpp' +DEBUG: MEK processed 81920 events across 123 channels { 112 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 64/64 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 104 + [XSECTION] ChannelId = 112 + [XSECTION] Cross section = 0.2093 [0.20930257969248320] fbridge_mode=1 + [UNWEIGHT] Wrote 17 events (found 331 events) + [COUNTERS] PROGRAM TOTAL : 13.7765s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0514s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.7220s for 81920 events => throughput is 6.99E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0030s + +*** (2-512z) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (0.20930257969248323) and cpp (0.20930257969248320) differ by less than 3E-14 (1.1102230246251565e-16) + +*** (2-512z) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+BRDHST/512z+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 7.083404e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 7.136126e+03 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 7.099846e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 7.093853e+03 ) sec^-1 *** (3-cuda) EXECUTE MADEVENT_CUDA x1 (create events.lhe) *** -------------------- @@ -334,10 +534,10 @@ DEBUG: MEK processed 8192 events across 123 channels { 112 : 8192 } [XSECTION] ChannelId = 112 [XSECTION] Cross section = 0.3314 [0.33144786561240192] fbridge_mode=1 [UNWEIGHT] Wrote 7 events (found 213 events) - [COUNTERS] PROGRAM TOTAL : 0.8110s - [COUNTERS] Fortran Overhead ( 0 ) : 0.7374s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0388s for 8192 events => throughput is 2.11E+05 events/s - [COUNTERS] CudaCpp HEL ( 3 ) : 0.0348s + [COUNTERS] PROGRAM TOTAL : 0.8115s + [COUNTERS] Fortran Overhead ( 0 ) : 0.7382s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0385s for 8192 events => throughput is 2.13E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0347s *** (3-cuda) Compare MADEVENT_CUDA x1 xsec to MADEVENT_FORTRAN xsec *** @@ -347,45 +547,80 @@ OK! xsec from fortran (0.33144786561240197) and cuda (0.33144786561240192) diffe OK! events.lhe.cuda.1 and events.lhe.ref.1 are identical +*** (3-cuda) EXECUTE MADEVENT_CUDA x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +104 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.cuda_d_inl0_hrd0/madevent_cuda < /tmp/avalassi/input_ggttgg_x10_cudacpp > /tmp/avalassi/output_ggttgg_x10_cudacpp' +DEBUG: MEK processed 81920 events across 123 channels { 112 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 64/64 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 104 + [XSECTION] ChannelId = 112 + [XSECTION] Cross section = 0.2093 [0.20930257969248336] fbridge_mode=1 + [UNWEIGHT] Wrote 17 events (found 331 events) + [COUNTERS] PROGRAM TOTAL : 2.8500s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4788s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.3364s for 81920 events => throughput is 2.44E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0348s + +*** (3-cuda) Compare MADEVENT_CUDA x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (0.20930257969248323) and cuda (0.20930257969248336) differ by less than 3E-14 (6.661338147750939e-16) + +*** (3-cuda) Compare MADEVENT_CUDA x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cuda.10 and events.lhe.ref.10 are identical + *** EXECUTE GCHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SM_GG_TTXGG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 2.149005e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.156379e+05 ) sec^-1 *** EXECUTE GCHECK(8192) -p 256 32 1 *** Process = SIGMA_SM_GG_TTXGG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 2.350783e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.359030e+05 ) sec^-1 *** EXECUTE GCHECK(MAX) -p 16384 32 1 --bridge *** Process = SIGMA_SM_GG_TTXGG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 4.129093e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 4.122995e+05 ) sec^-1 *** EXECUTE GCHECK(MAX) -p 16384 32 1 *** Process = SIGMA_SM_GG_TTXGG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 4.172100e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 4.162119e+05 ) sec^-1 *** EXECUTE GCHECK(MAX128THR) -p 4096 128 1 --bridge *** Process = SIGMA_SM_GG_TTXGG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 4.126645e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 4.117451e+05 ) sec^-1 *** EXECUTE GCHECK(MAX128THR) -p 4096 128 1 *** Process = SIGMA_SM_GG_TTXGG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 4.170032e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 4.171532e+05 ) sec^-1 *** EXECUTE GCHECK(MAX8THR) -p 65536 8 1 --bridge *** Process = SIGMA_SM_GG_TTXGG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 4.144552e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 4.116589e+05 ) sec^-1 *** EXECUTE GCHECK(MAX8THR) -p 65536 8 1 *** Process = SIGMA_SM_GG_TTXGG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.426547e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.418271e+05 ) sec^-1 *** (3-hip) WARNING! SKIP MADEVENT_HIP (hip is not supported on this node) *** diff --git a/epochX/cudacpp/tmad/logs_ggttgg_mad/log_ggttgg_mad_f_inl0_hrd0.txt b/epochX/cudacpp/tmad/logs_ggttgg_mad/log_ggttgg_mad_f_inl0_hrd0.txt index 78567e12c9..e59c7c1fe9 100644 --- a/epochX/cudacpp/tmad/logs_ggttgg_mad/log_ggttgg_mad_f_inl0_hrd0.txt +++ b/epochX/cudacpp/tmad/logs_ggttgg_mad/log_ggttgg_mad_f_inl0_hrd0.txt @@ -1,10 +1,10 @@ Working directory (build): /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg - make USEBUILDDIR=1 BACKEND=cuda -make USEBUILDDIR=1 BACKEND=cppnone + +make USEBUILDDIR=1 BACKEND=cppnone make USEBUILDDIR=1 BACKEND=cppsse4 make USEBUILDDIR=1 BACKEND=cppavx2 @@ -32,7 +32,7 @@ make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ OMP_NUM_THREADS= -DATE: 2024-10-06_10:08:49 +DATE: 2024-10-06_14:32:17 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: Working directory (run): /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg @@ -58,9 +58,9 @@ Executing ' ./madevent_fortran < /tmp/avalassi/input_ggttgg_x1_fortran > /tmp/av [XSECTION] ChannelId = 112 [XSECTION] Cross section = 0.3314 [0.33144786561240197] fbridge_mode=0 [UNWEIGHT] Wrote 7 events (found 223 events) - [COUNTERS] PROGRAM TOTAL : 4.5864s - [COUNTERS] Fortran Overhead ( 0 ) : 0.2940s - [COUNTERS] Fortran MEs ( 1 ) : 4.2923s for 8192 events => throughput is 1.91E+03 events/s + [COUNTERS] PROGRAM TOTAL : 4.5877s + [COUNTERS] Fortran Overhead ( 0 ) : 0.2933s + [COUNTERS] Fortran MEs ( 1 ) : 4.2944s for 8192 events => throughput is 1.91E+03 events/s *** (1) EXECUTE MADEVENT_FORTRAN x1 (create events.lhe) *** -------------------- @@ -83,9 +83,34 @@ Executing ' ./madevent_fortran < /tmp/avalassi/input_ggttgg_x1_fortran > /tmp/av [XSECTION] ChannelId = 112 [XSECTION] Cross section = 0.3314 [0.33144786561240197] fbridge_mode=0 [UNWEIGHT] Wrote 7 events (found 213 events) - [COUNTERS] PROGRAM TOTAL : 4.5924s - [COUNTERS] Fortran Overhead ( 0 ) : 0.2886s - [COUNTERS] Fortran MEs ( 1 ) : 4.3038s for 8192 events => throughput is 1.90E+03 events/s + [COUNTERS] PROGRAM TOTAL : 4.5772s + [COUNTERS] Fortran Overhead ( 0 ) : 0.2899s + [COUNTERS] Fortran MEs ( 1 ) : 4.2873s for 8192 events => throughput is 1.91E+03 events/s + +*** (1) EXECUTE MADEVENT_FORTRAN x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +104 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./madevent_fortran < /tmp/avalassi/input_ggttgg_x10_fortran > /tmp/avalassi/output_ggttgg_x10_fortran' + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 64/64 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 104 + [XSECTION] ChannelId = 112 + [XSECTION] Cross section = 0.2093 [0.20930257969248323] fbridge_mode=0 + [UNWEIGHT] Wrote 17 events (found 331 events) + [COUNTERS] PROGRAM TOTAL : 44.9817s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0436s + [COUNTERS] Fortran MEs ( 1 ) : 42.9381s for 81920 events => throughput is 1.91E+03 events/s *** (2-none) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -109,10 +134,10 @@ DEBUG: MEK processed 8192 events across 123 channels { 112 : 8192 } [XSECTION] ChannelId = 112 [XSECTION] Cross section = 0.3314 [0.33144941544531159] fbridge_mode=1 [UNWEIGHT] Wrote 7 events (found 213 events) - [COUNTERS] PROGRAM TOTAL : 4.6210s - [COUNTERS] Fortran Overhead ( 0 ) : 0.2941s - [COUNTERS] CudaCpp MEs ( 2 ) : 4.3184s for 8192 events => throughput is 1.90E+03 events/s - [COUNTERS] CudaCpp HEL ( 3 ) : 0.0085s + [COUNTERS] PROGRAM TOTAL : 4.6181s + [COUNTERS] Fortran Overhead ( 0 ) : 0.2922s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.3172s for 8192 events => throughput is 1.90E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0086s *** (2-none) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -122,15 +147,51 @@ OK! xsec from fortran (0.33144786561240197) and cpp (0.33144941544531159) differ OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-none) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +104 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.none_f_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_ggttgg_x10_cudacpp > /tmp/avalassi/output_ggttgg_x10_cudacpp' +INFO: The following Floating Point Exceptions have been reported: FE_UNDERFLOW +DEBUG: MEK processed 81920 events across 123 channels { 112 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 64/64 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 104 + [XSECTION] ChannelId = 112 + [XSECTION] Cross section = 0.2093 [0.20930329135137288] fbridge_mode=1 + [UNWEIGHT] Wrote 17 events (found 331 events) + [COUNTERS] PROGRAM TOTAL : 45.3202s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0443s + [COUNTERS] CudaCpp MEs ( 2 ) : 43.2673s for 81920 events => throughput is 1.89E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0086s + +*** (2-none) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (0.20930257969248323) and cpp (0.20930329135137288) differ by less than 4E-4 (3.400143900211816e-06) + +*** (2-none) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+BRDHST/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.957206e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.949107e+03 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.957921e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.953244e+03 ) sec^-1 *** (2-sse4) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -154,10 +215,10 @@ DEBUG: MEK processed 8192 events across 123 channels { 112 : 8192 } [XSECTION] ChannelId = 112 [XSECTION] Cross section = 0.3314 [0.33144937378275385] fbridge_mode=1 [UNWEIGHT] Wrote 7 events (found 213 events) - [COUNTERS] PROGRAM TOTAL : 1.4924s - [COUNTERS] Fortran Overhead ( 0 ) : 0.2933s - [COUNTERS] CudaCpp MEs ( 2 ) : 1.1966s for 8192 events => throughput is 6.85E+03 events/s - [COUNTERS] CudaCpp HEL ( 3 ) : 0.0025s + [COUNTERS] PROGRAM TOTAL : 1.4894s + [COUNTERS] Fortran Overhead ( 0 ) : 0.2881s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.1989s for 8192 events => throughput is 6.83E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0024s *** (2-sse4) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -167,15 +228,50 @@ OK! xsec from fortran (0.33144786561240197) and cpp (0.33144937378275385) differ OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-sse4) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +104 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.sse4_f_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_ggttgg_x10_cudacpp > /tmp/avalassi/output_ggttgg_x10_cudacpp' +DEBUG: MEK processed 81920 events across 123 channels { 112 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 64/64 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 104 + [XSECTION] ChannelId = 112 + [XSECTION] Cross section = 0.2093 [0.20930324959819654] fbridge_mode=1 + [UNWEIGHT] Wrote 17 events (found 331 events) + [COUNTERS] PROGRAM TOTAL : 14.0466s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0561s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.9880s for 81920 events => throughput is 6.83E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0025s + +*** (2-sse4) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (0.20930257969248323) and cpp (0.20930324959819654) differ by less than 4E-4 (3.2006567445286294e-06) + +*** (2-sse4) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+BRDHST/sse4+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 7.048957e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 7.007724e+03 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 7.041651e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 7.026251e+03 ) sec^-1 *** (2-avx2) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -199,10 +295,10 @@ DEBUG: MEK processed 8192 events across 123 channels { 112 : 8192 } [XSECTION] ChannelId = 112 [XSECTION] Cross section = 0.3314 [0.33144939353225550] fbridge_mode=1 [UNWEIGHT] Wrote 7 events (found 213 events) - [COUNTERS] PROGRAM TOTAL : 0.8128s - [COUNTERS] Fortran Overhead ( 0 ) : 0.2926s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.5189s for 8192 events => throughput is 1.58E+04 events/s - [COUNTERS] CudaCpp HEL ( 3 ) : 0.0013s + [COUNTERS] PROGRAM TOTAL : 0.8149s + [COUNTERS] Fortran Overhead ( 0 ) : 0.2942s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5193s for 8192 events => throughput is 1.58E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0014s *** (2-avx2) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -212,15 +308,50 @@ OK! xsec from fortran (0.33144786561240197) and cpp (0.33144939353225550) differ OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-avx2) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +104 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.avx2_f_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_ggttgg_x10_cudacpp > /tmp/avalassi/output_ggttgg_x10_cudacpp' +DEBUG: MEK processed 81920 events across 123 channels { 112 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 64/64 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 104 + [XSECTION] ChannelId = 112 + [XSECTION] Cross section = 0.2093 [0.20930327551379133] fbridge_mode=1 + [UNWEIGHT] Wrote 17 events (found 331 events) + [COUNTERS] PROGRAM TOTAL : 7.2486s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0505s + [COUNTERS] CudaCpp MEs ( 2 ) : 5.1968s for 81920 events => throughput is 1.58E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0013s + +*** (2-avx2) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (0.20930257969248323) and cpp (0.20930327551379133) differ by less than 4E-4 (3.3244755468508913e-06) + +*** (2-avx2) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+BRDHST/avx2+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.622272e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.611823e+04 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.613287e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.620649e+04 ) sec^-1 *** (2-512y) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -244,9 +375,9 @@ DEBUG: MEK processed 8192 events across 123 channels { 112 : 8192 } [XSECTION] ChannelId = 112 [XSECTION] Cross section = 0.3314 [0.33144939353225550] fbridge_mode=1 [UNWEIGHT] Wrote 7 events (found 213 events) - [COUNTERS] PROGRAM TOTAL : 0.7779s - [COUNTERS] Fortran Overhead ( 0 ) : 0.2950s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.4817s for 8192 events => throughput is 1.70E+04 events/s + [COUNTERS] PROGRAM TOTAL : 0.7669s + [COUNTERS] Fortran Overhead ( 0 ) : 0.2939s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4718s for 8192 events => throughput is 1.74E+04 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0012s *** (2-512y) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -257,15 +388,50 @@ OK! xsec from fortran (0.33144786561240197) and cpp (0.33144939353225550) differ OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-512y) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +104 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.512y_f_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_ggttgg_x10_cudacpp > /tmp/avalassi/output_ggttgg_x10_cudacpp' +DEBUG: MEK processed 81920 events across 123 channels { 112 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 64/64 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 104 + [XSECTION] ChannelId = 112 + [XSECTION] Cross section = 0.2093 [0.20930327551379133] fbridge_mode=1 + [UNWEIGHT] Wrote 17 events (found 331 events) + [COUNTERS] PROGRAM TOTAL : 6.7420s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0552s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.6857s for 81920 events => throughput is 1.75E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0011s + +*** (2-512y) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (0.20930257969248323) and cpp (0.20930327551379133) differ by less than 4E-4 (3.3244755468508913e-06) + +*** (2-512y) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+BRDHST/512y+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.826080e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.815680e+04 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.802534e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.825356e+04 ) sec^-1 *** (2-512z) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -289,9 +455,9 @@ DEBUG: MEK processed 8192 events across 123 channels { 112 : 8192 } [XSECTION] ChannelId = 112 [XSECTION] Cross section = 0.3314 [0.33144947551388249] fbridge_mode=1 [UNWEIGHT] Wrote 7 events (found 213 events) - [COUNTERS] PROGRAM TOTAL : 0.8771s - [COUNTERS] Fortran Overhead ( 0 ) : 0.2920s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.5836s for 8192 events => throughput is 1.40E+04 events/s + [COUNTERS] PROGRAM TOTAL : 0.8796s + [COUNTERS] Fortran Overhead ( 0 ) : 0.2902s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5879s for 8192 events => throughput is 1.39E+04 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0015s *** (2-512z) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -302,15 +468,50 @@ OK! xsec from fortran (0.33144786561240197) and cpp (0.33144947551388249) differ OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-512z) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +104 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.512z_f_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_ggttgg_x10_cudacpp > /tmp/avalassi/output_ggttgg_x10_cudacpp' +DEBUG: MEK processed 81920 events across 123 channels { 112 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 64/64 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 104 + [XSECTION] ChannelId = 112 + [XSECTION] Cross section = 0.2093 [0.20930331717025510] fbridge_mode=1 + [UNWEIGHT] Wrote 17 events (found 331 events) + [COUNTERS] PROGRAM TOTAL : 7.9126s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0553s + [COUNTERS] CudaCpp MEs ( 2 ) : 5.8560s for 81920 events => throughput is 1.40E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0013s + +*** (2-512z) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (0.20930257969248323) and cpp (0.20930331717025510) differ by less than 4E-4 (3.523500632152121e-06) + +*** (2-512z) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+BRDHST/512z+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.430502e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.430526e+04 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.421428e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.431690e+04 ) sec^-1 *** (3-cuda) EXECUTE MADEVENT_CUDA x1 (create events.lhe) *** -------------------- @@ -334,9 +535,9 @@ DEBUG: MEK processed 8192 events across 123 channels { 112 : 8192 } [XSECTION] ChannelId = 112 [XSECTION] Cross section = 0.3314 [0.33144955535316123] fbridge_mode=1 [UNWEIGHT] Wrote 7 events (found 213 events) - [COUNTERS] PROGRAM TOTAL : 0.7866s - [COUNTERS] Fortran Overhead ( 0 ) : 0.7350s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0271s for 8192 events => throughput is 3.02E+05 events/s + [COUNTERS] PROGRAM TOTAL : 0.7798s + [COUNTERS] Fortran Overhead ( 0 ) : 0.7283s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0270s for 8192 events => throughput is 3.03E+05 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0245s *** (3-cuda) Compare MADEVENT_CUDA x1 xsec to MADEVENT_FORTRAN xsec *** @@ -347,45 +548,80 @@ OK! xsec from fortran (0.33144786561240197) and cuda (0.33144955535316123) diffe OK! events.lhe.cuda.1 and events.lhe.ref.1 are identical +*** (3-cuda) EXECUTE MADEVENT_CUDA x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +104 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.cuda_f_inl0_hrd0/madevent_cuda < /tmp/avalassi/input_ggttgg_x10_cudacpp > /tmp/avalassi/output_ggttgg_x10_cudacpp' +DEBUG: MEK processed 81920 events across 123 channels { 112 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 64/64 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 104 + [XSECTION] ChannelId = 112 + [XSECTION] Cross section = 0.2093 [0.20930336562619947] fbridge_mode=1 + [UNWEIGHT] Wrote 17 events (found 331 events) + [COUNTERS] PROGRAM TOTAL : 2.7528s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4959s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2324s for 81920 events => throughput is 3.53E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0246s + +*** (3-cuda) Compare MADEVENT_CUDA x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (0.20930257969248323) and cuda (0.20930336562619947) differ by less than 4E-4 (3.755012085271403e-06) + +*** (3-cuda) Compare MADEVENT_CUDA x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cuda.10 and events.lhe.ref.10 are identical + *** EXECUTE GCHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SM_GG_TTXGG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.089397e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.120124e+05 ) sec^-1 *** EXECUTE GCHECK(8192) -p 256 32 1 *** Process = SIGMA_SM_GG_TTXGG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.388762e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.409750e+05 ) sec^-1 *** EXECUTE GCHECK(MAX) -p 16384 32 1 --bridge *** Process = SIGMA_SM_GG_TTXGG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 8.126017e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 8.096587e+05 ) sec^-1 *** EXECUTE GCHECK(MAX) -p 16384 32 1 *** Process = SIGMA_SM_GG_TTXGG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 8.254976e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 8.218896e+05 ) sec^-1 *** EXECUTE GCHECK(MAX128THR) -p 4096 128 1 --bridge *** Process = SIGMA_SM_GG_TTXGG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 8.087410e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 8.093287e+05 ) sec^-1 *** EXECUTE GCHECK(MAX128THR) -p 4096 128 1 *** Process = SIGMA_SM_GG_TTXGG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 8.221892e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 8.246943e+05 ) sec^-1 *** EXECUTE GCHECK(MAX8THR) -p 65536 8 1 --bridge *** Process = SIGMA_SM_GG_TTXGG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 8.084262e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 8.116341e+05 ) sec^-1 *** EXECUTE GCHECK(MAX8THR) -p 65536 8 1 *** Process = SIGMA_SM_GG_TTXGG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 2.392382e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.394183e+05 ) sec^-1 *** (3-hip) WARNING! SKIP MADEVENT_HIP (hip is not supported on this node) *** diff --git a/epochX/cudacpp/tmad/logs_ggttgg_mad/log_ggttgg_mad_m_inl0_hrd0.txt b/epochX/cudacpp/tmad/logs_ggttgg_mad/log_ggttgg_mad_m_inl0_hrd0.txt index 0f7d6f4131..68eccc20f3 100644 --- a/epochX/cudacpp/tmad/logs_ggttgg_mad/log_ggttgg_mad_m_inl0_hrd0.txt +++ b/epochX/cudacpp/tmad/logs_ggttgg_mad/log_ggttgg_mad_m_inl0_hrd0.txt @@ -20,9 +20,9 @@ make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg' make[1]: Nothing to be done for 'all'. -make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg' make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg' +make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg' make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg' make[1]: Nothing to be done for 'all'. @@ -32,7 +32,7 @@ make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ OMP_NUM_THREADS= -DATE: 2024-10-06_10:07:37 +DATE: 2024-10-06_14:28:26 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: Working directory (run): /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg @@ -58,9 +58,9 @@ Executing ' ./madevent_fortran < /tmp/avalassi/input_ggttgg_x1_fortran > /tmp/av [XSECTION] ChannelId = 112 [XSECTION] Cross section = 0.3314 [0.33144786561240197] fbridge_mode=0 [UNWEIGHT] Wrote 7 events (found 223 events) - [COUNTERS] PROGRAM TOTAL : 4.5989s - [COUNTERS] Fortran Overhead ( 0 ) : 0.2922s - [COUNTERS] Fortran MEs ( 1 ) : 4.3067s for 8192 events => throughput is 1.90E+03 events/s + [COUNTERS] PROGRAM TOTAL : 4.5759s + [COUNTERS] Fortran Overhead ( 0 ) : 0.2930s + [COUNTERS] Fortran MEs ( 1 ) : 4.2829s for 8192 events => throughput is 1.91E+03 events/s *** (1) EXECUTE MADEVENT_FORTRAN x1 (create events.lhe) *** -------------------- @@ -83,9 +83,34 @@ Executing ' ./madevent_fortran < /tmp/avalassi/input_ggttgg_x1_fortran > /tmp/av [XSECTION] ChannelId = 112 [XSECTION] Cross section = 0.3314 [0.33144786561240197] fbridge_mode=0 [UNWEIGHT] Wrote 7 events (found 213 events) - [COUNTERS] PROGRAM TOTAL : 4.6012s - [COUNTERS] Fortran Overhead ( 0 ) : 0.2899s - [COUNTERS] Fortran MEs ( 1 ) : 4.3113s for 8192 events => throughput is 1.90E+03 events/s + [COUNTERS] PROGRAM TOTAL : 4.5787s + [COUNTERS] Fortran Overhead ( 0 ) : 0.2896s + [COUNTERS] Fortran MEs ( 1 ) : 4.2891s for 8192 events => throughput is 1.91E+03 events/s + +*** (1) EXECUTE MADEVENT_FORTRAN x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +104 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./madevent_fortran < /tmp/avalassi/input_ggttgg_x10_fortran > /tmp/avalassi/output_ggttgg_x10_fortran' + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 64/64 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 104 + [XSECTION] ChannelId = 112 + [XSECTION] Cross section = 0.2093 [0.20930257969248323] fbridge_mode=0 + [UNWEIGHT] Wrote 17 events (found 331 events) + [COUNTERS] PROGRAM TOTAL : 45.0435s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0503s + [COUNTERS] Fortran MEs ( 1 ) : 42.9932s for 81920 events => throughput is 1.91E+03 events/s *** (2-none) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -109,9 +134,9 @@ DEBUG: MEK processed 8192 events across 123 channels { 112 : 8192 } [XSECTION] ChannelId = 112 [XSECTION] Cross section = 0.3314 [0.33144786734542164] fbridge_mode=1 [UNWEIGHT] Wrote 7 events (found 213 events) - [COUNTERS] PROGRAM TOTAL : 4.8059s - [COUNTERS] Fortran Overhead ( 0 ) : 0.2941s - [COUNTERS] CudaCpp MEs ( 2 ) : 4.5027s for 8192 events => throughput is 1.82E+03 events/s + [COUNTERS] PROGRAM TOTAL : 4.8241s + [COUNTERS] Fortran Overhead ( 0 ) : 0.2945s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.5205s for 8192 events => throughput is 1.81E+03 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0091s *** (2-none) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -122,15 +147,50 @@ OK! xsec from fortran (0.33144786561240197) and cpp (0.33144786734542164) differ OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-none) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +104 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.none_m_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_ggttgg_x10_cudacpp > /tmp/avalassi/output_ggttgg_x10_cudacpp' +DEBUG: MEK processed 81920 events across 123 channels { 112 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 64/64 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 104 + [XSECTION] ChannelId = 112 + [XSECTION] Cross section = 0.2093 [0.20930258048084049] fbridge_mode=1 + [UNWEIGHT] Wrote 17 events (found 331 events) + [COUNTERS] PROGRAM TOTAL : 47.2685s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0557s + [COUNTERS] CudaCpp MEs ( 2 ) : 45.2038s for 81920 events => throughput is 1.81E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0090s + +*** (2-none) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (0.20930257969248323) and cpp (0.20930258048084049) differ by less than 2E-4 (3.766591261111785e-09) + +*** (2-none) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+BRDHST/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.881337e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.880080e+03 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.867505e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.877249e+03 ) sec^-1 *** (2-sse4) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -154,10 +214,10 @@ DEBUG: MEK processed 8192 events across 123 channels { 112 : 8192 } [XSECTION] ChannelId = 112 [XSECTION] Cross section = 0.3314 [0.33144786651655289] fbridge_mode=1 [UNWEIGHT] Wrote 7 events (found 213 events) - [COUNTERS] PROGRAM TOTAL : 2.6829s - [COUNTERS] Fortran Overhead ( 0 ) : 0.2920s - [COUNTERS] CudaCpp MEs ( 2 ) : 2.3862s for 8192 events => throughput is 3.43E+03 events/s - [COUNTERS] CudaCpp HEL ( 3 ) : 0.0047s + [COUNTERS] PROGRAM TOTAL : 2.6811s + [COUNTERS] Fortran Overhead ( 0 ) : 0.2930s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.3832s for 8192 events => throughput is 3.44E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0048s *** (2-sse4) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -167,15 +227,50 @@ OK! xsec from fortran (0.33144786561240197) and cpp (0.33144786651655289) differ OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-sse4) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +104 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.sse4_m_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_ggttgg_x10_cudacpp > /tmp/avalassi/output_ggttgg_x10_cudacpp' +DEBUG: MEK processed 81920 events across 123 channels { 112 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 64/64 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 104 + [XSECTION] ChannelId = 112 + [XSECTION] Cross section = 0.2093 [0.20930258019984904] fbridge_mode=1 + [UNWEIGHT] Wrote 17 events (found 331 events) + [COUNTERS] PROGRAM TOTAL : 25.8741s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0484s + [COUNTERS] CudaCpp MEs ( 2 ) : 23.8206s for 81920 events => throughput is 3.44E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0050s + +*** (2-sse4) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (0.20930257969248323) and cpp (0.20930258019984904) differ by less than 2E-4 (2.424078271445751e-09) + +*** (2-sse4) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+BRDHST/sse4+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.548157e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.513925e+03 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.537868e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.507367e+03 ) sec^-1 *** (2-avx2) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -199,10 +294,10 @@ DEBUG: MEK processed 8192 events across 123 channels { 112 : 8192 } [XSECTION] ChannelId = 112 [XSECTION] Cross section = 0.3314 [0.33144786627894518] fbridge_mode=1 [UNWEIGHT] Wrote 7 events (found 213 events) - [COUNTERS] PROGRAM TOTAL : 1.3285s - [COUNTERS] Fortran Overhead ( 0 ) : 0.2936s - [COUNTERS] CudaCpp MEs ( 2 ) : 1.0325s for 8192 events => throughput is 7.93E+03 events/s - [COUNTERS] CudaCpp HEL ( 3 ) : 0.0024s + [COUNTERS] PROGRAM TOTAL : 1.3254s + [COUNTERS] Fortran Overhead ( 0 ) : 0.2919s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.0312s for 8192 events => throughput is 7.94E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0023s *** (2-avx2) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -212,15 +307,50 @@ OK! xsec from fortran (0.33144786561240197) and cpp (0.33144786627894518) differ OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-avx2) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +104 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.avx2_m_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_ggttgg_x10_cudacpp > /tmp/avalassi/output_ggttgg_x10_cudacpp' +DEBUG: MEK processed 81920 events across 123 channels { 112 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 64/64 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 104 + [XSECTION] ChannelId = 112 + [XSECTION] Cross section = 0.2093 [0.20930258019863174] fbridge_mode=1 + [UNWEIGHT] Wrote 17 events (found 331 events) + [COUNTERS] PROGRAM TOTAL : 12.4431s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0567s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.3841s for 81920 events => throughput is 7.89E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0024s + +*** (2-avx2) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (0.20930257969248323) and cpp (0.20930258019863174) differ by less than 2E-4 (2.4182622571089496e-09) + +*** (2-avx2) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+BRDHST/avx2+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 8.171504e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 8.123456e+03 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 8.183239e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 8.129449e+03 ) sec^-1 *** (2-512y) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -244,9 +374,9 @@ DEBUG: MEK processed 8192 events across 123 channels { 112 : 8192 } [XSECTION] ChannelId = 112 [XSECTION] Cross section = 0.3314 [0.33144786627894518] fbridge_mode=1 [UNWEIGHT] Wrote 7 events (found 213 events) - [COUNTERS] PROGRAM TOTAL : 1.1999s - [COUNTERS] Fortran Overhead ( 0 ) : 0.2927s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.9051s for 8192 events => throughput is 9.05E+03 events/s + [COUNTERS] PROGRAM TOTAL : 1.1939s + [COUNTERS] Fortran Overhead ( 0 ) : 0.2925s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8993s for 8192 events => throughput is 9.11E+03 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0022s *** (2-512y) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -257,15 +387,50 @@ OK! xsec from fortran (0.33144786561240197) and cpp (0.33144786627894518) differ OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-512y) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +104 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.512y_m_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_ggttgg_x10_cudacpp > /tmp/avalassi/output_ggttgg_x10_cudacpp' +DEBUG: MEK processed 81920 events across 123 channels { 112 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 64/64 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 104 + [XSECTION] ChannelId = 112 + [XSECTION] Cross section = 0.2093 [0.20930258019863174] fbridge_mode=1 + [UNWEIGHT] Wrote 17 events (found 331 events) + [COUNTERS] PROGRAM TOTAL : 11.0619s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0451s + [COUNTERS] CudaCpp MEs ( 2 ) : 9.0147s for 81920 events => throughput is 9.09E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0021s + +*** (2-512y) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (0.20930257969248323) and cpp (0.20930258019863174) differ by less than 2E-4 (2.4182622571089496e-09) + +*** (2-512y) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+BRDHST/512y+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 9.165581e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 9.316394e+03 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 9.350878e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 9.383818e+03 ) sec^-1 *** (2-512z) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -289,9 +454,9 @@ DEBUG: MEK processed 8192 events across 123 channels { 112 : 8192 } [XSECTION] ChannelId = 112 [XSECTION] Cross section = 0.3314 [0.33144786627894518] fbridge_mode=1 [UNWEIGHT] Wrote 7 events (found 213 events) - [COUNTERS] PROGRAM TOTAL : 1.4750s - [COUNTERS] Fortran Overhead ( 0 ) : 0.2928s - [COUNTERS] CudaCpp MEs ( 2 ) : 1.1793s for 8192 events => throughput is 6.95E+03 events/s + [COUNTERS] PROGRAM TOTAL : 1.4801s + [COUNTERS] Fortran Overhead ( 0 ) : 0.2911s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.1861s for 8192 events => throughput is 6.91E+03 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0029s *** (2-512z) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -302,15 +467,50 @@ OK! xsec from fortran (0.33144786561240197) and cpp (0.33144786627894518) differ OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-512z) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +104 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.512z_m_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_ggttgg_x10_cudacpp > /tmp/avalassi/output_ggttgg_x10_cudacpp' +DEBUG: MEK processed 81920 events across 123 channels { 112 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 64/64 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 104 + [XSECTION] ChannelId = 112 + [XSECTION] Cross section = 0.2093 [0.20930258019863174] fbridge_mode=1 + [UNWEIGHT] Wrote 17 events (found 331 events) + [COUNTERS] PROGRAM TOTAL : 13.8612s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0540s + [COUNTERS] CudaCpp MEs ( 2 ) : 11.8043s for 81920 events => throughput is 6.94E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0028s + +*** (2-512z) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (0.20930257969248323) and cpp (0.20930258019863174) differ by less than 2E-4 (2.4182622571089496e-09) + +*** (2-512z) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+BRDHST/512z+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 7.035517e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 7.041390e+03 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 6.843003e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 7.053393e+03 ) sec^-1 *** (3-cuda) EXECUTE MADEVENT_CUDA x1 (create events.lhe) *** -------------------- @@ -334,9 +534,9 @@ DEBUG: MEK processed 8192 events across 123 channels { 112 : 8192 } [XSECTION] ChannelId = 112 [XSECTION] Cross section = 0.3314 [0.33144786533876569] fbridge_mode=1 [UNWEIGHT] Wrote 7 events (found 213 events) - [COUNTERS] PROGRAM TOTAL : 0.8136s - [COUNTERS] Fortran Overhead ( 0 ) : 0.7401s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0389s for 8192 events => throughput is 2.11E+05 events/s + [COUNTERS] PROGRAM TOTAL : 0.8105s + [COUNTERS] Fortran Overhead ( 0 ) : 0.7372s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0385s for 8192 events => throughput is 2.13E+05 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0347s *** (3-cuda) Compare MADEVENT_CUDA x1 xsec to MADEVENT_FORTRAN xsec *** @@ -347,45 +547,80 @@ OK! xsec from fortran (0.33144786561240197) and cuda (0.33144786533876569) diffe OK! events.lhe.cuda.1 and events.lhe.ref.1 are identical +*** (3-cuda) EXECUTE MADEVENT_CUDA x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +104 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.cuda_m_inl0_hrd0/madevent_cuda < /tmp/avalassi/input_ggttgg_x10_cudacpp > /tmp/avalassi/output_ggttgg_x10_cudacpp' +DEBUG: MEK processed 81920 events across 123 channels { 112 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 64/64 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 104 + [XSECTION] ChannelId = 112 + [XSECTION] Cross section = 0.2093 [0.20930258003933860] fbridge_mode=1 + [UNWEIGHT] Wrote 17 events (found 331 events) + [COUNTERS] PROGRAM TOTAL : 2.8444s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4708s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.3391s for 81920 events => throughput is 2.42E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0345s + +*** (3-cuda) Compare MADEVENT_CUDA x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (0.20930257969248323) and cuda (0.20930258003933860) differ by less than 2E-4 (1.6571959360334176e-09) + +*** (3-cuda) Compare MADEVENT_CUDA x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cuda.10 and events.lhe.ref.10 are identical + *** EXECUTE GCHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SM_GG_TTXGG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:MIX+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 2.142259e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.157649e+05 ) sec^-1 *** EXECUTE GCHECK(8192) -p 256 32 1 *** Process = SIGMA_SM_GG_TTXGG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:MIX+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 2.350796e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.358771e+05 ) sec^-1 *** EXECUTE GCHECK(MAX) -p 16384 32 1 --bridge *** Process = SIGMA_SM_GG_TTXGG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:MIX+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 4.127674e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 4.125489e+05 ) sec^-1 *** EXECUTE GCHECK(MAX) -p 16384 32 1 *** Process = SIGMA_SM_GG_TTXGG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:MIX+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 4.154284e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 4.168513e+05 ) sec^-1 *** EXECUTE GCHECK(MAX128THR) -p 4096 128 1 --bridge *** Process = SIGMA_SM_GG_TTXGG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:MIX+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 4.123213e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 4.120413e+05 ) sec^-1 *** EXECUTE GCHECK(MAX128THR) -p 4096 128 1 *** Process = SIGMA_SM_GG_TTXGG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:MIX+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 4.173815e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 4.171027e+05 ) sec^-1 *** EXECUTE GCHECK(MAX8THR) -p 65536 8 1 --bridge *** Process = SIGMA_SM_GG_TTXGG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:MIX+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 4.121978e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 4.139470e+05 ) sec^-1 *** EXECUTE GCHECK(MAX8THR) -p 65536 8 1 *** Process = SIGMA_SM_GG_TTXGG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:MIX+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.416494e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.428946e+05 ) sec^-1 *** (3-hip) WARNING! SKIP MADEVENT_HIP (hip is not supported on this node) *** diff --git a/epochX/cudacpp/tmad/logs_ggttggg_mad/log_ggttggg_mad_d_inl0_hrd0.txt b/epochX/cudacpp/tmad/logs_ggttggg_mad/log_ggttggg_mad_d_inl0_hrd0.txt index 74862dd5f7..096671eecc 100644 --- a/epochX/cudacpp/tmad/logs_ggttggg_mad/log_ggttggg_mad_d_inl0_hrd0.txt +++ b/epochX/cudacpp/tmad/logs_ggttggg_mad/log_ggttggg_mad_d_inl0_hrd0.txt @@ -2,10 +2,10 @@ Working directory (build): /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/g make USEBUILDDIR=1 BACKEND=cuda +make USEBUILDDIR=1 BACKEND=cppnone -make USEBUILDDIR=1 BACKEND=cppsse4 -make USEBUILDDIR=1 BACKEND=cppnone +make USEBUILDDIR=1 BACKEND=cppsse4 make USEBUILDDIR=1 BACKEND=cppavx2 make USEBUILDDIR=1 BACKEND=cpp512y @@ -32,7 +32,7 @@ make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ OMP_NUM_THREADS= -DATE: 2024-10-06_10:10:25 +DATE: 2024-10-06_14:36:55 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: Working directory (run): /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttggg.mad/SubProcesses/P1_gg_ttxggg @@ -58,9 +58,9 @@ Executing ' ./madevent_fortran < /tmp/avalassi/input_ggttggg_x1_fortran > /tmp/a [XSECTION] ChannelId = 1 [XSECTION] Cross section = 2.357e-07 [2.3572561551282417E-007] fbridge_mode=0 [UNWEIGHT] Wrote 1 events (found 285 events) - [COUNTERS] PROGRAM TOTAL : 100.9475s - [COUNTERS] Fortran Overhead ( 0 ) : 0.5365s - [COUNTERS] Fortran MEs ( 1 ) : 100.4109s for 8192 events => throughput is 8.16E+01 events/s + [COUNTERS] PROGRAM TOTAL : 100.9965s + [COUNTERS] Fortran Overhead ( 0 ) : 0.5300s + [COUNTERS] Fortran MEs ( 1 ) : 100.4666s for 8192 events => throughput is 8.15E+01 events/s *** (1) EXECUTE MADEVENT_FORTRAN x1 (create events.lhe) *** -------------------- @@ -83,9 +83,34 @@ Executing ' ./madevent_fortran < /tmp/avalassi/input_ggttggg_x1_fortran > /tmp/a [XSECTION] ChannelId = 1 [XSECTION] Cross section = 2.357e-07 [2.3572561551282417E-007] fbridge_mode=0 [UNWEIGHT] Wrote 18 events (found 285 events) - [COUNTERS] PROGRAM TOTAL : 100.8105s - [COUNTERS] Fortran Overhead ( 0 ) : 0.5296s - [COUNTERS] Fortran MEs ( 1 ) : 100.2810s for 8192 events => throughput is 8.17E+01 events/s + [COUNTERS] PROGRAM TOTAL : 100.7427s + [COUNTERS] Fortran Overhead ( 0 ) : 0.5303s + [COUNTERS] Fortran MEs ( 1 ) : 100.2123s for 8192 events => throughput is 8.17E+01 events/s + +*** (1) EXECUTE MADEVENT_FORTRAN x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./madevent_fortran < /tmp/avalassi/input_ggttggg_x10_fortran > /tmp/avalassi/output_ggttggg_x10_fortran' + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 128/128 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 2.284e-07 [2.2842713115633741E-007] fbridge_mode=0 + [UNWEIGHT] Wrote 380 events (found 1707 events) + [COUNTERS] PROGRAM TOTAL : 1006.8475s + [COUNTERS] Fortran Overhead ( 0 ) : 4.5483s + [COUNTERS] Fortran MEs ( 1 ) : 1002.2992s for 81920 events => throughput is 8.17E+01 events/s *** (2-none) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -109,10 +134,10 @@ DEBUG: MEK processed 8192 events across 1240 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 2.357e-07 [2.3572561551282475E-007] fbridge_mode=1 [UNWEIGHT] Wrote 18 events (found 285 events) - [COUNTERS] PROGRAM TOTAL : 127.1376s - [COUNTERS] Fortran Overhead ( 0 ) : 0.5284s - [COUNTERS] CudaCpp MEs ( 2 ) : 126.4018s for 8192 events => throughput is 6.48E+01 events/s - [COUNTERS] CudaCpp HEL ( 3 ) : 0.2074s + [COUNTERS] PROGRAM TOTAL : 120.9629s + [COUNTERS] Fortran Overhead ( 0 ) : 0.5286s + [COUNTERS] CudaCpp MEs ( 2 ) : 120.2343s for 8192 events => throughput is 6.81E+01 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.2000s *** (2-none) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -122,15 +147,50 @@ OK! xsec from fortran (2.3572561551282417E-007) and cpp (2.3572561551282475E-007 OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-none) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.none_d_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_ggttggg_x10_cudacpp > /tmp/avalassi/output_ggttggg_x10_cudacpp' +DEBUG: MEK processed 81920 events across 1240 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 128/128 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 2.284e-07 [2.2842713115633775E-007] fbridge_mode=1 + [UNWEIGHT] Wrote 380 events (found 1707 events) + [COUNTERS] PROGRAM TOTAL : 1275.9041s + [COUNTERS] Fortran Overhead ( 0 ) : 4.4593s + [COUNTERS] CudaCpp MEs ( 2 ) : 1271.2430s for 81920 events => throughput is 6.44E+01 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.2017s + +*** (2-none) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (2.2842713115633741E-007) and cpp (2.2842713115633775E-007) differ by less than 3E-14 (1.5543122344752192e-15) + +*** (2-none) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SM_GG_TTXGGG_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+BRDHST/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 7.678586e+01 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 7.694624e+01 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_SM_GG_TTXGGG_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 7.694101e+01 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 7.661281e+01 ) sec^-1 *** (2-sse4) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -154,10 +214,10 @@ DEBUG: MEK processed 8192 events across 1240 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 2.357e-07 [2.3572561551282467E-007] fbridge_mode=1 [UNWEIGHT] Wrote 18 events (found 285 events) - [COUNTERS] PROGRAM TOTAL : 61.7097s - [COUNTERS] Fortran Overhead ( 0 ) : 0.5331s - [COUNTERS] CudaCpp MEs ( 2 ) : 61.0765s for 8192 events => throughput is 1.34E+02 events/s - [COUNTERS] CudaCpp HEL ( 3 ) : 0.1001s + [COUNTERS] PROGRAM TOTAL : 62.2118s + [COUNTERS] Fortran Overhead ( 0 ) : 0.5252s + [COUNTERS] CudaCpp MEs ( 2 ) : 61.5861s for 8192 events => throughput is 1.33E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1006s *** (2-sse4) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -167,15 +227,50 @@ OK! xsec from fortran (2.3572561551282417E-007) and cpp (2.3572561551282467E-007 OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-sse4) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.sse4_d_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_ggttggg_x10_cudacpp > /tmp/avalassi/output_ggttggg_x10_cudacpp' +DEBUG: MEK processed 81920 events across 1240 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 128/128 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 2.284e-07 [2.2842713115633781E-007] fbridge_mode=1 + [UNWEIGHT] Wrote 380 events (found 1707 events) + [COUNTERS] PROGRAM TOTAL : 611.1338s + [COUNTERS] Fortran Overhead ( 0 ) : 4.3705s + [COUNTERS] CudaCpp MEs ( 2 ) : 606.6620s for 81920 events => throughput is 1.35E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1012s + +*** (2-sse4) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (2.2842713115633741E-007) and cpp (2.2842713115633781E-007) differ by less than 3E-14 (1.7763568394002505e-15) + +*** (2-sse4) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SM_GG_TTXGGG_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+BRDHST/sse4+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.591189e+02 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.655107e+02 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_SM_GG_TTXGGG_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.580161e+02 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.652262e+02 ) sec^-1 *** (2-avx2) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -199,10 +294,10 @@ DEBUG: MEK processed 8192 events across 1240 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 2.357e-07 [2.3572561551282467E-007] fbridge_mode=1 [UNWEIGHT] Wrote 18 events (found 285 events) - [COUNTERS] PROGRAM TOTAL : 29.3577s - [COUNTERS] Fortran Overhead ( 0 ) : 0.5263s - [COUNTERS] CudaCpp MEs ( 2 ) : 28.7837s for 8192 events => throughput is 2.85E+02 events/s - [COUNTERS] CudaCpp HEL ( 3 ) : 0.0477s + [COUNTERS] PROGRAM TOTAL : 28.2174s + [COUNTERS] Fortran Overhead ( 0 ) : 0.5057s + [COUNTERS] CudaCpp MEs ( 2 ) : 27.6658s for 8192 events => throughput is 2.96E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0459s *** (2-avx2) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -212,15 +307,50 @@ OK! xsec from fortran (2.3572561551282417E-007) and cpp (2.3572561551282467E-007 OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-avx2) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.avx2_d_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_ggttggg_x10_cudacpp > /tmp/avalassi/output_ggttggg_x10_cudacpp' +DEBUG: MEK processed 81920 events across 1240 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 128/128 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 2.284e-07 [2.2842713115633781E-007] fbridge_mode=1 + [UNWEIGHT] Wrote 380 events (found 1707 events) + [COUNTERS] PROGRAM TOTAL : 281.9236s + [COUNTERS] Fortran Overhead ( 0 ) : 4.3390s + [COUNTERS] CudaCpp MEs ( 2 ) : 277.5388s for 81920 events => throughput is 2.95E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0458s + +*** (2-avx2) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (2.2842713115633741E-007) and cpp (2.2842713115633781E-007) differ by less than 3E-14 (1.7763568394002505e-15) + +*** (2-avx2) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SM_GG_TTXGGG_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+BRDHST/avx2+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.407090e+02 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.547863e+02 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_SM_GG_TTXGGG_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.415212e+02 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.560509e+02 ) sec^-1 *** (2-512y) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -244,10 +374,10 @@ DEBUG: MEK processed 8192 events across 1240 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 2.357e-07 [2.3572561551282467E-007] fbridge_mode=1 [UNWEIGHT] Wrote 18 events (found 285 events) - [COUNTERS] PROGRAM TOTAL : 26.2469s - [COUNTERS] Fortran Overhead ( 0 ) : 0.5271s - [COUNTERS] CudaCpp MEs ( 2 ) : 25.6788s for 8192 events => throughput is 3.19E+02 events/s - [COUNTERS] CudaCpp HEL ( 3 ) : 0.0410s + [COUNTERS] PROGRAM TOTAL : 26.5460s + [COUNTERS] Fortran Overhead ( 0 ) : 0.5084s + [COUNTERS] CudaCpp MEs ( 2 ) : 25.9973s for 8192 events => throughput is 3.15E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0403s *** (2-512y) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -257,15 +387,50 @@ OK! xsec from fortran (2.3572561551282417E-007) and cpp (2.3572561551282467E-007 OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-512y) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.512y_d_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_ggttggg_x10_cudacpp > /tmp/avalassi/output_ggttggg_x10_cudacpp' +DEBUG: MEK processed 81920 events across 1240 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 128/128 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 2.284e-07 [2.2842713115633781E-007] fbridge_mode=1 + [UNWEIGHT] Wrote 380 events (found 1707 events) + [COUNTERS] PROGRAM TOTAL : 263.4329s + [COUNTERS] Fortran Overhead ( 0 ) : 4.3171s + [COUNTERS] CudaCpp MEs ( 2 ) : 259.0753s for 81920 events => throughput is 3.16E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0404s + +*** (2-512y) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (2.2842713115633741E-007) and cpp (2.2842713115633781E-007) differ by less than 3E-14 (1.7763568394002505e-15) + +*** (2-512y) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SM_GG_TTXGGG_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+BRDHST/512y+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.913687e+02 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.987108e+02 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_SM_GG_TTXGGG_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.895964e+02 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.970312e+02 ) sec^-1 *** (2-512z) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -289,10 +454,10 @@ DEBUG: MEK processed 8192 events across 1240 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 2.357e-07 [2.3572561551282467E-007] fbridge_mode=1 [UNWEIGHT] Wrote 18 events (found 285 events) - [COUNTERS] PROGRAM TOTAL : 26.1607s - [COUNTERS] Fortran Overhead ( 0 ) : 0.5255s - [COUNTERS] CudaCpp MEs ( 2 ) : 25.5871s for 8192 events => throughput is 3.20E+02 events/s - [COUNTERS] CudaCpp HEL ( 3 ) : 0.0481s + [COUNTERS] PROGRAM TOTAL : 24.8758s + [COUNTERS] Fortran Overhead ( 0 ) : 0.5213s + [COUNTERS] CudaCpp MEs ( 2 ) : 24.3056s for 8192 events => throughput is 3.37E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0488s *** (2-512z) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -302,15 +467,50 @@ OK! xsec from fortran (2.3572561551282417E-007) and cpp (2.3572561551282467E-007 OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-512z) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.512z_d_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_ggttggg_x10_cudacpp > /tmp/avalassi/output_ggttggg_x10_cudacpp' +DEBUG: MEK processed 81920 events across 1240 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 128/128 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 2.284e-07 [2.2842713115633781E-007] fbridge_mode=1 + [UNWEIGHT] Wrote 380 events (found 1707 events) + [COUNTERS] PROGRAM TOTAL : 245.3750s + [COUNTERS] Fortran Overhead ( 0 ) : 4.2852s + [COUNTERS] CudaCpp MEs ( 2 ) : 241.0443s for 81920 events => throughput is 3.40E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0456s + +*** (2-512z) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (2.2842713115633741E-007) and cpp (2.2842713115633781E-007) differ by less than 3E-14 (1.7763568394002505e-15) + +*** (2-512z) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SM_GG_TTXGGG_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+BRDHST/512z+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.408791e+02 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.681291e+02 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_SM_GG_TTXGGG_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.444614e+02 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.662358e+02 ) sec^-1 *** (3-cuda) EXECUTE MADEVENT_CUDA x1 (create events.lhe) *** -------------------- @@ -334,10 +534,10 @@ DEBUG: MEK processed 8192 events across 1240 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 2.357e-07 [2.3572561551282475E-007] fbridge_mode=1 [UNWEIGHT] Wrote 18 events (found 285 events) - [COUNTERS] PROGRAM TOTAL : 3.3131s - [COUNTERS] Fortran Overhead ( 0 ) : 1.1215s - [COUNTERS] CudaCpp MEs ( 2 ) : 1.1040s for 8192 events => throughput is 7.42E+03 events/s - [COUNTERS] CudaCpp HEL ( 3 ) : 1.0875s + [COUNTERS] PROGRAM TOTAL : 3.1859s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0006s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.0970s for 8192 events => throughput is 7.47E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 1.0883s *** (3-cuda) Compare MADEVENT_CUDA x1 xsec to MADEVENT_FORTRAN xsec *** @@ -347,45 +547,80 @@ OK! xsec from fortran (2.3572561551282417E-007) and cuda (2.3572561551282475E-00 OK! events.lhe.cuda.1 and events.lhe.ref.1 are identical +*** (3-cuda) EXECUTE MADEVENT_CUDA x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.cuda_d_inl0_hrd0/madevent_cuda < /tmp/avalassi/input_ggttggg_x10_cudacpp > /tmp/avalassi/output_ggttggg_x10_cudacpp' +DEBUG: MEK processed 81920 events across 1240 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 128/128 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 2.284e-07 [2.2842713115633791E-007] fbridge_mode=1 + [UNWEIGHT] Wrote 380 events (found 1707 events) + [COUNTERS] PROGRAM TOTAL : 16.7355s + [COUNTERS] Fortran Overhead ( 0 ) : 4.8020s + [COUNTERS] CudaCpp MEs ( 2 ) : 10.8469s for 81920 events => throughput is 7.55E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 1.0867s + +*** (3-cuda) Compare MADEVENT_CUDA x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (2.2842713115633741E-007) and cuda (2.2842713115633791E-007) differ by less than 3E-14 (2.220446049250313e-15) + +*** (3-cuda) Compare MADEVENT_CUDA x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cuda.10 and events.lhe.ref.10 are identical + *** EXECUTE GCHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SM_GG_TTXGGG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 7.491511e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 7.519704e+03 ) sec^-1 *** EXECUTE GCHECK(8192) -p 256 32 1 *** Process = SIGMA_SM_GG_TTXGGG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 9.275455e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 9.248750e+03 ) sec^-1 *** EXECUTE GCHECK(MAX) -p 512 32 1 --bridge *** Process = SIGMA_SM_GG_TTXGGG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 9.282089e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 9.237963e+03 ) sec^-1 *** EXECUTE GCHECK(MAX) -p 512 32 1 *** Process = SIGMA_SM_GG_TTXGGG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 9.552042e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 9.548393e+03 ) sec^-1 *** EXECUTE GCHECK(MAX128THR) -p 128 128 1 --bridge *** Process = SIGMA_SM_GG_TTXGGG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 9.301465e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 9.237735e+03 ) sec^-1 *** EXECUTE GCHECK(MAX128THR) -p 128 128 1 *** Process = SIGMA_SM_GG_TTXGGG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 9.448921e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 9.444076e+03 ) sec^-1 *** EXECUTE GCHECK(MAX8THR) -p 2048 8 1 --bridge *** Process = SIGMA_SM_GG_TTXGGG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 9.252906e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 9.216168e+03 ) sec^-1 *** EXECUTE GCHECK(MAX8THR) -p 2048 8 1 *** Process = SIGMA_SM_GG_TTXGGG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.241973e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.240361e+03 ) sec^-1 *** (3-hip) WARNING! SKIP MADEVENT_HIP (hip is not supported on this node) *** diff --git a/epochX/cudacpp/tmad/logs_ggttggg_mad/log_ggttggg_mad_f_inl0_hrd0.txt b/epochX/cudacpp/tmad/logs_ggttggg_mad/log_ggttggg_mad_f_inl0_hrd0.txt index bfa4b4cda4..37e3e168af 100644 --- a/epochX/cudacpp/tmad/logs_ggttggg_mad/log_ggttggg_mad_f_inl0_hrd0.txt +++ b/epochX/cudacpp/tmad/logs_ggttggg_mad/log_ggttggg_mad_f_inl0_hrd0.txt @@ -5,9 +5,9 @@ make USEBUILDDIR=1 BACKEND=cuda make USEBUILDDIR=1 BACKEND=cppnone - make USEBUILDDIR=1 BACKEND=cppsse4 make USEBUILDDIR=1 BACKEND=cppavx2 + make USEBUILDDIR=1 BACKEND=cpp512y make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttggg.mad/SubProcesses/P1_gg_ttxggg' make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttggg.mad/SubProcesses/P1_gg_ttxggg' @@ -32,7 +32,7 @@ make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ OMP_NUM_THREADS= -DATE: 2024-10-06_10:42:40 +DATE: 2024-10-06_17:09:10 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: Working directory (run): /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttggg.mad/SubProcesses/P1_gg_ttxggg @@ -58,9 +58,9 @@ Executing ' ./madevent_fortran < /tmp/avalassi/input_ggttggg_x1_fortran > /tmp/a [XSECTION] ChannelId = 1 [XSECTION] Cross section = 2.357e-07 [2.3572561551282417E-007] fbridge_mode=0 [UNWEIGHT] Wrote 1 events (found 285 events) - [COUNTERS] PROGRAM TOTAL : 100.8152s - [COUNTERS] Fortran Overhead ( 0 ) : 0.5282s - [COUNTERS] Fortran MEs ( 1 ) : 100.2871s for 8192 events => throughput is 8.17E+01 events/s + [COUNTERS] PROGRAM TOTAL : 98.3897s + [COUNTERS] Fortran Overhead ( 0 ) : 0.5153s + [COUNTERS] Fortran MEs ( 1 ) : 97.8744s for 8192 events => throughput is 8.37E+01 events/s *** (1) EXECUTE MADEVENT_FORTRAN x1 (create events.lhe) *** -------------------- @@ -83,9 +83,34 @@ Executing ' ./madevent_fortran < /tmp/avalassi/input_ggttggg_x1_fortran > /tmp/a [XSECTION] ChannelId = 1 [XSECTION] Cross section = 2.357e-07 [2.3572561551282417E-007] fbridge_mode=0 [UNWEIGHT] Wrote 18 events (found 285 events) - [COUNTERS] PROGRAM TOTAL : 100.7247s - [COUNTERS] Fortran Overhead ( 0 ) : 0.5322s - [COUNTERS] Fortran MEs ( 1 ) : 100.1925s for 8192 events => throughput is 8.18E+01 events/s + [COUNTERS] PROGRAM TOTAL : 98.5961s + [COUNTERS] Fortran Overhead ( 0 ) : 0.5204s + [COUNTERS] Fortran MEs ( 1 ) : 98.0757s for 8192 events => throughput is 8.35E+01 events/s + +*** (1) EXECUTE MADEVENT_FORTRAN x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./madevent_fortran < /tmp/avalassi/input_ggttggg_x10_fortran > /tmp/avalassi/output_ggttggg_x10_fortran' + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 128/128 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 2.284e-07 [2.2842713115633741E-007] fbridge_mode=0 + [UNWEIGHT] Wrote 380 events (found 1707 events) + [COUNTERS] PROGRAM TOTAL : 979.7427s + [COUNTERS] Fortran Overhead ( 0 ) : 4.4333s + [COUNTERS] Fortran MEs ( 1 ) : 975.3094s for 81920 events => throughput is 8.40E+01 events/s *** (2-none) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -110,10 +135,10 @@ DEBUG: MEK processed 8192 events across 1240 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 2.358e-07 [2.3575849446922190E-007] fbridge_mode=1 [UNWEIGHT] Wrote 18 events (found 285 events) - [COUNTERS] PROGRAM TOTAL : 112.7914s - [COUNTERS] Fortran Overhead ( 0 ) : 0.5240s - [COUNTERS] CudaCpp MEs ( 2 ) : 112.0829s for 8192 events => throughput is 7.31E+01 events/s - [COUNTERS] CudaCpp HEL ( 3 ) : 0.1845s + [COUNTERS] PROGRAM TOTAL : 110.0306s + [COUNTERS] Fortran Overhead ( 0 ) : 0.5186s + [COUNTERS] CudaCpp MEs ( 2 ) : 109.3329s for 8192 events => throughput is 7.49E+01 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1790s *** (2-none) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -123,15 +148,51 @@ OK! xsec from fortran (2.3572561551282417E-007) and cpp (2.3575849446922190E-007 OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-none) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.none_f_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_ggttggg_x10_cudacpp > /tmp/avalassi/output_ggttggg_x10_cudacpp' +INFO: The following Floating Point Exceptions have been reported: FE_UNDERFLOW +DEBUG: MEK processed 81920 events across 1240 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 128/128 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 2.285e-07 [2.2845954405861011E-007] fbridge_mode=1 + [UNWEIGHT] Wrote 380 events (found 1707 events) + [COUNTERS] PROGRAM TOTAL : 1110.7095s + [COUNTERS] Fortran Overhead ( 0 ) : 4.3349s + [COUNTERS] CudaCpp MEs ( 2 ) : 1106.1951s for 81920 events => throughput is 7.41E+01 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1795s + +*** (2-none) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (2.2842713115633741E-007) and cpp (2.2845954405861011E-007) differ by less than 4E-4 (0.00014189602657355138) + +*** (2-none) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SM_GG_TTXGGG_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+BRDHST/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 8.631916e+01 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 8.897566e+01 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_SM_GG_TTXGGG_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 8.625132e+01 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 8.895980e+01 ) sec^-1 *** (2-sse4) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -156,9 +217,9 @@ DEBUG: MEK processed 8192 events across 1240 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 2.358e-07 [2.3575845178322101E-007] fbridge_mode=1 [UNWEIGHT] Wrote 18 events (found 285 events) - [COUNTERS] PROGRAM TOTAL : 28.7980s - [COUNTERS] Fortran Overhead ( 0 ) : 0.5271s - [COUNTERS] CudaCpp MEs ( 2 ) : 28.2235s for 8192 events => throughput is 2.90E+02 events/s + [COUNTERS] PROGRAM TOTAL : 27.6885s + [COUNTERS] Fortran Overhead ( 0 ) : 0.5126s + [COUNTERS] CudaCpp MEs ( 2 ) : 27.1285s for 8192 events => throughput is 3.02E+02 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0474s *** (2-sse4) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -169,15 +230,51 @@ OK! xsec from fortran (2.3572561551282417E-007) and cpp (2.3575845178322101E-007 OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-sse4) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.sse4_f_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_ggttggg_x10_cudacpp > /tmp/avalassi/output_ggttggg_x10_cudacpp' +INFO: The following Floating Point Exceptions have been reported: FE_UNDERFLOW +DEBUG: MEK processed 81920 events across 1240 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 128/128 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 2.285e-07 [2.2845949484525033E-007] fbridge_mode=1 + [UNWEIGHT] Wrote 380 events (found 1707 events) + [COUNTERS] PROGRAM TOTAL : 274.4000s + [COUNTERS] Fortran Overhead ( 0 ) : 4.3092s + [COUNTERS] CudaCpp MEs ( 2 ) : 270.0455s for 81920 events => throughput is 3.03E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0453s + +*** (2-sse4) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (2.2842713115633741E-007) and cpp (2.2845949484525033E-007) differ by less than 4E-4 (0.00014168058211416756) + +*** (2-sse4) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SM_GG_TTXGGG_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+BRDHST/sse4+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.386203e+02 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.512103e+02 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_SM_GG_TTXGGG_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.374145e+02 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.405421e+02 ) sec^-1 *** (2-avx2) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -202,10 +299,10 @@ DEBUG: MEK processed 8192 events across 1240 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 2.358e-07 [2.3575845169411084E-007] fbridge_mode=1 [UNWEIGHT] Wrote 18 events (found 285 events) - [COUNTERS] PROGRAM TOTAL : 14.8120s - [COUNTERS] Fortran Overhead ( 0 ) : 0.5245s - [COUNTERS] CudaCpp MEs ( 2 ) : 14.2638s for 8192 events => throughput is 5.74E+02 events/s - [COUNTERS] CudaCpp HEL ( 3 ) : 0.0236s + [COUNTERS] PROGRAM TOTAL : 14.6517s + [COUNTERS] Fortran Overhead ( 0 ) : 0.5123s + [COUNTERS] CudaCpp MEs ( 2 ) : 14.1165s for 8192 events => throughput is 5.80E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0229s *** (2-avx2) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -215,15 +312,51 @@ OK! xsec from fortran (2.3572561551282417E-007) and cpp (2.3575845169411084E-007 OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-avx2) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.avx2_f_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_ggttggg_x10_cudacpp > /tmp/avalassi/output_ggttggg_x10_cudacpp' +INFO: The following Floating Point Exceptions have been reported: FE_UNDERFLOW +DEBUG: MEK processed 81920 events across 1240 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 128/128 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 2.285e-07 [2.2845940747287339E-007] fbridge_mode=1 + [UNWEIGHT] Wrote 380 events (found 1707 events) + [COUNTERS] PROGRAM TOTAL : 145.6803s + [COUNTERS] Fortran Overhead ( 0 ) : 4.3408s + [COUNTERS] CudaCpp MEs ( 2 ) : 141.3166s for 81920 events => throughput is 5.80E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0229s + +*** (2-avx2) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (2.2842713115633741E-007) and cpp (2.2845940747287339E-007) differ by less than 4E-4 (0.0001412980864952118) + +*** (2-avx2) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SM_GG_TTXGGG_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+BRDHST/avx2+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 6.872770e+02 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 7.131299e+02 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_SM_GG_TTXGGG_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 6.864576e+02 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 7.010979e+02 ) sec^-1 *** (2-512y) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -248,10 +381,10 @@ DEBUG: MEK processed 8192 events across 1240 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 2.358e-07 [2.3575845169411084E-007] fbridge_mode=1 [UNWEIGHT] Wrote 18 events (found 285 events) - [COUNTERS] PROGRAM TOTAL : 13.3091s - [COUNTERS] Fortran Overhead ( 0 ) : 0.5262s - [COUNTERS] CudaCpp MEs ( 2 ) : 12.7618s for 8192 events => throughput is 6.42E+02 events/s - [COUNTERS] CudaCpp HEL ( 3 ) : 0.0211s + [COUNTERS] PROGRAM TOTAL : 12.9851s + [COUNTERS] Fortran Overhead ( 0 ) : 0.5121s + [COUNTERS] CudaCpp MEs ( 2 ) : 12.4529s for 8192 events => throughput is 6.58E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0201s *** (2-512y) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -261,15 +394,51 @@ OK! xsec from fortran (2.3572561551282417E-007) and cpp (2.3575845169411084E-007 OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-512y) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.512y_f_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_ggttggg_x10_cudacpp > /tmp/avalassi/output_ggttggg_x10_cudacpp' +INFO: The following Floating Point Exceptions have been reported: FE_UNDERFLOW +DEBUG: MEK processed 81920 events across 1240 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 128/128 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 2.285e-07 [2.2845940747287339E-007] fbridge_mode=1 + [UNWEIGHT] Wrote 380 events (found 1707 events) + [COUNTERS] PROGRAM TOTAL : 129.6728s + [COUNTERS] Fortran Overhead ( 0 ) : 4.3138s + [COUNTERS] CudaCpp MEs ( 2 ) : 125.3389s for 81920 events => throughput is 6.54E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0201s + +*** (2-512y) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (2.2842713115633741E-007) and cpp (2.2845940747287339E-007) differ by less than 4E-4 (0.0001412980864952118) + +*** (2-512y) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SM_GG_TTXGGG_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+BRDHST/512y+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 7.728743e+02 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 8.002756e+02 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_SM_GG_TTXGGG_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 7.768099e+02 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 7.984157e+02 ) sec^-1 *** (2-512z) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -294,10 +463,10 @@ DEBUG: MEK processed 8192 events across 1240 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 2.358e-07 [2.3575850859831750E-007] fbridge_mode=1 [UNWEIGHT] Wrote 18 events (found 285 events) - [COUNTERS] PROGRAM TOTAL : 13.2286s - [COUNTERS] Fortran Overhead ( 0 ) : 0.5280s - [COUNTERS] CudaCpp MEs ( 2 ) : 12.6780s for 8192 events => throughput is 6.46E+02 events/s - [COUNTERS] CudaCpp HEL ( 3 ) : 0.0225s + [COUNTERS] PROGRAM TOTAL : 12.7534s + [COUNTERS] Fortran Overhead ( 0 ) : 0.5092s + [COUNTERS] CudaCpp MEs ( 2 ) : 12.2224s for 8192 events => throughput is 6.70E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0219s *** (2-512z) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -307,15 +476,51 @@ OK! xsec from fortran (2.3572561551282417E-007) and cpp (2.3575850859831750E-007 OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-512z) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.512z_f_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_ggttggg_x10_cudacpp > /tmp/avalassi/output_ggttggg_x10_cudacpp' +INFO: The following Floating Point Exceptions have been reported: FE_UNDERFLOW +DEBUG: MEK processed 81920 events across 1240 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 128/128 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 2.285e-07 [2.2845946568145136E-007] fbridge_mode=1 + [UNWEIGHT] Wrote 380 events (found 1707 events) + [COUNTERS] PROGRAM TOTAL : 126.5588s + [COUNTERS] Fortran Overhead ( 0 ) : 4.3206s + [COUNTERS] CudaCpp MEs ( 2 ) : 122.2138s for 81920 events => throughput is 6.70E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0245s + +*** (2-512z) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (2.2842713115633741E-007) and cpp (2.2845946568145136E-007) differ by less than 4E-4 (0.00014155290989403824) + +*** (2-512z) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SM_GG_TTXGGG_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+BRDHST/512z+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 6.948019e+02 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 7.320629e+02 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_SM_GG_TTXGGG_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 6.969717e+02 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 7.292846e+02 ) sec^-1 *** (3-cuda) EXECUTE MADEVENT_CUDA x1 (create events.lhe) *** -------------------- @@ -339,10 +544,10 @@ DEBUG: MEK processed 8192 events across 1240 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 2.358e-07 [2.3575862304433055E-007] fbridge_mode=1 [UNWEIGHT] Wrote 18 events (found 285 events) - [COUNTERS] PROGRAM TOTAL : 2.2079s - [COUNTERS] Fortran Overhead ( 0 ) : 1.1084s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.5470s for 8192 events => throughput is 1.50E+04 events/s - [COUNTERS] CudaCpp HEL ( 3 ) : 0.5524s + [COUNTERS] PROGRAM TOTAL : 2.0844s + [COUNTERS] Fortran Overhead ( 0 ) : 0.9971s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5396s for 8192 events => throughput is 1.52E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.5477s *** (3-cuda) Compare MADEVENT_CUDA x1 xsec to MADEVENT_FORTRAN xsec *** @@ -352,45 +557,80 @@ OK! xsec from fortran (2.3572561551282417E-007) and cuda (2.3575862304433055E-00 OK! events.lhe.cuda.1 and events.lhe.ref.1 are identical +*** (3-cuda) EXECUTE MADEVENT_CUDA x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.cuda_f_inl0_hrd0/madevent_cuda < /tmp/avalassi/input_ggttggg_x10_cudacpp > /tmp/avalassi/output_ggttggg_x10_cudacpp' +DEBUG: MEK processed 81920 events across 1240 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 128/128 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 2.285e-07 [2.2845959888250639E-007] fbridge_mode=1 + [UNWEIGHT] Wrote 380 events (found 1707 events) + [COUNTERS] PROGRAM TOTAL : 10.6076s + [COUNTERS] Fortran Overhead ( 0 ) : 4.7881s + [COUNTERS] CudaCpp MEs ( 2 ) : 5.2672s for 81920 events => throughput is 1.56E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.5523s + +*** (3-cuda) Compare MADEVENT_CUDA x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (2.2842713115633741E-007) and cuda (2.2845959888250639E-007) differ by less than 4E-4 (0.0001421360326359089) + +*** (3-cuda) Compare MADEVENT_CUDA x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cuda.10 and events.lhe.ref.10 are identical + *** EXECUTE GCHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SM_GG_TTXGGG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.517499e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.528336e+04 ) sec^-1 *** EXECUTE GCHECK(8192) -p 256 32 1 *** Process = SIGMA_SM_GG_TTXGGG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.545233e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.528484e+04 ) sec^-1 *** EXECUTE GCHECK(MAX) -p 512 32 1 --bridge *** Process = SIGMA_SM_GG_TTXGGG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 2.140576e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.106514e+04 ) sec^-1 *** EXECUTE GCHECK(MAX) -p 512 32 1 *** Process = SIGMA_SM_GG_TTXGGG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 2.181453e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.210496e+04 ) sec^-1 *** EXECUTE GCHECK(MAX128THR) -p 128 128 1 --bridge *** Process = SIGMA_SM_GG_TTXGGG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 2.126165e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.100878e+04 ) sec^-1 *** EXECUTE GCHECK(MAX128THR) -p 128 128 1 *** Process = SIGMA_SM_GG_TTXGGG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 2.164632e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.173287e+04 ) sec^-1 *** EXECUTE GCHECK(MAX8THR) -p 2048 8 1 --bridge *** Process = SIGMA_SM_GG_TTXGGG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 2.163932e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.153456e+04 ) sec^-1 *** EXECUTE GCHECK(MAX8THR) -p 2048 8 1 *** Process = SIGMA_SM_GG_TTXGGG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 6.073078e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 5.943566e+03 ) sec^-1 *** (3-hip) WARNING! SKIP MADEVENT_HIP (hip is not supported on this node) *** diff --git a/epochX/cudacpp/tmad/logs_ggttggg_mad/log_ggttggg_mad_m_inl0_hrd0.txt b/epochX/cudacpp/tmad/logs_ggttggg_mad/log_ggttggg_mad_m_inl0_hrd0.txt index 3a68950921..84be1ff3b5 100644 --- a/epochX/cudacpp/tmad/logs_ggttggg_mad/log_ggttggg_mad_m_inl0_hrd0.txt +++ b/epochX/cudacpp/tmad/logs_ggttggg_mad/log_ggttggg_mad_m_inl0_hrd0.txt @@ -14,8 +14,8 @@ make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttggg.mad/SubProcesses/P1_gg_ttxggg' make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttggg.mad/SubProcesses/P1_gg_ttxggg' -make USEBUILDDIR=1 BACKEND=cpp512z make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttggg.mad/SubProcesses/P1_gg_ttxggg' +make USEBUILDDIR=1 BACKEND=cpp512z make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttggg.mad/SubProcesses/P1_gg_ttxggg' make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttggg.mad/SubProcesses/P1_gg_ttxggg' @@ -32,7 +32,7 @@ make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ OMP_NUM_THREADS= -DATE: 2024-10-06_10:26:37 +DATE: 2024-10-06_15:54:34 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: Working directory (run): /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttggg.mad/SubProcesses/P1_gg_ttxggg @@ -58,9 +58,9 @@ Executing ' ./madevent_fortran < /tmp/avalassi/input_ggttggg_x1_fortran > /tmp/a [XSECTION] ChannelId = 1 [XSECTION] Cross section = 2.357e-07 [2.3572561551282417E-007] fbridge_mode=0 [UNWEIGHT] Wrote 1 events (found 285 events) - [COUNTERS] PROGRAM TOTAL : 101.1381s - [COUNTERS] Fortran Overhead ( 0 ) : 0.5302s - [COUNTERS] Fortran MEs ( 1 ) : 100.6080s for 8192 events => throughput is 8.14E+01 events/s + [COUNTERS] PROGRAM TOTAL : 98.0763s + [COUNTERS] Fortran Overhead ( 0 ) : 0.5154s + [COUNTERS] Fortran MEs ( 1 ) : 97.5609s for 8192 events => throughput is 8.40E+01 events/s *** (1) EXECUTE MADEVENT_FORTRAN x1 (create events.lhe) *** -------------------- @@ -83,9 +83,34 @@ Executing ' ./madevent_fortran < /tmp/avalassi/input_ggttggg_x1_fortran > /tmp/a [XSECTION] ChannelId = 1 [XSECTION] Cross section = 2.357e-07 [2.3572561551282417E-007] fbridge_mode=0 [UNWEIGHT] Wrote 18 events (found 285 events) - [COUNTERS] PROGRAM TOTAL : 100.8808s - [COUNTERS] Fortran Overhead ( 0 ) : 0.5357s - [COUNTERS] Fortran MEs ( 1 ) : 100.3451s for 8192 events => throughput is 8.16E+01 events/s + [COUNTERS] PROGRAM TOTAL : 97.9993s + [COUNTERS] Fortran Overhead ( 0 ) : 0.5140s + [COUNTERS] Fortran MEs ( 1 ) : 97.4853s for 8192 events => throughput is 8.40E+01 events/s + +*** (1) EXECUTE MADEVENT_FORTRAN x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./madevent_fortran < /tmp/avalassi/input_ggttggg_x10_fortran > /tmp/avalassi/output_ggttggg_x10_fortran' + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 128/128 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 2.284e-07 [2.2842713115633741E-007] fbridge_mode=0 + [UNWEIGHT] Wrote 380 events (found 1707 events) + [COUNTERS] PROGRAM TOTAL : 975.8244s + [COUNTERS] Fortran Overhead ( 0 ) : 4.4045s + [COUNTERS] Fortran MEs ( 1 ) : 971.4199s for 81920 events => throughput is 8.43E+01 events/s *** (2-none) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -109,10 +134,10 @@ DEBUG: MEK processed 8192 events across 1240 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 2.357e-07 [2.3572561678995975E-007] fbridge_mode=1 [UNWEIGHT] Wrote 18 events (found 285 events) - [COUNTERS] PROGRAM TOTAL : 123.7239s - [COUNTERS] Fortran Overhead ( 0 ) : 0.5356s - [COUNTERS] CudaCpp MEs ( 2 ) : 122.9787s for 8192 events => throughput is 6.66E+01 events/s - [COUNTERS] CudaCpp HEL ( 3 ) : 0.2095s + [COUNTERS] PROGRAM TOTAL : 120.4240s + [COUNTERS] Fortran Overhead ( 0 ) : 0.5095s + [COUNTERS] CudaCpp MEs ( 2 ) : 119.7109s for 8192 events => throughput is 6.84E+01 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.2036s *** (2-none) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -122,15 +147,50 @@ OK! xsec from fortran (2.3572561551282417E-007) and cpp (2.3572561678995975E-007 OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-none) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.none_m_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_ggttggg_x10_cudacpp > /tmp/avalassi/output_ggttggg_x10_cudacpp' +DEBUG: MEK processed 81920 events across 1240 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 128/128 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 2.284e-07 [2.2842713238614534E-007] fbridge_mode=1 + [UNWEIGHT] Wrote 380 events (found 1707 events) + [COUNTERS] PROGRAM TOTAL : 1179.4291s + [COUNTERS] Fortran Overhead ( 0 ) : 4.3299s + [COUNTERS] CudaCpp MEs ( 2 ) : 1174.8960s for 81920 events => throughput is 6.97E+01 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.2032s + +*** (2-none) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (2.2842713115633741E-007) and cpp (2.2842713238614534E-007) differ by less than 2E-4 (5.38380851011766e-09) + +*** (2-none) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SM_GG_TTXGGG_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+BRDHST/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 7.634632e+01 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 7.858740e+01 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_SM_GG_TTXGGG_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 7.608909e+01 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 7.851010e+01 ) sec^-1 *** (2-sse4) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -154,10 +214,10 @@ DEBUG: MEK processed 8192 events across 1240 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 2.357e-07 [2.3572561701257335E-007] fbridge_mode=1 [UNWEIGHT] Wrote 18 events (found 285 events) - [COUNTERS] PROGRAM TOTAL : 64.5975s - [COUNTERS] Fortran Overhead ( 0 ) : 0.5274s - [COUNTERS] CudaCpp MEs ( 2 ) : 63.9661s for 8192 events => throughput is 1.28E+02 events/s - [COUNTERS] CudaCpp HEL ( 3 ) : 0.1041s + [COUNTERS] PROGRAM TOTAL : 61.8165s + [COUNTERS] Fortran Overhead ( 0 ) : 0.5107s + [COUNTERS] CudaCpp MEs ( 2 ) : 61.2072s for 8192 events => throughput is 1.34E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0986s *** (2-sse4) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -167,15 +227,50 @@ OK! xsec from fortran (2.3572561551282417E-007) and cpp (2.3572561701257335E-007 OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-sse4) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.sse4_m_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_ggttggg_x10_cudacpp > /tmp/avalassi/output_ggttggg_x10_cudacpp' +DEBUG: MEK processed 81920 events across 1240 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 128/128 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 2.284e-07 [2.2842713242471448E-007] fbridge_mode=1 + [UNWEIGHT] Wrote 380 events (found 1707 events) + [COUNTERS] PROGRAM TOTAL : 621.0159s + [COUNTERS] Fortran Overhead ( 0 ) : 4.3165s + [COUNTERS] CudaCpp MEs ( 2 ) : 616.5993s for 81920 events => throughput is 1.33E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.1000s + +*** (2-sse4) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (2.2842713115633741E-007) and cpp (2.2842713242471448E-007) differ by less than 2E-4 (5.552655002460938e-09) + +*** (2-sse4) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SM_GG_TTXGGG_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+BRDHST/sse4+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.549992e+02 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.603742e+02 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_SM_GG_TTXGGG_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.544779e+02 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.602424e+02 ) sec^-1 *** (2-avx2) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -199,10 +294,10 @@ DEBUG: MEK processed 8192 events across 1240 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 2.357e-07 [2.3572561705911026E-007] fbridge_mode=1 [UNWEIGHT] Wrote 18 events (found 285 events) - [COUNTERS] PROGRAM TOTAL : 28.6856s - [COUNTERS] Fortran Overhead ( 0 ) : 0.5254s - [COUNTERS] CudaCpp MEs ( 2 ) : 28.1150s for 8192 events => throughput is 2.91E+02 events/s - [COUNTERS] CudaCpp HEL ( 3 ) : 0.0453s + [COUNTERS] PROGRAM TOTAL : 27.5989s + [COUNTERS] Fortran Overhead ( 0 ) : 0.5097s + [COUNTERS] CudaCpp MEs ( 2 ) : 27.0447s for 8192 events => throughput is 3.03E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0444s *** (2-avx2) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -212,15 +307,50 @@ OK! xsec from fortran (2.3572561551282417E-007) and cpp (2.3572561705911026E-007 OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-avx2) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.avx2_m_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_ggttggg_x10_cudacpp > /tmp/avalassi/output_ggttggg_x10_cudacpp' +DEBUG: MEK processed 81920 events across 1240 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 128/128 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 2.284e-07 [2.2842713241239113E-007] fbridge_mode=1 + [UNWEIGHT] Wrote 380 events (found 1707 events) + [COUNTERS] PROGRAM TOTAL : 270.6414s + [COUNTERS] Fortran Overhead ( 0 ) : 4.3226s + [COUNTERS] CudaCpp MEs ( 2 ) : 266.2754s for 81920 events => throughput is 3.08E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0433s + +*** (2-avx2) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (2.2842713115633741E-007) and cpp (2.2842713241239113E-007) differ by less than 2E-4 (5.498706379114537e-09) + +*** (2-avx2) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SM_GG_TTXGGG_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+BRDHST/avx2+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.581303e+02 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.706421e+02 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_SM_GG_TTXGGG_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.574698e+02 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.700217e+02 ) sec^-1 *** (2-512y) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -244,10 +374,10 @@ DEBUG: MEK processed 8192 events across 1240 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 2.357e-07 [2.3572561705911026E-007] fbridge_mode=1 [UNWEIGHT] Wrote 18 events (found 285 events) - [COUNTERS] PROGRAM TOTAL : 24.6205s - [COUNTERS] Fortran Overhead ( 0 ) : 0.5315s - [COUNTERS] CudaCpp MEs ( 2 ) : 24.0503s for 8192 events => throughput is 3.41E+02 events/s - [COUNTERS] CudaCpp HEL ( 3 ) : 0.0387s + [COUNTERS] PROGRAM TOTAL : 23.7834s + [COUNTERS] Fortran Overhead ( 0 ) : 0.5143s + [COUNTERS] CudaCpp MEs ( 2 ) : 23.2307s for 8192 events => throughput is 3.53E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0384s *** (2-512y) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -257,15 +387,50 @@ OK! xsec from fortran (2.3572561551282417E-007) and cpp (2.3572561705911026E-007 OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-512y) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.512y_m_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_ggttggg_x10_cudacpp > /tmp/avalassi/output_ggttggg_x10_cudacpp' +DEBUG: MEK processed 81920 events across 1240 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 128/128 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 2.284e-07 [2.2842713241239113E-007] fbridge_mode=1 + [UNWEIGHT] Wrote 380 events (found 1707 events) + [COUNTERS] PROGRAM TOTAL : 236.8204s + [COUNTERS] Fortran Overhead ( 0 ) : 4.3009s + [COUNTERS] CudaCpp MEs ( 2 ) : 232.4820s for 81920 events => throughput is 3.52E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0375s + +*** (2-512y) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (2.2842713115633741E-007) and cpp (2.2842713241239113E-007) differ by less than 2E-4 (5.498706379114537e-09) + +*** (2-512y) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SM_GG_TTXGGG_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+BRDHST/512y+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 4.161373e+02 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 4.319263e+02 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_SM_GG_TTXGGG_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 4.184852e+02 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 4.320326e+02 ) sec^-1 *** (2-512z) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -289,10 +454,10 @@ DEBUG: MEK processed 8192 events across 1240 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 2.357e-07 [2.3572561705911026E-007] fbridge_mode=1 [UNWEIGHT] Wrote 18 events (found 285 events) - [COUNTERS] PROGRAM TOTAL : 25.7441s - [COUNTERS] Fortran Overhead ( 0 ) : 0.5280s - [COUNTERS] CudaCpp MEs ( 2 ) : 25.1699s for 8192 events => throughput is 3.25E+02 events/s - [COUNTERS] CudaCpp HEL ( 3 ) : 0.0462s + [COUNTERS] PROGRAM TOTAL : 24.2369s + [COUNTERS] Fortran Overhead ( 0 ) : 0.5094s + [COUNTERS] CudaCpp MEs ( 2 ) : 23.6833s for 8192 events => throughput is 3.46E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0442s *** (2-512z) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -302,15 +467,50 @@ OK! xsec from fortran (2.3572561551282417E-007) and cpp (2.3572561705911026E-007 OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-512z) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.512z_m_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_ggttggg_x10_cudacpp > /tmp/avalassi/output_ggttggg_x10_cudacpp' +DEBUG: MEK processed 81920 events across 1240 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 128/128 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 2.284e-07 [2.2842713241239113E-007] fbridge_mode=1 + [UNWEIGHT] Wrote 380 events (found 1707 events) + [COUNTERS] PROGRAM TOTAL : 244.3932s + [COUNTERS] Fortran Overhead ( 0 ) : 4.3419s + [COUNTERS] CudaCpp MEs ( 2 ) : 240.0087s for 81920 events => throughput is 3.41E+02 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0427s + +*** (2-512z) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (2.2842713115633741E-007) and cpp (2.2842713241239113E-007) differ by less than 2E-4 (5.498706379114537e-09) + +*** (2-512z) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SM_GG_TTXGGG_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+BRDHST/512z+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.516660e+02 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.708667e+02 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_SM_GG_TTXGGG_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.515216e+02 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.707701e+02 ) sec^-1 *** (3-cuda) EXECUTE MADEVENT_CUDA x1 (create events.lhe) *** -------------------- @@ -334,10 +534,10 @@ DEBUG: MEK processed 8192 events across 1240 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 2.357e-07 [2.3572561518129465E-007] fbridge_mode=1 [UNWEIGHT] Wrote 18 events (found 285 events) - [COUNTERS] PROGRAM TOTAL : 2.8461s - [COUNTERS] Fortran Overhead ( 0 ) : 1.0822s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.8795s for 8192 events => throughput is 9.31E+03 events/s - [COUNTERS] CudaCpp HEL ( 3 ) : 0.8844s + [COUNTERS] PROGRAM TOTAL : 2.7460s + [COUNTERS] Fortran Overhead ( 0 ) : 1.0013s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.8778s for 8192 events => throughput is 9.33E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8669s *** (3-cuda) Compare MADEVENT_CUDA x1 xsec to MADEVENT_FORTRAN xsec *** @@ -347,45 +547,80 @@ OK! xsec from fortran (2.3572561551282417E-007) and cuda (2.3572561518129465E-00 OK! events.lhe.cuda.1 and events.lhe.ref.1 are identical +*** (3-cuda) EXECUTE MADEVENT_CUDA x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.cuda_m_inl0_hrd0/madevent_cuda < /tmp/avalassi/input_ggttggg_x10_cudacpp > /tmp/avalassi/output_ggttggg_x10_cudacpp' +DEBUG: MEK processed 81920 events across 1240 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 128/128 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 2.284e-07 [2.2842713109538129E-007] fbridge_mode=1 + [UNWEIGHT] Wrote 380 events (found 1707 events) + [COUNTERS] PROGRAM TOTAL : 14.3456s + [COUNTERS] Fortran Overhead ( 0 ) : 4.8414s + [COUNTERS] CudaCpp MEs ( 2 ) : 8.6294s for 81920 events => throughput is 9.49E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.8748s + +*** (3-cuda) Compare MADEVENT_CUDA x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (2.2842713115633741E-007) and cuda (2.2842713109538129E-007) differ by less than 2E-4 (2.668514298420632e-10) + +*** (3-cuda) Compare MADEVENT_CUDA x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cuda.10 and events.lhe.ref.10 are identical + *** EXECUTE GCHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SM_GG_TTXGGG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:MIX+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 9.415473e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 9.428940e+03 ) sec^-1 *** EXECUTE GCHECK(8192) -p 256 32 1 *** Process = SIGMA_SM_GG_TTXGGG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:MIX+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.080771e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.076364e+04 ) sec^-1 *** EXECUTE GCHECK(MAX) -p 512 32 1 --bridge *** Process = SIGMA_SM_GG_TTXGGG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:MIX+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.106752e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.105217e+04 ) sec^-1 *** EXECUTE GCHECK(MAX) -p 512 32 1 *** Process = SIGMA_SM_GG_TTXGGG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:MIX+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.156598e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.154050e+04 ) sec^-1 *** EXECUTE GCHECK(MAX128THR) -p 128 128 1 --bridge *** Process = SIGMA_SM_GG_TTXGGG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:MIX+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.106849e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.104561e+04 ) sec^-1 *** EXECUTE GCHECK(MAX128THR) -p 128 128 1 *** Process = SIGMA_SM_GG_TTXGGG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:MIX+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.103409e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.112910e+04 ) sec^-1 *** EXECUTE GCHECK(MAX8THR) -p 2048 8 1 --bridge *** Process = SIGMA_SM_GG_TTXGGG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:MIX+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.111142e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.105446e+04 ) sec^-1 *** EXECUTE GCHECK(MAX8THR) -p 2048 8 1 *** Process = SIGMA_SM_GG_TTXGGG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:MIX+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.667428e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.665840e+03 ) sec^-1 *** (3-hip) WARNING! SKIP MADEVENT_HIP (hip is not supported on this node) *** diff --git a/epochX/cudacpp/tmad/logs_gqttq_mad/log_gqttq_mad_d_inl0_hrd0.txt b/epochX/cudacpp/tmad/logs_gqttq_mad/log_gqttq_mad_d_inl0_hrd0.txt index 7310cfc72a..ddb8b20de1 100644 --- a/epochX/cudacpp/tmad/logs_gqttq_mad/log_gqttq_mad_d_inl0_hrd0.txt +++ b/epochX/cudacpp/tmad/logs_gqttq_mad/log_gqttq_mad_d_inl0_hrd0.txt @@ -3,20 +3,20 @@ Working directory (build): /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/g make USEBUILDDIR=1 BACKEND=cuda -make USEBUILDDIR=1 BACKEND=cppnone - +make USEBUILDDIR=1 BACKEND=cppnone make USEBUILDDIR=1 BACKEND=cppsse4 + make USEBUILDDIR=1 BACKEND=cppavx2 make USEBUILDDIR=1 BACKEND=cpp512y make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gq_ttq.mad/SubProcesses/P1_gu_ttxu' make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gq_ttq.mad/SubProcesses/P1_gu_ttxu' make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gq_ttq.mad/SubProcesses/P1_gu_ttxu' make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gq_ttq.mad/SubProcesses/P1_gu_ttxu' -make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gq_ttq.mad/SubProcesses/P1_gu_ttxu' make USEBUILDDIR=1 BACKEND=cpp512z make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gq_ttq.mad/SubProcesses/P1_gu_ttxu' +make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gq_ttq.mad/SubProcesses/P1_gu_ttxu' make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gq_ttq.mad/SubProcesses/P1_gu_ttxu' make[1]: Nothing to be done for 'all'. @@ -32,7 +32,7 @@ make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ OMP_NUM_THREADS= -DATE: 2024-10-06_10:09:42 +DATE: 2024-10-06_14:35:20 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: Working directory (run): /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gq_ttq.mad/SubProcesses/P1_gu_ttxu @@ -58,9 +58,9 @@ Executing ' ./madevent_fortran < /tmp/avalassi/input_gqttq_x1_fortran > /tmp/ava [XSECTION] ChannelId = 1 [XSECTION] Cross section = 0.2031 [0.20313504505737126] fbridge_mode=0 [UNWEIGHT] Wrote 506 events (found 1943 events) - [COUNTERS] PROGRAM TOTAL : 0.5319s - [COUNTERS] Fortran Overhead ( 0 ) : 0.4597s - [COUNTERS] Fortran MEs ( 1 ) : 0.0722s for 8192 events => throughput is 1.13E+05 events/s + [COUNTERS] PROGRAM TOTAL : 0.5231s + [COUNTERS] Fortran Overhead ( 0 ) : 0.4515s + [COUNTERS] Fortran MEs ( 1 ) : 0.0716s for 8192 events => throughput is 1.14E+05 events/s *** (1) EXECUTE MADEVENT_FORTRAN x1 (create events.lhe) *** -------------------- @@ -83,9 +83,34 @@ Executing ' ./madevent_fortran < /tmp/avalassi/input_gqttq_x1_fortran > /tmp/ava [XSECTION] ChannelId = 1 [XSECTION] Cross section = 0.2031 [0.20313504505737126] fbridge_mode=0 [UNWEIGHT] Wrote 499 events (found 1502 events) - [COUNTERS] PROGRAM TOTAL : 0.4765s - [COUNTERS] Fortran Overhead ( 0 ) : 0.4047s - [COUNTERS] Fortran MEs ( 1 ) : 0.0718s for 8192 events => throughput is 1.14E+05 events/s + [COUNTERS] PROGRAM TOTAL : 0.4780s + [COUNTERS] Fortran Overhead ( 0 ) : 0.4059s + [COUNTERS] Fortran MEs ( 1 ) : 0.0721s for 8192 events => throughput is 1.14E+05 events/s + +*** (1) EXECUTE MADEVENT_FORTRAN x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./madevent_fortran < /tmp/avalassi/input_gqttq_x10_fortran > /tmp/avalassi/output_gqttq_x10_fortran' + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 16/32 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 0.211 [0.21095842877427595] fbridge_mode=0 + [UNWEIGHT] Wrote 2259 events (found 2264 events) + [COUNTERS] PROGRAM TOTAL : 2.5826s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8698s + [COUNTERS] Fortran MEs ( 1 ) : 0.7128s for 81920 events => throughput is 1.15E+05 events/s *** (2-none) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -109,9 +134,9 @@ DEBUG: MEK processed 8192 events across 5 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 0.2031 [0.20313504505737132] fbridge_mode=1 [UNWEIGHT] Wrote 499 events (found 1502 events) - [COUNTERS] PROGRAM TOTAL : 0.4865s - [COUNTERS] Fortran Overhead ( 0 ) : 0.4077s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0781s for 8192 events => throughput is 1.05E+05 events/s + [COUNTERS] PROGRAM TOTAL : 0.4912s + [COUNTERS] Fortran Overhead ( 0 ) : 0.4131s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0774s for 8192 events => throughput is 1.06E+05 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0007s *** (2-none) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -122,15 +147,50 @@ OK! xsec from fortran (0.20313504505737126) and cpp (0.20313504505737132) differ OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-none) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.none_d_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_gqttq_x10_cudacpp > /tmp/avalassi/output_gqttq_x10_cudacpp' +DEBUG: MEK processed 81920 events across 5 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 16/32 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 0.211 [0.21095842877427598] fbridge_mode=1 + [UNWEIGHT] Wrote 2259 events (found 2264 events) + [COUNTERS] PROGRAM TOTAL : 2.6809s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9011s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.7791s for 81920 events => throughput is 1.05E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0007s + +*** (2-none) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (0.21095842877427595) and cpp (0.21095842877427598) differ by less than 3E-14 (2.220446049250313e-16) + +*** (2-none) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SM_GU_TTXU_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+BRDHST/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.073164e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.070062e+05 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_SM_GU_TTXU_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.079140e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.074883e+05 ) sec^-1 *** (2-sse4) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -154,9 +214,9 @@ DEBUG: MEK processed 8192 events across 5 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 0.2031 [0.20313504505737170] fbridge_mode=1 [UNWEIGHT] Wrote 499 events (found 1502 events) - [COUNTERS] PROGRAM TOTAL : 0.4492s - [COUNTERS] Fortran Overhead ( 0 ) : 0.4062s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0425s for 8192 events => throughput is 1.93E+05 events/s + [COUNTERS] PROGRAM TOTAL : 0.4541s + [COUNTERS] Fortran Overhead ( 0 ) : 0.4107s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0430s for 8192 events => throughput is 1.91E+05 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0005s *** (2-sse4) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -167,15 +227,50 @@ OK! xsec from fortran (0.20313504505737126) and cpp (0.20313504505737170) differ OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-sse4) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.sse4_d_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_gqttq_x10_cudacpp > /tmp/avalassi/output_gqttq_x10_cudacpp' +DEBUG: MEK processed 81920 events across 5 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 16/32 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 0.211 [0.21095842877427590] fbridge_mode=1 + [UNWEIGHT] Wrote 2259 events (found 2264 events) + [COUNTERS] PROGRAM TOTAL : 2.3268s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8960s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4303s for 81920 events => throughput is 1.90E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0005s + +*** (2-sse4) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (0.21095842877427595) and cpp (0.21095842877427590) differ by less than 3E-14 (2.220446049250313e-16) + +*** (2-sse4) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SM_GU_TTXU_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+BRDHST/sse4+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.895347e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.919864e+05 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_SM_GU_TTXU_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.917908e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.911269e+05 ) sec^-1 *** (2-avx2) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -199,9 +294,9 @@ DEBUG: MEK processed 8192 events across 5 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 0.2031 [0.20313504505737162] fbridge_mode=1 [UNWEIGHT] Wrote 499 events (found 1502 events) - [COUNTERS] PROGRAM TOTAL : 0.4396s - [COUNTERS] Fortran Overhead ( 0 ) : 0.4134s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0257s for 8192 events => throughput is 3.19E+05 events/s + [COUNTERS] PROGRAM TOTAL : 0.4334s + [COUNTERS] Fortran Overhead ( 0 ) : 0.4078s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0251s for 8192 events => throughput is 3.27E+05 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0005s *** (2-avx2) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -212,15 +307,50 @@ OK! xsec from fortran (0.20313504505737126) and cpp (0.20313504505737162) differ OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-avx2) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.avx2_d_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_gqttq_x10_cudacpp > /tmp/avalassi/output_gqttq_x10_cudacpp' +DEBUG: MEK processed 81920 events across 5 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 16/32 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 0.211 [0.21095842877427592] fbridge_mode=1 + [UNWEIGHT] Wrote 2259 events (found 2264 events) + [COUNTERS] PROGRAM TOTAL : 2.1355s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8866s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2485s for 81920 events => throughput is 3.30E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0005s + +*** (2-avx2) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (0.21095842877427595) and cpp (0.21095842877427592) differ by less than 3E-14 (1.1102230246251565e-16) + +*** (2-avx2) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SM_GU_TTXU_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+BRDHST/avx2+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.340027e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.299158e+05 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_SM_GU_TTXU_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.307491e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.317113e+05 ) sec^-1 *** (2-512y) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -244,9 +374,9 @@ DEBUG: MEK processed 8192 events across 5 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 0.2031 [0.20313504505737162] fbridge_mode=1 [UNWEIGHT] Wrote 499 events (found 1502 events) - [COUNTERS] PROGRAM TOTAL : 0.4310s - [COUNTERS] Fortran Overhead ( 0 ) : 0.4082s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0224s for 8192 events => throughput is 3.66E+05 events/s + [COUNTERS] PROGRAM TOTAL : 0.4299s + [COUNTERS] Fortran Overhead ( 0 ) : 0.4068s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0226s for 8192 events => throughput is 3.62E+05 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0005s *** (2-512y) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -257,15 +387,50 @@ OK! xsec from fortran (0.20313504505737126) and cpp (0.20313504505737162) differ OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-512y) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.512y_d_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_gqttq_x10_cudacpp > /tmp/avalassi/output_gqttq_x10_cudacpp' +DEBUG: MEK processed 81920 events across 5 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 16/32 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 0.211 [0.21095842877427592] fbridge_mode=1 + [UNWEIGHT] Wrote 2259 events (found 2264 events) + [COUNTERS] PROGRAM TOTAL : 2.1180s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8958s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2218s for 81920 events => throughput is 3.69E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0004s + +*** (2-512y) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (0.21095842877427595) and cpp (0.21095842877427592) differ by less than 3E-14 (1.1102230246251565e-16) + +*** (2-512y) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SM_GU_TTXU_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+BRDHST/512y+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.693677e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.701684e+05 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_SM_GU_TTXU_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.718907e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.730304e+05 ) sec^-1 *** (2-512z) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -289,8 +454,8 @@ DEBUG: MEK processed 8192 events across 5 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 0.2031 [0.20313504505737162] fbridge_mode=1 [UNWEIGHT] Wrote 499 events (found 1502 events) - [COUNTERS] PROGRAM TOTAL : 0.4438s - [COUNTERS] Fortran Overhead ( 0 ) : 0.4093s + [COUNTERS] PROGRAM TOTAL : 0.4420s + [COUNTERS] Fortran Overhead ( 0 ) : 0.4076s [COUNTERS] CudaCpp MEs ( 2 ) : 0.0340s for 8192 events => throughput is 2.41E+05 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0005s @@ -302,15 +467,50 @@ OK! xsec from fortran (0.20313504505737126) and cpp (0.20313504505737162) differ OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-512z) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.512z_d_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_gqttq_x10_cudacpp > /tmp/avalassi/output_gqttq_x10_cudacpp' +DEBUG: MEK processed 81920 events across 5 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 16/32 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 0.211 [0.21095842877427592] fbridge_mode=1 + [UNWEIGHT] Wrote 2259 events (found 2264 events) + [COUNTERS] PROGRAM TOTAL : 2.2367s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8967s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.3396s for 81920 events => throughput is 2.41E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0005s + +*** (2-512z) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (0.21095842877427595) and cpp (0.21095842877427592) differ by less than 3E-14 (1.1102230246251565e-16) + +*** (2-512z) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SM_GU_TTXU_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+BRDHST/512z+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 2.386493e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.370528e+05 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_SM_GU_TTXU_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 2.395890e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.400828e+05 ) sec^-1 *** (3-cuda) EXECUTE MADEVENT_CUDA x1 (create events.lhe) *** -------------------- @@ -334,9 +534,9 @@ DEBUG: MEK processed 8192 events across 5 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 0.2031 [0.20313504505737173] fbridge_mode=1 [UNWEIGHT] Wrote 499 events (found 1502 events) - [COUNTERS] PROGRAM TOTAL : 0.8495s - [COUNTERS] Fortran Overhead ( 0 ) : 0.8451s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0033s for 8192 events => throughput is 2.52E+06 events/s + [COUNTERS] PROGRAM TOTAL : 0.8466s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8424s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0031s for 8192 events => throughput is 2.65E+06 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0011s *** (3-cuda) Compare MADEVENT_CUDA x1 xsec to MADEVENT_FORTRAN xsec *** @@ -347,45 +547,80 @@ OK! xsec from fortran (0.20313504505737126) and cuda (0.20313504505737173) diffe OK! events.lhe.cuda.1 and events.lhe.ref.1 are identical +*** (3-cuda) EXECUTE MADEVENT_CUDA x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.cuda_d_inl0_hrd0/madevent_cuda < /tmp/avalassi/input_gqttq_x10_cudacpp > /tmp/avalassi/output_gqttq_x10_cudacpp' +DEBUG: MEK processed 81920 events across 5 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 16/32 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 0.211 [0.21095842877427598] fbridge_mode=1 + [UNWEIGHT] Wrote 2259 events (found 2264 events) + [COUNTERS] PROGRAM TOTAL : 2.3513s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3401s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0101s for 81920 events => throughput is 8.12E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0011s + +*** (3-cuda) Compare MADEVENT_CUDA x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (0.21095842877427595) and cuda (0.21095842877427598) differ by less than 3E-14 (2.220446049250313e-16) + +*** (3-cuda) Compare MADEVENT_CUDA x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cuda.10 and events.lhe.ref.10 are identical + *** EXECUTE GCHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SM_GU_TTXU_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 2.777000e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.977651e+06 ) sec^-1 *** EXECUTE GCHECK(8192) -p 256 32 1 *** Process = SIGMA_SM_GU_TTXU_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.265214e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.384250e+06 ) sec^-1 *** EXECUTE GCHECK(MAX) -p 16384 32 1 --bridge *** Process = SIGMA_SM_GU_TTXU_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 2.327919e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.353404e+07 ) sec^-1 *** EXECUTE GCHECK(MAX) -p 16384 32 1 *** Process = SIGMA_SM_GU_TTXU_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 5.161258e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 5.179131e+07 ) sec^-1 *** EXECUTE GCHECK(MAX128THR) -p 4096 128 1 --bridge *** Process = SIGMA_SM_GU_TTXU_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 2.316740e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.342453e+07 ) sec^-1 *** EXECUTE GCHECK(MAX128THR) -p 4096 128 1 *** Process = SIGMA_SM_GU_TTXU_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 5.319766e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 5.296616e+07 ) sec^-1 *** EXECUTE GCHECK(MAX8THR) -p 65536 8 1 --bridge *** Process = SIGMA_SM_GU_TTXU_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 2.323054e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.358679e+07 ) sec^-1 *** EXECUTE GCHECK(MAX8THR) -p 65536 8 1 *** Process = SIGMA_SM_GU_TTXU_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.646948e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.657932e+07 ) sec^-1 *** (3-hip) WARNING! SKIP MADEVENT_HIP (hip is not supported on this node) *** diff --git a/epochX/cudacpp/tmad/logs_gqttq_mad/log_gqttq_mad_f_inl0_hrd0.txt b/epochX/cudacpp/tmad/logs_gqttq_mad/log_gqttq_mad_f_inl0_hrd0.txt index 748c92b28c..f90bddc289 100644 --- a/epochX/cudacpp/tmad/logs_gqttq_mad/log_gqttq_mad_f_inl0_hrd0.txt +++ b/epochX/cudacpp/tmad/logs_gqttq_mad/log_gqttq_mad_f_inl0_hrd0.txt @@ -2,26 +2,26 @@ Working directory (build): /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/g make USEBUILDDIR=1 BACKEND=cuda - make USEBUILDDIR=1 BACKEND=cppnone + + make USEBUILDDIR=1 BACKEND=cppsse4 make USEBUILDDIR=1 BACKEND=cppavx2 - make USEBUILDDIR=1 BACKEND=cpp512y make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gq_ttq.mad/SubProcesses/P1_gu_ttxu' make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gq_ttq.mad/SubProcesses/P1_gu_ttxu' make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gq_ttq.mad/SubProcesses/P1_gu_ttxu' make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gq_ttq.mad/SubProcesses/P1_gu_ttxu' -make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gq_ttq.mad/SubProcesses/P1_gu_ttxu' +make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gq_ttq.mad/SubProcesses/P1_gu_ttxu' make USEBUILDDIR=1 BACKEND=cpp512z make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gq_ttq.mad/SubProcesses/P1_gu_ttxu' make[1]: Nothing to be done for 'all'. -make[1]: Nothing to be done for 'all'. -make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gq_ttq.mad/SubProcesses/P1_gu_ttxu' make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gq_ttq.mad/SubProcesses/P1_gu_ttxu' make[1]: Nothing to be done for 'all'. +make[1]: Nothing to be done for 'all'. +make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gq_ttq.mad/SubProcesses/P1_gu_ttxu' make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gq_ttq.mad/SubProcesses/P1_gu_ttxu' make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gq_ttq.mad/SubProcesses/P1_gu_ttxu' @@ -32,7 +32,7 @@ make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ OMP_NUM_THREADS= -DATE: 2024-10-06_10:10:11 +DATE: 2024-10-06_14:36:24 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: Working directory (run): /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gq_ttq.mad/SubProcesses/P1_gu_ttxu @@ -58,8 +58,8 @@ Executing ' ./madevent_fortran < /tmp/avalassi/input_gqttq_x1_fortran > /tmp/ava [XSECTION] ChannelId = 1 [XSECTION] Cross section = 0.2031 [0.20313504505737126] fbridge_mode=0 [UNWEIGHT] Wrote 506 events (found 1943 events) - [COUNTERS] PROGRAM TOTAL : 0.5240s - [COUNTERS] Fortran Overhead ( 0 ) : 0.4523s + [COUNTERS] PROGRAM TOTAL : 0.5253s + [COUNTERS] Fortran Overhead ( 0 ) : 0.4535s [COUNTERS] Fortran MEs ( 1 ) : 0.0718s for 8192 events => throughput is 1.14E+05 events/s *** (1) EXECUTE MADEVENT_FORTRAN x1 (create events.lhe) *** @@ -83,9 +83,34 @@ Executing ' ./madevent_fortran < /tmp/avalassi/input_gqttq_x1_fortran > /tmp/ava [XSECTION] ChannelId = 1 [XSECTION] Cross section = 0.2031 [0.20313504505737126] fbridge_mode=0 [UNWEIGHT] Wrote 499 events (found 1502 events) - [COUNTERS] PROGRAM TOTAL : 0.4796s - [COUNTERS] Fortran Overhead ( 0 ) : 0.4074s - [COUNTERS] Fortran MEs ( 1 ) : 0.0721s for 8192 events => throughput is 1.14E+05 events/s + [COUNTERS] PROGRAM TOTAL : 0.4743s + [COUNTERS] Fortran Overhead ( 0 ) : 0.4024s + [COUNTERS] Fortran MEs ( 1 ) : 0.0719s for 8192 events => throughput is 1.14E+05 events/s + +*** (1) EXECUTE MADEVENT_FORTRAN x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./madevent_fortran < /tmp/avalassi/input_gqttq_x10_fortran > /tmp/avalassi/output_gqttq_x10_fortran' + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 16/32 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 0.211 [0.21095842877427595] fbridge_mode=0 + [UNWEIGHT] Wrote 2259 events (found 2264 events) + [COUNTERS] PROGRAM TOTAL : 2.6047s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8867s + [COUNTERS] Fortran MEs ( 1 ) : 0.7179s for 81920 events => throughput is 1.14E+05 events/s *** (2-none) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -109,9 +134,9 @@ DEBUG: MEK processed 8192 events across 5 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 0.2031 [0.20313506133732837] fbridge_mode=1 [UNWEIGHT] Wrote 499 events (found 1502 events) - [COUNTERS] PROGRAM TOTAL : 0.4786s - [COUNTERS] Fortran Overhead ( 0 ) : 0.4057s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0723s for 8192 events => throughput is 1.13E+05 events/s + [COUNTERS] PROGRAM TOTAL : 0.4827s + [COUNTERS] Fortran Overhead ( 0 ) : 0.4088s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0732s for 8192 events => throughput is 1.12E+05 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0006s *** (2-none) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -122,15 +147,50 @@ OK! xsec from fortran (0.20313504505737126) and cpp (0.20313506133732837) differ OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-none) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.none_f_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_gqttq_x10_cudacpp > /tmp/avalassi/output_gqttq_x10_cudacpp' +DEBUG: MEK processed 81920 events across 5 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 16/32 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 0.211 [0.21095842907143103] fbridge_mode=1 + [UNWEIGHT] Wrote 2259 events (found 2264 events) + [COUNTERS] PROGRAM TOTAL : 2.6321s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9019s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.7296s for 81920 events => throughput is 1.12E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0005s + +*** (2-none) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (0.21095842877427595) and cpp (0.21095842907143103) differ by less than 4E-4 (1.4085954624931674e-09) + +*** (2-none) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SM_GU_TTXU_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+BRDHST/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.132089e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.134226e+05 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_SM_GU_TTXU_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.123977e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.126246e+05 ) sec^-1 *** (2-sse4) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -154,10 +214,10 @@ DEBUG: MEK processed 8192 events across 5 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 0.2031 [0.20313502997679400] fbridge_mode=1 [UNWEIGHT] Wrote 499 events (found 1502 events) - [COUNTERS] PROGRAM TOTAL : 0.4346s - [COUNTERS] Fortran Overhead ( 0 ) : 0.4073s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0269s for 8192 events => throughput is 3.04E+05 events/s - [COUNTERS] CudaCpp HEL ( 3 ) : 0.0004s + [COUNTERS] PROGRAM TOTAL : 0.4341s + [COUNTERS] Fortran Overhead ( 0 ) : 0.4072s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0266s for 8192 events => throughput is 3.08E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0003s *** (2-sse4) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -167,15 +227,50 @@ OK! xsec from fortran (0.20313504505737126) and cpp (0.20313502997679400) differ OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-sse4) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.sse4_f_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_gqttq_x10_cudacpp > /tmp/avalassi/output_gqttq_x10_cudacpp' +DEBUG: MEK processed 81920 events across 5 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 16/32 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 0.211 [0.21095839656505114] fbridge_mode=1 + [UNWEIGHT] Wrote 2259 events (found 2264 events) + [COUNTERS] PROGRAM TOTAL : 2.1480s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8821s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2655s for 81920 events => throughput is 3.09E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0004s + +*** (2-sse4) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (0.21095842877427595) and cpp (0.21095839656505114) differ by less than 4E-4 (1.5268043562777223e-07) + +*** (2-sse4) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SM_GU_TTXU_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+BRDHST/sse4+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.016574e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.972813e+05 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_SM_GU_TTXU_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.049161e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.044687e+05 ) sec^-1 *** (2-avx2) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -199,9 +294,9 @@ DEBUG: MEK processed 8192 events across 5 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 0.2031 [0.20313502619857851] fbridge_mode=1 [UNWEIGHT] Wrote 499 events (found 1502 events) - [COUNTERS] PROGRAM TOTAL : 0.4231s - [COUNTERS] Fortran Overhead ( 0 ) : 0.4098s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0130s for 8192 events => throughput is 6.28E+05 events/s + [COUNTERS] PROGRAM TOTAL : 0.4199s + [COUNTERS] Fortran Overhead ( 0 ) : 0.4064s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0132s for 8192 events => throughput is 6.19E+05 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0003s *** (2-avx2) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -212,15 +307,50 @@ OK! xsec from fortran (0.20313504505737126) and cpp (0.20313502619857851) differ OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-avx2) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.avx2_f_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_gqttq_x10_cudacpp > /tmp/avalassi/output_gqttq_x10_cudacpp' +DEBUG: MEK processed 81920 events across 5 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 16/32 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 0.211 [0.21095839412856376] fbridge_mode=1 + [UNWEIGHT] Wrote 2259 events (found 2264 events) + [COUNTERS] PROGRAM TOTAL : 2.0192s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8870s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1319s for 81920 events => throughput is 6.21E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0003s + +*** (2-avx2) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (0.21095842877427595) and cpp (0.21095839412856376) differ by less than 4E-4 (1.6423004467469582e-07) + +*** (2-avx2) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SM_GU_TTXU_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+BRDHST/avx2+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 6.215183e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 6.161250e+05 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_SM_GU_TTXU_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 6.201945e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 6.234782e+05 ) sec^-1 *** (2-512y) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -244,9 +374,9 @@ DEBUG: MEK processed 8192 events across 5 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 0.2031 [0.20313502619857851] fbridge_mode=1 [UNWEIGHT] Wrote 499 events (found 1502 events) - [COUNTERS] PROGRAM TOTAL : 0.4177s - [COUNTERS] Fortran Overhead ( 0 ) : 0.4051s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0122s for 8192 events => throughput is 6.69E+05 events/s + [COUNTERS] PROGRAM TOTAL : 0.4231s + [COUNTERS] Fortran Overhead ( 0 ) : 0.4102s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0126s for 8192 events => throughput is 6.49E+05 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0003s *** (2-512y) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -257,15 +387,50 @@ OK! xsec from fortran (0.20313504505737126) and cpp (0.20313502619857851) differ OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-512y) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.512y_f_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_gqttq_x10_cudacpp > /tmp/avalassi/output_gqttq_x10_cudacpp' +DEBUG: MEK processed 81920 events across 5 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 16/32 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 0.211 [0.21095839412856376] fbridge_mode=1 + [UNWEIGHT] Wrote 2259 events (found 2264 events) + [COUNTERS] PROGRAM TOTAL : 2.0186s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8942s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1241s for 81920 events => throughput is 6.60E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0003s + +*** (2-512y) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (0.21095842877427595) and cpp (0.21095839412856376) differ by less than 4E-4 (1.6423004467469582e-07) + +*** (2-512y) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SM_GU_TTXU_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+BRDHST/512y+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 6.557168e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 6.717986e+05 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_SM_GU_TTXU_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 6.659565e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 6.727989e+05 ) sec^-1 *** (2-512z) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -289,9 +454,9 @@ DEBUG: MEK processed 8192 events across 5 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 0.2031 [0.20313505300145301] fbridge_mode=1 [UNWEIGHT] Wrote 499 events (found 1502 events) - [COUNTERS] PROGRAM TOTAL : 0.4231s - [COUNTERS] Fortran Overhead ( 0 ) : 0.4064s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0163s for 8192 events => throughput is 5.02E+05 events/s + [COUNTERS] PROGRAM TOTAL : 0.4200s + [COUNTERS] Fortran Overhead ( 0 ) : 0.4036s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0160s for 8192 events => throughput is 5.11E+05 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0004s *** (2-512z) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -302,15 +467,50 @@ OK! xsec from fortran (0.20313504505737126) and cpp (0.20313505300145301) differ OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-512z) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.512z_f_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_gqttq_x10_cudacpp > /tmp/avalassi/output_gqttq_x10_cudacpp' +DEBUG: MEK processed 81920 events across 5 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 16/32 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 0.211 [0.21095842133012335] fbridge_mode=1 + [UNWEIGHT] Wrote 2259 events (found 2264 events) + [COUNTERS] PROGRAM TOTAL : 2.0522s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8816s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1702s for 81920 events => throughput is 4.81E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0004s + +*** (2-512z) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (0.21095842877427595) and cpp (0.21095842133012335) differ by less than 4E-4 (3.528729641821826e-08) + +*** (2-512z) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SM_GU_TTXU_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+BRDHST/512z+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 4.736521e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 4.690255e+05 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_SM_GU_TTXU_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 4.799657e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 4.784909e+05 ) sec^-1 *** (3-cuda) EXECUTE MADEVENT_CUDA x1 (create events.lhe) *** -------------------- @@ -334,10 +534,10 @@ DEBUG: MEK processed 8192 events across 5 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 0.2031 [0.20313508590887899] fbridge_mode=1 [UNWEIGHT] Wrote 499 events (found 1502 events) - [COUNTERS] PROGRAM TOTAL : 0.8496s - [COUNTERS] Fortran Overhead ( 0 ) : 0.8457s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0031s for 8192 events => throughput is 2.64E+06 events/s - [COUNTERS] CudaCpp HEL ( 3 ) : 0.0009s + [COUNTERS] PROGRAM TOTAL : 0.8529s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8491s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0030s for 8192 events => throughput is 2.77E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0008s *** (3-cuda) Compare MADEVENT_CUDA x1 xsec to MADEVENT_FORTRAN xsec *** @@ -347,45 +547,80 @@ OK! xsec from fortran (0.20313504505737126) and cuda (0.20313508590887899) diffe OK! events.lhe.cuda.1 and events.lhe.ref.1 are identical +*** (3-cuda) EXECUTE MADEVENT_CUDA x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.cuda_f_inl0_hrd0/madevent_cuda < /tmp/avalassi/input_gqttq_x10_cudacpp > /tmp/avalassi/output_gqttq_x10_cudacpp' +DEBUG: MEK processed 81920 events across 5 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 16/32 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 0.211 [0.21095846337765808] fbridge_mode=1 + [UNWEIGHT] Wrote 2259 events (found 2264 events) + [COUNTERS] PROGRAM TOTAL : 2.3445s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3350s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0086s for 81920 events => throughput is 9.55E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0008s + +*** (3-cuda) Compare MADEVENT_CUDA x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (0.21095842877427595) and cuda (0.21095846337765808) differ by less than 4E-4 (1.640293887383848e-07) + +*** (3-cuda) Compare MADEVENT_CUDA x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cuda.10 and events.lhe.ref.10 are identical + *** EXECUTE GCHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SM_GU_TTXU_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.049327e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.281487e+06 ) sec^-1 *** EXECUTE GCHECK(8192) -p 256 32 1 *** Process = SIGMA_SM_GU_TTXU_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.339018e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.456047e+06 ) sec^-1 *** EXECUTE GCHECK(MAX) -p 16384 32 1 --bridge *** Process = SIGMA_SM_GU_TTXU_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 4.110522e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 4.191834e+07 ) sec^-1 *** EXECUTE GCHECK(MAX) -p 16384 32 1 *** Process = SIGMA_SM_GU_TTXU_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 9.423874e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 9.703390e+07 ) sec^-1 *** EXECUTE GCHECK(MAX128THR) -p 4096 128 1 --bridge *** Process = SIGMA_SM_GU_TTXU_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 4.090502e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 4.269264e+07 ) sec^-1 *** EXECUTE GCHECK(MAX128THR) -p 4096 128 1 *** Process = SIGMA_SM_GU_TTXU_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 9.757351e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 9.774521e+07 ) sec^-1 *** EXECUTE GCHECK(MAX8THR) -p 65536 8 1 --bridge *** Process = SIGMA_SM_GU_TTXU_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.720065e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.844507e+07 ) sec^-1 *** EXECUTE GCHECK(MAX8THR) -p 65536 8 1 *** Process = SIGMA_SM_GU_TTXU_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 4.206204e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 4.252023e+07 ) sec^-1 *** (3-hip) WARNING! SKIP MADEVENT_HIP (hip is not supported on this node) *** diff --git a/epochX/cudacpp/tmad/logs_gqttq_mad/log_gqttq_mad_m_inl0_hrd0.txt b/epochX/cudacpp/tmad/logs_gqttq_mad/log_gqttq_mad_m_inl0_hrd0.txt index dd13a39319..c79d2bd61a 100644 --- a/epochX/cudacpp/tmad/logs_gqttq_mad/log_gqttq_mad_m_inl0_hrd0.txt +++ b/epochX/cudacpp/tmad/logs_gqttq_mad/log_gqttq_mad_m_inl0_hrd0.txt @@ -2,13 +2,12 @@ Working directory (build): /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/g make USEBUILDDIR=1 BACKEND=cuda -make USEBUILDDIR=1 BACKEND=cppnone +make USEBUILDDIR=1 BACKEND=cppnone make USEBUILDDIR=1 BACKEND=cppsse4 make USEBUILDDIR=1 BACKEND=cppavx2 -make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gq_ttq.mad/SubProcesses/P1_gu_ttxu' make USEBUILDDIR=1 BACKEND=cpp512y make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gq_ttq.mad/SubProcesses/P1_gu_ttxu' make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gq_ttq.mad/SubProcesses/P1_gu_ttxu' @@ -17,6 +16,7 @@ make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp make USEBUILDDIR=1 BACKEND=cpp512z make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gq_ttq.mad/SubProcesses/P1_gu_ttxu' +make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gq_ttq.mad/SubProcesses/P1_gu_ttxu' make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gq_ttq.mad/SubProcesses/P1_gu_ttxu' make[1]: Nothing to be done for 'all'. @@ -32,7 +32,7 @@ make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ OMP_NUM_THREADS= -DATE: 2024-10-06_10:09:56 +DATE: 2024-10-06_14:35:52 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: Working directory (run): /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gq_ttq.mad/SubProcesses/P1_gu_ttxu @@ -58,8 +58,8 @@ Executing ' ./madevent_fortran < /tmp/avalassi/input_gqttq_x1_fortran > /tmp/ava [XSECTION] ChannelId = 1 [XSECTION] Cross section = 0.2031 [0.20313504505737126] fbridge_mode=0 [UNWEIGHT] Wrote 506 events (found 1943 events) - [COUNTERS] PROGRAM TOTAL : 0.5254s - [COUNTERS] Fortran Overhead ( 0 ) : 0.4537s + [COUNTERS] PROGRAM TOTAL : 0.5218s + [COUNTERS] Fortran Overhead ( 0 ) : 0.4501s [COUNTERS] Fortran MEs ( 1 ) : 0.0717s for 8192 events => throughput is 1.14E+05 events/s *** (1) EXECUTE MADEVENT_FORTRAN x1 (create events.lhe) *** @@ -83,9 +83,34 @@ Executing ' ./madevent_fortran < /tmp/avalassi/input_gqttq_x1_fortran > /tmp/ava [XSECTION] ChannelId = 1 [XSECTION] Cross section = 0.2031 [0.20313504505737126] fbridge_mode=0 [UNWEIGHT] Wrote 499 events (found 1502 events) - [COUNTERS] PROGRAM TOTAL : 0.4842s - [COUNTERS] Fortran Overhead ( 0 ) : 0.4128s - [COUNTERS] Fortran MEs ( 1 ) : 0.0714s for 8192 events => throughput is 1.15E+05 events/s + [COUNTERS] PROGRAM TOTAL : 0.4806s + [COUNTERS] Fortran Overhead ( 0 ) : 0.4091s + [COUNTERS] Fortran MEs ( 1 ) : 0.0716s for 8192 events => throughput is 1.14E+05 events/s + +*** (1) EXECUTE MADEVENT_FORTRAN x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./madevent_fortran < /tmp/avalassi/input_gqttq_x10_fortran > /tmp/avalassi/output_gqttq_x10_fortran' + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 16/32 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 0.211 [0.21095842877427595] fbridge_mode=0 + [UNWEIGHT] Wrote 2259 events (found 2264 events) + [COUNTERS] PROGRAM TOTAL : 2.6012s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8838s + [COUNTERS] Fortran MEs ( 1 ) : 0.7174s for 81920 events => throughput is 1.14E+05 events/s *** (2-none) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -109,10 +134,10 @@ DEBUG: MEK processed 8192 events across 5 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 0.2031 [0.20313504495344831] fbridge_mode=1 [UNWEIGHT] Wrote 499 events (found 1502 events) - [COUNTERS] PROGRAM TOTAL : 0.4899s - [COUNTERS] Fortran Overhead ( 0 ) : 0.4117s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0776s for 8192 events => throughput is 1.06E+05 events/s - [COUNTERS] CudaCpp HEL ( 3 ) : 0.0006s + [COUNTERS] PROGRAM TOTAL : 0.4854s + [COUNTERS] Fortran Overhead ( 0 ) : 0.4072s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0775s for 8192 events => throughput is 1.06E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0007s *** (2-none) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -122,15 +147,50 @@ OK! xsec from fortran (0.20313504505737126) and cpp (0.20313504495344831) differ OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-none) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.none_m_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_gqttq_x10_cudacpp > /tmp/avalassi/output_gqttq_x10_cudacpp' +DEBUG: MEK processed 81920 events across 5 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 16/32 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 0.211 [0.21095842877343590] fbridge_mode=1 + [UNWEIGHT] Wrote 2259 events (found 2264 events) + [COUNTERS] PROGRAM TOTAL : 2.6792s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9003s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.7782s for 81920 events => throughput is 1.05E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0007s + +*** (2-none) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (0.21095842877427595) and cpp (0.21095842877343590) differ by less than 2E-4 (3.982036922423049e-12) + +*** (2-none) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SM_GU_TTXU_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+BRDHST/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.073352e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.070156e+05 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_SM_GU_TTXU_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.073996e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.073149e+05 ) sec^-1 *** (2-sse4) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -154,9 +214,9 @@ DEBUG: MEK processed 8192 events across 5 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 0.2031 [0.20313504495344833] fbridge_mode=1 [UNWEIGHT] Wrote 499 events (found 1502 events) - [COUNTERS] PROGRAM TOTAL : 0.4513s - [COUNTERS] Fortran Overhead ( 0 ) : 0.4086s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0421s for 8192 events => throughput is 1.94E+05 events/s + [COUNTERS] PROGRAM TOTAL : 0.4514s + [COUNTERS] Fortran Overhead ( 0 ) : 0.4089s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0420s for 8192 events => throughput is 1.95E+05 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0005s *** (2-sse4) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -167,15 +227,50 @@ OK! xsec from fortran (0.20313504505737126) and cpp (0.20313504495344833) differ OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-sse4) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.sse4_m_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_gqttq_x10_cudacpp > /tmp/avalassi/output_gqttq_x10_cudacpp' +DEBUG: MEK processed 81920 events across 5 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 16/32 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 0.211 [0.21095842877343590] fbridge_mode=1 + [UNWEIGHT] Wrote 2259 events (found 2264 events) + [COUNTERS] PROGRAM TOTAL : 2.3288s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9025s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4257s for 81920 events => throughput is 1.92E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0005s + +*** (2-sse4) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (0.21095842877427595) and cpp (0.21095842877343590) differ by less than 2E-4 (3.982036922423049e-12) + +*** (2-sse4) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SM_GU_TTXU_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+BRDHST/sse4+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.886911e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.919480e+05 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_SM_GU_TTXU_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.898728e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.944879e+05 ) sec^-1 *** (2-avx2) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -199,9 +294,9 @@ DEBUG: MEK processed 8192 events across 5 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 0.2031 [0.20313504510700500] fbridge_mode=1 [UNWEIGHT] Wrote 499 events (found 1502 events) - [COUNTERS] PROGRAM TOTAL : 0.4424s - [COUNTERS] Fortran Overhead ( 0 ) : 0.4165s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0255s for 8192 events => throughput is 3.22E+05 events/s + [COUNTERS] PROGRAM TOTAL : 0.4321s + [COUNTERS] Fortran Overhead ( 0 ) : 0.4072s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0244s for 8192 events => throughput is 3.36E+05 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0005s *** (2-avx2) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -212,15 +307,50 @@ OK! xsec from fortran (0.20313504505737126) and cpp (0.20313504510700500) differ OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-avx2) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.avx2_m_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_gqttq_x10_cudacpp > /tmp/avalassi/output_gqttq_x10_cudacpp' +DEBUG: MEK processed 81920 events across 5 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 16/32 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 0.211 [0.21095842875361914] fbridge_mode=1 + [UNWEIGHT] Wrote 2259 events (found 2264 events) + [COUNTERS] PROGRAM TOTAL : 2.1440s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8984s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2451s for 81920 events => throughput is 3.34E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0005s + +*** (2-avx2) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (0.21095842877427595) and cpp (0.21095842875361914) differ by less than 2E-4 (9.791889521437724e-11) + +*** (2-avx2) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SM_GU_TTXU_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+BRDHST/avx2+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.243245e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.392741e+05 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_SM_GU_TTXU_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.311888e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.405961e+05 ) sec^-1 *** (2-512y) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -244,9 +374,9 @@ DEBUG: MEK processed 8192 events across 5 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 0.2031 [0.20313504510700500] fbridge_mode=1 [UNWEIGHT] Wrote 499 events (found 1502 events) - [COUNTERS] PROGRAM TOTAL : 0.4308s - [COUNTERS] Fortran Overhead ( 0 ) : 0.4086s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0217s for 8192 events => throughput is 3.78E+05 events/s + [COUNTERS] PROGRAM TOTAL : 0.4315s + [COUNTERS] Fortran Overhead ( 0 ) : 0.4088s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0222s for 8192 events => throughput is 3.69E+05 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0005s *** (2-512y) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -257,15 +387,50 @@ OK! xsec from fortran (0.20313504505737126) and cpp (0.20313504510700500) differ OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-512y) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.512y_m_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_gqttq_x10_cudacpp > /tmp/avalassi/output_gqttq_x10_cudacpp' +DEBUG: MEK processed 81920 events across 5 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 16/32 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 0.211 [0.21095842875361914] fbridge_mode=1 + [UNWEIGHT] Wrote 2259 events (found 2264 events) + [COUNTERS] PROGRAM TOTAL : 2.1103s + [COUNTERS] Fortran Overhead ( 0 ) : 1.8914s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2185s for 81920 events => throughput is 3.75E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0004s + +*** (2-512y) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (0.21095842877427595) and cpp (0.21095842875361914) differ by less than 2E-4 (9.791889521437724e-11) + +*** (2-512y) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SM_GU_TTXU_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+BRDHST/512y+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.793279e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.787571e+05 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_SM_GU_TTXU_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.775522e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.751557e+05 ) sec^-1 *** (2-512z) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -289,9 +454,9 @@ DEBUG: MEK processed 8192 events across 5 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 0.2031 [0.20313504510700500] fbridge_mode=1 [UNWEIGHT] Wrote 499 events (found 1502 events) - [COUNTERS] PROGRAM TOTAL : 0.4486s - [COUNTERS] Fortran Overhead ( 0 ) : 0.4131s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0350s for 8192 events => throughput is 2.34E+05 events/s + [COUNTERS] PROGRAM TOTAL : 0.4458s + [COUNTERS] Fortran Overhead ( 0 ) : 0.4105s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0348s for 8192 events => throughput is 2.36E+05 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0005s *** (2-512z) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -302,15 +467,50 @@ OK! xsec from fortran (0.20313504505737126) and cpp (0.20313504510700500) differ OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-512z) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.512z_m_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_gqttq_x10_cudacpp > /tmp/avalassi/output_gqttq_x10_cudacpp' +DEBUG: MEK processed 81920 events across 5 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 16/32 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 0.211 [0.21095842875361914] fbridge_mode=1 + [UNWEIGHT] Wrote 2259 events (found 2264 events) + [COUNTERS] PROGRAM TOTAL : 2.2524s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9020s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.3499s for 81920 events => throughput is 2.34E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0005s + +*** (2-512z) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (0.21095842877427595) and cpp (0.21095842875361914) differ by less than 2E-4 (9.791889521437724e-11) + +*** (2-512z) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SM_GU_TTXU_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+BRDHST/512z+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 2.316706e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.315061e+05 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_SM_GU_TTXU_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 2.334216e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.341669e+05 ) sec^-1 *** (3-cuda) EXECUTE MADEVENT_CUDA x1 (create events.lhe) *** -------------------- @@ -334,9 +534,9 @@ DEBUG: MEK processed 8192 events across 5 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 0.2031 [0.20313504512110778] fbridge_mode=1 [UNWEIGHT] Wrote 499 events (found 1502 events) - [COUNTERS] PROGRAM TOTAL : 0.8511s - [COUNTERS] Fortran Overhead ( 0 ) : 0.8469s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0031s for 8192 events => throughput is 2.63E+06 events/s + [COUNTERS] PROGRAM TOTAL : 0.8502s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8461s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0031s for 8192 events => throughput is 2.67E+06 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0011s *** (3-cuda) Compare MADEVENT_CUDA x1 xsec to MADEVENT_FORTRAN xsec *** @@ -347,45 +547,80 @@ OK! xsec from fortran (0.20313504505737126) and cuda (0.20313504512110778) diffe OK! events.lhe.cuda.1 and events.lhe.ref.1 are identical +*** (3-cuda) EXECUTE MADEVENT_CUDA x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.cuda_m_inl0_hrd0/madevent_cuda < /tmp/avalassi/input_gqttq_x10_cudacpp > /tmp/avalassi/output_gqttq_x10_cudacpp' +DEBUG: MEK processed 81920 events across 5 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 16/32 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 0.211 [0.21095842873460982] fbridge_mode=1 + [UNWEIGHT] Wrote 2259 events (found 2264 events) + [COUNTERS] PROGRAM TOTAL : 2.3490s + [COUNTERS] Fortran Overhead ( 0 ) : 2.3379s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0100s for 81920 events => throughput is 8.16E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0011s + +*** (3-cuda) Compare MADEVENT_CUDA x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (0.21095842877427595) and cuda (0.21095842873460982) differ by less than 2E-4 (1.8802814860663375e-10) + +*** (3-cuda) Compare MADEVENT_CUDA x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cuda.10 and events.lhe.ref.10 are identical + *** EXECUTE GCHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SM_GU_TTXU_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:MIX+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 2.929266e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.056451e+06 ) sec^-1 *** EXECUTE GCHECK(8192) -p 256 32 1 *** Process = SIGMA_SM_GU_TTXU_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:MIX+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.319589e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.321272e+06 ) sec^-1 *** EXECUTE GCHECK(MAX) -p 16384 32 1 --bridge *** Process = SIGMA_SM_GU_TTXU_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:MIX+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 2.340652e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.338600e+07 ) sec^-1 *** EXECUTE GCHECK(MAX) -p 16384 32 1 *** Process = SIGMA_SM_GU_TTXU_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:MIX+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 5.169068e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 5.234660e+07 ) sec^-1 *** EXECUTE GCHECK(MAX128THR) -p 4096 128 1 --bridge *** Process = SIGMA_SM_GU_TTXU_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:MIX+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 2.326566e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.340604e+07 ) sec^-1 *** EXECUTE GCHECK(MAX128THR) -p 4096 128 1 *** Process = SIGMA_SM_GU_TTXU_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:MIX+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 5.337296e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 5.358253e+07 ) sec^-1 *** EXECUTE GCHECK(MAX8THR) -p 65536 8 1 --bridge *** Process = SIGMA_SM_GU_TTXU_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:MIX+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 2.337938e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.356398e+07 ) sec^-1 *** EXECUTE GCHECK(MAX8THR) -p 65536 8 1 *** Process = SIGMA_SM_GU_TTXU_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:MIX+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.656612e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.659236e+07 ) sec^-1 *** (3-hip) WARNING! SKIP MADEVENT_HIP (hip is not supported on this node) *** diff --git a/epochX/cudacpp/tmad/logs_heftggbb_mad/log_heftggbb_mad_d_inl0_hrd0.txt b/epochX/cudacpp/tmad/logs_heftggbb_mad/log_heftggbb_mad_d_inl0_hrd0.txt index d2a669114e..b7144f5b16 100644 --- a/epochX/cudacpp/tmad/logs_heftggbb_mad/log_heftggbb_mad_d_inl0_hrd0.txt +++ b/epochX/cudacpp/tmad/logs_heftggbb_mad/log_heftggbb_mad_d_inl0_hrd0.txt @@ -1,8 +1,8 @@ Working directory (build): /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/heft_gg_bb.mad/SubProcesses/P1_gg_bbx -make USEBUILDDIR=1 BACKEND=cuda +make USEBUILDDIR=1 BACKEND=cuda make USEBUILDDIR=1 BACKEND=cppnone make USEBUILDDIR=1 BACKEND=cppsse4 @@ -13,10 +13,10 @@ make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/heft_gg_bb.mad/SubProcesses/P1_gg_bbx' make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/heft_gg_bb.mad/SubProcesses/P1_gg_bbx' make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/heft_gg_bb.mad/SubProcesses/P1_gg_bbx' +make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/heft_gg_bb.mad/SubProcesses/P1_gg_bbx' make USEBUILDDIR=1 BACKEND=cpp512z make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/heft_gg_bb.mad/SubProcesses/P1_gg_bbx' -make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/heft_gg_bb.mad/SubProcesses/P1_gg_bbx' make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/heft_gg_bb.mad/SubProcesses/P1_gg_bbx' make[1]: Nothing to be done for 'all'. @@ -32,7 +32,7 @@ make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ OMP_NUM_THREADS= -DATE: 2024-10-06_10:54:40 +DATE: 2024-10-06_18:07:11 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: Working directory (run): /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/heft_gg_bb.mad/SubProcesses/P1_gg_bbx @@ -58,9 +58,9 @@ Executing ' ./madevent_fortran < /tmp/avalassi/input_heftggbb_x1_fortran > /tmp/ [XSECTION] ChannelId = 1 [XSECTION] Cross section = 2.016 [2.0160081479755183] fbridge_mode=0 [UNWEIGHT] Wrote 3371 events (found 6399 events) - [COUNTERS] PROGRAM TOTAL : 0.9766s - [COUNTERS] Fortran Overhead ( 0 ) : 0.9291s - [COUNTERS] Fortran MEs ( 1 ) : 0.0475s for 8192 events => throughput is 1.72E+05 events/s + [COUNTERS] PROGRAM TOTAL : 0.9294s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8831s + [COUNTERS] Fortran MEs ( 1 ) : 0.0464s for 8192 events => throughput is 1.77E+05 events/s *** (1) EXECUTE MADEVENT_FORTRAN x1 (create events.lhe) *** -------------------- @@ -83,9 +83,34 @@ Executing ' ./madevent_fortran < /tmp/avalassi/input_heftggbb_x1_fortran > /tmp/ [XSECTION] ChannelId = 1 [XSECTION] Cross section = 2.016 [2.0160081479755183] fbridge_mode=0 [UNWEIGHT] Wrote 1652 events (found 1657 events) - [COUNTERS] PROGRAM TOTAL : 0.4581s - [COUNTERS] Fortran Overhead ( 0 ) : 0.4105s - [COUNTERS] Fortran MEs ( 1 ) : 0.0476s for 8192 events => throughput is 1.72E+05 events/s + [COUNTERS] PROGRAM TOTAL : 0.4431s + [COUNTERS] Fortran Overhead ( 0 ) : 0.3959s + [COUNTERS] Fortran MEs ( 1 ) : 0.0472s for 8192 events => throughput is 1.74E+05 events/s + +*** (1) EXECUTE MADEVENT_FORTRAN x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./madevent_fortran < /tmp/avalassi/input_heftggbb_x10_fortran > /tmp/avalassi/output_heftggbb_x10_fortran' + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 16/16 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 2.034 [2.0336713375865285] fbridge_mode=0 + [UNWEIGHT] Wrote 1707 events (found 1712 events) + [COUNTERS] PROGRAM TOTAL : 1.9894s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5193s + [COUNTERS] Fortran MEs ( 1 ) : 0.4700s for 81920 events => throughput is 1.74E+05 events/s *** (2-none) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -109,9 +134,9 @@ DEBUG: MEK processed 8192 events across 4 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 2.016 [2.0160081479755170] fbridge_mode=1 [UNWEIGHT] Wrote 1652 events (found 1657 events) - [COUNTERS] PROGRAM TOTAL : 0.4592s - [COUNTERS] Fortran Overhead ( 0 ) : 0.4086s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0501s for 8192 events => throughput is 1.63E+05 events/s + [COUNTERS] PROGRAM TOTAL : 0.4456s + [COUNTERS] Fortran Overhead ( 0 ) : 0.3954s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0497s for 8192 events => throughput is 1.65E+05 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0004s *** (2-none) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -122,15 +147,50 @@ OK! xsec from fortran (2.0160081479755183) and cpp (2.0160081479755170) differ b OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-none) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.none_d_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_heftggbb_x10_cudacpp > /tmp/avalassi/output_heftggbb_x10_cudacpp' +DEBUG: MEK processed 81920 events across 4 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 16/16 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 2.034 [2.0336713375865285] fbridge_mode=1 + [UNWEIGHT] Wrote 1707 events (found 1712 events) + [COUNTERS] PROGRAM TOTAL : 2.0136s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5203s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4929s for 81920 events => throughput is 1.66E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0004s + +*** (2-none) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (2.0336713375865285) and cpp (2.0336713375865285) differ by less than 3E-14 (0.0) + +*** (2-none) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_HEFT_GG_BBX_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+BRDHST/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.648377e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.718509e+05 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_HEFT_GG_BBX_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.642355e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.691982e+05 ) sec^-1 *** (2-sse4) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -154,9 +214,9 @@ DEBUG: MEK processed 8192 events across 4 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 2.016 [2.0160081479755183] fbridge_mode=1 [UNWEIGHT] Wrote 1652 events (found 1657 events) - [COUNTERS] PROGRAM TOTAL : 0.4344s - [COUNTERS] Fortran Overhead ( 0 ) : 0.4065s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0275s for 8192 events => throughput is 2.98E+05 events/s + [COUNTERS] PROGRAM TOTAL : 0.4293s + [COUNTERS] Fortran Overhead ( 0 ) : 0.4023s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0267s for 8192 events => throughput is 3.07E+05 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0004s *** (2-sse4) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -167,15 +227,50 @@ OK! xsec from fortran (2.0160081479755183) and cpp (2.0160081479755183) differ b OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-sse4) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.sse4_d_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_heftggbb_x10_cudacpp > /tmp/avalassi/output_heftggbb_x10_cudacpp' +DEBUG: MEK processed 81920 events across 4 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 16/16 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 2.034 [2.0336713375865285] fbridge_mode=1 + [UNWEIGHT] Wrote 1707 events (found 1712 events) + [COUNTERS] PROGRAM TOTAL : 1.7864s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5194s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2667s for 81920 events => throughput is 3.07E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0003s + +*** (2-sse4) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (2.0336713375865285) and cpp (2.0336713375865285) differ by less than 3E-14 (0.0) + +*** (2-sse4) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_HEFT_GG_BBX_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+BRDHST/sse4+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 2.984151e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.050993e+05 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_HEFT_GG_BBX_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.017550e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.049211e+05 ) sec^-1 *** (2-avx2) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -199,9 +294,9 @@ DEBUG: MEK processed 8192 events across 4 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 2.016 [2.0160081479755165] fbridge_mode=1 [UNWEIGHT] Wrote 1652 events (found 1657 events) - [COUNTERS] PROGRAM TOTAL : 0.4261s - [COUNTERS] Fortran Overhead ( 0 ) : 0.4086s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0171s for 8192 events => throughput is 4.80E+05 events/s + [COUNTERS] PROGRAM TOTAL : 0.4153s + [COUNTERS] Fortran Overhead ( 0 ) : 0.3985s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0164s for 8192 events => throughput is 5.00E+05 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0004s *** (2-avx2) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -212,15 +307,50 @@ OK! xsec from fortran (2.0160081479755183) and cpp (2.0160081479755165) differ b OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-avx2) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.avx2_d_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_heftggbb_x10_cudacpp > /tmp/avalassi/output_heftggbb_x10_cudacpp' +DEBUG: MEK processed 81920 events across 4 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 16/16 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 2.034 [2.0336713375865476] fbridge_mode=1 + [UNWEIGHT] Wrote 1707 events (found 1712 events) + [COUNTERS] PROGRAM TOTAL : 1.6852s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5225s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1623s for 81920 events => throughput is 5.05E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0003s + +*** (2-avx2) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (2.0336713375865285) and cpp (2.0336713375865476) differ by less than 3E-14 (9.325873406851315e-15) + +*** (2-avx2) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_HEFT_GG_BBX_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+BRDHST/avx2+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 4.938014e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 4.816981e+05 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_HEFT_GG_BBX_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 4.942444e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 4.908466e+05 ) sec^-1 *** (2-512y) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -244,9 +374,9 @@ DEBUG: MEK processed 8192 events across 4 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 2.016 [2.0160081479755165] fbridge_mode=1 [UNWEIGHT] Wrote 1652 events (found 1657 events) - [COUNTERS] PROGRAM TOTAL : 0.4299s - [COUNTERS] Fortran Overhead ( 0 ) : 0.4143s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0153s for 8192 events => throughput is 5.37E+05 events/s + [COUNTERS] PROGRAM TOTAL : 0.4258s + [COUNTERS] Fortran Overhead ( 0 ) : 0.4093s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0161s for 8192 events => throughput is 5.09E+05 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0004s *** (2-512y) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -257,15 +387,50 @@ OK! xsec from fortran (2.0160081479755183) and cpp (2.0160081479755165) differ b OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-512y) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.512y_d_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_heftggbb_x10_cudacpp > /tmp/avalassi/output_heftggbb_x10_cudacpp' +DEBUG: MEK processed 81920 events across 4 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 16/16 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 2.034 [2.0336713375865476] fbridge_mode=1 + [UNWEIGHT] Wrote 1707 events (found 1712 events) + [COUNTERS] PROGRAM TOTAL : 1.6773s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5279s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1491s for 81920 events => throughput is 5.49E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0003s + +*** (2-512y) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (2.0336713375865285) and cpp (2.0336713375865476) differ by less than 3E-14 (9.325873406851315e-15) + +*** (2-512y) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_HEFT_GG_BBX_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+BRDHST/512y+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 5.398535e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 5.399166e+05 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_HEFT_GG_BBX_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 5.466636e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 5.455309e+05 ) sec^-1 *** (2-512z) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -289,9 +454,9 @@ DEBUG: MEK processed 8192 events across 4 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 2.016 [2.0160081479755179] fbridge_mode=1 [UNWEIGHT] Wrote 1652 events (found 1657 events) - [COUNTERS] PROGRAM TOTAL : 0.4391s - [COUNTERS] Fortran Overhead ( 0 ) : 0.4149s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0238s for 8192 events => throughput is 3.45E+05 events/s + [COUNTERS] PROGRAM TOTAL : 0.4162s + [COUNTERS] Fortran Overhead ( 0 ) : 0.3938s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0220s for 8192 events => throughput is 3.72E+05 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0004s *** (2-512z) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -302,15 +467,50 @@ OK! xsec from fortran (2.0160081479755183) and cpp (2.0160081479755179) differ b OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-512z) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.512z_d_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_heftggbb_x10_cudacpp > /tmp/avalassi/output_heftggbb_x10_cudacpp' +DEBUG: MEK processed 81920 events across 4 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 16/16 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 2.034 [2.0336713375865285] fbridge_mode=1 + [UNWEIGHT] Wrote 1707 events (found 1712 events) + [COUNTERS] PROGRAM TOTAL : 1.7773s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5509s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2260s for 81920 events => throughput is 3.63E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0004s + +*** (2-512z) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (2.0336713375865285) and cpp (2.0336713375865285) differ by less than 3E-14 (0.0) + +*** (2-512z) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_HEFT_GG_BBX_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+BRDHST/512z+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.480162e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.535219e+05 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_HEFT_GG_BBX_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.526547e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.560847e+05 ) sec^-1 *** (3-cuda) EXECUTE MADEVENT_CUDA x1 (create events.lhe) *** -------------------- @@ -334,9 +534,9 @@ DEBUG: MEK processed 8192 events across 4 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 2.016 [2.0160081479755192] fbridge_mode=1 [UNWEIGHT] Wrote 1652 events (found 1657 events) - [COUNTERS] PROGRAM TOTAL : 0.8532s - [COUNTERS] Fortran Overhead ( 0 ) : 0.8493s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0031s for 8192 events => throughput is 2.62E+06 events/s + [COUNTERS] PROGRAM TOTAL : 0.8635s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8597s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0031s for 8192 events => throughput is 2.67E+06 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0008s *** (3-cuda) Compare MADEVENT_CUDA x1 xsec to MADEVENT_FORTRAN xsec *** @@ -347,45 +547,80 @@ OK! xsec from fortran (2.0160081479755183) and cuda (2.0160081479755192) differ OK! events.lhe.cuda.1 and events.lhe.ref.1 are identical +*** (3-cuda) EXECUTE MADEVENT_CUDA x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.cuda_d_inl0_hrd0/madevent_cuda < /tmp/avalassi/input_heftggbb_x10_cudacpp > /tmp/avalassi/output_heftggbb_x10_cudacpp' +DEBUG: MEK processed 81920 events across 4 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 16/16 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 2.034 [2.0336713375865294] fbridge_mode=1 + [UNWEIGHT] Wrote 1707 events (found 1712 events) + [COUNTERS] PROGRAM TOTAL : 1.9821s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9724s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0089s for 81920 events => throughput is 9.21E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0008s + +*** (3-cuda) Compare MADEVENT_CUDA x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (2.0336713375865285) and cuda (2.0336713375865294) differ by less than 3E-14 (4.440892098500626e-16) + +*** (3-cuda) Compare MADEVENT_CUDA x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cuda.10 and events.lhe.ref.10 are identical + *** EXECUTE GCHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_HEFT_GG_BBX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 2.920216e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.059528e+06 ) sec^-1 *** EXECUTE GCHECK(8192) -p 256 32 1 *** Process = SIGMA_HEFT_GG_BBX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.457557e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.467293e+06 ) sec^-1 *** EXECUTE GCHECK(MAX) -p 16384 32 1 --bridge *** Process = SIGMA_HEFT_GG_BBX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 2.816989e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.850145e+07 ) sec^-1 *** EXECUTE GCHECK(MAX) -p 16384 32 1 *** Process = SIGMA_HEFT_GG_BBX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 7.149758e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 7.204356e+07 ) sec^-1 *** EXECUTE GCHECK(MAX128THR) -p 4096 128 1 --bridge *** Process = SIGMA_HEFT_GG_BBX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 2.802618e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.864075e+07 ) sec^-1 *** EXECUTE GCHECK(MAX128THR) -p 4096 128 1 *** Process = SIGMA_HEFT_GG_BBX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 7.511448e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 7.461303e+07 ) sec^-1 *** EXECUTE GCHECK(MAX8THR) -p 65536 8 1 --bridge *** Process = SIGMA_HEFT_GG_BBX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 2.832166e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.870591e+07 ) sec^-1 *** EXECUTE GCHECK(MAX8THR) -p 65536 8 1 *** Process = SIGMA_HEFT_GG_BBX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 2.514724e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.543680e+07 ) sec^-1 *** (3-hip) WARNING! SKIP MADEVENT_HIP (hip is not supported on this node) *** diff --git a/epochX/cudacpp/tmad/logs_heftggbb_mad/log_heftggbb_mad_f_inl0_hrd0.txt b/epochX/cudacpp/tmad/logs_heftggbb_mad/log_heftggbb_mad_f_inl0_hrd0.txt index 483bc4166c..dbed5e8998 100644 --- a/epochX/cudacpp/tmad/logs_heftggbb_mad/log_heftggbb_mad_f_inl0_hrd0.txt +++ b/epochX/cudacpp/tmad/logs_heftggbb_mad/log_heftggbb_mad_f_inl0_hrd0.txt @@ -1,13 +1,13 @@ Working directory (build): /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/heft_gg_bb.mad/SubProcesses/P1_gg_bbx -make USEBUILDDIR=1 BACKEND=cuda +make USEBUILDDIR=1 BACKEND=cuda make USEBUILDDIR=1 BACKEND=cppnone make USEBUILDDIR=1 BACKEND=cppsse4 - make USEBUILDDIR=1 BACKEND=cppavx2 + make USEBUILDDIR=1 BACKEND=cpp512y make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/heft_gg_bb.mad/SubProcesses/P1_gg_bbx' make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/heft_gg_bb.mad/SubProcesses/P1_gg_bbx' @@ -32,7 +32,7 @@ make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ OMP_NUM_THREADS= -DATE: 2024-10-06_10:55:09 +DATE: 2024-10-06_18:08:08 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: Working directory (run): /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/heft_gg_bb.mad/SubProcesses/P1_gg_bbx @@ -58,9 +58,9 @@ Executing ' ./madevent_fortran < /tmp/avalassi/input_heftggbb_x1_fortran > /tmp/ [XSECTION] ChannelId = 1 [XSECTION] Cross section = 2.016 [2.0160081479755183] fbridge_mode=0 [UNWEIGHT] Wrote 3371 events (found 6399 events) - [COUNTERS] PROGRAM TOTAL : 0.9638s - [COUNTERS] Fortran Overhead ( 0 ) : 0.9156s - [COUNTERS] Fortran MEs ( 1 ) : 0.0482s for 8192 events => throughput is 1.70E+05 events/s + [COUNTERS] PROGRAM TOTAL : 0.9295s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8835s + [COUNTERS] Fortran MEs ( 1 ) : 0.0459s for 8192 events => throughput is 1.78E+05 events/s *** (1) EXECUTE MADEVENT_FORTRAN x1 (create events.lhe) *** -------------------- @@ -83,9 +83,34 @@ Executing ' ./madevent_fortran < /tmp/avalassi/input_heftggbb_x1_fortran > /tmp/ [XSECTION] ChannelId = 1 [XSECTION] Cross section = 2.016 [2.0160081479755183] fbridge_mode=0 [UNWEIGHT] Wrote 1652 events (found 1657 events) - [COUNTERS] PROGRAM TOTAL : 0.4563s - [COUNTERS] Fortran Overhead ( 0 ) : 0.4085s - [COUNTERS] Fortran MEs ( 1 ) : 0.0478s for 8192 events => throughput is 1.71E+05 events/s + [COUNTERS] PROGRAM TOTAL : 0.4410s + [COUNTERS] Fortran Overhead ( 0 ) : 0.3958s + [COUNTERS] Fortran MEs ( 1 ) : 0.0452s for 8192 events => throughput is 1.81E+05 events/s + +*** (1) EXECUTE MADEVENT_FORTRAN x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./madevent_fortran < /tmp/avalassi/input_heftggbb_x10_fortran > /tmp/avalassi/output_heftggbb_x10_fortran' + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 16/16 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 2.034 [2.0336713375865285] fbridge_mode=0 + [UNWEIGHT] Wrote 1707 events (found 1712 events) + [COUNTERS] PROGRAM TOTAL : 1.9835s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5188s + [COUNTERS] Fortran MEs ( 1 ) : 0.4647s for 81920 events => throughput is 1.76E+05 events/s *** (2-none) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -109,9 +134,9 @@ DEBUG: MEK processed 8192 events across 4 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 2.016 [2.0160406825242951] fbridge_mode=1 [UNWEIGHT] Wrote 1653 events (found 1658 events) - [COUNTERS] PROGRAM TOTAL : 0.4552s - [COUNTERS] Fortran Overhead ( 0 ) : 0.4076s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0473s for 8192 events => throughput is 1.73E+05 events/s + [COUNTERS] PROGRAM TOTAL : 0.4427s + [COUNTERS] Fortran Overhead ( 0 ) : 0.3973s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0451s for 8192 events => throughput is 1.82E+05 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0003s *** (2-none) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** diff --git a/epochX/cudacpp/tmad/logs_heftggbb_mad/log_heftggbb_mad_m_inl0_hrd0.txt b/epochX/cudacpp/tmad/logs_heftggbb_mad/log_heftggbb_mad_m_inl0_hrd0.txt index b61563e796..219f877310 100644 --- a/epochX/cudacpp/tmad/logs_heftggbb_mad/log_heftggbb_mad_m_inl0_hrd0.txt +++ b/epochX/cudacpp/tmad/logs_heftggbb_mad/log_heftggbb_mad_m_inl0_hrd0.txt @@ -1,13 +1,13 @@ Working directory (build): /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/heft_gg_bb.mad/SubProcesses/P1_gg_bbx - make USEBUILDDIR=1 BACKEND=cuda make USEBUILDDIR=1 BACKEND=cppnone + + make USEBUILDDIR=1 BACKEND=cppsse4 make USEBUILDDIR=1 BACKEND=cppavx2 - make USEBUILDDIR=1 BACKEND=cpp512y make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/heft_gg_bb.mad/SubProcesses/P1_gg_bbx' make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/heft_gg_bb.mad/SubProcesses/P1_gg_bbx' @@ -32,7 +32,7 @@ make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ OMP_NUM_THREADS= -DATE: 2024-10-06_10:54:54 +DATE: 2024-10-06_18:07:39 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: Working directory (run): /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/heft_gg_bb.mad/SubProcesses/P1_gg_bbx @@ -58,9 +58,9 @@ Executing ' ./madevent_fortran < /tmp/avalassi/input_heftggbb_x1_fortran > /tmp/ [XSECTION] ChannelId = 1 [XSECTION] Cross section = 2.016 [2.0160081479755183] fbridge_mode=0 [UNWEIGHT] Wrote 3371 events (found 6399 events) - [COUNTERS] PROGRAM TOTAL : 0.9594s - [COUNTERS] Fortran Overhead ( 0 ) : 0.9118s - [COUNTERS] Fortran MEs ( 1 ) : 0.0475s for 8192 events => throughput is 1.72E+05 events/s + [COUNTERS] PROGRAM TOTAL : 0.9312s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8847s + [COUNTERS] Fortran MEs ( 1 ) : 0.0465s for 8192 events => throughput is 1.76E+05 events/s *** (1) EXECUTE MADEVENT_FORTRAN x1 (create events.lhe) *** -------------------- @@ -83,9 +83,34 @@ Executing ' ./madevent_fortran < /tmp/avalassi/input_heftggbb_x1_fortran > /tmp/ [XSECTION] ChannelId = 1 [XSECTION] Cross section = 2.016 [2.0160081479755183] fbridge_mode=0 [UNWEIGHT] Wrote 1652 events (found 1657 events) - [COUNTERS] PROGRAM TOTAL : 0.4589s - [COUNTERS] Fortran Overhead ( 0 ) : 0.4111s - [COUNTERS] Fortran MEs ( 1 ) : 0.0478s for 8192 events => throughput is 1.71E+05 events/s + [COUNTERS] PROGRAM TOTAL : 0.4435s + [COUNTERS] Fortran Overhead ( 0 ) : 0.3974s + [COUNTERS] Fortran MEs ( 1 ) : 0.0461s for 8192 events => throughput is 1.78E+05 events/s + +*** (1) EXECUTE MADEVENT_FORTRAN x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./madevent_fortran < /tmp/avalassi/input_heftggbb_x10_fortran > /tmp/avalassi/output_heftggbb_x10_fortran' + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 16/16 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 2.034 [2.0336713375865285] fbridge_mode=0 + [UNWEIGHT] Wrote 1707 events (found 1712 events) + [COUNTERS] PROGRAM TOTAL : 1.9811s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5183s + [COUNTERS] Fortran MEs ( 1 ) : 0.4628s for 81920 events => throughput is 1.77E+05 events/s *** (2-none) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -110,9 +135,9 @@ DEBUG: MEK processed 8192 events across 4 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 2.016 [2.0160081964453331] fbridge_mode=1 [UNWEIGHT] Wrote 1652 events (found 1657 events) - [COUNTERS] PROGRAM TOTAL : 0.4600s - [COUNTERS] Fortran Overhead ( 0 ) : 0.4089s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0507s for 8192 events => throughput is 1.62E+05 events/s + [COUNTERS] PROGRAM TOTAL : 0.4425s + [COUNTERS] Fortran Overhead ( 0 ) : 0.3928s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0493s for 8192 events => throughput is 1.66E+05 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0004s *** (2-none) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -123,18 +148,54 @@ OK! xsec from fortran (2.0160081479755183) and cpp (2.0160081964453331) differ b OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-none) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.none_m_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_heftggbb_x10_cudacpp > /tmp/avalassi/output_heftggbb_x10_cudacpp' +INFO: The following Floating Point Exceptions have been reported: FE_UNDERFLOW +DEBUG: MEK processed 81920 events across 4 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 16/16 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 2.034 [2.0336713843200420] fbridge_mode=1 + [UNWEIGHT] Wrote 1707 events (found 1712 events) + [COUNTERS] PROGRAM TOTAL : 2.0147s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5206s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4937s for 81920 events => throughput is 1.66E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0004s + +*** (2-none) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (2.0336713375865285) and cpp (2.0336713843200420) differ by less than 2E-4 (2.2979875113904313e-08) + +*** (2-none) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** INFO: The following Floating Point Exceptions have been reported: FE_UNDERFLOW INFO: The following Floating Point Exceptions have been reported: FE_UNDERFLOW Process = SIGMA_HEFT_GG_BBX_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+BRDHST/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.539881e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.575311e+05 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** INFO: The following Floating Point Exceptions have been reported: FE_UNDERFLOW Process = SIGMA_HEFT_GG_BBX_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.532971e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.577018e+05 ) sec^-1 *** (2-sse4) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -159,9 +220,9 @@ DEBUG: MEK processed 8192 events across 4 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 2.016 [2.0160081964453336] fbridge_mode=1 [UNWEIGHT] Wrote 1652 events (found 1657 events) - [COUNTERS] PROGRAM TOTAL : 0.4363s - [COUNTERS] Fortran Overhead ( 0 ) : 0.4080s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0279s for 8192 events => throughput is 2.94E+05 events/s + [COUNTERS] PROGRAM TOTAL : 0.4314s + [COUNTERS] Fortran Overhead ( 0 ) : 0.4029s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0281s for 8192 events => throughput is 2.91E+05 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0004s *** (2-sse4) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -172,18 +233,54 @@ OK! xsec from fortran (2.0160081479755183) and cpp (2.0160081964453336) differ b OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-sse4) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.sse4_m_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_heftggbb_x10_cudacpp > /tmp/avalassi/output_heftggbb_x10_cudacpp' +INFO: The following Floating Point Exceptions have been reported: FE_UNDERFLOW +DEBUG: MEK processed 81920 events across 4 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 16/16 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 2.034 [2.0336713843200425] fbridge_mode=1 + [UNWEIGHT] Wrote 1707 events (found 1712 events) + [COUNTERS] PROGRAM TOTAL : 1.8049s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5342s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2703s for 81920 events => throughput is 3.03E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0004s + +*** (2-sse4) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (2.0336713375865285) and cpp (2.0336713843200425) differ by less than 2E-4 (2.2979875335948918e-08) + +*** (2-sse4) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** INFO: The following Floating Point Exceptions have been reported: FE_UNDERFLOW INFO: The following Floating Point Exceptions have been reported: FE_UNDERFLOW Process = SIGMA_HEFT_GG_BBX_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+BRDHST/sse4+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 2.824636e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.909582e+05 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** INFO: The following Floating Point Exceptions have been reported: FE_UNDERFLOW Process = SIGMA_HEFT_GG_BBX_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 2.869373e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.897797e+05 ) sec^-1 *** (2-avx2) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -208,9 +305,9 @@ DEBUG: MEK processed 8192 events across 4 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 2.016 [2.0160081962974745] fbridge_mode=1 [UNWEIGHT] Wrote 1652 events (found 1657 events) - [COUNTERS] PROGRAM TOTAL : 0.4311s - [COUNTERS] Fortran Overhead ( 0 ) : 0.4138s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0169s for 8192 events => throughput is 4.85E+05 events/s + [COUNTERS] PROGRAM TOTAL : 0.4140s + [COUNTERS] Fortran Overhead ( 0 ) : 0.3971s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0165s for 8192 events => throughput is 4.95E+05 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0004s *** (2-avx2) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -221,18 +318,54 @@ OK! xsec from fortran (2.0160081479755183) and cpp (2.0160081962974745) differ b OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-avx2) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.avx2_m_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_heftggbb_x10_cudacpp > /tmp/avalassi/output_heftggbb_x10_cudacpp' +INFO: The following Floating Point Exceptions have been reported: FE_UNDERFLOW +DEBUG: MEK processed 81920 events across 4 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 16/16 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 2.034 [2.0336713836598665] fbridge_mode=1 + [UNWEIGHT] Wrote 1707 events (found 1712 events) + [COUNTERS] PROGRAM TOTAL : 1.6788s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5143s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1641s for 81920 events => throughput is 4.99E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0003s + +*** (2-avx2) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (2.0336713375865285) and cpp (2.0336713836598665) differ by less than 2E-4 (2.265525278488667e-08) + +*** (2-avx2) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** INFO: The following Floating Point Exceptions have been reported: FE_UNDERFLOW INFO: The following Floating Point Exceptions have been reported: FE_UNDERFLOW Process = SIGMA_HEFT_GG_BBX_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+BRDHST/avx2+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 4.809707e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 4.748372e+05 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** INFO: The following Floating Point Exceptions have been reported: FE_UNDERFLOW Process = SIGMA_HEFT_GG_BBX_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 4.724204e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 4.806903e+05 ) sec^-1 *** (2-512y) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -257,9 +390,9 @@ DEBUG: MEK processed 8192 events across 4 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 2.016 [2.0160081962974745] fbridge_mode=1 [UNWEIGHT] Wrote 1652 events (found 1657 events) - [COUNTERS] PROGRAM TOTAL : 0.4252s - [COUNTERS] Fortran Overhead ( 0 ) : 0.4093s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0155s for 8192 events => throughput is 5.28E+05 events/s + [COUNTERS] PROGRAM TOTAL : 0.4089s + [COUNTERS] Fortran Overhead ( 0 ) : 0.3931s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0154s for 8192 events => throughput is 5.33E+05 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0004s *** (2-512y) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -270,18 +403,54 @@ OK! xsec from fortran (2.0160081479755183) and cpp (2.0160081962974745) differ b OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-512y) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.512y_m_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_heftggbb_x10_cudacpp > /tmp/avalassi/output_heftggbb_x10_cudacpp' +INFO: The following Floating Point Exceptions have been reported: FE_UNDERFLOW +DEBUG: MEK processed 81920 events across 4 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 16/16 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 2.034 [2.0336713836598665] fbridge_mode=1 + [UNWEIGHT] Wrote 1707 events (found 1712 events) + [COUNTERS] PROGRAM TOTAL : 1.6673s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5162s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1507s for 81920 events => throughput is 5.44E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0004s + +*** (2-512y) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (2.0336713375865285) and cpp (2.0336713836598665) differ by less than 2E-4 (2.265525278488667e-08) + +*** (2-512y) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** INFO: The following Floating Point Exceptions have been reported: FE_UNDERFLOW INFO: The following Floating Point Exceptions have been reported: FE_UNDERFLOW Process = SIGMA_HEFT_GG_BBX_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+BRDHST/512y+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 5.163712e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 5.165866e+05 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** INFO: The following Floating Point Exceptions have been reported: FE_UNDERFLOW Process = SIGMA_HEFT_GG_BBX_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 5.204514e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 5.214443e+05 ) sec^-1 *** (2-512z) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -306,9 +475,9 @@ DEBUG: MEK processed 8192 events across 4 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 2.016 [2.0160081962970020] fbridge_mode=1 [UNWEIGHT] Wrote 1652 events (found 1657 events) - [COUNTERS] PROGRAM TOTAL : 0.4306s - [COUNTERS] Fortran Overhead ( 0 ) : 0.4058s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0244s for 8192 events => throughput is 3.36E+05 events/s + [COUNTERS] PROGRAM TOTAL : 0.4204s + [COUNTERS] Fortran Overhead ( 0 ) : 0.3964s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0236s for 8192 events => throughput is 3.48E+05 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0004s *** (2-512z) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -319,18 +488,54 @@ OK! xsec from fortran (2.0160081479755183) and cpp (2.0160081962970020) differ b OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-512z) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.512z_m_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_heftggbb_x10_cudacpp > /tmp/avalassi/output_heftggbb_x10_cudacpp' +INFO: The following Floating Point Exceptions have been reported: FE_UNDERFLOW +DEBUG: MEK processed 81920 events across 4 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 16/16 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 2.034 [2.0336713836598515] fbridge_mode=1 + [UNWEIGHT] Wrote 1707 events (found 1712 events) + [COUNTERS] PROGRAM TOTAL : 1.7552s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5275s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2273s for 81920 events => throughput is 3.60E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0004s + +*** (2-512z) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (2.0336713375865285) and cpp (2.0336713836598515) differ by less than 2E-4 (2.2655245235370103e-08) + +*** (2-512z) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** INFO: The following Floating Point Exceptions have been reported: FE_UNDERFLOW INFO: The following Floating Point Exceptions have been reported: FE_UNDERFLOW Process = SIGMA_HEFT_GG_BBX_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+BRDHST/512z+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.121651e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.362727e+05 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** INFO: The following Floating Point Exceptions have been reported: FE_UNDERFLOW Process = SIGMA_HEFT_GG_BBX_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.119023e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.208876e+05 ) sec^-1 *** (3-cuda) EXECUTE MADEVENT_CUDA x1 (create events.lhe) *** -------------------- @@ -354,9 +559,9 @@ DEBUG: MEK processed 8192 events across 4 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 2.016 [2.0160081483021330] fbridge_mode=1 [UNWEIGHT] Wrote 1652 events (found 1657 events) - [COUNTERS] PROGRAM TOTAL : 0.8574s - [COUNTERS] Fortran Overhead ( 0 ) : 0.8536s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0031s for 8192 events => throughput is 2.63E+06 events/s + [COUNTERS] PROGRAM TOTAL : 0.8367s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8329s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0031s for 8192 events => throughput is 2.67E+06 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0008s *** (3-cuda) Compare MADEVENT_CUDA x1 xsec to MADEVENT_FORTRAN xsec *** @@ -367,45 +572,80 @@ OK! xsec from fortran (2.0160081479755183) and cuda (2.0160081483021330) differ OK! events.lhe.cuda.1 and events.lhe.ref.1 are identical +*** (3-cuda) EXECUTE MADEVENT_CUDA x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.cuda_m_inl0_hrd0/madevent_cuda < /tmp/avalassi/input_heftggbb_x10_cudacpp > /tmp/avalassi/output_heftggbb_x10_cudacpp' +DEBUG: MEK processed 81920 events across 4 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 16/16 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 2.034 [2.0336713380111449] fbridge_mode=1 + [UNWEIGHT] Wrote 1707 events (found 1712 events) + [COUNTERS] PROGRAM TOTAL : 1.9728s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9632s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0089s for 81920 events => throughput is 9.22E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0008s + +*** (3-cuda) Compare MADEVENT_CUDA x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (2.0336713375865285) and cuda (2.0336713380111449) differ by less than 2E-4 (2.0879298290310544e-10) + +*** (3-cuda) Compare MADEVENT_CUDA x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cuda.10 and events.lhe.ref.10 are identical + *** EXECUTE GCHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_HEFT_GG_BBX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:MIX+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.018963e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.971854e+06 ) sec^-1 *** EXECUTE GCHECK(8192) -p 256 32 1 *** Process = SIGMA_HEFT_GG_BBX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:MIX+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.363694e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.494992e+06 ) sec^-1 *** EXECUTE GCHECK(MAX) -p 16384 32 1 --bridge *** Process = SIGMA_HEFT_GG_BBX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:MIX+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 2.820757e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.855479e+07 ) sec^-1 *** EXECUTE GCHECK(MAX) -p 16384 32 1 *** Process = SIGMA_HEFT_GG_BBX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:MIX+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 7.067644e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 7.243059e+07 ) sec^-1 *** EXECUTE GCHECK(MAX128THR) -p 4096 128 1 --bridge *** Process = SIGMA_HEFT_GG_BBX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:MIX+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 2.797704e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.875636e+07 ) sec^-1 *** EXECUTE GCHECK(MAX128THR) -p 4096 128 1 *** Process = SIGMA_HEFT_GG_BBX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:MIX+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 7.465309e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 7.500162e+07 ) sec^-1 *** EXECUTE GCHECK(MAX8THR) -p 65536 8 1 --bridge *** Process = SIGMA_HEFT_GG_BBX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:MIX+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 2.821262e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.884903e+07 ) sec^-1 *** EXECUTE GCHECK(MAX8THR) -p 65536 8 1 *** Process = SIGMA_HEFT_GG_BBX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:MIX+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 2.503862e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.482605e+07 ) sec^-1 *** (3-hip) WARNING! SKIP MADEVENT_HIP (hip is not supported on this node) *** diff --git a/epochX/cudacpp/tmad/logs_smeftggtttt_mad/log_smeftggtttt_mad_d_inl0_hrd0.txt b/epochX/cudacpp/tmad/logs_smeftggtttt_mad/log_smeftggtttt_mad_d_inl0_hrd0.txt index d3cb91b8cd..9cafa8a5e0 100644 --- a/epochX/cudacpp/tmad/logs_smeftggtttt_mad/log_smeftggtttt_mad_d_inl0_hrd0.txt +++ b/epochX/cudacpp/tmad/logs_smeftggtttt_mad/log_smeftggtttt_mad_d_inl0_hrd0.txt @@ -1,13 +1,13 @@ Working directory (build): /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/smeft_gg_tttt.mad/SubProcesses/P1_gg_ttxttx + make USEBUILDDIR=1 BACKEND=cuda make USEBUILDDIR=1 BACKEND=cppnone - make USEBUILDDIR=1 BACKEND=cppsse4 - make USEBUILDDIR=1 BACKEND=cppavx2 + make USEBUILDDIR=1 BACKEND=cpp512y make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/smeft_gg_tttt.mad/SubProcesses/P1_gg_ttxttx' make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/smeft_gg_tttt.mad/SubProcesses/P1_gg_ttxttx' @@ -32,7 +32,7 @@ make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ OMP_NUM_THREADS= -DATE: 2024-10-06_10:56:37 +DATE: 2024-10-06_18:10:57 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: Working directory (run): /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/smeft_gg_tttt.mad/SubProcesses/P1_gg_ttxttx @@ -58,9 +58,9 @@ Executing ' ./madevent_fortran < /tmp/avalassi/input_smeftggtttt_x1_fortran > /t [XSECTION] ChannelId = 1 [XSECTION] Cross section = 7.638e-07 [7.6381610362728588E-007] fbridge_mode=0 [UNWEIGHT] Wrote 1 events (found 902 events) - [COUNTERS] PROGRAM TOTAL : 2.6766s - [COUNTERS] Fortran Overhead ( 0 ) : 0.3643s - [COUNTERS] Fortran MEs ( 1 ) : 2.3123s for 8192 events => throughput is 3.54E+03 events/s + [COUNTERS] PROGRAM TOTAL : 2.6034s + [COUNTERS] Fortran Overhead ( 0 ) : 0.3569s + [COUNTERS] Fortran MEs ( 1 ) : 2.2465s for 8192 events => throughput is 3.65E+03 events/s *** (1) EXECUTE MADEVENT_FORTRAN x1 (create events.lhe) *** -------------------- @@ -83,9 +83,34 @@ Executing ' ./madevent_fortran < /tmp/avalassi/input_smeftggtttt_x1_fortran > /t [XSECTION] ChannelId = 1 [XSECTION] Cross section = 7.638e-07 [7.6381610362728588E-007] fbridge_mode=0 [UNWEIGHT] Wrote 230 events (found 851 events) - [COUNTERS] PROGRAM TOTAL : 2.6640s - [COUNTERS] Fortran Overhead ( 0 ) : 0.3633s - [COUNTERS] Fortran MEs ( 1 ) : 2.3007s for 8192 events => throughput is 3.56E+03 events/s + [COUNTERS] PROGRAM TOTAL : 2.5896s + [COUNTERS] Fortran Overhead ( 0 ) : 0.3537s + [COUNTERS] Fortran MEs ( 1 ) : 2.2359s for 8192 events => throughput is 3.66E+03 events/s + +*** (1) EXECUTE MADEVENT_FORTRAN x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./madevent_fortran < /tmp/avalassi/input_smeftggtttt_x10_fortran > /tmp/avalassi/output_smeftggtttt_x10_fortran' + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 64/64 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 7.654e-07 [7.6542926582898148E-007] fbridge_mode=0 + [UNWEIGHT] Wrote 1679 events (found 1684 events) + [COUNTERS] PROGRAM TOTAL : 24.3444s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0268s + [COUNTERS] Fortran MEs ( 1 ) : 22.3176s for 81920 events => throughput is 3.67E+03 events/s *** (2-none) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -109,10 +134,10 @@ DEBUG: MEK processed 8192 events across 72 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 7.638e-07 [7.6381610362728588E-007] fbridge_mode=1 [UNWEIGHT] Wrote 230 events (found 851 events) - [COUNTERS] PROGRAM TOTAL : 2.8505s - [COUNTERS] Fortran Overhead ( 0 ) : 0.3633s - [COUNTERS] CudaCpp MEs ( 2 ) : 2.4822s for 8192 events => throughput is 3.30E+03 events/s - [COUNTERS] CudaCpp HEL ( 3 ) : 0.0050s + [COUNTERS] PROGRAM TOTAL : 2.7750s + [COUNTERS] Fortran Overhead ( 0 ) : 0.3558s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.4144s for 8192 events => throughput is 3.39E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0048s *** (2-none) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -122,15 +147,50 @@ OK! xsec from fortran (7.6381610362728588E-007) and cpp (7.6381610362728588E-007 OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-none) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.none_d_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_smeftggtttt_x10_cudacpp > /tmp/avalassi/output_smeftggtttt_x10_cudacpp' +DEBUG: MEK processed 81920 events across 72 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 64/64 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 7.654e-07 [7.6542926582898148E-007] fbridge_mode=1 + [UNWEIGHT] Wrote 1679 events (found 1684 events) + [COUNTERS] PROGRAM TOTAL : 25.9894s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0219s + [COUNTERS] CudaCpp MEs ( 2 ) : 23.9626s for 81920 events => throughput is 3.42E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0049s + +*** (2-none) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (7.6542926582898148E-007) and cpp (7.6542926582898148E-007) differ by less than 3E-14 (0.0) + +*** (2-none) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SMEFTSIM_TOPU3L_MWSCHEME_UFO_GG_TTXTTX_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+BRDHST/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.457369e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.557284e+03 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_SMEFTSIM_TOPU3L_MWSCHEME_UFO_GG_TTXTTX_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.441555e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.555640e+03 ) sec^-1 *** (2-sse4) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -154,9 +214,9 @@ DEBUG: MEK processed 8192 events across 72 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 7.638e-07 [7.6381610362728610E-007] fbridge_mode=1 [UNWEIGHT] Wrote 230 events (found 851 events) - [COUNTERS] PROGRAM TOTAL : 1.6655s - [COUNTERS] Fortran Overhead ( 0 ) : 0.3645s - [COUNTERS] CudaCpp MEs ( 2 ) : 1.2984s for 8192 events => throughput is 6.31E+03 events/s + [COUNTERS] PROGRAM TOTAL : 1.6246s + [COUNTERS] Fortran Overhead ( 0 ) : 0.3556s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.2662s for 8192 events => throughput is 6.47E+03 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0027s *** (2-sse4) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -167,15 +227,50 @@ OK! xsec from fortran (7.6381610362728588E-007) and cpp (7.6381610362728610E-007 OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-sse4) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.sse4_d_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_smeftggtttt_x10_cudacpp > /tmp/avalassi/output_smeftggtttt_x10_cudacpp' +DEBUG: MEK processed 81920 events across 72 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 64/64 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 7.654e-07 [7.6542926582898191E-007] fbridge_mode=1 + [UNWEIGHT] Wrote 1679 events (found 1684 events) + [COUNTERS] PROGRAM TOTAL : 14.5871s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0315s + [COUNTERS] CudaCpp MEs ( 2 ) : 12.5529s for 81920 events => throughput is 6.53E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0027s + +*** (2-sse4) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (7.6542926582898148E-007) and cpp (7.6542926582898191E-007) differ by less than 3E-14 (4.440892098500626e-16) + +*** (2-sse4) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SMEFTSIM_TOPU3L_MWSCHEME_UFO_GG_TTXTTX_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+BRDHST/sse4+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 6.514132e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 6.761070e+03 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_SMEFTSIM_TOPU3L_MWSCHEME_UFO_GG_TTXTTX_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 6.544925e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 6.650372e+03 ) sec^-1 *** (2-avx2) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -199,9 +294,9 @@ DEBUG: MEK processed 8192 events across 72 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 7.638e-07 [7.6381610362728588E-007] fbridge_mode=1 [UNWEIGHT] Wrote 230 events (found 851 events) - [COUNTERS] PROGRAM TOTAL : 0.9435s - [COUNTERS] Fortran Overhead ( 0 ) : 0.3668s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.5751s for 8192 events => throughput is 1.42E+04 events/s + [COUNTERS] PROGRAM TOTAL : 0.9149s + [COUNTERS] Fortran Overhead ( 0 ) : 0.3592s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5543s for 8192 events => throughput is 1.48E+04 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0015s *** (2-avx2) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -212,15 +307,50 @@ OK! xsec from fortran (7.6381610362728588E-007) and cpp (7.6381610362728588E-007 OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-avx2) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.avx2_d_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_smeftggtttt_x10_cudacpp > /tmp/avalassi/output_smeftggtttt_x10_cudacpp' +DEBUG: MEK processed 81920 events across 72 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 64/64 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 7.654e-07 [7.6542926582898201E-007] fbridge_mode=1 + [UNWEIGHT] Wrote 1679 events (found 1684 events) + [COUNTERS] PROGRAM TOTAL : 7.5864s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0278s + [COUNTERS] CudaCpp MEs ( 2 ) : 5.5571s for 81920 events => throughput is 1.47E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0015s + +*** (2-avx2) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (7.6542926582898148E-007) and cpp (7.6542926582898201E-007) differ by less than 3E-14 (6.661338147750939e-16) + +*** (2-avx2) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SMEFTSIM_TOPU3L_MWSCHEME_UFO_GG_TTXTTX_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+BRDHST/avx2+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.460459e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.481311e+04 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_SMEFTSIM_TOPU3L_MWSCHEME_UFO_GG_TTXTTX_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.466853e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.495361e+04 ) sec^-1 *** (2-512y) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -244,9 +374,9 @@ DEBUG: MEK processed 8192 events across 72 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 7.638e-07 [7.6381610362728588E-007] fbridge_mode=1 [UNWEIGHT] Wrote 230 events (found 851 events) - [COUNTERS] PROGRAM TOTAL : 0.8804s - [COUNTERS] Fortran Overhead ( 0 ) : 0.3647s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.5144s for 8192 events => throughput is 1.59E+04 events/s + [COUNTERS] PROGRAM TOTAL : 0.8703s + [COUNTERS] Fortran Overhead ( 0 ) : 0.3627s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5062s for 8192 events => throughput is 1.62E+04 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0014s *** (2-512y) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -257,15 +387,50 @@ OK! xsec from fortran (7.6381610362728588E-007) and cpp (7.6381610362728588E-007 OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-512y) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.512y_d_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_smeftggtttt_x10_cudacpp > /tmp/avalassi/output_smeftggtttt_x10_cudacpp' +DEBUG: MEK processed 81920 events across 72 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 64/64 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 7.654e-07 [7.6542926582898201E-007] fbridge_mode=1 + [UNWEIGHT] Wrote 1679 events (found 1684 events) + [COUNTERS] PROGRAM TOTAL : 7.0138s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0271s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.9852s for 81920 events => throughput is 1.64E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0014s + +*** (2-512y) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (7.6542926582898148E-007) and cpp (7.6542926582898201E-007) differ by less than 3E-14 (6.661338147750939e-16) + +*** (2-512y) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SMEFTSIM_TOPU3L_MWSCHEME_UFO_GG_TTXTTX_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+BRDHST/512y+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.641494e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.708405e+04 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_SMEFTSIM_TOPU3L_MWSCHEME_UFO_GG_TTXTTX_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.655223e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.694620e+04 ) sec^-1 *** (2-512z) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -289,9 +454,9 @@ DEBUG: MEK processed 8192 events across 72 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 7.638e-07 [7.6381610362728588E-007] fbridge_mode=1 [UNWEIGHT] Wrote 230 events (found 851 events) - [COUNTERS] PROGRAM TOTAL : 1.0440s - [COUNTERS] Fortran Overhead ( 0 ) : 0.3665s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.6757s for 8192 events => throughput is 1.21E+04 events/s + [COUNTERS] PROGRAM TOTAL : 0.9990s + [COUNTERS] Fortran Overhead ( 0 ) : 0.3585s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6387s for 8192 events => throughput is 1.28E+04 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0018s *** (2-512z) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -302,15 +467,50 @@ OK! xsec from fortran (7.6381610362728588E-007) and cpp (7.6381610362728588E-007 OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-512z) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.512z_d_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_smeftggtttt_x10_cudacpp > /tmp/avalassi/output_smeftggtttt_x10_cudacpp' +DEBUG: MEK processed 81920 events across 72 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 64/64 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 7.654e-07 [7.6542926582898201E-007] fbridge_mode=1 + [UNWEIGHT] Wrote 1679 events (found 1684 events) + [COUNTERS] PROGRAM TOTAL : 8.4156s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0169s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.3971s for 81920 events => throughput is 1.28E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0016s + +*** (2-512z) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (7.6542926582898148E-007) and cpp (7.6542926582898201E-007) differ by less than 3E-14 (6.661338147750939e-16) + +*** (2-512z) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SMEFTSIM_TOPU3L_MWSCHEME_UFO_GG_TTXTTX_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+BRDHST/512z+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.221115e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.269051e+04 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_SMEFTSIM_TOPU3L_MWSCHEME_UFO_GG_TTXTTX_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.225553e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.304837e+04 ) sec^-1 *** (3-cuda) EXECUTE MADEVENT_CUDA x1 (create events.lhe) *** -------------------- @@ -334,10 +534,10 @@ DEBUG: MEK processed 8192 events across 72 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 7.638e-07 [7.6381610362728578E-007] fbridge_mode=1 [UNWEIGHT] Wrote 230 events (found 851 events) - [COUNTERS] PROGRAM TOTAL : 0.8457s - [COUNTERS] Fortran Overhead ( 0 ) : 0.8061s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0199s for 8192 events => throughput is 4.13E+05 events/s - [COUNTERS] CudaCpp HEL ( 3 ) : 0.0198s + [COUNTERS] PROGRAM TOTAL : 0.8288s + [COUNTERS] Fortran Overhead ( 0 ) : 0.7895s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0196s for 8192 events => throughput is 4.17E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0197s *** (3-cuda) Compare MADEVENT_CUDA x1 xsec to MADEVENT_FORTRAN xsec *** @@ -347,45 +547,80 @@ OK! xsec from fortran (7.6381610362728588E-007) and cuda (7.6381610362728578E-00 OK! events.lhe.cuda.1 and events.lhe.ref.1 are identical +*** (3-cuda) EXECUTE MADEVENT_CUDA x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.cuda_d_inl0_hrd0/madevent_cuda < /tmp/avalassi/input_smeftggtttt_x10_cudacpp > /tmp/avalassi/output_smeftggtttt_x10_cudacpp' +DEBUG: MEK processed 81920 events across 72 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 64/64 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 7.654e-07 [7.6542926582898201E-007] fbridge_mode=1 + [UNWEIGHT] Wrote 1679 events (found 1684 events) + [COUNTERS] PROGRAM TOTAL : 2.6489s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4533s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1760s for 81920 events => throughput is 4.66E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0196s + +*** (3-cuda) Compare MADEVENT_CUDA x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (7.6542926582898148E-007) and cuda (7.6542926582898201E-007) differ by less than 3E-14 (6.661338147750939e-16) + +*** (3-cuda) Compare MADEVENT_CUDA x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cuda.10 and events.lhe.ref.10 are identical + *** EXECUTE GCHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SMEFTSIM_TOPU3L_MWSCHEME_UFO_GG_TTXTTX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 4.230611e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 4.244258e+05 ) sec^-1 *** EXECUTE GCHECK(8192) -p 256 32 1 *** Process = SIGMA_SMEFTSIM_TOPU3L_MWSCHEME_UFO_GG_TTXTTX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 4.541816e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 4.557216e+05 ) sec^-1 *** EXECUTE GCHECK(MAX) -p 16384 32 1 --bridge *** Process = SIGMA_SMEFTSIM_TOPU3L_MWSCHEME_UFO_GG_TTXTTX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 8.854537e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 8.727992e+05 ) sec^-1 *** EXECUTE GCHECK(MAX) -p 16384 32 1 *** Process = SIGMA_SMEFTSIM_TOPU3L_MWSCHEME_UFO_GG_TTXTTX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 9.229320e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 9.211409e+05 ) sec^-1 *** EXECUTE GCHECK(MAX128THR) -p 4096 128 1 --bridge *** Process = SIGMA_SMEFTSIM_TOPU3L_MWSCHEME_UFO_GG_TTXTTX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 8.859903e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 8.826370e+05 ) sec^-1 *** EXECUTE GCHECK(MAX128THR) -p 4096 128 1 *** Process = SIGMA_SMEFTSIM_TOPU3L_MWSCHEME_UFO_GG_TTXTTX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 9.225591e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 9.205506e+05 ) sec^-1 *** EXECUTE GCHECK(MAX8THR) -p 65536 8 1 --bridge *** Process = SIGMA_SMEFTSIM_TOPU3L_MWSCHEME_UFO_GG_TTXTTX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 8.850975e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 8.830521e+05 ) sec^-1 *** EXECUTE GCHECK(MAX8THR) -p 65536 8 1 *** Process = SIGMA_SMEFTSIM_TOPU3L_MWSCHEME_UFO_GG_TTXTTX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 2.687847e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.684747e+05 ) sec^-1 *** (3-hip) WARNING! SKIP MADEVENT_HIP (hip is not supported on this node) *** diff --git a/epochX/cudacpp/tmad/logs_smeftggtttt_mad/log_smeftggtttt_mad_f_inl0_hrd0.txt b/epochX/cudacpp/tmad/logs_smeftggtttt_mad/log_smeftggtttt_mad_f_inl0_hrd0.txt index 10c15cf9d1..77f09f6b6a 100644 --- a/epochX/cudacpp/tmad/logs_smeftggtttt_mad/log_smeftggtttt_mad_f_inl0_hrd0.txt +++ b/epochX/cudacpp/tmad/logs_smeftggtttt_mad/log_smeftggtttt_mad_f_inl0_hrd0.txt @@ -1,11 +1,11 @@ Working directory (build): /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/smeft_gg_tttt.mad/SubProcesses/P1_gg_ttxttx - make USEBUILDDIR=1 BACKEND=cuda + + make USEBUILDDIR=1 BACKEND=cppnone make USEBUILDDIR=1 BACKEND=cppsse4 - make USEBUILDDIR=1 BACKEND=cppavx2 make USEBUILDDIR=1 BACKEND=cpp512y @@ -13,10 +13,10 @@ make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/smeft_gg_tttt.mad/SubProcesses/P1_gg_ttxttx' make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/smeft_gg_tttt.mad/SubProcesses/P1_gg_ttxttx' make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/smeft_gg_tttt.mad/SubProcesses/P1_gg_ttxttx' +make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/smeft_gg_tttt.mad/SubProcesses/P1_gg_ttxttx' make USEBUILDDIR=1 BACKEND=cpp512z make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/smeft_gg_tttt.mad/SubProcesses/P1_gg_ttxttx' -make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/smeft_gg_tttt.mad/SubProcesses/P1_gg_ttxttx' make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/smeft_gg_tttt.mad/SubProcesses/P1_gg_ttxttx' make[1]: Nothing to be done for 'all'. @@ -32,7 +32,7 @@ make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ OMP_NUM_THREADS= -DATE: 2024-10-06_10:58:07 +DATE: 2024-10-06_18:15:29 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: Working directory (run): /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/smeft_gg_tttt.mad/SubProcesses/P1_gg_ttxttx @@ -58,9 +58,9 @@ Executing ' ./madevent_fortran < /tmp/avalassi/input_smeftggtttt_x1_fortran > /t [XSECTION] ChannelId = 1 [XSECTION] Cross section = 7.638e-07 [7.6381610362728588E-007] fbridge_mode=0 [UNWEIGHT] Wrote 1 events (found 902 events) - [COUNTERS] PROGRAM TOTAL : 2.6755s - [COUNTERS] Fortran Overhead ( 0 ) : 0.3603s - [COUNTERS] Fortran MEs ( 1 ) : 2.3152s for 8192 events => throughput is 3.54E+03 events/s + [COUNTERS] PROGRAM TOTAL : 2.5806s + [COUNTERS] Fortran Overhead ( 0 ) : 0.3458s + [COUNTERS] Fortran MEs ( 1 ) : 2.2349s for 8192 events => throughput is 3.67E+03 events/s *** (1) EXECUTE MADEVENT_FORTRAN x1 (create events.lhe) *** -------------------- @@ -83,9 +83,34 @@ Executing ' ./madevent_fortran < /tmp/avalassi/input_smeftggtttt_x1_fortran > /t [XSECTION] ChannelId = 1 [XSECTION] Cross section = 7.638e-07 [7.6381610362728588E-007] fbridge_mode=0 [UNWEIGHT] Wrote 230 events (found 851 events) - [COUNTERS] PROGRAM TOTAL : 2.6754s - [COUNTERS] Fortran Overhead ( 0 ) : 0.3698s - [COUNTERS] Fortran MEs ( 1 ) : 2.3056s for 8192 events => throughput is 3.55E+03 events/s + [COUNTERS] PROGRAM TOTAL : 2.6116s + [COUNTERS] Fortran Overhead ( 0 ) : 0.3560s + [COUNTERS] Fortran MEs ( 1 ) : 2.2556s for 8192 events => throughput is 3.63E+03 events/s + +*** (1) EXECUTE MADEVENT_FORTRAN x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./madevent_fortran < /tmp/avalassi/input_smeftggtttt_x10_fortran > /tmp/avalassi/output_smeftggtttt_x10_fortran' + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 64/64 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 7.654e-07 [7.6542926582898148E-007] fbridge_mode=0 + [UNWEIGHT] Wrote 1679 events (found 1684 events) + [COUNTERS] PROGRAM TOTAL : 24.2905s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0100s + [COUNTERS] Fortran MEs ( 1 ) : 22.2805s for 81920 events => throughput is 3.68E+03 events/s *** (2-none) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -109,10 +134,10 @@ DEBUG: MEK processed 8192 events across 72 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 7.638e-07 [7.6381686438954397E-007] fbridge_mode=1 [UNWEIGHT] Wrote 230 events (found 851 events) - [COUNTERS] PROGRAM TOTAL : 2.8067s - [COUNTERS] Fortran Overhead ( 0 ) : 0.3659s - [COUNTERS] CudaCpp MEs ( 2 ) : 2.4358s for 8192 events => throughput is 3.36E+03 events/s - [COUNTERS] CudaCpp HEL ( 3 ) : 0.0050s + [COUNTERS] PROGRAM TOTAL : 2.7259s + [COUNTERS] Fortran Overhead ( 0 ) : 0.3562s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.3648s for 8192 events => throughput is 3.46E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0049s *** (2-none) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -122,15 +147,50 @@ OK! xsec from fortran (7.6381610362728588E-007) and cpp (7.6381686438954397E-007 OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-none) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.none_f_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_smeftggtttt_x10_cudacpp > /tmp/avalassi/output_smeftggtttt_x10_cudacpp' +DEBUG: MEK processed 81920 events across 72 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 64/64 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 7.654e-07 [7.6542978900095690E-007] fbridge_mode=1 + [UNWEIGHT] Wrote 1679 events (found 1684 events) + [COUNTERS] PROGRAM TOTAL : 25.5851s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0236s + [COUNTERS] CudaCpp MEs ( 2 ) : 23.5567s for 81920 events => throughput is 3.48E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0048s + +*** (2-none) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (7.6542926582898148E-007) and cpp (7.6542978900095690E-007) differ by less than 4E-4 (6.835014008110818e-07) + +*** (2-none) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SMEFTSIM_TOPU3L_MWSCHEME_UFO_GG_TTXTTX_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+BRDHST/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.485505e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.599026e+03 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_SMEFTSIM_TOPU3L_MWSCHEME_UFO_GG_TTXTTX_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.473644e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.606431e+03 ) sec^-1 *** (2-sse4) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -154,10 +214,10 @@ DEBUG: MEK processed 8192 events across 72 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 7.638e-07 [7.6381671483253128E-007] fbridge_mode=1 [UNWEIGHT] Wrote 230 events (found 851 events) - [COUNTERS] PROGRAM TOTAL : 1.0546s - [COUNTERS] Fortran Overhead ( 0 ) : 0.3688s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.6842s for 8192 events => throughput is 1.20E+04 events/s - [COUNTERS] CudaCpp HEL ( 3 ) : 0.0015s + [COUNTERS] PROGRAM TOTAL : 1.0097s + [COUNTERS] Fortran Overhead ( 0 ) : 0.3560s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6524s for 8192 events => throughput is 1.26E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0014s *** (2-sse4) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -167,15 +227,50 @@ OK! xsec from fortran (7.6381610362728588E-007) and cpp (7.6381671483253128E-007 OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-sse4) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.sse4_f_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_smeftggtttt_x10_cudacpp > /tmp/avalassi/output_smeftggtttt_x10_cudacpp' +DEBUG: MEK processed 81920 events across 72 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 64/64 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 7.654e-07 [7.6542962735029303E-007] fbridge_mode=1 + [UNWEIGHT] Wrote 1679 events (found 1684 events) + [COUNTERS] PROGRAM TOTAL : 8.5983s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0295s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.5674s for 81920 events => throughput is 1.25E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0014s + +*** (2-sse4) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (7.6542926582898148E-007) and cpp (7.6542962735029303E-007) differ by less than 4E-4 (4.7231184874263477e-07) + +*** (2-sse4) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SMEFTSIM_TOPU3L_MWSCHEME_UFO_GG_TTXTTX_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+BRDHST/sse4+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.232148e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.282117e+04 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_SMEFTSIM_TOPU3L_MWSCHEME_UFO_GG_TTXTTX_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.242719e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.285869e+04 ) sec^-1 *** (2-avx2) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -199,9 +294,9 @@ DEBUG: MEK processed 8192 events across 72 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 7.638e-07 [7.6381672175647812E-007] fbridge_mode=1 [UNWEIGHT] Wrote 230 events (found 851 events) - [COUNTERS] PROGRAM TOTAL : 0.6626s - [COUNTERS] Fortran Overhead ( 0 ) : 0.3670s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.2947s for 8192 events => throughput is 2.78E+04 events/s + [COUNTERS] PROGRAM TOTAL : 0.6393s + [COUNTERS] Fortran Overhead ( 0 ) : 0.3575s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2809s for 8192 events => throughput is 2.92E+04 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0009s *** (2-avx2) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -212,15 +307,50 @@ OK! xsec from fortran (7.6381610362728588E-007) and cpp (7.6381672175647812E-007 OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-avx2) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.avx2_f_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_smeftggtttt_x10_cudacpp > /tmp/avalassi/output_smeftggtttt_x10_cudacpp' +DEBUG: MEK processed 81920 events across 72 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 64/64 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 7.654e-07 [7.6542989697352719E-007] fbridge_mode=1 + [UNWEIGHT] Wrote 1679 events (found 1684 events) + [COUNTERS] PROGRAM TOTAL : 4.8764s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0252s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.8503s for 81920 events => throughput is 2.87E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0009s + +*** (2-avx2) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (7.6542926582898148E-007) and cpp (7.6542989697352719E-007) differ by less than 4E-4 (8.245628615455303e-07) + +*** (2-avx2) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SMEFTSIM_TOPU3L_MWSCHEME_UFO_GG_TTXTTX_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+BRDHST/avx2+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 2.866680e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.947300e+04 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_SMEFTSIM_TOPU3L_MWSCHEME_UFO_GG_TTXTTX_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 2.814611e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.970579e+04 ) sec^-1 *** (2-512y) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -244,9 +374,9 @@ DEBUG: MEK processed 8192 events across 72 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 7.638e-07 [7.6381672175647812E-007] fbridge_mode=1 [UNWEIGHT] Wrote 230 events (found 851 events) - [COUNTERS] PROGRAM TOTAL : 0.6345s - [COUNTERS] Fortran Overhead ( 0 ) : 0.3672s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.2664s for 8192 events => throughput is 3.07E+04 events/s + [COUNTERS] PROGRAM TOTAL : 0.6113s + [COUNTERS] Fortran Overhead ( 0 ) : 0.3531s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2573s for 8192 events => throughput is 3.18E+04 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0008s *** (2-512y) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -257,15 +387,50 @@ OK! xsec from fortran (7.6381610362728588E-007) and cpp (7.6381672175647812E-007 OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-512y) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.512y_f_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_smeftggtttt_x10_cudacpp > /tmp/avalassi/output_smeftggtttt_x10_cudacpp' +DEBUG: MEK processed 81920 events across 72 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 64/64 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 7.654e-07 [7.6542989697352719E-007] fbridge_mode=1 + [UNWEIGHT] Wrote 1679 events (found 1684 events) + [COUNTERS] PROGRAM TOTAL : 4.5792s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0162s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.5622s for 81920 events => throughput is 3.20E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0008s + +*** (2-512y) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (7.6542926582898148E-007) and cpp (7.6542989697352719E-007) differ by less than 4E-4 (8.245628615455303e-07) + +*** (2-512y) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SMEFTSIM_TOPU3L_MWSCHEME_UFO_GG_TTXTTX_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+BRDHST/512y+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.183014e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.285965e+04 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_SMEFTSIM_TOPU3L_MWSCHEME_UFO_GG_TTXTTX_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.199503e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.270360e+04 ) sec^-1 *** (2-512z) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -289,10 +454,10 @@ DEBUG: MEK processed 8192 events across 72 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 7.638e-07 [7.6381686320975603E-007] fbridge_mode=1 [UNWEIGHT] Wrote 230 events (found 851 events) - [COUNTERS] PROGRAM TOTAL : 0.7045s - [COUNTERS] Fortran Overhead ( 0 ) : 0.3656s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.3379s for 8192 events => throughput is 2.42E+04 events/s - [COUNTERS] CudaCpp HEL ( 3 ) : 0.0010s + [COUNTERS] PROGRAM TOTAL : 0.6754s + [COUNTERS] Fortran Overhead ( 0 ) : 0.3532s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.3213s for 8192 events => throughput is 2.55E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0009s *** (2-512z) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -302,15 +467,50 @@ OK! xsec from fortran (7.6381610362728588E-007) and cpp (7.6381686320975603E-007 OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-512z) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.512z_f_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_smeftggtttt_x10_cudacpp > /tmp/avalassi/output_smeftggtttt_x10_cudacpp' +DEBUG: MEK processed 81920 events across 72 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 64/64 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 7.654e-07 [7.6543004237976207E-007] fbridge_mode=1 + [UNWEIGHT] Wrote 1679 events (found 1684 events) + [COUNTERS] PROGRAM TOTAL : 5.2474s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0268s + [COUNTERS] CudaCpp MEs ( 2 ) : 3.2197s for 81920 events => throughput is 2.54E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0009s + +*** (2-512z) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (7.6542926582898148E-007) and cpp (7.6543004237976207E-007) differ by less than 4E-4 (1.014529774634454e-06) + +*** (2-512z) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SMEFTSIM_TOPU3L_MWSCHEME_UFO_GG_TTXTTX_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+BRDHST/512z+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 2.460974e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.586234e+04 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_SMEFTSIM_TOPU3L_MWSCHEME_UFO_GG_TTXTTX_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 2.436294e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.606911e+04 ) sec^-1 *** (3-cuda) EXECUTE MADEVENT_CUDA x1 (create events.lhe) *** -------------------- @@ -334,9 +534,9 @@ DEBUG: MEK processed 8192 events across 72 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 7.638e-07 [7.6381711031958629E-007] fbridge_mode=1 [UNWEIGHT] Wrote 230 events (found 851 events) - [COUNTERS] PROGRAM TOTAL : 0.8419s - [COUNTERS] Fortran Overhead ( 0 ) : 0.8049s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0197s for 8192 events => throughput is 4.15E+05 events/s + [COUNTERS] PROGRAM TOTAL : 0.8330s + [COUNTERS] Fortran Overhead ( 0 ) : 0.7963s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0195s for 8192 events => throughput is 4.20E+05 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0172s *** (3-cuda) Compare MADEVENT_CUDA x1 xsec to MADEVENT_FORTRAN xsec *** @@ -347,45 +547,80 @@ OK! xsec from fortran (7.6381610362728588E-007) and cuda (7.6381711031958629E-00 OK! events.lhe.cuda.1 and events.lhe.ref.1 are identical +*** (3-cuda) EXECUTE MADEVENT_CUDA x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.cuda_f_inl0_hrd0/madevent_cuda < /tmp/avalassi/input_smeftggtttt_x10_cudacpp > /tmp/avalassi/output_smeftggtttt_x10_cudacpp' +DEBUG: MEK processed 81920 events across 72 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 64/64 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 7.654e-07 [7.6543026921346333E-007] fbridge_mode=1 + [UNWEIGHT] Wrote 1679 events (found 1684 events) + [COUNTERS] PROGRAM TOTAL : 2.6150s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4387s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1592s for 81920 events => throughput is 5.15E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0171s + +*** (3-cuda) Compare MADEVENT_CUDA x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (7.6542926582898148E-007) and cuda (7.6543026921346333E-007) differ by less than 4E-4 (1.3108781262705094e-06) + +*** (3-cuda) Compare MADEVENT_CUDA x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cuda.10 and events.lhe.ref.10 are identical + *** EXECUTE GCHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SMEFTSIM_TOPU3L_MWSCHEME_UFO_GG_TTXTTX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 4.233915e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 4.226927e+05 ) sec^-1 *** EXECUTE GCHECK(8192) -p 256 32 1 *** Process = SIGMA_SMEFTSIM_TOPU3L_MWSCHEME_UFO_GG_TTXTTX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 4.454452e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 4.470299e+05 ) sec^-1 *** EXECUTE GCHECK(MAX) -p 16384 32 1 --bridge *** Process = SIGMA_SMEFTSIM_TOPU3L_MWSCHEME_UFO_GG_TTXTTX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.300238e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.314881e+06 ) sec^-1 *** EXECUTE GCHECK(MAX) -p 16384 32 1 *** Process = SIGMA_SMEFTSIM_TOPU3L_MWSCHEME_UFO_GG_TTXTTX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.323216e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.339500e+06 ) sec^-1 *** EXECUTE GCHECK(MAX128THR) -p 4096 128 1 --bridge *** Process = SIGMA_SMEFTSIM_TOPU3L_MWSCHEME_UFO_GG_TTXTTX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.294935e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.314839e+06 ) sec^-1 *** EXECUTE GCHECK(MAX128THR) -p 4096 128 1 *** Process = SIGMA_SMEFTSIM_TOPU3L_MWSCHEME_UFO_GG_TTXTTX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.322990e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.336913e+06 ) sec^-1 *** EXECUTE GCHECK(MAX8THR) -p 65536 8 1 --bridge *** Process = SIGMA_SMEFTSIM_TOPU3L_MWSCHEME_UFO_GG_TTXTTX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.292471e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.307875e+06 ) sec^-1 *** EXECUTE GCHECK(MAX8THR) -p 65536 8 1 *** Process = SIGMA_SMEFTSIM_TOPU3L_MWSCHEME_UFO_GG_TTXTTX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.654983e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.641571e+05 ) sec^-1 *** (3-hip) WARNING! SKIP MADEVENT_HIP (hip is not supported on this node) *** diff --git a/epochX/cudacpp/tmad/logs_smeftggtttt_mad/log_smeftggtttt_mad_m_inl0_hrd0.txt b/epochX/cudacpp/tmad/logs_smeftggtttt_mad/log_smeftggtttt_mad_m_inl0_hrd0.txt index 9cff3d3d2c..0efe905d14 100644 --- a/epochX/cudacpp/tmad/logs_smeftggtttt_mad/log_smeftggtttt_mad_m_inl0_hrd0.txt +++ b/epochX/cudacpp/tmad/logs_smeftggtttt_mad/log_smeftggtttt_mad_m_inl0_hrd0.txt @@ -1,22 +1,22 @@ Working directory (build): /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/smeft_gg_tttt.mad/SubProcesses/P1_gg_ttxttx - make USEBUILDDIR=1 BACKEND=cuda + make USEBUILDDIR=1 BACKEND=cppnone make USEBUILDDIR=1 BACKEND=cppsse4 -make USEBUILDDIR=1 BACKEND=cppavx2 -make USEBUILDDIR=1 BACKEND=cpp512y -make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/smeft_gg_tttt.mad/SubProcesses/P1_gg_ttxttx' -make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/smeft_gg_tttt.mad/SubProcesses/P1_gg_ttxttx' +make USEBUILDDIR=1 BACKEND=cppavx2 make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/smeft_gg_tttt.mad/SubProcesses/P1_gg_ttxttx' +make USEBUILDDIR=1 BACKEND=cpp512y make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/smeft_gg_tttt.mad/SubProcesses/P1_gg_ttxttx' make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/smeft_gg_tttt.mad/SubProcesses/P1_gg_ttxttx' make USEBUILDDIR=1 BACKEND=cpp512z make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/smeft_gg_tttt.mad/SubProcesses/P1_gg_ttxttx' +make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/smeft_gg_tttt.mad/SubProcesses/P1_gg_ttxttx' +make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/smeft_gg_tttt.mad/SubProcesses/P1_gg_ttxttx' make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/smeft_gg_tttt.mad/SubProcesses/P1_gg_ttxttx' make[1]: Nothing to be done for 'all'. @@ -32,7 +32,7 @@ make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ OMP_NUM_THREADS= -DATE: 2024-10-06_10:57:22 +DATE: 2024-10-06_18:13:13 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: Working directory (run): /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/smeft_gg_tttt.mad/SubProcesses/P1_gg_ttxttx @@ -58,9 +58,9 @@ Executing ' ./madevent_fortran < /tmp/avalassi/input_smeftggtttt_x1_fortran > /t [XSECTION] ChannelId = 1 [XSECTION] Cross section = 7.638e-07 [7.6381610362728588E-007] fbridge_mode=0 [UNWEIGHT] Wrote 1 events (found 902 events) - [COUNTERS] PROGRAM TOTAL : 2.6661s - [COUNTERS] Fortran Overhead ( 0 ) : 0.3588s - [COUNTERS] Fortran MEs ( 1 ) : 2.3072s for 8192 events => throughput is 3.55E+03 events/s + [COUNTERS] PROGRAM TOTAL : 2.5783s + [COUNTERS] Fortran Overhead ( 0 ) : 0.3506s + [COUNTERS] Fortran MEs ( 1 ) : 2.2277s for 8192 events => throughput is 3.68E+03 events/s *** (1) EXECUTE MADEVENT_FORTRAN x1 (create events.lhe) *** -------------------- @@ -83,9 +83,34 @@ Executing ' ./madevent_fortran < /tmp/avalassi/input_smeftggtttt_x1_fortran > /t [XSECTION] ChannelId = 1 [XSECTION] Cross section = 7.638e-07 [7.6381610362728588E-007] fbridge_mode=0 [UNWEIGHT] Wrote 230 events (found 851 events) - [COUNTERS] PROGRAM TOTAL : 2.6664s - [COUNTERS] Fortran Overhead ( 0 ) : 0.3632s - [COUNTERS] Fortran MEs ( 1 ) : 2.3031s for 8192 events => throughput is 3.56E+03 events/s + [COUNTERS] PROGRAM TOTAL : 2.5911s + [COUNTERS] Fortran Overhead ( 0 ) : 0.3581s + [COUNTERS] Fortran MEs ( 1 ) : 2.2329s for 8192 events => throughput is 3.67E+03 events/s + +*** (1) EXECUTE MADEVENT_FORTRAN x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./madevent_fortran < /tmp/avalassi/input_smeftggtttt_x10_fortran > /tmp/avalassi/output_smeftggtttt_x10_fortran' + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 64/64 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 7.654e-07 [7.6542926582898148E-007] fbridge_mode=0 + [UNWEIGHT] Wrote 1679 events (found 1684 events) + [COUNTERS] PROGRAM TOTAL : 24.3887s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0276s + [COUNTERS] Fortran MEs ( 1 ) : 22.3611s for 81920 events => throughput is 3.66E+03 events/s *** (2-none) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -109,10 +134,10 @@ DEBUG: MEK processed 8192 events across 72 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 7.638e-07 [7.6381608764955655E-007] fbridge_mode=1 [UNWEIGHT] Wrote 230 events (found 851 events) - [COUNTERS] PROGRAM TOTAL : 2.8757s - [COUNTERS] Fortran Overhead ( 0 ) : 0.3651s - [COUNTERS] CudaCpp MEs ( 2 ) : 2.5054s for 8192 events => throughput is 3.27E+03 events/s - [COUNTERS] CudaCpp HEL ( 3 ) : 0.0052s + [COUNTERS] PROGRAM TOTAL : 2.8003s + [COUNTERS] Fortran Overhead ( 0 ) : 0.3556s + [COUNTERS] CudaCpp MEs ( 2 ) : 2.4399s for 8192 events => throughput is 3.36E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0048s *** (2-none) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -122,15 +147,50 @@ OK! xsec from fortran (7.6381610362728588E-007) and cpp (7.6381608764955655E-007 OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-none) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.none_m_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_smeftggtttt_x10_cudacpp > /tmp/avalassi/output_smeftggtttt_x10_cudacpp' +DEBUG: MEK processed 81920 events across 72 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 64/64 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 7.654e-07 [7.6542925018181681E-007] fbridge_mode=1 + [UNWEIGHT] Wrote 1679 events (found 1684 events) + [COUNTERS] PROGRAM TOTAL : 26.3208s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0277s + [COUNTERS] CudaCpp MEs ( 2 ) : 24.2882s for 81920 events => throughput is 3.37E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0049s + +*** (2-none) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (7.6542926582898148E-007) and cpp (7.6542925018181681E-007) differ by less than 2E-4 (2.044233915476923e-08) + +*** (2-none) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SMEFTSIM_TOPU3L_MWSCHEME_UFO_GG_TTXTTX_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+BRDHST/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.427512e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.547575e+03 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_SMEFTSIM_TOPU3L_MWSCHEME_UFO_GG_TTXTTX_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.426484e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.547035e+03 ) sec^-1 *** (2-sse4) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -154,10 +214,10 @@ DEBUG: MEK processed 8192 events across 72 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 7.638e-07 [7.6381608686521600E-007] fbridge_mode=1 [UNWEIGHT] Wrote 230 events (found 851 events) - [COUNTERS] PROGRAM TOTAL : 1.6394s - [COUNTERS] Fortran Overhead ( 0 ) : 0.3662s - [COUNTERS] CudaCpp MEs ( 2 ) : 1.2706s for 8192 events => throughput is 6.45E+03 events/s - [COUNTERS] CudaCpp HEL ( 3 ) : 0.0027s + [COUNTERS] PROGRAM TOTAL : 1.6105s + [COUNTERS] Fortran Overhead ( 0 ) : 0.3596s + [COUNTERS] CudaCpp MEs ( 2 ) : 1.2484s for 8192 events => throughput is 6.56E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0026s *** (2-sse4) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -167,15 +227,50 @@ OK! xsec from fortran (7.6381610362728588E-007) and cpp (7.6381608686521600E-007 OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-sse4) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.sse4_m_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_smeftggtttt_x10_cudacpp > /tmp/avalassi/output_smeftggtttt_x10_cudacpp' +DEBUG: MEK processed 81920 events across 72 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 64/64 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 7.654e-07 [7.6542924921991264E-007] fbridge_mode=1 + [UNWEIGHT] Wrote 1679 events (found 1684 events) + [COUNTERS] PROGRAM TOTAL : 14.4021s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0285s + [COUNTERS] CudaCpp MEs ( 2 ) : 12.3711s for 81920 events => throughput is 6.62E+03 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0026s + +*** (2-sse4) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (7.6542926582898148E-007) and cpp (7.6542924921991264E-007) differ by less than 2E-4 (2.1699025132271288e-08) + +*** (2-sse4) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SMEFTSIM_TOPU3L_MWSCHEME_UFO_GG_TTXTTX_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+BRDHST/sse4+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 6.733385e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 6.830689e+03 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_SMEFTSIM_TOPU3L_MWSCHEME_UFO_GG_TTXTTX_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 6.780255e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 6.942570e+03 ) sec^-1 *** (2-avx2) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -199,10 +294,10 @@ DEBUG: MEK processed 8192 events across 72 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 7.638e-07 [7.6381608826200266E-007] fbridge_mode=1 [UNWEIGHT] Wrote 230 events (found 851 events) - [COUNTERS] PROGRAM TOTAL : 0.9411s - [COUNTERS] Fortran Overhead ( 0 ) : 0.3649s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.5747s for 8192 events => throughput is 1.43E+04 events/s - [COUNTERS] CudaCpp HEL ( 3 ) : 0.0015s + [COUNTERS] PROGRAM TOTAL : 0.9059s + [COUNTERS] Fortran Overhead ( 0 ) : 0.3543s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.5502s for 8192 events => throughput is 1.49E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0014s *** (2-avx2) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -212,15 +307,50 @@ OK! xsec from fortran (7.6381610362728588E-007) and cpp (7.6381608826200266E-007 OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-avx2) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.avx2_m_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_smeftggtttt_x10_cudacpp > /tmp/avalassi/output_smeftggtttt_x10_cudacpp' +DEBUG: MEK processed 81920 events across 72 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 64/64 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 7.654e-07 [7.6542925056010437E-007] fbridge_mode=1 + [UNWEIGHT] Wrote 1679 events (found 1684 events) + [COUNTERS] PROGRAM TOTAL : 7.5332s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0208s + [COUNTERS] CudaCpp MEs ( 2 ) : 5.5109s for 81920 events => throughput is 1.49E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0015s + +*** (2-avx2) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (7.6542926582898148E-007) and cpp (7.6542925056010437E-007) differ by less than 2E-4 (1.994812293126813e-08) + +*** (2-avx2) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SMEFTSIM_TOPU3L_MWSCHEME_UFO_GG_TTXTTX_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+BRDHST/avx2+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.446717e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.514230e+04 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_SMEFTSIM_TOPU3L_MWSCHEME_UFO_GG_TTXTTX_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.473262e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.491905e+04 ) sec^-1 *** (2-512y) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -244,9 +374,9 @@ DEBUG: MEK processed 8192 events across 72 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 7.638e-07 [7.6381608826200266E-007] fbridge_mode=1 [UNWEIGHT] Wrote 230 events (found 851 events) - [COUNTERS] PROGRAM TOTAL : 0.8685s - [COUNTERS] Fortran Overhead ( 0 ) : 0.3656s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.5015s for 8192 events => throughput is 1.63E+04 events/s + [COUNTERS] PROGRAM TOTAL : 0.8394s + [COUNTERS] Fortran Overhead ( 0 ) : 0.3535s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4845s for 8192 events => throughput is 1.69E+04 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0014s *** (2-512y) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -257,15 +387,50 @@ OK! xsec from fortran (7.6381610362728588E-007) and cpp (7.6381608826200266E-007 OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-512y) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.512y_m_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_smeftggtttt_x10_cudacpp > /tmp/avalassi/output_smeftggtttt_x10_cudacpp' +DEBUG: MEK processed 81920 events across 72 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 64/64 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 7.654e-07 [7.6542925056010437E-007] fbridge_mode=1 + [UNWEIGHT] Wrote 1679 events (found 1684 events) + [COUNTERS] PROGRAM TOTAL : 6.8748s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0145s + [COUNTERS] CudaCpp MEs ( 2 ) : 4.8589s for 81920 events => throughput is 1.69E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0014s + +*** (2-512y) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (7.6542926582898148E-007) and cpp (7.6542925056010437E-007) differ by less than 2E-4 (1.994812293126813e-08) + +*** (2-512y) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SMEFTSIM_TOPU3L_MWSCHEME_UFO_GG_TTXTTX_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+BRDHST/512y+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.681650e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.756015e+04 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_SMEFTSIM_TOPU3L_MWSCHEME_UFO_GG_TTXTTX_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.668117e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.738512e+04 ) sec^-1 *** (2-512z) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -289,9 +454,9 @@ DEBUG: MEK processed 8192 events across 72 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 7.638e-07 [7.6381608826200266E-007] fbridge_mode=1 [UNWEIGHT] Wrote 230 events (found 851 events) - [COUNTERS] PROGRAM TOTAL : 1.0574s - [COUNTERS] Fortran Overhead ( 0 ) : 0.3699s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.6857s for 8192 events => throughput is 1.19E+04 events/s + [COUNTERS] PROGRAM TOTAL : 1.0024s + [COUNTERS] Fortran Overhead ( 0 ) : 0.3557s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.6450s for 8192 events => throughput is 1.27E+04 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0018s *** (2-512z) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -302,15 +467,50 @@ OK! xsec from fortran (7.6381610362728588E-007) and cpp (7.6381608826200266E-007 OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-512z) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.512z_m_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_smeftggtttt_x10_cudacpp > /tmp/avalassi/output_smeftggtttt_x10_cudacpp' +DEBUG: MEK processed 81920 events across 72 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 64/64 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 7.654e-07 [7.6542925056010437E-007] fbridge_mode=1 + [UNWEIGHT] Wrote 1679 events (found 1684 events) + [COUNTERS] PROGRAM TOTAL : 8.5513s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0357s + [COUNTERS] CudaCpp MEs ( 2 ) : 6.5138s for 81920 events => throughput is 1.26E+04 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0018s + +*** (2-512z) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (7.6542926582898148E-007) and cpp (7.6542925056010437E-007) differ by less than 2E-4 (1.994812293126813e-08) + +*** (2-512z) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SMEFTSIM_TOPU3L_MWSCHEME_UFO_GG_TTXTTX_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+BRDHST/512z+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.232369e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.274766e+04 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_SMEFTSIM_TOPU3L_MWSCHEME_UFO_GG_TTXTTX_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.216790e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.289107e+04 ) sec^-1 *** (3-cuda) EXECUTE MADEVENT_CUDA x1 (create events.lhe) *** -------------------- @@ -334,10 +534,10 @@ DEBUG: MEK processed 8192 events across 72 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 7.638e-07 [7.6381610372590318E-007] fbridge_mode=1 [UNWEIGHT] Wrote 230 events (found 851 events) - [COUNTERS] PROGRAM TOTAL : 0.8397s - [COUNTERS] Fortran Overhead ( 0 ) : 0.8000s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0198s for 8192 events => throughput is 4.13E+05 events/s - [COUNTERS] CudaCpp HEL ( 3 ) : 0.0198s + [COUNTERS] PROGRAM TOTAL : 0.8320s + [COUNTERS] Fortran Overhead ( 0 ) : 0.7926s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0197s for 8192 events => throughput is 4.15E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0197s *** (3-cuda) Compare MADEVENT_CUDA x1 xsec to MADEVENT_FORTRAN xsec *** @@ -347,45 +547,80 @@ OK! xsec from fortran (7.6381610362728588E-007) and cuda (7.6381610372590318E-00 OK! events.lhe.cuda.1 and events.lhe.ref.1 are identical +*** (3-cuda) EXECUTE MADEVENT_CUDA x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.cuda_m_inl0_hrd0/madevent_cuda < /tmp/avalassi/input_smeftggtttt_x10_cudacpp > /tmp/avalassi/output_smeftggtttt_x10_cudacpp' +DEBUG: MEK processed 81920 events across 72 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 64/64 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 7.654e-07 [7.6542926581386226E-007] fbridge_mode=1 + [UNWEIGHT] Wrote 1679 events (found 1684 events) + [COUNTERS] PROGRAM TOTAL : 2.6486s + [COUNTERS] Fortran Overhead ( 0 ) : 2.4524s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1765s for 81920 events => throughput is 4.64E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0198s + +*** (3-cuda) Compare MADEVENT_CUDA x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (7.6542926582898148E-007) and cuda (7.6542926581386226E-007) differ by less than 2E-4 (1.9752643964920935e-11) + +*** (3-cuda) Compare MADEVENT_CUDA x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cuda.10 and events.lhe.ref.10 are identical + *** EXECUTE GCHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_SMEFTSIM_TOPU3L_MWSCHEME_UFO_GG_TTXTTX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:MIX+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 4.219575e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 4.231112e+05 ) sec^-1 *** EXECUTE GCHECK(8192) -p 256 32 1 *** Process = SIGMA_SMEFTSIM_TOPU3L_MWSCHEME_UFO_GG_TTXTTX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:MIX+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 4.527801e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 4.555311e+05 ) sec^-1 *** EXECUTE GCHECK(MAX) -p 16384 32 1 --bridge *** Process = SIGMA_SMEFTSIM_TOPU3L_MWSCHEME_UFO_GG_TTXTTX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:MIX+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 8.836972e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 8.739908e+05 ) sec^-1 *** EXECUTE GCHECK(MAX) -p 16384 32 1 *** Process = SIGMA_SMEFTSIM_TOPU3L_MWSCHEME_UFO_GG_TTXTTX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:MIX+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 9.176072e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 9.178500e+05 ) sec^-1 *** EXECUTE GCHECK(MAX128THR) -p 4096 128 1 --bridge *** Process = SIGMA_SMEFTSIM_TOPU3L_MWSCHEME_UFO_GG_TTXTTX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:MIX+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 8.835271e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 8.804483e+05 ) sec^-1 *** EXECUTE GCHECK(MAX128THR) -p 4096 128 1 *** Process = SIGMA_SMEFTSIM_TOPU3L_MWSCHEME_UFO_GG_TTXTTX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:MIX+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 9.206917e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 9.205758e+05 ) sec^-1 *** EXECUTE GCHECK(MAX8THR) -p 65536 8 1 --bridge *** Process = SIGMA_SMEFTSIM_TOPU3L_MWSCHEME_UFO_GG_TTXTTX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:MIX+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 8.823749e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 8.805286e+05 ) sec^-1 *** EXECUTE GCHECK(MAX8THR) -p 65536 8 1 *** Process = SIGMA_SMEFTSIM_TOPU3L_MWSCHEME_UFO_GG_TTXTTX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:MIX+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 2.671807e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.671942e+05 ) sec^-1 *** (3-hip) WARNING! SKIP MADEVENT_HIP (hip is not supported on this node) *** diff --git a/epochX/cudacpp/tmad/logs_susyggt1t1_mad/log_susyggt1t1_mad_d_inl0_hrd0.txt b/epochX/cudacpp/tmad/logs_susyggt1t1_mad/log_susyggt1t1_mad_d_inl0_hrd0.txt index f18eaf3551..835e510ba3 100644 --- a/epochX/cudacpp/tmad/logs_susyggt1t1_mad/log_susyggt1t1_mad_d_inl0_hrd0.txt +++ b/epochX/cudacpp/tmad/logs_susyggt1t1_mad/log_susyggt1t1_mad_d_inl0_hrd0.txt @@ -2,12 +2,12 @@ Working directory (build): /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/s make USEBUILDDIR=1 BACKEND=cuda - make USEBUILDDIR=1 BACKEND=cppnone -make USEBUILDDIR=1 BACKEND=cppsse4 +make USEBUILDDIR=1 BACKEND=cppsse4 make USEBUILDDIR=1 BACKEND=cppavx2 + make USEBUILDDIR=1 BACKEND=cpp512y make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_t1t1.mad/SubProcesses/P1_gg_t1t1x' make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_t1t1.mad/SubProcesses/P1_gg_t1t1x' @@ -22,9 +22,9 @@ make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_t1t1.mad/SubProcesses/P1_gg_t1t1x' make[1]: Nothing to be done for 'all'. -make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_t1t1.mad/SubProcesses/P1_gg_t1t1x' make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_t1t1.mad/SubProcesses/P1_gg_t1t1x' +make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_t1t1.mad/SubProcesses/P1_gg_t1t1x' make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_t1t1.mad/SubProcesses/P1_gg_t1t1x' make[1]: Nothing to be done for 'all'. @@ -32,7 +32,7 @@ make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ OMP_NUM_THREADS= -DATE: 2024-10-06_10:55:56 +DATE: 2024-10-06_18:09:38 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: Working directory (run): /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_t1t1.mad/SubProcesses/P1_gg_t1t1x @@ -58,9 +58,9 @@ Executing ' ./madevent_fortran < /tmp/avalassi/input_susyggt1t1_x1_fortran > /tm [XSECTION] ChannelId = 3 [XSECTION] Cross section = 0.3045 [0.30449452343426120] fbridge_mode=0 [UNWEIGHT] Wrote 1732 events (found 4297 events) - [COUNTERS] PROGRAM TOTAL : 0.6925s - [COUNTERS] Fortran Overhead ( 0 ) : 0.6838s - [COUNTERS] Fortran MEs ( 1 ) : 0.0087s for 8192 events => throughput is 9.43E+05 events/s + [COUNTERS] PROGRAM TOTAL : 0.6709s + [COUNTERS] Fortran Overhead ( 0 ) : 0.6627s + [COUNTERS] Fortran MEs ( 1 ) : 0.0082s for 8192 events => throughput is 1.00E+06 events/s *** (1) EXECUTE MADEVENT_FORTRAN x1 (create events.lhe) *** -------------------- @@ -83,9 +83,34 @@ Executing ' ./madevent_fortran < /tmp/avalassi/input_susyggt1t1_x1_fortran > /tm [XSECTION] ChannelId = 3 [XSECTION] Cross section = 0.3045 [0.30449452343426120] fbridge_mode=0 [UNWEIGHT] Wrote 1612 events (found 1617 events) - [COUNTERS] PROGRAM TOTAL : 0.4263s - [COUNTERS] Fortran Overhead ( 0 ) : 0.4177s - [COUNTERS] Fortran MEs ( 1 ) : 0.0085s for 8192 events => throughput is 9.62E+05 events/s + [COUNTERS] PROGRAM TOTAL : 0.4114s + [COUNTERS] Fortran Overhead ( 0 ) : 0.4032s + [COUNTERS] Fortran MEs ( 1 ) : 0.0082s for 8192 events => throughput is 9.96E+05 events/s + +*** (1) EXECUTE MADEVENT_FORTRAN x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +2 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./madevent_fortran < /tmp/avalassi/input_susyggt1t1_x10_fortran > /tmp/avalassi/output_susyggt1t1_x10_fortran' + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 4/4 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 2 + [XSECTION] ChannelId = 3 + [XSECTION] Cross section = 0.3075 [0.30747310722207288] fbridge_mode=0 + [UNWEIGHT] Wrote 1631 events (found 1636 events) + [COUNTERS] PROGRAM TOTAL : 1.6338s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5519s + [COUNTERS] Fortran MEs ( 1 ) : 0.0819s for 81920 events => throughput is 1.00E+06 events/s *** (2-none) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -109,9 +134,9 @@ DEBUG: MEK processed 8192 events across 6 channels { 3 : 8192 } [XSECTION] ChannelId = 3 [XSECTION] Cross section = 0.3045 [0.30449452343426120] fbridge_mode=1 [UNWEIGHT] Wrote 1612 events (found 1617 events) - [COUNTERS] PROGRAM TOTAL : 0.4276s - [COUNTERS] Fortran Overhead ( 0 ) : 0.4188s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0084s for 8192 events => throughput is 9.73E+05 events/s + [COUNTERS] PROGRAM TOTAL : 0.4172s + [COUNTERS] Fortran Overhead ( 0 ) : 0.4089s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0080s for 8192 events => throughput is 1.02E+06 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0003s *** (2-none) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -122,15 +147,50 @@ OK! xsec from fortran (0.30449452343426120) and cpp (0.30449452343426120) differ OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-none) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +2 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.none_d_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_susyggt1t1_x10_cudacpp > /tmp/avalassi/output_susyggt1t1_x10_cudacpp' +DEBUG: MEK processed 81920 events across 6 channels { 3 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 4/4 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 2 + [XSECTION] ChannelId = 3 + [XSECTION] Cross section = 0.3075 [0.30747310722207283] fbridge_mode=1 + [UNWEIGHT] Wrote 1631 events (found 1636 events) + [COUNTERS] PROGRAM TOTAL : 1.6434s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5612s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0819s for 81920 events => throughput is 1.00E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0003s + +*** (2-none) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (0.30747310722207288) and cpp (0.30747310722207283) differ by less than 3E-14 (2.220446049250313e-16) + +*** (2-none) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_MSSM_SLHA2_GG_T1T1X_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+BRDHST/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 9.916439e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.021532e+06 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_MSSM_SLHA2_GG_T1T1X_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.017065e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.010436e+06 ) sec^-1 *** (2-sse4) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -154,9 +214,9 @@ DEBUG: MEK processed 8192 events across 6 channels { 3 : 8192 } [XSECTION] ChannelId = 3 [XSECTION] Cross section = 0.3045 [0.30449452343426120] fbridge_mode=1 [UNWEIGHT] Wrote 1612 events (found 1617 events) - [COUNTERS] PROGRAM TOTAL : 0.4290s - [COUNTERS] Fortran Overhead ( 0 ) : 0.4241s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0046s for 8192 events => throughput is 1.79E+06 events/s + [COUNTERS] PROGRAM TOTAL : 0.4112s + [COUNTERS] Fortran Overhead ( 0 ) : 0.4064s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0045s for 8192 events => throughput is 1.84E+06 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0003s *** (2-sse4) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -167,15 +227,50 @@ OK! xsec from fortran (0.30449452343426120) and cpp (0.30449452343426120) differ OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-sse4) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +2 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.sse4_d_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_susyggt1t1_x10_cudacpp > /tmp/avalassi/output_susyggt1t1_x10_cudacpp' +DEBUG: MEK processed 81920 events across 6 channels { 3 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 4/4 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 2 + [XSECTION] ChannelId = 3 + [XSECTION] Cross section = 0.3075 [0.30747310722207283] fbridge_mode=1 + [UNWEIGHT] Wrote 1631 events (found 1636 events) + [COUNTERS] PROGRAM TOTAL : 1.6176s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5727s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0445s for 81920 events => throughput is 1.84E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0003s + +*** (2-sse4) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (0.30747310722207288) and cpp (0.30747310722207283) differ by less than 3E-14 (2.220446049250313e-16) + +*** (2-sse4) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_MSSM_SLHA2_GG_T1T1X_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+BRDHST/sse4+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.913729e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.948167e+06 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_MSSM_SLHA2_GG_T1T1X_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.928329e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.943954e+06 ) sec^-1 *** (2-avx2) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -199,10 +294,10 @@ DEBUG: MEK processed 8192 events across 6 channels { 3 : 8192 } [XSECTION] ChannelId = 3 [XSECTION] Cross section = 0.3045 [0.30449452343426114] fbridge_mode=1 [UNWEIGHT] Wrote 1612 events (found 1617 events) - [COUNTERS] PROGRAM TOTAL : 0.4231s - [COUNTERS] Fortran Overhead ( 0 ) : 0.4198s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0029s for 8192 events => throughput is 2.80E+06 events/s - [COUNTERS] CudaCpp HEL ( 3 ) : 0.0004s + [COUNTERS] PROGRAM TOTAL : 0.4087s + [COUNTERS] Fortran Overhead ( 0 ) : 0.4056s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0027s for 8192 events => throughput is 3.03E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0003s *** (2-avx2) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -212,15 +307,50 @@ OK! xsec from fortran (0.30449452343426120) and cpp (0.30449452343426114) differ OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-avx2) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +2 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.avx2_d_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_susyggt1t1_x10_cudacpp > /tmp/avalassi/output_susyggt1t1_x10_cudacpp' +DEBUG: MEK processed 81920 events across 6 channels { 3 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 4/4 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 2 + [XSECTION] ChannelId = 3 + [XSECTION] Cross section = 0.3075 [0.30747310722207283] fbridge_mode=1 + [UNWEIGHT] Wrote 1631 events (found 1636 events) + [COUNTERS] PROGRAM TOTAL : 1.6008s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5730s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0275s for 81920 events => throughput is 2.98E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0003s + +*** (2-avx2) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (0.30747310722207288) and cpp (0.30747310722207283) differ by less than 3E-14 (2.220446049250313e-16) + +*** (2-avx2) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_MSSM_SLHA2_GG_T1T1X_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+BRDHST/avx2+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.118646e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.133606e+06 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_MSSM_SLHA2_GG_T1T1X_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.327279e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.361099e+06 ) sec^-1 *** (2-512y) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -244,9 +374,9 @@ DEBUG: MEK processed 8192 events across 6 channels { 3 : 8192 } [XSECTION] ChannelId = 3 [XSECTION] Cross section = 0.3045 [0.30449452343426114] fbridge_mode=1 [UNWEIGHT] Wrote 1612 events (found 1617 events) - [COUNTERS] PROGRAM TOTAL : 0.4242s - [COUNTERS] Fortran Overhead ( 0 ) : 0.4212s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0027s for 8192 events => throughput is 3.07E+06 events/s + [COUNTERS] PROGRAM TOTAL : 0.4155s + [COUNTERS] Fortran Overhead ( 0 ) : 0.4126s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0026s for 8192 events => throughput is 3.17E+06 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0003s *** (2-512y) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -257,15 +387,50 @@ OK! xsec from fortran (0.30449452343426120) and cpp (0.30449452343426114) differ OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-512y) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +2 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.512y_d_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_susyggt1t1_x10_cudacpp > /tmp/avalassi/output_susyggt1t1_x10_cudacpp' +DEBUG: MEK processed 81920 events across 6 channels { 3 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 4/4 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 2 + [XSECTION] ChannelId = 3 + [XSECTION] Cross section = 0.3075 [0.30747310722207283] fbridge_mode=1 + [UNWEIGHT] Wrote 1631 events (found 1636 events) + [COUNTERS] PROGRAM TOTAL : 1.5970s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5702s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0265s for 81920 events => throughput is 3.09E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0004s + +*** (2-512y) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (0.30747310722207288) and cpp (0.30747310722207283) differ by less than 3E-14 (2.220446049250313e-16) + +*** (2-512y) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_MSSM_SLHA2_GG_T1T1X_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+BRDHST/512y+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.142389e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.291024e+06 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_MSSM_SLHA2_GG_T1T1X_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.418661e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.514451e+06 ) sec^-1 *** (2-512z) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -289,9 +454,9 @@ DEBUG: MEK processed 8192 events across 6 channels { 3 : 8192 } [XSECTION] ChannelId = 3 [XSECTION] Cross section = 0.3045 [0.30449452343426114] fbridge_mode=1 [UNWEIGHT] Wrote 1612 events (found 1617 events) - [COUNTERS] PROGRAM TOTAL : 0.4276s - [COUNTERS] Fortran Overhead ( 0 ) : 0.4241s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0031s for 8192 events => throughput is 2.67E+06 events/s + [COUNTERS] PROGRAM TOTAL : 0.4133s + [COUNTERS] Fortran Overhead ( 0 ) : 0.4098s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0031s for 8192 events => throughput is 2.63E+06 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0004s *** (2-512z) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -302,15 +467,50 @@ OK! xsec from fortran (0.30449452343426120) and cpp (0.30449452343426114) differ OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-512z) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +2 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.512z_d_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_susyggt1t1_x10_cudacpp > /tmp/avalassi/output_susyggt1t1_x10_cudacpp' +DEBUG: MEK processed 81920 events across 6 channels { 3 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 4/4 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 2 + [XSECTION] ChannelId = 3 + [XSECTION] Cross section = 0.3075 [0.30747310722207283] fbridge_mode=1 + [UNWEIGHT] Wrote 1631 events (found 1636 events) + [COUNTERS] PROGRAM TOTAL : 1.5948s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5650s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0294s for 81920 events => throughput is 2.79E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0004s + +*** (2-512z) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (0.30747310722207288) and cpp (0.30747310722207283) differ by less than 3E-14 (2.220446049250313e-16) + +*** (2-512z) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_MSSM_SLHA2_GG_T1T1X_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+BRDHST/512z+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 2.810680e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.802325e+06 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_MSSM_SLHA2_GG_T1T1X_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.123505e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.016828e+06 ) sec^-1 *** (3-cuda) EXECUTE MADEVENT_CUDA x1 (create events.lhe) *** -------------------- @@ -334,9 +534,9 @@ DEBUG: MEK processed 8192 events across 6 channels { 3 : 8192 } [XSECTION] ChannelId = 3 [XSECTION] Cross section = 0.3045 [0.30449452343426109] fbridge_mode=1 [UNWEIGHT] Wrote 1612 events (found 1617 events) - [COUNTERS] PROGRAM TOTAL : 0.8704s - [COUNTERS] Fortran Overhead ( 0 ) : 0.8668s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0031s for 8192 events => throughput is 2.69E+06 events/s + [COUNTERS] PROGRAM TOTAL : 0.8463s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8429s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0029s for 8192 events => throughput is 2.84E+06 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0005s *** (3-cuda) Compare MADEVENT_CUDA x1 xsec to MADEVENT_FORTRAN xsec *** @@ -347,45 +547,80 @@ OK! xsec from fortran (0.30449452343426120) and cuda (0.30449452343426109) diffe OK! events.lhe.cuda.1 and events.lhe.ref.1 are identical +*** (3-cuda) EXECUTE MADEVENT_CUDA x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +2 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.cuda_d_inl0_hrd0/madevent_cuda < /tmp/avalassi/input_susyggt1t1_x10_cudacpp > /tmp/avalassi/output_susyggt1t1_x10_cudacpp' +DEBUG: MEK processed 81920 events across 6 channels { 3 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 4/4 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 2 + [XSECTION] ChannelId = 3 + [XSECTION] Cross section = 0.3075 [0.30747310722207283] fbridge_mode=1 + [UNWEIGHT] Wrote 1631 events (found 1636 events) + [COUNTERS] PROGRAM TOTAL : 2.0024s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9942s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0076s for 81920 events => throughput is 1.08E+07 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0005s + +*** (3-cuda) Compare MADEVENT_CUDA x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (0.30747310722207288) and cuda (0.30747310722207283) differ by less than 3E-14 (2.220446049250313e-16) + +*** (3-cuda) Compare MADEVENT_CUDA x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cuda.10 and events.lhe.ref.10 are identical + *** EXECUTE GCHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_MSSM_SLHA2_GG_T1T1X_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.094441e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.171661e+06 ) sec^-1 *** EXECUTE GCHECK(8192) -p 256 32 1 *** Process = SIGMA_MSSM_SLHA2_GG_T1T1X_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.576690e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.563530e+06 ) sec^-1 *** EXECUTE GCHECK(MAX) -p 16384 32 1 --bridge *** Process = SIGMA_MSSM_SLHA2_GG_T1T1X_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.540792e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.567045e+07 ) sec^-1 *** EXECUTE GCHECK(MAX) -p 16384 32 1 *** Process = SIGMA_MSSM_SLHA2_GG_T1T1X_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.885377e+08 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.939388e+08 ) sec^-1 *** EXECUTE GCHECK(MAX128THR) -p 4096 128 1 --bridge *** Process = SIGMA_MSSM_SLHA2_GG_T1T1X_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.486109e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.588917e+07 ) sec^-1 *** EXECUTE GCHECK(MAX128THR) -p 4096 128 1 *** Process = SIGMA_MSSM_SLHA2_GG_T1T1X_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.914518e+08 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.977584e+08 ) sec^-1 *** EXECUTE GCHECK(MAX8THR) -p 65536 8 1 --bridge *** Process = SIGMA_MSSM_SLHA2_GG_T1T1X_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.512059e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.599494e+07 ) sec^-1 *** EXECUTE GCHECK(MAX8THR) -p 65536 8 1 *** Process = SIGMA_MSSM_SLHA2_GG_T1T1X_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.224875e+08 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.271013e+08 ) sec^-1 *** (3-hip) WARNING! SKIP MADEVENT_HIP (hip is not supported on this node) *** diff --git a/epochX/cudacpp/tmad/logs_susyggt1t1_mad/log_susyggt1t1_mad_f_inl0_hrd0.txt b/epochX/cudacpp/tmad/logs_susyggt1t1_mad/log_susyggt1t1_mad_f_inl0_hrd0.txt index 9cee2ab297..4859848d49 100644 --- a/epochX/cudacpp/tmad/logs_susyggt1t1_mad/log_susyggt1t1_mad_f_inl0_hrd0.txt +++ b/epochX/cudacpp/tmad/logs_susyggt1t1_mad/log_susyggt1t1_mad_f_inl0_hrd0.txt @@ -5,8 +5,8 @@ make USEBUILDDIR=1 BACKEND=cuda make USEBUILDDIR=1 BACKEND=cppnone -make USEBUILDDIR=1 BACKEND=cppsse4 +make USEBUILDDIR=1 BACKEND=cppsse4 make USEBUILDDIR=1 BACKEND=cppavx2 make USEBUILDDIR=1 BACKEND=cpp512y make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_t1t1.mad/SubProcesses/P1_gg_t1t1x' @@ -18,8 +18,8 @@ make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp make USEBUILDDIR=1 BACKEND=cpp512z make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_t1t1.mad/SubProcesses/P1_gg_t1t1x' make[1]: Nothing to be done for 'all'. -make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_t1t1.mad/SubProcesses/P1_gg_t1t1x' +make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_t1t1.mad/SubProcesses/P1_gg_t1t1x' make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_t1t1.mad/SubProcesses/P1_gg_t1t1x' @@ -32,7 +32,7 @@ make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ OMP_NUM_THREADS= -DATE: 2024-10-06_10:56:23 +DATE: 2024-10-06_18:10:31 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: Working directory (run): /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_t1t1.mad/SubProcesses/P1_gg_t1t1x @@ -58,9 +58,9 @@ Executing ' ./madevent_fortran < /tmp/avalassi/input_susyggt1t1_x1_fortran > /tm [XSECTION] ChannelId = 3 [XSECTION] Cross section = 0.3045 [0.30449452343426120] fbridge_mode=0 [UNWEIGHT] Wrote 1732 events (found 4297 events) - [COUNTERS] PROGRAM TOTAL : 0.6965s - [COUNTERS] Fortran Overhead ( 0 ) : 0.6879s - [COUNTERS] Fortran MEs ( 1 ) : 0.0086s for 8192 events => throughput is 9.48E+05 events/s + [COUNTERS] PROGRAM TOTAL : 0.6672s + [COUNTERS] Fortran Overhead ( 0 ) : 0.6588s + [COUNTERS] Fortran MEs ( 1 ) : 0.0084s for 8192 events => throughput is 9.81E+05 events/s *** (1) EXECUTE MADEVENT_FORTRAN x1 (create events.lhe) *** -------------------- @@ -83,9 +83,34 @@ Executing ' ./madevent_fortran < /tmp/avalassi/input_susyggt1t1_x1_fortran > /tm [XSECTION] ChannelId = 3 [XSECTION] Cross section = 0.3045 [0.30449452343426120] fbridge_mode=0 [UNWEIGHT] Wrote 1612 events (found 1617 events) - [COUNTERS] PROGRAM TOTAL : 0.4263s - [COUNTERS] Fortran Overhead ( 0 ) : 0.4177s - [COUNTERS] Fortran MEs ( 1 ) : 0.0087s for 8192 events => throughput is 9.43E+05 events/s + [COUNTERS] PROGRAM TOTAL : 0.4124s + [COUNTERS] Fortran Overhead ( 0 ) : 0.4040s + [COUNTERS] Fortran MEs ( 1 ) : 0.0084s for 8192 events => throughput is 9.73E+05 events/s + +*** (1) EXECUTE MADEVENT_FORTRAN x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +2 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./madevent_fortran < /tmp/avalassi/input_susyggt1t1_x10_fortran > /tmp/avalassi/output_susyggt1t1_x10_fortran' + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 4/4 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 2 + [XSECTION] ChannelId = 3 + [XSECTION] Cross section = 0.3075 [0.30747310722207288] fbridge_mode=0 + [UNWEIGHT] Wrote 1631 events (found 1636 events) + [COUNTERS] PROGRAM TOTAL : 1.6465s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5653s + [COUNTERS] Fortran MEs ( 1 ) : 0.0812s for 81920 events => throughput is 1.01E+06 events/s *** (2-none) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -109,9 +134,9 @@ DEBUG: MEK processed 8192 events across 6 channels { 3 : 8192 } [XSECTION] ChannelId = 3 [XSECTION] Cross section = 0.3045 [0.30449446496609361] fbridge_mode=1 [UNWEIGHT] Wrote 1612 events (found 1617 events) - [COUNTERS] PROGRAM TOTAL : 0.4268s - [COUNTERS] Fortran Overhead ( 0 ) : 0.4183s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0082s for 8192 events => throughput is 9.95E+05 events/s + [COUNTERS] PROGRAM TOTAL : 0.4139s + [COUNTERS] Fortran Overhead ( 0 ) : 0.4051s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0086s for 8192 events => throughput is 9.52E+05 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0003s *** (2-none) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -122,15 +147,50 @@ OK! xsec from fortran (0.30449452343426120) and cpp (0.30449446496609361) differ OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-none) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +2 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.none_f_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_susyggt1t1_x10_cudacpp > /tmp/avalassi/output_susyggt1t1_x10_cudacpp' +DEBUG: MEK processed 81920 events across 6 channels { 3 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 4/4 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 2 + [XSECTION] ChannelId = 3 + [XSECTION] Cross section = 0.3075 [0.30747305007079218] fbridge_mode=1 + [UNWEIGHT] Wrote 1631 events (found 1636 events) + [COUNTERS] PROGRAM TOTAL : 1.6490s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5686s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0801s for 81920 events => throughput is 1.02E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0003s + +*** (2-none) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (0.30747310722207288) and cpp (0.30747305007079218) differ by less than 4E-4 (1.858740792393121e-07) + +*** (2-none) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_MSSM_SLHA2_GG_T1T1X_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+BRDHST/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.006620e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.020301e+06 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_MSSM_SLHA2_GG_T1T1X_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.012762e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.016148e+06 ) sec^-1 *** (2-sse4) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -154,9 +214,9 @@ DEBUG: MEK processed 8192 events across 6 channels { 3 : 8192 } [XSECTION] ChannelId = 3 [XSECTION] Cross section = 0.3045 [0.30449446369440458] fbridge_mode=1 [UNWEIGHT] Wrote 1612 events (found 1617 events) - [COUNTERS] PROGRAM TOTAL : 0.4190s - [COUNTERS] Fortran Overhead ( 0 ) : 0.4159s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0028s for 8192 events => throughput is 2.94E+06 events/s + [COUNTERS] PROGRAM TOTAL : 0.4064s + [COUNTERS] Fortran Overhead ( 0 ) : 0.4034s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0028s for 8192 events => throughput is 2.97E+06 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0002s *** (2-sse4) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -167,15 +227,50 @@ OK! xsec from fortran (0.30449452343426120) and cpp (0.30449446369440458) differ OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-sse4) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +2 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.sse4_f_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_susyggt1t1_x10_cudacpp > /tmp/avalassi/output_susyggt1t1_x10_cudacpp' +DEBUG: MEK processed 81920 events across 6 channels { 3 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 4/4 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 2 + [XSECTION] ChannelId = 3 + [XSECTION] Cross section = 0.3075 [0.30747304961041555] fbridge_mode=1 + [UNWEIGHT] Wrote 1631 events (found 1636 events) + [COUNTERS] PROGRAM TOTAL : 1.5930s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5659s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0269s for 81920 events => throughput is 3.04E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0002s + +*** (2-sse4) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (0.30747310722207288) and cpp (0.30747304961041555) differ by less than 4E-4 (1.8737136997515336e-07) + +*** (2-sse4) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_MSSM_SLHA2_GG_T1T1X_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+BRDHST/sse4+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.282555e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.256149e+06 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_MSSM_SLHA2_GG_T1T1X_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.369793e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.296355e+06 ) sec^-1 *** (2-avx2) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -199,9 +294,9 @@ DEBUG: MEK processed 8192 events across 6 channels { 3 : 8192 } [XSECTION] ChannelId = 3 [XSECTION] Cross section = 0.3045 [0.30449446614968528] fbridge_mode=1 [UNWEIGHT] Wrote 1612 events (found 1617 events) - [COUNTERS] PROGRAM TOTAL : 0.4206s - [COUNTERS] Fortran Overhead ( 0 ) : 0.4183s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0020s for 8192 events => throughput is 4.08E+06 events/s + [COUNTERS] PROGRAM TOTAL : 0.4061s + [COUNTERS] Fortran Overhead ( 0 ) : 0.4041s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0017s for 8192 events => throughput is 4.71E+06 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0002s *** (2-avx2) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -212,15 +307,50 @@ OK! xsec from fortran (0.30449452343426120) and cpp (0.30449446614968528) differ OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-avx2) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +2 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.avx2_f_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_susyggt1t1_x10_cudacpp > /tmp/avalassi/output_susyggt1t1_x10_cudacpp' +DEBUG: MEK processed 81920 events across 6 channels { 3 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 4/4 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 2 + [XSECTION] ChannelId = 3 + [XSECTION] Cross section = 0.3075 [0.30747305065199410] fbridge_mode=1 + [UNWEIGHT] Wrote 1631 events (found 1636 events) + [COUNTERS] PROGRAM TOTAL : 1.5698s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5513s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0182s for 81920 events => throughput is 4.51E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0002s + +*** (2-avx2) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (0.30747310722207288) and cpp (0.30747305065199410) differ by less than 4E-4 (1.839838263961724e-07) + +*** (2-avx2) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_MSSM_SLHA2_GG_T1T1X_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+BRDHST/avx2+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 4.872977e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 5.032788e+06 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_MSSM_SLHA2_GG_T1T1X_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 5.148892e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 5.286120e+06 ) sec^-1 *** (2-512y) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -244,9 +374,9 @@ DEBUG: MEK processed 8192 events across 6 channels { 3 : 8192 } [XSECTION] ChannelId = 3 [XSECTION] Cross section = 0.3045 [0.30449446614968528] fbridge_mode=1 [UNWEIGHT] Wrote 1612 events (found 1617 events) - [COUNTERS] PROGRAM TOTAL : 0.4201s - [COUNTERS] Fortran Overhead ( 0 ) : 0.4180s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0018s for 8192 events => throughput is 4.44E+06 events/s + [COUNTERS] PROGRAM TOTAL : 0.4083s + [COUNTERS] Fortran Overhead ( 0 ) : 0.4062s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0018s for 8192 events => throughput is 4.55E+06 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0002s *** (2-512y) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -257,15 +387,50 @@ OK! xsec from fortran (0.30449452343426120) and cpp (0.30449446614968528) differ OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-512y) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +2 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.512y_f_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_susyggt1t1_x10_cudacpp > /tmp/avalassi/output_susyggt1t1_x10_cudacpp' +DEBUG: MEK processed 81920 events across 6 channels { 3 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 4/4 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 2 + [XSECTION] ChannelId = 3 + [XSECTION] Cross section = 0.3075 [0.30747305065199410] fbridge_mode=1 + [UNWEIGHT] Wrote 1631 events (found 1636 events) + [COUNTERS] PROGRAM TOTAL : 1.5768s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5592s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0173s for 81920 events => throughput is 4.73E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0002s + +*** (2-512y) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (0.30747310722207288) and cpp (0.30747305065199410) differ by less than 4E-4 (1.839838263961724e-07) + +*** (2-512y) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_MSSM_SLHA2_GG_T1T1X_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+BRDHST/512y+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 4.886846e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 5.257389e+06 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_MSSM_SLHA2_GG_T1T1X_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 5.506416e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 5.564923e+06 ) sec^-1 *** (2-512z) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -289,9 +454,9 @@ DEBUG: MEK processed 8192 events across 6 channels { 3 : 8192 } [XSECTION] ChannelId = 3 [XSECTION] Cross section = 0.3045 [0.30449447031649013] fbridge_mode=1 [UNWEIGHT] Wrote 1612 events (found 1617 events) - [COUNTERS] PROGRAM TOTAL : 0.4202s - [COUNTERS] Fortran Overhead ( 0 ) : 0.4176s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0023s for 8192 events => throughput is 3.54E+06 events/s + [COUNTERS] PROGRAM TOTAL : 0.4078s + [COUNTERS] Fortran Overhead ( 0 ) : 0.4052s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0023s for 8192 events => throughput is 3.62E+06 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0003s *** (2-512z) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -302,15 +467,50 @@ OK! xsec from fortran (0.30449452343426120) and cpp (0.30449447031649013) differ OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-512z) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +2 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.512z_f_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_susyggt1t1_x10_cudacpp > /tmp/avalassi/output_susyggt1t1_x10_cudacpp' +DEBUG: MEK processed 81920 events across 6 channels { 3 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 4/4 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 2 + [XSECTION] ChannelId = 3 + [XSECTION] Cross section = 0.3075 [0.30747305508949557] fbridge_mode=1 + [UNWEIGHT] Wrote 1631 events (found 1636 events) + [COUNTERS] PROGRAM TOTAL : 1.6096s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5881s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0212s for 81920 events => throughput is 3.87E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0003s + +*** (2-512z) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (0.30747310722207288) and cpp (0.30747305508949557) differ by less than 4E-4 (1.6955166515231213e-07) + +*** (2-512z) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_MSSM_SLHA2_GG_T1T1X_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+BRDHST/512z+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 4.376595e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 4.426359e+06 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_MSSM_SLHA2_GG_T1T1X_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 4.863933e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 4.632334e+06 ) sec^-1 *** (3-cuda) EXECUTE MADEVENT_CUDA x1 (create events.lhe) *** -------------------- @@ -334,9 +534,9 @@ DEBUG: MEK processed 8192 events across 6 channels { 3 : 8192 } [XSECTION] ChannelId = 3 [XSECTION] Cross section = 0.3045 [0.30449447352014630] fbridge_mode=1 [UNWEIGHT] Wrote 1612 events (found 1617 events) - [COUNTERS] PROGRAM TOTAL : 0.8576s - [COUNTERS] Fortran Overhead ( 0 ) : 0.8540s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0031s for 8192 events => throughput is 2.64E+06 events/s + [COUNTERS] PROGRAM TOTAL : 0.8458s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8423s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0030s for 8192 events => throughput is 2.77E+06 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0005s *** (3-cuda) Compare MADEVENT_CUDA x1 xsec to MADEVENT_FORTRAN xsec *** @@ -347,45 +547,80 @@ OK! xsec from fortran (0.30449452343426120) and cuda (0.30449447352014630) diffe OK! events.lhe.cuda.1 and events.lhe.ref.1 are identical +*** (3-cuda) EXECUTE MADEVENT_CUDA x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +2 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.cuda_f_inl0_hrd0/madevent_cuda < /tmp/avalassi/input_susyggt1t1_x10_cudacpp > /tmp/avalassi/output_susyggt1t1_x10_cudacpp' +DEBUG: MEK processed 81920 events across 6 channels { 3 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 4/4 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 2 + [XSECTION] ChannelId = 3 + [XSECTION] Cross section = 0.3075 [0.30747305761315818] fbridge_mode=1 + [UNWEIGHT] Wrote 1631 events (found 1636 events) + [COUNTERS] PROGRAM TOTAL : 2.0164s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0084s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0075s for 81920 events => throughput is 1.10E+07 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0005s + +*** (3-cuda) Compare MADEVENT_CUDA x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (0.30747310722207288) and cuda (0.30747305761315818) differ by less than 4E-4 (1.6134391445099538e-07) + +*** (3-cuda) Compare MADEVENT_CUDA x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cuda.10 and events.lhe.ref.10 are identical + *** EXECUTE GCHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_MSSM_SLHA2_GG_T1T1X_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.209039e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.340122e+06 ) sec^-1 *** EXECUTE GCHECK(8192) -p 256 32 1 *** Process = SIGMA_MSSM_SLHA2_GG_T1T1X_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.497762e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.659548e+06 ) sec^-1 *** EXECUTE GCHECK(MAX) -p 16384 32 1 --bridge *** Process = SIGMA_MSSM_SLHA2_GG_T1T1X_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 5.599688e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 5.892680e+07 ) sec^-1 *** EXECUTE GCHECK(MAX) -p 16384 32 1 *** Process = SIGMA_MSSM_SLHA2_GG_T1T1X_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 2.103544e+08 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.153395e+08 ) sec^-1 *** EXECUTE GCHECK(MAX128THR) -p 4096 128 1 --bridge *** Process = SIGMA_MSSM_SLHA2_GG_T1T1X_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 5.606706e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 5.844938e+07 ) sec^-1 *** EXECUTE GCHECK(MAX128THR) -p 4096 128 1 *** Process = SIGMA_MSSM_SLHA2_GG_T1T1X_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 2.131283e+08 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.201627e+08 ) sec^-1 *** EXECUTE GCHECK(MAX8THR) -p 65536 8 1 --bridge *** Process = SIGMA_MSSM_SLHA2_GG_T1T1X_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 5.229812e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 5.444586e+07 ) sec^-1 *** EXECUTE GCHECK(MAX8THR) -p 65536 8 1 *** Process = SIGMA_MSSM_SLHA2_GG_T1T1X_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.664371e+08 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.701431e+08 ) sec^-1 *** (3-hip) WARNING! SKIP MADEVENT_HIP (hip is not supported on this node) *** diff --git a/epochX/cudacpp/tmad/logs_susyggt1t1_mad/log_susyggt1t1_mad_m_inl0_hrd0.txt b/epochX/cudacpp/tmad/logs_susyggt1t1_mad/log_susyggt1t1_mad_m_inl0_hrd0.txt index 782fee34a5..dde4d607f7 100644 --- a/epochX/cudacpp/tmad/logs_susyggt1t1_mad/log_susyggt1t1_mad_m_inl0_hrd0.txt +++ b/epochX/cudacpp/tmad/logs_susyggt1t1_mad/log_susyggt1t1_mad_m_inl0_hrd0.txt @@ -4,8 +4,8 @@ Working directory (build): /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/s make USEBUILDDIR=1 BACKEND=cuda make USEBUILDDIR=1 BACKEND=cppnone - make USEBUILDDIR=1 BACKEND=cppsse4 + make USEBUILDDIR=1 BACKEND=cppavx2 make USEBUILDDIR=1 BACKEND=cpp512y @@ -32,7 +32,7 @@ make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ OMP_NUM_THREADS= -DATE: 2024-10-06_10:56:09 +DATE: 2024-10-06_18:10:04 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: Working directory (run): /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_t1t1.mad/SubProcesses/P1_gg_t1t1x @@ -58,9 +58,9 @@ Executing ' ./madevent_fortran < /tmp/avalassi/input_susyggt1t1_x1_fortran > /tm [XSECTION] ChannelId = 3 [XSECTION] Cross section = 0.3045 [0.30449452343426120] fbridge_mode=0 [UNWEIGHT] Wrote 1732 events (found 4297 events) - [COUNTERS] PROGRAM TOTAL : 0.6953s - [COUNTERS] Fortran Overhead ( 0 ) : 0.6866s - [COUNTERS] Fortran MEs ( 1 ) : 0.0086s for 8192 events => throughput is 9.49E+05 events/s + [COUNTERS] PROGRAM TOTAL : 0.6696s + [COUNTERS] Fortran Overhead ( 0 ) : 0.6614s + [COUNTERS] Fortran MEs ( 1 ) : 0.0082s for 8192 events => throughput is 9.96E+05 events/s *** (1) EXECUTE MADEVENT_FORTRAN x1 (create events.lhe) *** -------------------- @@ -83,9 +83,34 @@ Executing ' ./madevent_fortran < /tmp/avalassi/input_susyggt1t1_x1_fortran > /tm [XSECTION] ChannelId = 3 [XSECTION] Cross section = 0.3045 [0.30449452343426120] fbridge_mode=0 [UNWEIGHT] Wrote 1612 events (found 1617 events) - [COUNTERS] PROGRAM TOTAL : 0.4236s - [COUNTERS] Fortran Overhead ( 0 ) : 0.4152s - [COUNTERS] Fortran MEs ( 1 ) : 0.0085s for 8192 events => throughput is 9.68E+05 events/s + [COUNTERS] PROGRAM TOTAL : 0.4125s + [COUNTERS] Fortran Overhead ( 0 ) : 0.4043s + [COUNTERS] Fortran MEs ( 1 ) : 0.0082s for 8192 events => throughput is 9.99E+05 events/s + +*** (1) EXECUTE MADEVENT_FORTRAN x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +2 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./madevent_fortran < /tmp/avalassi/input_susyggt1t1_x10_fortran > /tmp/avalassi/output_susyggt1t1_x10_fortran' + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 4/4 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 2 + [XSECTION] ChannelId = 3 + [XSECTION] Cross section = 0.3075 [0.30747310722207288] fbridge_mode=0 + [UNWEIGHT] Wrote 1631 events (found 1636 events) + [COUNTERS] PROGRAM TOTAL : 1.6323s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5511s + [COUNTERS] Fortran MEs ( 1 ) : 0.0812s for 81920 events => throughput is 1.01E+06 events/s *** (2-none) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -109,9 +134,9 @@ DEBUG: MEK processed 8192 events across 6 channels { 3 : 8192 } [XSECTION] ChannelId = 3 [XSECTION] Cross section = 0.3045 [0.30449453160892032] fbridge_mode=1 [UNWEIGHT] Wrote 1612 events (found 1617 events) - [COUNTERS] PROGRAM TOTAL : 0.4277s - [COUNTERS] Fortran Overhead ( 0 ) : 0.4187s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0086s for 8192 events => throughput is 9.54E+05 events/s + [COUNTERS] PROGRAM TOTAL : 0.4147s + [COUNTERS] Fortran Overhead ( 0 ) : 0.4060s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0084s for 8192 events => throughput is 9.77E+05 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0003s *** (2-none) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -122,15 +147,50 @@ OK! xsec from fortran (0.30449452343426120) and cpp (0.30449453160892032) differ OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-none) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +2 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.none_m_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_susyggt1t1_x10_cudacpp > /tmp/avalassi/output_susyggt1t1_x10_cudacpp' +DEBUG: MEK processed 81920 events across 6 channels { 3 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 4/4 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 2 + [XSECTION] ChannelId = 3 + [XSECTION] Cross section = 0.3075 [0.30747311535940236] fbridge_mode=1 + [UNWEIGHT] Wrote 1631 events (found 1636 events) + [COUNTERS] PROGRAM TOTAL : 1.6627s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5790s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0834s for 81920 events => throughput is 9.83E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0003s + +*** (2-none) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (0.30747310722207288) and cpp (0.30747311535940236) differ by less than 2E-4 (2.6465174718381945e-08) + +*** (2-none) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_MSSM_SLHA2_GG_T1T1X_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+BRDHST/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 9.831908e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 9.798196e+05 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_MSSM_SLHA2_GG_T1T1X_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 9.918457e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 9.894804e+05 ) sec^-1 *** (2-sse4) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -154,9 +214,9 @@ DEBUG: MEK processed 8192 events across 6 channels { 3 : 8192 } [XSECTION] ChannelId = 3 [XSECTION] Cross section = 0.3045 [0.30449453160892032] fbridge_mode=1 [UNWEIGHT] Wrote 1612 events (found 1617 events) - [COUNTERS] PROGRAM TOTAL : 0.4287s - [COUNTERS] Fortran Overhead ( 0 ) : 0.4239s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0045s for 8192 events => throughput is 1.83E+06 events/s + [COUNTERS] PROGRAM TOTAL : 0.4077s + [COUNTERS] Fortran Overhead ( 0 ) : 0.4031s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0043s for 8192 events => throughput is 1.93E+06 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0003s *** (2-sse4) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -167,15 +227,50 @@ OK! xsec from fortran (0.30449452343426120) and cpp (0.30449453160892032) differ OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-sse4) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +2 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.sse4_m_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_susyggt1t1_x10_cudacpp > /tmp/avalassi/output_susyggt1t1_x10_cudacpp' +DEBUG: MEK processed 81920 events across 6 channels { 3 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 4/4 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 2 + [XSECTION] ChannelId = 3 + [XSECTION] Cross section = 0.3075 [0.30747311535940236] fbridge_mode=1 + [UNWEIGHT] Wrote 1631 events (found 1636 events) + [COUNTERS] PROGRAM TOTAL : 1.6130s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5699s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0428s for 81920 events => throughput is 1.91E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0003s + +*** (2-sse4) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (0.30747310722207288) and cpp (0.30747311535940236) differ by less than 2E-4 (2.6465174718381945e-08) + +*** (2-sse4) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_MSSM_SLHA2_GG_T1T1X_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+BRDHST/sse4+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.892977e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.928131e+06 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_MSSM_SLHA2_GG_T1T1X_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.974211e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.961757e+06 ) sec^-1 *** (2-avx2) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -199,9 +294,9 @@ DEBUG: MEK processed 8192 events across 6 channels { 3 : 8192 } [XSECTION] ChannelId = 3 [XSECTION] Cross section = 0.3045 [0.30449453255288433] fbridge_mode=1 [UNWEIGHT] Wrote 1612 events (found 1617 events) - [COUNTERS] PROGRAM TOTAL : 0.4216s - [COUNTERS] Fortran Overhead ( 0 ) : 0.4183s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0029s for 8192 events => throughput is 2.84E+06 events/s + [COUNTERS] PROGRAM TOTAL : 0.4095s + [COUNTERS] Fortran Overhead ( 0 ) : 0.4064s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0027s for 8192 events => throughput is 3.03E+06 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0004s *** (2-avx2) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -212,15 +307,50 @@ OK! xsec from fortran (0.30449452343426120) and cpp (0.30449453255288433) differ OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-avx2) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +2 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.avx2_m_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_susyggt1t1_x10_cudacpp > /tmp/avalassi/output_susyggt1t1_x10_cudacpp' +DEBUG: MEK processed 81920 events across 6 channels { 3 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 4/4 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 2 + [XSECTION] ChannelId = 3 + [XSECTION] Cross section = 0.3075 [0.30747311619894635] fbridge_mode=1 + [UNWEIGHT] Wrote 1631 events (found 1636 events) + [COUNTERS] PROGRAM TOTAL : 1.5942s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5670s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0268s for 81920 events => throughput is 3.06E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0004s + +*** (2-avx2) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (0.30747310722207288) and cpp (0.30747311619894635) differ by less than 2E-4 (2.9195637685219822e-08) + +*** (2-avx2) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_MSSM_SLHA2_GG_T1T1X_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+BRDHST/avx2+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.237521e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.239411e+06 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_MSSM_SLHA2_GG_T1T1X_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.477152e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.406770e+06 ) sec^-1 *** (2-512y) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -244,10 +374,10 @@ DEBUG: MEK processed 8192 events across 6 channels { 3 : 8192 } [XSECTION] ChannelId = 3 [XSECTION] Cross section = 0.3045 [0.30449453255288433] fbridge_mode=1 [UNWEIGHT] Wrote 1612 events (found 1617 events) - [COUNTERS] PROGRAM TOTAL : 0.4270s - [COUNTERS] Fortran Overhead ( 0 ) : 0.4239s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0027s for 8192 events => throughput is 3.00E+06 events/s - [COUNTERS] CudaCpp HEL ( 3 ) : 0.0004s + [COUNTERS] PROGRAM TOTAL : 0.4110s + [COUNTERS] Fortran Overhead ( 0 ) : 0.4081s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0025s for 8192 events => throughput is 3.27E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0003s *** (2-512y) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -257,15 +387,50 @@ OK! xsec from fortran (0.30449452343426120) and cpp (0.30449453255288433) differ OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-512y) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +2 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.512y_m_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_susyggt1t1_x10_cudacpp > /tmp/avalassi/output_susyggt1t1_x10_cudacpp' +DEBUG: MEK processed 81920 events across 6 channels { 3 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 4/4 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 2 + [XSECTION] ChannelId = 3 + [XSECTION] Cross section = 0.3075 [0.30747311619894635] fbridge_mode=1 + [UNWEIGHT] Wrote 1631 events (found 1636 events) + [COUNTERS] PROGRAM TOTAL : 1.5814s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5560s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0250s for 81920 events => throughput is 3.27E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0003s + +*** (2-512y) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (0.30747310722207288) and cpp (0.30747311619894635) differ by less than 2E-4 (2.9195637685219822e-08) + +*** (2-512y) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_MSSM_SLHA2_GG_T1T1X_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+BRDHST/512y+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.311234e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.378245e+06 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_MSSM_SLHA2_GG_T1T1X_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.507028e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.598418e+06 ) sec^-1 *** (2-512z) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -289,9 +454,9 @@ DEBUG: MEK processed 8192 events across 6 channels { 3 : 8192 } [XSECTION] ChannelId = 3 [XSECTION] Cross section = 0.3045 [0.30449453255288433] fbridge_mode=1 [UNWEIGHT] Wrote 1612 events (found 1617 events) - [COUNTERS] PROGRAM TOTAL : 0.4247s - [COUNTERS] Fortran Overhead ( 0 ) : 0.4212s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0032s for 8192 events => throughput is 2.59E+06 events/s + [COUNTERS] PROGRAM TOTAL : 0.4117s + [COUNTERS] Fortran Overhead ( 0 ) : 0.4085s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0028s for 8192 events => throughput is 2.91E+06 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0004s *** (2-512z) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -302,15 +467,50 @@ OK! xsec from fortran (0.30449452343426120) and cpp (0.30449453255288433) differ OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-512z) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +2 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.512z_m_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_susyggt1t1_x10_cudacpp > /tmp/avalassi/output_susyggt1t1_x10_cudacpp' +DEBUG: MEK processed 81920 events across 6 channels { 3 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 4/4 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 2 + [XSECTION] ChannelId = 3 + [XSECTION] Cross section = 0.3075 [0.30747311619894635] fbridge_mode=1 + [UNWEIGHT] Wrote 1631 events (found 1636 events) + [COUNTERS] PROGRAM TOTAL : 1.6143s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5830s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0309s for 81920 events => throughput is 2.65E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0004s + +*** (2-512z) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (0.30747310722207288) and cpp (0.30747311619894635) differ by less than 2E-4 (2.9195637685219822e-08) + +*** (2-512z) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_MSSM_SLHA2_GG_T1T1X_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+BRDHST/512z+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 2.926715e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.930903e+06 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_MSSM_SLHA2_GG_T1T1X_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.198931e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.192075e+06 ) sec^-1 *** (3-cuda) EXECUTE MADEVENT_CUDA x1 (create events.lhe) *** -------------------- @@ -334,9 +534,9 @@ DEBUG: MEK processed 8192 events across 6 channels { 3 : 8192 } [XSECTION] ChannelId = 3 [XSECTION] Cross section = 0.3045 [0.30449452360186230] fbridge_mode=1 [UNWEIGHT] Wrote 1612 events (found 1617 events) - [COUNTERS] PROGRAM TOTAL : 0.8627s - [COUNTERS] Fortran Overhead ( 0 ) : 0.8591s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0030s for 8192 events => throughput is 2.69E+06 events/s + [COUNTERS] PROGRAM TOTAL : 0.8460s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8426s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0029s for 8192 events => throughput is 2.79E+06 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0005s *** (3-cuda) Compare MADEVENT_CUDA x1 xsec to MADEVENT_FORTRAN xsec *** @@ -347,45 +547,80 @@ OK! xsec from fortran (0.30449452343426120) and cuda (0.30449452360186230) diffe OK! events.lhe.cuda.1 and events.lhe.ref.1 are identical +*** (3-cuda) EXECUTE MADEVENT_CUDA x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +2 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.cuda_m_inl0_hrd0/madevent_cuda < /tmp/avalassi/input_susyggt1t1_x10_cudacpp > /tmp/avalassi/output_susyggt1t1_x10_cudacpp' +DEBUG: MEK processed 81920 events across 6 channels { 3 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 4/4 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 2 + [XSECTION] ChannelId = 3 + [XSECTION] Cross section = 0.3075 [0.30747310720557364] fbridge_mode=1 + [UNWEIGHT] Wrote 1631 events (found 1636 events) + [COUNTERS] PROGRAM TOTAL : 2.0019s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9938s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0076s for 81920 events => throughput is 1.08E+07 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0005s + +*** (3-cuda) Compare MADEVENT_CUDA x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (0.30747310722207288) and cuda (0.30747310720557364) differ by less than 2E-4 (5.366074251611508e-11) + +*** (3-cuda) Compare MADEVENT_CUDA x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cuda.10 and events.lhe.ref.10 are identical + *** EXECUTE GCHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_MSSM_SLHA2_GG_T1T1X_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:MIX+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.206349e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.240557e+06 ) sec^-1 *** EXECUTE GCHECK(8192) -p 256 32 1 *** Process = SIGMA_MSSM_SLHA2_GG_T1T1X_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:MIX+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.536038e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.675592e+06 ) sec^-1 *** EXECUTE GCHECK(MAX) -p 16384 32 1 --bridge *** Process = SIGMA_MSSM_SLHA2_GG_T1T1X_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:MIX+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.506637e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.606272e+07 ) sec^-1 *** EXECUTE GCHECK(MAX) -p 16384 32 1 *** Process = SIGMA_MSSM_SLHA2_GG_T1T1X_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:MIX+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.900315e+08 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.966350e+08 ) sec^-1 *** EXECUTE GCHECK(MAX128THR) -p 4096 128 1 --bridge *** Process = SIGMA_MSSM_SLHA2_GG_T1T1X_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:MIX+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.486873e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.545591e+07 ) sec^-1 *** EXECUTE GCHECK(MAX128THR) -p 4096 128 1 *** Process = SIGMA_MSSM_SLHA2_GG_T1T1X_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:MIX+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.921916e+08 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.931391e+08 ) sec^-1 *** EXECUTE GCHECK(MAX8THR) -p 65536 8 1 --bridge *** Process = SIGMA_MSSM_SLHA2_GG_T1T1X_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:MIX+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.466467e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.585991e+07 ) sec^-1 *** EXECUTE GCHECK(MAX8THR) -p 65536 8 1 *** Process = SIGMA_MSSM_SLHA2_GG_T1T1X_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:MIX+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.235205e+08 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.263289e+08 ) sec^-1 *** (3-hip) WARNING! SKIP MADEVENT_HIP (hip is not supported on this node) *** diff --git a/epochX/cudacpp/tmad/logs_susyggtt_mad/log_susyggtt_mad_d_inl0_hrd0.txt b/epochX/cudacpp/tmad/logs_susyggtt_mad/log_susyggtt_mad_d_inl0_hrd0.txt index bebebe43ae..4561a2da73 100644 --- a/epochX/cudacpp/tmad/logs_susyggtt_mad/log_susyggtt_mad_d_inl0_hrd0.txt +++ b/epochX/cudacpp/tmad/logs_susyggtt_mad/log_susyggtt_mad_d_inl0_hrd0.txt @@ -1,13 +1,13 @@ Working directory (build): /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_tt.mad/SubProcesses/P1_gg_ttx + make USEBUILDDIR=1 BACKEND=cuda make USEBUILDDIR=1 BACKEND=cppnone - - make USEBUILDDIR=1 BACKEND=cppsse4 make USEBUILDDIR=1 BACKEND=cppavx2 + make USEBUILDDIR=1 BACKEND=cpp512y make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_tt.mad/SubProcesses/P1_gg_ttx' make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_tt.mad/SubProcesses/P1_gg_ttx' @@ -32,7 +32,7 @@ make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ OMP_NUM_THREADS= -DATE: 2024-10-06_10:55:13 +DATE: 2024-10-06_18:08:14 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: Working directory (run): /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_tt.mad/SubProcesses/P1_gg_ttx @@ -58,9 +58,9 @@ Executing ' ./madevent_fortran < /tmp/avalassi/input_susyggtt_x1_fortran > /tmp/ [XSECTION] ChannelId = 1 [XSECTION] Cross section = 44.64 [44.641911695846957] fbridge_mode=0 [UNWEIGHT] Wrote 2625 events (found 5368 events) - [COUNTERS] PROGRAM TOTAL : 0.8496s - [COUNTERS] Fortran Overhead ( 0 ) : 0.8074s - [COUNTERS] Fortran MEs ( 1 ) : 0.0422s for 8192 events => throughput is 1.94E+05 events/s + [COUNTERS] PROGRAM TOTAL : 0.8197s + [COUNTERS] Fortran Overhead ( 0 ) : 0.7787s + [COUNTERS] Fortran MEs ( 1 ) : 0.0410s for 8192 events => throughput is 2.00E+05 events/s *** (1) EXECUTE MADEVENT_FORTRAN x1 (create events.lhe) *** -------------------- @@ -83,9 +83,34 @@ Executing ' ./madevent_fortran < /tmp/avalassi/input_susyggtt_x1_fortran > /tmp/ [XSECTION] ChannelId = 1 [XSECTION] Cross section = 44.64 [44.641911695846957] fbridge_mode=0 [UNWEIGHT] Wrote 1617 events (found 1622 events) - [COUNTERS] PROGRAM TOTAL : 0.4529s - [COUNTERS] Fortran Overhead ( 0 ) : 0.4105s - [COUNTERS] Fortran MEs ( 1 ) : 0.0424s for 8192 events => throughput is 1.93E+05 events/s + [COUNTERS] PROGRAM TOTAL : 0.4373s + [COUNTERS] Fortran Overhead ( 0 ) : 0.3965s + [COUNTERS] Fortran MEs ( 1 ) : 0.0408s for 8192 events => throughput is 2.01E+05 events/s + +*** (1) EXECUTE MADEVENT_FORTRAN x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./madevent_fortran < /tmp/avalassi/input_susyggtt_x10_fortran > /tmp/avalassi/output_susyggtt_x10_fortran' + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 16/16 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 44.47 [44.473264592444671] fbridge_mode=0 + [UNWEIGHT] Wrote 1622 events (found 1627 events) + [COUNTERS] PROGRAM TOTAL : 1.9419s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5348s + [COUNTERS] Fortran MEs ( 1 ) : 0.4071s for 81920 events => throughput is 2.01E+05 events/s *** (2-none) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -109,9 +134,9 @@ DEBUG: MEK processed 8192 events across 3 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 44.64 [44.641911695846964] fbridge_mode=1 [UNWEIGHT] Wrote 1617 events (found 1622 events) - [COUNTERS] PROGRAM TOTAL : 0.4621s - [COUNTERS] Fortran Overhead ( 0 ) : 0.4167s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0450s for 8192 events => throughput is 1.82E+05 events/s + [COUNTERS] PROGRAM TOTAL : 0.4492s + [COUNTERS] Fortran Overhead ( 0 ) : 0.4055s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0433s for 8192 events => throughput is 1.89E+05 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0004s *** (2-none) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -122,15 +147,50 @@ OK! xsec from fortran (44.641911695846957) and cpp (44.641911695846964) differ b OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-none) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.none_d_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_susyggtt_x10_cudacpp > /tmp/avalassi/output_susyggtt_x10_cudacpp' +DEBUG: MEK processed 81920 events across 3 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 16/16 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 44.47 [44.473264592444664] fbridge_mode=1 + [UNWEIGHT] Wrote 1622 events (found 1627 events) + [COUNTERS] PROGRAM TOTAL : 1.9887s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5545s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4337s for 81920 events => throughput is 1.89E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0004s + +*** (2-none) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (44.473264592444671) and cpp (44.473264592444664) differ by less than 3E-14 (1.1102230246251565e-16) + +*** (2-none) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_MSSM_SLHA2_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+BRDHST/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.859940e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.887580e+05 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_MSSM_SLHA2_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.839978e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.916192e+05 ) sec^-1 *** (2-sse4) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -154,10 +214,10 @@ DEBUG: MEK processed 8192 events across 3 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 44.64 [44.641911695846957] fbridge_mode=1 [UNWEIGHT] Wrote 1617 events (found 1622 events) - [COUNTERS] PROGRAM TOTAL : 0.4403s - [COUNTERS] Fortran Overhead ( 0 ) : 0.4153s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0246s for 8192 events => throughput is 3.33E+05 events/s - [COUNTERS] CudaCpp HEL ( 3 ) : 0.0004s + [COUNTERS] PROGRAM TOTAL : 0.4310s + [COUNTERS] Fortran Overhead ( 0 ) : 0.4056s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0250s for 8192 events => throughput is 3.27E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0003s *** (2-sse4) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -167,15 +227,50 @@ OK! xsec from fortran (44.641911695846957) and cpp (44.641911695846957) differ b OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-sse4) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.sse4_d_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_susyggtt_x10_cudacpp > /tmp/avalassi/output_susyggtt_x10_cudacpp' +DEBUG: MEK processed 81920 events across 3 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 16/16 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 44.47 [44.473264592444671] fbridge_mode=1 + [UNWEIGHT] Wrote 1622 events (found 1627 events) + [COUNTERS] PROGRAM TOTAL : 1.7867s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5432s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2431s for 81920 events => throughput is 3.37E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0004s + +*** (2-sse4) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (44.473264592444671) and cpp (44.473264592444671) differ by less than 3E-14 (0.0) + +*** (2-sse4) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_MSSM_SLHA2_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+BRDHST/sse4+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.243144e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.310196e+05 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_MSSM_SLHA2_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.273347e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.294349e+05 ) sec^-1 *** (2-avx2) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -199,10 +294,10 @@ DEBUG: MEK processed 8192 events across 3 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 44.64 [44.641911695846950] fbridge_mode=1 [UNWEIGHT] Wrote 1617 events (found 1622 events) - [COUNTERS] PROGRAM TOTAL : 0.4324s - [COUNTERS] Fortran Overhead ( 0 ) : 0.4161s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0159s for 8192 events => throughput is 5.14E+05 events/s - [COUNTERS] CudaCpp HEL ( 3 ) : 0.0004s + [COUNTERS] PROGRAM TOTAL : 0.4215s + [COUNTERS] Fortran Overhead ( 0 ) : 0.4055s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0157s for 8192 events => throughput is 5.23E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0003s *** (2-avx2) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -212,15 +307,50 @@ OK! xsec from fortran (44.641911695846957) and cpp (44.641911695846950) differ b OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-avx2) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.avx2_d_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_susyggtt_x10_cudacpp > /tmp/avalassi/output_susyggtt_x10_cudacpp' +DEBUG: MEK processed 81920 events across 3 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 16/16 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 44.47 [44.473264592444679] fbridge_mode=1 + [UNWEIGHT] Wrote 1622 events (found 1627 events) + [COUNTERS] PROGRAM TOTAL : 1.7045s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5519s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1522s for 81920 events => throughput is 5.38E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0003s + +*** (2-avx2) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (44.473264592444671) and cpp (44.473264592444679) differ by less than 3E-14 (2.220446049250313e-16) + +*** (2-avx2) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_MSSM_SLHA2_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+BRDHST/avx2+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 5.210364e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 5.304681e+05 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_MSSM_SLHA2_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 5.310117e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 5.417189e+05 ) sec^-1 *** (2-512y) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -244,9 +374,9 @@ DEBUG: MEK processed 8192 events across 3 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 44.64 [44.641911695846950] fbridge_mode=1 [UNWEIGHT] Wrote 1617 events (found 1622 events) - [COUNTERS] PROGRAM TOTAL : 0.4350s - [COUNTERS] Fortran Overhead ( 0 ) : 0.4201s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0146s for 8192 events => throughput is 5.62E+05 events/s + [COUNTERS] PROGRAM TOTAL : 0.4180s + [COUNTERS] Fortran Overhead ( 0 ) : 0.4036s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0140s for 8192 events => throughput is 5.83E+05 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0003s *** (2-512y) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -257,15 +387,50 @@ OK! xsec from fortran (44.641911695846957) and cpp (44.641911695846950) differ b OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-512y) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.512y_d_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_susyggtt_x10_cudacpp > /tmp/avalassi/output_susyggtt_x10_cudacpp' +DEBUG: MEK processed 81920 events across 3 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 16/16 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 44.47 [44.473264592444679] fbridge_mode=1 + [UNWEIGHT] Wrote 1622 events (found 1627 events) + [COUNTERS] PROGRAM TOTAL : 1.7055s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5641s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1411s for 81920 events => throughput is 5.80E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0003s + +*** (2-512y) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (44.473264592444671) and cpp (44.473264592444679) differ by less than 3E-14 (2.220446049250313e-16) + +*** (2-512y) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_MSSM_SLHA2_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+BRDHST/512y+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 5.704117e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 5.709971e+05 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_MSSM_SLHA2_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 5.793092e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 5.838608e+05 ) sec^-1 *** (2-512z) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -289,9 +454,9 @@ DEBUG: MEK processed 8192 events across 3 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 44.64 [44.641911695846950] fbridge_mode=1 [UNWEIGHT] Wrote 1617 events (found 1622 events) - [COUNTERS] PROGRAM TOTAL : 0.4434s - [COUNTERS] Fortran Overhead ( 0 ) : 0.4195s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0234s for 8192 events => throughput is 3.49E+05 events/s + [COUNTERS] PROGRAM TOTAL : 0.4364s + [COUNTERS] Fortran Overhead ( 0 ) : 0.4135s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0225s for 8192 events => throughput is 3.64E+05 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0004s *** (2-512z) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -302,15 +467,50 @@ OK! xsec from fortran (44.641911695846957) and cpp (44.641911695846950) differ b OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-512z) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.512z_d_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_susyggtt_x10_cudacpp > /tmp/avalassi/output_susyggtt_x10_cudacpp' +DEBUG: MEK processed 81920 events across 3 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 16/16 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 44.47 [44.473264592444671] fbridge_mode=1 + [UNWEIGHT] Wrote 1622 events (found 1627 events) + [COUNTERS] PROGRAM TOTAL : 1.7719s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5544s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2171s for 81920 events => throughput is 3.77E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0004s + +*** (2-512z) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (44.473264592444671) and cpp (44.473264592444671) differ by less than 3E-14 (0.0) + +*** (2-512z) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_MSSM_SLHA2_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+BRDHST/512z+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.552376e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.654419e+05 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_MSSM_SLHA2_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.639783e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.603592e+05 ) sec^-1 *** (3-cuda) EXECUTE MADEVENT_CUDA x1 (create events.lhe) *** -------------------- @@ -334,10 +534,10 @@ DEBUG: MEK processed 8192 events across 3 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 44.64 [44.641911695846950] fbridge_mode=1 [UNWEIGHT] Wrote 1617 events (found 1622 events) - [COUNTERS] PROGRAM TOTAL : 0.8650s - [COUNTERS] Fortran Overhead ( 0 ) : 0.8612s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0031s for 8192 events => throughput is 2.66E+06 events/s - [COUNTERS] CudaCpp HEL ( 3 ) : 0.0008s + [COUNTERS] PROGRAM TOTAL : 0.8411s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8373s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0030s for 8192 events => throughput is 2.73E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0007s *** (3-cuda) Compare MADEVENT_CUDA x1 xsec to MADEVENT_FORTRAN xsec *** @@ -347,45 +547,80 @@ OK! xsec from fortran (44.641911695846957) and cuda (44.641911695846950) differ OK! events.lhe.cuda.1 and events.lhe.ref.1 are identical +*** (3-cuda) EXECUTE MADEVENT_CUDA x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.cuda_d_inl0_hrd0/madevent_cuda < /tmp/avalassi/input_susyggtt_x10_cudacpp > /tmp/avalassi/output_susyggtt_x10_cudacpp' +DEBUG: MEK processed 81920 events across 3 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 16/16 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 44.47 [44.473264592444679] fbridge_mode=1 + [UNWEIGHT] Wrote 1622 events (found 1627 events) + [COUNTERS] PROGRAM TOTAL : 1.9948s + [COUNTERS] Fortran Overhead ( 0 ) : 1.9853s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0088s for 81920 events => throughput is 9.35E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0008s + +*** (3-cuda) Compare MADEVENT_CUDA x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (44.473264592444671) and cuda (44.473264592444679) differ by less than 3E-14 (2.220446049250313e-16) + +*** (3-cuda) Compare MADEVENT_CUDA x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cuda.10 and events.lhe.ref.10 are identical + *** EXECUTE GCHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_MSSM_SLHA2_GG_TTX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.043338e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.105270e+06 ) sec^-1 *** EXECUTE GCHECK(8192) -p 256 32 1 *** Process = SIGMA_MSSM_SLHA2_GG_TTX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.325784e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.424891e+06 ) sec^-1 *** EXECUTE GCHECK(MAX) -p 16384 32 1 --bridge *** Process = SIGMA_MSSM_SLHA2_GG_TTX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 2.871559e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.916373e+07 ) sec^-1 *** EXECUTE GCHECK(MAX) -p 16384 32 1 *** Process = SIGMA_MSSM_SLHA2_GG_TTX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 6.143094e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 7.628699e+07 ) sec^-1 *** EXECUTE GCHECK(MAX128THR) -p 4096 128 1 --bridge *** Process = SIGMA_MSSM_SLHA2_GG_TTX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 2.865333e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.932895e+07 ) sec^-1 *** EXECUTE GCHECK(MAX128THR) -p 4096 128 1 *** Process = SIGMA_MSSM_SLHA2_GG_TTX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 8.020534e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 8.036591e+07 ) sec^-1 *** EXECUTE GCHECK(MAX8THR) -p 65536 8 1 --bridge *** Process = SIGMA_MSSM_SLHA2_GG_TTX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 2.868423e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.908058e+07 ) sec^-1 *** EXECUTE GCHECK(MAX8THR) -p 65536 8 1 *** Process = SIGMA_MSSM_SLHA2_GG_TTX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 2.708181e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.714200e+07 ) sec^-1 *** (3-hip) WARNING! SKIP MADEVENT_HIP (hip is not supported on this node) *** diff --git a/epochX/cudacpp/tmad/logs_susyggtt_mad/log_susyggtt_mad_f_inl0_hrd0.txt b/epochX/cudacpp/tmad/logs_susyggtt_mad/log_susyggtt_mad_f_inl0_hrd0.txt index 2a76a737ac..24f573d0f0 100644 --- a/epochX/cudacpp/tmad/logs_susyggtt_mad/log_susyggtt_mad_f_inl0_hrd0.txt +++ b/epochX/cudacpp/tmad/logs_susyggtt_mad/log_susyggtt_mad_f_inl0_hrd0.txt @@ -1,13 +1,14 @@ Working directory (build): /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_tt.mad/SubProcesses/P1_gg_ttx -make USEBUILDDIR=1 BACKEND=cuda make USEBUILDDIR=1 BACKEND=cppnone - - make USEBUILDDIR=1 BACKEND=cppsse4 + make USEBUILDDIR=1 BACKEND=cppavx2 + +make USEBUILDDIR=1 BACKEND=cuda +make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_tt.mad/SubProcesses/P1_gg_ttx' make USEBUILDDIR=1 BACKEND=cpp512y make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_tt.mad/SubProcesses/P1_gg_ttx' make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_tt.mad/SubProcesses/P1_gg_ttx' @@ -16,7 +17,6 @@ make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp make USEBUILDDIR=1 BACKEND=cpp512z make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_tt.mad/SubProcesses/P1_gg_ttx' -make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_tt.mad/SubProcesses/P1_gg_ttx' make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_tt.mad/SubProcesses/P1_gg_ttx' make[1]: Nothing to be done for 'all'. @@ -32,7 +32,7 @@ make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ OMP_NUM_THREADS= -DATE: 2024-10-06_10:55:42 +DATE: 2024-10-06_18:09:10 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: Working directory (run): /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_tt.mad/SubProcesses/P1_gg_ttx @@ -58,9 +58,9 @@ Executing ' ./madevent_fortran < /tmp/avalassi/input_susyggtt_x1_fortran > /tmp/ [XSECTION] ChannelId = 1 [XSECTION] Cross section = 44.64 [44.641911695846957] fbridge_mode=0 [UNWEIGHT] Wrote 2625 events (found 5368 events) - [COUNTERS] PROGRAM TOTAL : 0.8397s - [COUNTERS] Fortran Overhead ( 0 ) : 0.7971s - [COUNTERS] Fortran MEs ( 1 ) : 0.0425s for 8192 events => throughput is 1.93E+05 events/s + [COUNTERS] PROGRAM TOTAL : 0.8131s + [COUNTERS] Fortran Overhead ( 0 ) : 0.7723s + [COUNTERS] Fortran MEs ( 1 ) : 0.0408s for 8192 events => throughput is 2.01E+05 events/s *** (1) EXECUTE MADEVENT_FORTRAN x1 (create events.lhe) *** -------------------- @@ -83,9 +83,34 @@ Executing ' ./madevent_fortran < /tmp/avalassi/input_susyggtt_x1_fortran > /tmp/ [XSECTION] ChannelId = 1 [XSECTION] Cross section = 44.64 [44.641911695846957] fbridge_mode=0 [UNWEIGHT] Wrote 1617 events (found 1622 events) - [COUNTERS] PROGRAM TOTAL : 0.4553s - [COUNTERS] Fortran Overhead ( 0 ) : 0.4122s - [COUNTERS] Fortran MEs ( 1 ) : 0.0430s for 8192 events => throughput is 1.90E+05 events/s + [COUNTERS] PROGRAM TOTAL : 0.4409s + [COUNTERS] Fortran Overhead ( 0 ) : 0.3997s + [COUNTERS] Fortran MEs ( 1 ) : 0.0412s for 8192 events => throughput is 1.99E+05 events/s + +*** (1) EXECUTE MADEVENT_FORTRAN x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./madevent_fortran < /tmp/avalassi/input_susyggtt_x10_fortran > /tmp/avalassi/output_susyggtt_x10_fortran' + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 16/16 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 44.47 [44.473264592444671] fbridge_mode=0 + [UNWEIGHT] Wrote 1622 events (found 1627 events) + [COUNTERS] PROGRAM TOTAL : 1.9434s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5341s + [COUNTERS] Fortran MEs ( 1 ) : 0.4094s for 81920 events => throughput is 2.00E+05 events/s *** (2-none) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -109,9 +134,9 @@ DEBUG: MEK processed 8192 events across 3 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 44.64 [44.641906072918047] fbridge_mode=1 [UNWEIGHT] Wrote 1617 events (found 1622 events) - [COUNTERS] PROGRAM TOTAL : 0.4624s - [COUNTERS] Fortran Overhead ( 0 ) : 0.4193s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0429s for 8192 events => throughput is 1.91E+05 events/s + [COUNTERS] PROGRAM TOTAL : 0.4438s + [COUNTERS] Fortran Overhead ( 0 ) : 0.4027s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0408s for 8192 events => throughput is 2.01E+05 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0003s *** (2-none) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -122,15 +147,50 @@ OK! xsec from fortran (44.641911695846957) and cpp (44.641906072918047) differ b OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-none) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.none_f_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_susyggtt_x10_cudacpp > /tmp/avalassi/output_susyggtt_x10_cudacpp' +DEBUG: MEK processed 81920 events across 3 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 16/16 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 44.47 [44.473258789404959] fbridge_mode=1 + [UNWEIGHT] Wrote 1622 events (found 1627 events) + [COUNTERS] PROGRAM TOTAL : 1.9551s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5502s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4046s for 81920 events => throughput is 2.02E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0003s + +*** (2-none) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (44.473264592444671) and cpp (44.473258789404959) differ by less than 4E-4 (1.3048378089131063e-07) + +*** (2-none) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_MSSM_SLHA2_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+BRDHST/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.972969e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.021962e+05 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_MSSM_SLHA2_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.987350e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.997473e+05 ) sec^-1 *** (2-sse4) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -154,10 +214,10 @@ DEBUG: MEK processed 8192 events across 3 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 44.64 [44.641902189470080] fbridge_mode=1 [UNWEIGHT] Wrote 1617 events (found 1622 events) - [COUNTERS] PROGRAM TOTAL : 0.4356s - [COUNTERS] Fortran Overhead ( 0 ) : 0.4182s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0171s for 8192 events => throughput is 4.78E+05 events/s - [COUNTERS] CudaCpp HEL ( 3 ) : 0.0003s + [COUNTERS] PROGRAM TOTAL : 0.4198s + [COUNTERS] Fortran Overhead ( 0 ) : 0.4031s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0165s for 8192 events => throughput is 4.95E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0002s *** (2-sse4) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -167,15 +227,50 @@ OK! xsec from fortran (44.641911695846957) and cpp (44.641902189470080) differ b OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-sse4) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.sse4_f_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_susyggtt_x10_cudacpp > /tmp/avalassi/output_susyggtt_x10_cudacpp' +DEBUG: MEK processed 81920 events across 3 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 16/16 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 44.47 [44.473255074265531] fbridge_mode=1 + [UNWEIGHT] Wrote 1622 events (found 1627 events) + [COUNTERS] PROGRAM TOTAL : 1.7390s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5706s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1682s for 81920 events => throughput is 4.87E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0002s + +*** (2-sse4) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (44.473264592444671) and cpp (44.473255074265531) differ by less than 4E-4 (2.1402024852346102e-07) + +*** (2-sse4) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_MSSM_SLHA2_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+BRDHST/sse4+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 4.748983e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 4.765231e+05 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_MSSM_SLHA2_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 4.695429e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 4.735495e+05 ) sec^-1 *** (2-avx2) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -199,9 +294,9 @@ DEBUG: MEK processed 8192 events across 3 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 44.64 [44.641902360436738] fbridge_mode=1 [UNWEIGHT] Wrote 1617 events (found 1622 events) - [COUNTERS] PROGRAM TOTAL : 0.4310s - [COUNTERS] Fortran Overhead ( 0 ) : 0.4214s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0093s for 8192 events => throughput is 8.83E+05 events/s + [COUNTERS] PROGRAM TOTAL : 0.4136s + [COUNTERS] Fortran Overhead ( 0 ) : 0.4047s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0086s for 8192 events => throughput is 9.50E+05 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0002s *** (2-avx2) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -212,15 +307,50 @@ OK! xsec from fortran (44.641911695846957) and cpp (44.641902360436738) differ b OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-avx2) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.avx2_f_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_susyggtt_x10_cudacpp > /tmp/avalassi/output_susyggtt_x10_cudacpp' +DEBUG: MEK processed 81920 events across 3 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 16/16 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 44.47 [44.473254628666531] fbridge_mode=1 + [UNWEIGHT] Wrote 1622 events (found 1627 events) + [COUNTERS] PROGRAM TOTAL : 1.6391s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5518s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0871s for 81920 events => throughput is 9.41E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0002s + +*** (2-avx2) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (44.473264592444671) and cpp (44.473254628666531) differ by less than 4E-4 (2.240397288799656e-07) + +*** (2-avx2) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_MSSM_SLHA2_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+BRDHST/avx2+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 9.169652e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 9.354796e+05 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_MSSM_SLHA2_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 9.239468e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 9.227206e+05 ) sec^-1 *** (2-512y) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -244,9 +374,9 @@ DEBUG: MEK processed 8192 events across 3 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 44.64 [44.641902360436738] fbridge_mode=1 [UNWEIGHT] Wrote 1617 events (found 1622 events) - [COUNTERS] PROGRAM TOTAL : 0.4221s - [COUNTERS] Fortran Overhead ( 0 ) : 0.4132s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0086s for 8192 events => throughput is 9.50E+05 events/s + [COUNTERS] PROGRAM TOTAL : 0.4109s + [COUNTERS] Fortran Overhead ( 0 ) : 0.4025s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0082s for 8192 events => throughput is 1.00E+06 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0002s *** (2-512y) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -257,15 +387,50 @@ OK! xsec from fortran (44.641911695846957) and cpp (44.641902360436738) differ b OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-512y) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.512y_f_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_susyggtt_x10_cudacpp > /tmp/avalassi/output_susyggtt_x10_cudacpp' +DEBUG: MEK processed 81920 events across 3 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 16/16 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 44.47 [44.473254628666531] fbridge_mode=1 + [UNWEIGHT] Wrote 1622 events (found 1627 events) + [COUNTERS] PROGRAM TOTAL : 1.6350s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5511s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0836s for 81920 events => throughput is 9.80E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0002s + +*** (2-512y) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (44.473264592444671) and cpp (44.473254628666531) differ by less than 4E-4 (2.240397288799656e-07) + +*** (2-512y) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_MSSM_SLHA2_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+BRDHST/512y+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 9.627165e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 9.894174e+05 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_MSSM_SLHA2_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 9.935546e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 9.843989e+05 ) sec^-1 *** (2-512z) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -289,9 +454,9 @@ DEBUG: MEK processed 8192 events across 3 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 44.64 [44.641906399820272] fbridge_mode=1 [UNWEIGHT] Wrote 1617 events (found 1622 events) - [COUNTERS] PROGRAM TOTAL : 0.4293s - [COUNTERS] Fortran Overhead ( 0 ) : 0.4169s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0121s for 8192 events => throughput is 6.75E+05 events/s + [COUNTERS] PROGRAM TOTAL : 0.4221s + [COUNTERS] Fortran Overhead ( 0 ) : 0.4106s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0111s for 8192 events => throughput is 7.36E+05 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0003s *** (2-512z) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -302,15 +467,50 @@ OK! xsec from fortran (44.641911695846957) and cpp (44.641906399820272) differ b OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-512z) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.512z_f_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_susyggtt_x10_cudacpp > /tmp/avalassi/output_susyggtt_x10_cudacpp' +DEBUG: MEK processed 81920 events across 3 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 16/16 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 44.47 [44.473258854390501] fbridge_mode=1 + [UNWEIGHT] Wrote 1622 events (found 1627 events) + [COUNTERS] PROGRAM TOTAL : 1.6742s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5546s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1193s for 81920 events => throughput is 6.87E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0003s + +*** (2-512z) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (44.473264592444671) and cpp (44.473258854390501) differ by less than 4E-4 (1.2902255375202287e-07) + +*** (2-512z) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_MSSM_SLHA2_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+BRDHST/512z+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 6.774461e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 6.976210e+05 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_MSSM_SLHA2_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 6.994273e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 6.991318e+05 ) sec^-1 *** (3-cuda) EXECUTE MADEVENT_CUDA x1 (create events.lhe) *** -------------------- @@ -334,9 +534,9 @@ DEBUG: MEK processed 8192 events across 3 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 44.64 [44.641910992291372] fbridge_mode=1 [UNWEIGHT] Wrote 1617 events (found 1622 events) - [COUNTERS] PROGRAM TOTAL : 0.8577s - [COUNTERS] Fortran Overhead ( 0 ) : 0.8540s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0030s for 8192 events => throughput is 2.69E+06 events/s + [COUNTERS] PROGRAM TOTAL : 0.8518s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8482s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0030s for 8192 events => throughput is 2.72E+06 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0006s *** (3-cuda) Compare MADEVENT_CUDA x1 xsec to MADEVENT_FORTRAN xsec *** @@ -347,45 +547,80 @@ OK! xsec from fortran (44.641911695846957) and cuda (44.641910992291372) differ OK! events.lhe.cuda.1 and events.lhe.ref.1 are identical +*** (3-cuda) EXECUTE MADEVENT_CUDA x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.cuda_f_inl0_hrd0/madevent_cuda < /tmp/avalassi/input_susyggtt_x10_cudacpp > /tmp/avalassi/output_susyggtt_x10_cudacpp' +DEBUG: MEK processed 81920 events across 3 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 16/16 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 44.47 [44.473262664842089] fbridge_mode=1 + [UNWEIGHT] Wrote 1622 events (found 1627 events) + [COUNTERS] PROGRAM TOTAL : 2.0289s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0202s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0081s for 81920 events => throughput is 1.01E+07 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0006s + +*** (3-cuda) Compare MADEVENT_CUDA x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (44.473264592444671) and cuda (44.473262664842089) differ by less than 4E-4 (4.334295222729878e-08) + +*** (3-cuda) Compare MADEVENT_CUDA x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cuda.10 and events.lhe.ref.10 are identical + *** EXECUTE GCHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_MSSM_SLHA2_GG_TTX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.201092e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.362675e+06 ) sec^-1 *** EXECUTE GCHECK(8192) -p 256 32 1 *** Process = SIGMA_MSSM_SLHA2_GG_TTX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.452650e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.447769e+06 ) sec^-1 *** EXECUTE GCHECK(MAX) -p 16384 32 1 --bridge *** Process = SIGMA_MSSM_SLHA2_GG_TTX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 4.883185e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 5.149699e+07 ) sec^-1 *** EXECUTE GCHECK(MAX) -p 16384 32 1 *** Process = SIGMA_MSSM_SLHA2_GG_TTX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.341479e+08 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.377907e+08 ) sec^-1 *** EXECUTE GCHECK(MAX128THR) -p 4096 128 1 --bridge *** Process = SIGMA_MSSM_SLHA2_GG_TTX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 4.843740e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 5.140712e+07 ) sec^-1 *** EXECUTE GCHECK(MAX128THR) -p 4096 128 1 *** Process = SIGMA_MSSM_SLHA2_GG_TTX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.360831e+08 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.394859e+08 ) sec^-1 *** EXECUTE GCHECK(MAX8THR) -p 65536 8 1 --bridge *** Process = SIGMA_MSSM_SLHA2_GG_TTX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 4.608054e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 4.808864e+07 ) sec^-1 *** EXECUTE GCHECK(MAX8THR) -p 65536 8 1 *** Process = SIGMA_MSSM_SLHA2_GG_TTX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 7.014740e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 7.086331e+07 ) sec^-1 *** (3-hip) WARNING! SKIP MADEVENT_HIP (hip is not supported on this node) *** diff --git a/epochX/cudacpp/tmad/logs_susyggtt_mad/log_susyggtt_mad_m_inl0_hrd0.txt b/epochX/cudacpp/tmad/logs_susyggtt_mad/log_susyggtt_mad_m_inl0_hrd0.txt index 449e459bdc..f3460a2288 100644 --- a/epochX/cudacpp/tmad/logs_susyggtt_mad/log_susyggtt_mad_m_inl0_hrd0.txt +++ b/epochX/cudacpp/tmad/logs_susyggtt_mad/log_susyggtt_mad_m_inl0_hrd0.txt @@ -6,17 +6,17 @@ make USEBUILDDIR=1 BACKEND=cuda make USEBUILDDIR=1 BACKEND=cppnone make USEBUILDDIR=1 BACKEND=cppsse4 - make USEBUILDDIR=1 BACKEND=cppavx2 + make USEBUILDDIR=1 BACKEND=cpp512y make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_tt.mad/SubProcesses/P1_gg_ttx' make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_tt.mad/SubProcesses/P1_gg_ttx' make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_tt.mad/SubProcesses/P1_gg_ttx' make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_tt.mad/SubProcesses/P1_gg_ttx' +make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_tt.mad/SubProcesses/P1_gg_ttx' make USEBUILDDIR=1 BACKEND=cpp512z make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_tt.mad/SubProcesses/P1_gg_ttx' -make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_tt.mad/SubProcesses/P1_gg_ttx' make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_tt.mad/SubProcesses/P1_gg_ttx' make[1]: Nothing to be done for 'all'. @@ -32,7 +32,7 @@ make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ OMP_NUM_THREADS= -DATE: 2024-10-06_10:55:27 +DATE: 2024-10-06_18:08:42 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: Working directory (run): /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_tt.mad/SubProcesses/P1_gg_ttx @@ -58,9 +58,9 @@ Executing ' ./madevent_fortran < /tmp/avalassi/input_susyggtt_x1_fortran > /tmp/ [XSECTION] ChannelId = 1 [XSECTION] Cross section = 44.64 [44.641911695846957] fbridge_mode=0 [UNWEIGHT] Wrote 2625 events (found 5368 events) - [COUNTERS] PROGRAM TOTAL : 0.8469s - [COUNTERS] Fortran Overhead ( 0 ) : 0.8046s - [COUNTERS] Fortran MEs ( 1 ) : 0.0423s for 8192 events => throughput is 1.94E+05 events/s + [COUNTERS] PROGRAM TOTAL : 0.8183s + [COUNTERS] Fortran Overhead ( 0 ) : 0.7777s + [COUNTERS] Fortran MEs ( 1 ) : 0.0406s for 8192 events => throughput is 2.02E+05 events/s *** (1) EXECUTE MADEVENT_FORTRAN x1 (create events.lhe) *** -------------------- @@ -83,9 +83,34 @@ Executing ' ./madevent_fortran < /tmp/avalassi/input_susyggtt_x1_fortran > /tmp/ [XSECTION] ChannelId = 1 [XSECTION] Cross section = 44.64 [44.641911695846957] fbridge_mode=0 [UNWEIGHT] Wrote 1617 events (found 1622 events) - [COUNTERS] PROGRAM TOTAL : 0.4519s - [COUNTERS] Fortran Overhead ( 0 ) : 0.4098s - [COUNTERS] Fortran MEs ( 1 ) : 0.0422s for 8192 events => throughput is 1.94E+05 events/s + [COUNTERS] PROGRAM TOTAL : 0.4424s + [COUNTERS] Fortran Overhead ( 0 ) : 0.4014s + [COUNTERS] Fortran MEs ( 1 ) : 0.0410s for 8192 events => throughput is 2.00E+05 events/s + +*** (1) EXECUTE MADEVENT_FORTRAN x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./madevent_fortran < /tmp/avalassi/input_susyggtt_x10_fortran > /tmp/avalassi/output_susyggtt_x10_fortran' + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 16/16 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 44.47 [44.473264592444671] fbridge_mode=0 + [UNWEIGHT] Wrote 1622 events (found 1627 events) + [COUNTERS] PROGRAM TOTAL : 1.9609s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5453s + [COUNTERS] Fortran MEs ( 1 ) : 0.4156s for 81920 events => throughput is 1.97E+05 events/s *** (2-none) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -109,9 +134,9 @@ DEBUG: MEK processed 8192 events across 3 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 44.64 [44.641912938404218] fbridge_mode=1 [UNWEIGHT] Wrote 1617 events (found 1622 events) - [COUNTERS] PROGRAM TOTAL : 0.4618s - [COUNTERS] Fortran Overhead ( 0 ) : 0.4159s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0455s for 8192 events => throughput is 1.80E+05 events/s + [COUNTERS] PROGRAM TOTAL : 0.4459s + [COUNTERS] Fortran Overhead ( 0 ) : 0.4018s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0437s for 8192 events => throughput is 1.88E+05 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0004s *** (2-none) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -122,15 +147,50 @@ OK! xsec from fortran (44.641911695846957) and cpp (44.641912938404218) differ b OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-none) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.none_m_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_susyggtt_x10_cudacpp > /tmp/avalassi/output_susyggtt_x10_cudacpp' +DEBUG: MEK processed 81920 events across 3 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 16/16 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 44.47 [44.473265850735231] fbridge_mode=1 + [UNWEIGHT] Wrote 1622 events (found 1627 events) + [COUNTERS] PROGRAM TOTAL : 1.9966s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5597s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.4365s for 81920 events => throughput is 1.88E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0004s + +*** (2-none) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (44.473264592444671) and cpp (44.473265850735231) differ by less than 2E-4 (2.8293190679207214e-08) + +*** (2-none) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_MSSM_SLHA2_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+BRDHST/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.833802e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.826841e+05 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_MSSM_SLHA2_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 1.834236e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.884900e+05 ) sec^-1 *** (2-sse4) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -154,10 +214,10 @@ DEBUG: MEK processed 8192 events across 3 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 44.64 [44.641912938404218] fbridge_mode=1 [UNWEIGHT] Wrote 1617 events (found 1622 events) - [COUNTERS] PROGRAM TOTAL : 0.4439s - [COUNTERS] Fortran Overhead ( 0 ) : 0.4188s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0248s for 8192 events => throughput is 3.30E+05 events/s - [COUNTERS] CudaCpp HEL ( 3 ) : 0.0004s + [COUNTERS] PROGRAM TOTAL : 0.4268s + [COUNTERS] Fortran Overhead ( 0 ) : 0.4032s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0233s for 8192 events => throughput is 3.51E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0003s *** (2-sse4) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -167,15 +227,50 @@ OK! xsec from fortran (44.641911695846957) and cpp (44.641912938404218) differ b OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-sse4) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.sse4_m_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_susyggtt_x10_cudacpp > /tmp/avalassi/output_susyggtt_x10_cudacpp' +DEBUG: MEK processed 81920 events across 3 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 16/16 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 44.47 [44.473265850735231] fbridge_mode=1 + [UNWEIGHT] Wrote 1622 events (found 1627 events) + [COUNTERS] PROGRAM TOTAL : 1.7883s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5499s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2381s for 81920 events => throughput is 3.44E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0003s + +*** (2-sse4) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (44.473264592444671) and cpp (44.473265850735231) differ by less than 2E-4 (2.8293190679207214e-08) + +*** (2-sse4) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_MSSM_SLHA2_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+BRDHST/sse4+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.367073e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.406133e+05 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_MSSM_SLHA2_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.340820e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.384785e+05 ) sec^-1 *** (2-avx2) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -199,9 +294,9 @@ DEBUG: MEK processed 8192 events across 3 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 44.64 [44.641912970378179] fbridge_mode=1 [UNWEIGHT] Wrote 1617 events (found 1622 events) - [COUNTERS] PROGRAM TOTAL : 0.4302s - [COUNTERS] Fortran Overhead ( 0 ) : 0.4144s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0154s for 8192 events => throughput is 5.33E+05 events/s + [COUNTERS] PROGRAM TOTAL : 0.4181s + [COUNTERS] Fortran Overhead ( 0 ) : 0.4019s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0158s for 8192 events => throughput is 5.17E+05 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0004s *** (2-avx2) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -212,15 +307,50 @@ OK! xsec from fortran (44.641911695846957) and cpp (44.641912970378179) differ b OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-avx2) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.avx2_m_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_susyggtt_x10_cudacpp > /tmp/avalassi/output_susyggtt_x10_cudacpp' +DEBUG: MEK processed 81920 events across 3 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 16/16 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 44.47 [44.473265889684782] fbridge_mode=1 + [UNWEIGHT] Wrote 1622 events (found 1627 events) + [COUNTERS] PROGRAM TOTAL : 1.6937s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5438s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1496s for 81920 events => throughput is 5.48E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0003s + +*** (2-avx2) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (44.473264592444671) and cpp (44.473265889684782) differ by less than 2E-4 (2.9168987669692115e-08) + +*** (2-avx2) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_MSSM_SLHA2_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+BRDHST/avx2+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 5.283261e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 5.319808e+05 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_MSSM_SLHA2_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 5.353744e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 5.377912e+05 ) sec^-1 *** (2-512y) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -244,9 +374,9 @@ DEBUG: MEK processed 8192 events across 3 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 44.64 [44.641912970378179] fbridge_mode=1 [UNWEIGHT] Wrote 1617 events (found 1622 events) - [COUNTERS] PROGRAM TOTAL : 0.4297s - [COUNTERS] Fortran Overhead ( 0 ) : 0.4151s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0143s for 8192 events => throughput is 5.75E+05 events/s + [COUNTERS] PROGRAM TOTAL : 0.4153s + [COUNTERS] Fortran Overhead ( 0 ) : 0.4014s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0136s for 8192 events => throughput is 6.04E+05 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0003s *** (2-512y) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -257,15 +387,50 @@ OK! xsec from fortran (44.641911695846957) and cpp (44.641912970378179) differ b OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-512y) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.512y_m_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_susyggtt_x10_cudacpp > /tmp/avalassi/output_susyggtt_x10_cudacpp' +DEBUG: MEK processed 81920 events across 3 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 16/16 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 44.47 [44.473265889684782] fbridge_mode=1 + [UNWEIGHT] Wrote 1622 events (found 1627 events) + [COUNTERS] PROGRAM TOTAL : 1.6786s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5421s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.1361s for 81920 events => throughput is 6.02E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0003s + +*** (2-512y) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (44.473264592444671) and cpp (44.473265889684782) differ by less than 2E-4 (2.9168987669692115e-08) + +*** (2-512y) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_MSSM_SLHA2_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+BRDHST/512y+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 5.825518e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 5.840887e+05 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_MSSM_SLHA2_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 5.928231e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 5.966307e+05 ) sec^-1 *** (2-512z) EXECUTE MADEVENT_CPP x1 (create events.lhe) *** -------------------- @@ -289,9 +454,9 @@ DEBUG: MEK processed 8192 events across 3 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 44.64 [44.641912970378179] fbridge_mode=1 [UNWEIGHT] Wrote 1617 events (found 1622 events) - [COUNTERS] PROGRAM TOTAL : 0.4383s - [COUNTERS] Fortran Overhead ( 0 ) : 0.4159s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0220s for 8192 events => throughput is 3.73E+05 events/s + [COUNTERS] PROGRAM TOTAL : 0.4254s + [COUNTERS] Fortran Overhead ( 0 ) : 0.4037s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0213s for 8192 events => throughput is 3.84E+05 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0004s *** (2-512z) Compare MADEVENT_CPP x1 xsec to MADEVENT_FORTRAN xsec *** @@ -302,15 +467,50 @@ OK! xsec from fortran (44.641911695846957) and cpp (44.641912970378179) differ b OK! events.lhe.cpp.1 and events.lhe.ref.1 are identical +*** (2-512z) EXECUTE MADEVENT_CPP x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.512z_m_inl0_hrd0/madevent_cpp < /tmp/avalassi/input_susyggtt_x10_cudacpp > /tmp/avalassi/output_susyggtt_x10_cudacpp' +DEBUG: MEK processed 81920 events across 3 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 16/16 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 44.47 [44.473265889684782] fbridge_mode=1 + [UNWEIGHT] Wrote 1622 events (found 1627 events) + [COUNTERS] PROGRAM TOTAL : 1.7715s + [COUNTERS] Fortran Overhead ( 0 ) : 1.5586s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.2126s for 81920 events => throughput is 3.85E+05 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0004s + +*** (2-512z) Compare MADEVENT_CPP x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (44.473264592444671) and cpp (44.473265889684782) differ by less than 2E-4 (2.9168987669692115e-08) + +*** (2-512z) Compare MADEVENT_CPP x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cpp.10 and events.lhe.ref.10 are identical + *** EXECUTE CHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_MSSM_SLHA2_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+BRDHST/512z+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.615578e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.737651e+05 ) sec^-1 *** EXECUTE CHECK(8192) -p 256 32 1 *** Process = SIGMA_MSSM_SLHA2_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.732261e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.783927e+05 ) sec^-1 *** (3-cuda) EXECUTE MADEVENT_CUDA x1 (create events.lhe) *** -------------------- @@ -334,9 +534,9 @@ DEBUG: MEK processed 8192 events across 3 channels { 1 : 8192 } [XSECTION] ChannelId = 1 [XSECTION] Cross section = 44.64 [44.641911674225568] fbridge_mode=1 [UNWEIGHT] Wrote 1617 events (found 1622 events) - [COUNTERS] PROGRAM TOTAL : 0.8598s - [COUNTERS] Fortran Overhead ( 0 ) : 0.8560s - [COUNTERS] CudaCpp MEs ( 2 ) : 0.0030s for 8192 events => throughput is 2.69E+06 events/s + [COUNTERS] PROGRAM TOTAL : 0.8377s + [COUNTERS] Fortran Overhead ( 0 ) : 0.8340s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0030s for 8192 events => throughput is 2.76E+06 events/s [COUNTERS] CudaCpp HEL ( 3 ) : 0.0007s *** (3-cuda) Compare MADEVENT_CUDA x1 xsec to MADEVENT_FORTRAN xsec *** @@ -347,45 +547,80 @@ OK! xsec from fortran (44.641911695846957) and cuda (44.641911674225568) differ OK! events.lhe.cuda.1 and events.lhe.ref.1 are identical +*** (3-cuda) EXECUTE MADEVENT_CUDA x10 (create events.lhe) *** +-------------------- +CUDACPP_RUNTIME_FBRIDGEMODE = (not set) +CUDACPP_RUNTIME_VECSIZEUSED = 8192 +-------------------- +81920 1 1 ! Number of events and max and min iterations +0.000001 ! Accuracy (ignored because max iterations = min iterations) +0 ! Grid Adjustment 0=none, 2=adjust (NB if = 0, ftn26 will still be used if present) +1 ! Suppress Amplitude 1=yes (i.e. use MadEvent single-diagram enhancement) +0 ! Helicity Sum/event 0=exact +1 ! ICONFIG number (1-N) for single-diagram enhancement multi-channel (NB used even if suppress amplitude is 0!) +-------------------- +Executing ' ./build.cuda_m_inl0_hrd0/madevent_cuda < /tmp/avalassi/input_susyggtt_x10_cudacpp > /tmp/avalassi/output_susyggtt_x10_cudacpp' +DEBUG: MEK processed 81920 events across 3 channels { 1 : 81920 } + [OPENMPTH] omp_get_max_threads/nproc = 1/4 + [NGOODHEL] ngoodhel/ncomb = 16/16 + [XSECTION] VECSIZE_USED = 8192 + [XSECTION] MultiChannel = TRUE + [XSECTION] Configuration = 1 + [XSECTION] ChannelId = 1 + [XSECTION] Cross section = 44.47 [44.473264587763374] fbridge_mode=1 + [UNWEIGHT] Wrote 1622 events (found 1627 events) + [COUNTERS] PROGRAM TOTAL : 2.0098s + [COUNTERS] Fortran Overhead ( 0 ) : 2.0001s + [COUNTERS] CudaCpp MEs ( 2 ) : 0.0090s for 81920 events => throughput is 9.10E+06 events/s + [COUNTERS] CudaCpp HEL ( 3 ) : 0.0007s + +*** (3-cuda) Compare MADEVENT_CUDA x10 xsec to MADEVENT_FORTRAN xsec *** + +OK! xsec from fortran (44.473264592444671) and cuda (44.473264587763374) differ by less than 2E-4 (1.0526091109852587e-10) + +*** (3-cuda) Compare MADEVENT_CUDA x10 events.lhe to MADEVENT_FORTRAN events.lhe reference (including colors and helicities) *** + +OK! events.lhe.cuda.10 and events.lhe.ref.10 are identical + *** EXECUTE GCHECK(8192) -p 256 32 1 --bridge *** Process = SIGMA_MSSM_SLHA2_GG_TTX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:MIX+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 2.907482e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.030484e+06 ) sec^-1 *** EXECUTE GCHECK(8192) -p 256 32 1 *** Process = SIGMA_MSSM_SLHA2_GG_TTX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:MIX+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 3.361691e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.534287e+06 ) sec^-1 *** EXECUTE GCHECK(MAX) -p 16384 32 1 --bridge *** Process = SIGMA_MSSM_SLHA2_GG_TTX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:MIX+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 2.875077e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.937990e+07 ) sec^-1 *** EXECUTE GCHECK(MAX) -p 16384 32 1 *** Process = SIGMA_MSSM_SLHA2_GG_TTX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:MIX+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 7.567905e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 7.678146e+07 ) sec^-1 *** EXECUTE GCHECK(MAX128THR) -p 4096 128 1 --bridge *** Process = SIGMA_MSSM_SLHA2_GG_TTX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:MIX+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 2.865156e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.928533e+07 ) sec^-1 *** EXECUTE GCHECK(MAX128THR) -p 4096 128 1 *** Process = SIGMA_MSSM_SLHA2_GG_TTX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:MIX+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 7.911973e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 7.901462e+07 ) sec^-1 *** EXECUTE GCHECK(MAX8THR) -p 65536 8 1 --bridge *** Process = SIGMA_MSSM_SLHA2_GG_TTX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:MIX+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 2.881287e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.916021e+07 ) sec^-1 *** EXECUTE GCHECK(MAX8THR) -p 65536 8 1 *** Process = SIGMA_MSSM_SLHA2_GG_TTX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:MIX+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK -EvtsPerSec[MECalcOnly] (3a) = ( 2.733673e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.735846e+07 ) sec^-1 *** (3-hip) WARNING! SKIP MADEVENT_HIP (hip is not supported on this node) *** diff --git a/epochX/cudacpp/tput/logs_eemumu_mad/log_eemumu_mad_d_inl0_hrd0.txt b/epochX/cudacpp/tput/logs_eemumu_mad/log_eemumu_mad_d_inl0_hrd0.txt index 2396150f34..4a411cbc88 100644 --- a/epochX/cudacpp/tput/logs_eemumu_mad/log_eemumu_mad_d_inl0_hrd0.txt +++ b/epochX/cudacpp/tput/logs_eemumu_mad/log_eemumu_mad_d_inl0_hrd0.txt @@ -40,7 +40,7 @@ make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum' -DATE: 2024-10-06_08:54:52 +DATE: 2024-10-06_13:24:09 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: ========================================================================= @@ -49,15 +49,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_SM_EPEM_MUPMUM_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 5.715157e+07 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.495446e+08 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 4.756115e+08 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 7.341650e+07 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.744480e+08 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 4.922276e+08 ) sec^-1 MeanMatrixElemValue = ( 1.371706e-02 +- 3.270315e-06 ) GeV^0 -TOTAL : 0.722104 sec +TOTAL : 0.667074 sec INFO: No Floating Point Exceptions have been reported - 2,722,047,064 cycles # 2.855 GHz - 4,240,638,296 instructions # 1.56 insn per cycle - 1.034081868 seconds time elapsed + 2,606,525,896 cycles # 2.888 GHz + 4,104,502,442 instructions # 1.57 insn per cycle + 0.963580267 seconds time elapsed runNcu /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum/build.cuda_d_inl0_hrd0/check_cuda.exe -p 2048 256 1 ==PROF== Profiling "sigmaKin": launch__registers_per_thread 166 ==PROF== Profiling "sigmaKin": sm__sass_average_branch_targets_threads_uniform.pct 100% @@ -85,15 +85,15 @@ Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=0 Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/none+NAVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = SCALAR ('none': ~vector[1], no SIMD) -EvtsPerSec[Rmb+ME] (23) = ( 1.013288e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.182482e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.182482e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.040028e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.211212e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.211212e+06 ) sec^-1 MeanMatrixElemValue = ( 1.371706e-02 +- 3.270315e-06 ) GeV^0 -TOTAL : 6.662912 sec +TOTAL : 6.457819 sec INFO: No Floating Point Exceptions have been reported - 19,208,633,801 cycles # 2.880 GHz - 46,193,026,925 instructions # 2.40 insn per cycle - 6.677929994 seconds time elapsed + 18,993,599,427 cycles # 2.939 GHz + 46,088,835,161 instructions # 2.43 insn per cycle + 6.462956654 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 463) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum/build.none_d_inl0_hrd0/runTest_cpp.exe @@ -117,15 +117,15 @@ Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=0 Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[2] ('sse4': SSE4.2, 128bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 1.534189e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 2.004053e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 2.004053e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.574138e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 2.051803e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.051803e+06 ) sec^-1 MeanMatrixElemValue = ( 1.371706e-02 +- 3.270315e-06 ) GeV^0 -TOTAL : 4.563122 sec +TOTAL : 4.410849 sec INFO: No Floating Point Exceptions have been reported - 13,135,626,695 cycles # 2.874 GHz - 31,728,680,952 instructions # 2.42 insn per cycle - 4.573724377 seconds time elapsed + 12,942,080,920 cycles # 2.931 GHz + 31,624,050,183 instructions # 2.44 insn per cycle + 4.416041936 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 1662) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum/build.sse4_d_inl0_hrd0/runTest_cpp.exe @@ -149,15 +149,15 @@ Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=0 Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('avx2': AVX2, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 1.938790e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 2.711147e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 2.711147e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.933986e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 2.683413e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.683413e+06 ) sec^-1 MeanMatrixElemValue = ( 1.371706e-02 +- 3.270315e-06 ) GeV^0 -TOTAL : 3.694529 sec +TOTAL : 3.666044 sec INFO: No Floating Point Exceptions have been reported - 10,256,024,954 cycles # 2.769 GHz - 19,694,743,800 instructions # 1.92 insn per cycle - 3.707450749 seconds time elapsed + 10,029,008,550 cycles # 2.733 GHz + 19,589,375,169 instructions # 1.95 insn per cycle + 3.671326060 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 1909) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum/build.avx2_d_inl0_hrd0/runTest_cpp.exe @@ -181,15 +181,15 @@ Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=0 Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('512y': AVX512, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 1.944800e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 2.743029e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 2.743029e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 2.029975e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 2.844785e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.844785e+06 ) sec^-1 MeanMatrixElemValue = ( 1.371706e-02 +- 3.270315e-06 ) GeV^0 -TOTAL : 3.690433 sec +TOTAL : 3.504987 sec INFO: No Floating Point Exceptions have been reported - 10,133,821,420 cycles # 2.743 GHz - 19,357,887,145 instructions # 1.91 insn per cycle - 3.703105135 seconds time elapsed + 9,859,399,705 cycles # 2.810 GHz + 19,260,956,961 instructions # 1.95 insn per cycle + 3.510164054 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 1647) (512y: 180) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum/build.512y_d_inl0_hrd0/runTest_cpp.exe @@ -213,15 +213,15 @@ Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=0 Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('512z': AVX512, 512bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 1.663763e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 2.201339e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 2.201339e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.748106e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 2.315485e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.315485e+06 ) sec^-1 MeanMatrixElemValue = ( 1.371706e-02 +- 3.270315e-06 ) GeV^0 -TOTAL : 4.240790 sec +TOTAL : 4.006449 sec INFO: No Floating Point Exceptions have been reported - 8,791,817,571 cycles # 2.072 GHz - 15,864,118,825 instructions # 1.80 insn per cycle - 4.252718180 seconds time elapsed + 8,598,658,893 cycles # 2.144 GHz + 15,755,796,078 instructions # 1.83 insn per cycle + 4.011782646 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 871) (512y: 156) (512z: 1258) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum/build.512z_d_inl0_hrd0/runTest_cpp.exe diff --git a/epochX/cudacpp/tput/logs_eemumu_mad/log_eemumu_mad_d_inl0_hrd0_bridge.txt b/epochX/cudacpp/tput/logs_eemumu_mad/log_eemumu_mad_d_inl0_hrd0_bridge.txt index 97960252e7..e27955511e 100644 --- a/epochX/cudacpp/tput/logs_eemumu_mad/log_eemumu_mad_d_inl0_hrd0_bridge.txt +++ b/epochX/cudacpp/tput/logs_eemumu_mad/log_eemumu_mad_d_inl0_hrd0_bridge.txt @@ -40,7 +40,7 @@ make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum' -DATE: 2024-10-06_09:36:32 +DATE: 2024-10-06_13:55:06 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: ========================================================================= @@ -53,16 +53,16 @@ WARNING! Set grid in Bridge (nevt=524288, gpublocks=2048, gputhreads=256, gpublo Process = SIGMA_SM_EPEM_MUPMUM_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 4.729675e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.983590e+07 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.983590e+07 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 4.695757e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.862042e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.862042e+07 ) sec^-1 MeanMatrixElemValue = ( 1.371706e-02 +- 3.270315e-06 ) GeV^0 -TOTAL : 2.228883 sec +TOTAL : 2.244276 sec INFO: No Floating Point Exceptions have been reported INFO: No Floating Point Exceptions have been reported - 7,241,513,211 cycles # 2.923 GHz - 12,978,693,777 instructions # 1.79 insn per cycle - 2.533005072 seconds time elapsed + 7,224,938,449 cycles # 2.901 GHz + 12,954,563,213 instructions # 1.79 insn per cycle + 2.548150554 seconds time elapsed runNcu /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum/build.cuda_d_inl0_hrd0/check_cuda.exe -p 2048 256 1 --bridge WARNING! Bridge selected: cannot use RamboDevice, will use RamboHost WARNING! RamboHost selected: cannot use CurandDevice, will use CurandHost @@ -95,16 +95,16 @@ Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=0 Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+BRDHST/none+NAVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = SCALAR ('none': ~vector[1], no SIMD) -EvtsPerSec[Rmb+ME] (23) = ( 9.954014e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.154803e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.154803e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.008395e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.167973e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.167973e+06 ) sec^-1 MeanMatrixElemValue = ( 1.371706e-02 +- 3.270315e-06 ) GeV^0 -TOTAL : 6.972350 sec +TOTAL : 6.854283 sec INFO: No Floating Point Exceptions have been reported INFO: No Floating Point Exceptions have been reported - 20,384,148,235 cycles # 2.919 GHz - 46,410,615,309 instructions # 2.28 insn per cycle - 6.984536194 seconds time elapsed + 20,242,406,248 cycles # 2.951 GHz + 46,317,273,619 instructions # 2.29 insn per cycle + 6.861171447 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 463) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum/build.none_d_inl0_hrd0/runTest_cpp.exe @@ -129,16 +129,16 @@ Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=0 Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+BRDHST/sse4+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[2] ('sse4': SSE4.2, 128bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 1.493408e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.921090e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.921090e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.499738e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.928392e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.928392e+06 ) sec^-1 MeanMatrixElemValue = ( 1.371706e-02 +- 3.270315e-06 ) GeV^0 -TOTAL : 4.877492 sec +TOTAL : 4.829377 sec INFO: No Floating Point Exceptions have been reported INFO: No Floating Point Exceptions have been reported - 14,402,886,877 cycles # 2.946 GHz - 32,567,021,239 instructions # 2.26 insn per cycle - 4.890045852 seconds time elapsed + 14,233,780,469 cycles # 2.944 GHz + 32,470,408,920 instructions # 2.28 insn per cycle + 4.836366589 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 1662) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum/build.sse4_d_inl0_hrd0/runTest_cpp.exe @@ -163,16 +163,16 @@ Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=0 Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+BRDHST/avx2+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('avx2': AVX2, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 1.864025e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 2.539449e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 2.539449e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.870609e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 2.550871e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.550871e+06 ) sec^-1 MeanMatrixElemValue = ( 1.371706e-02 +- 3.270315e-06 ) GeV^0 -TOTAL : 4.048395 sec +TOTAL : 3.996114 sec INFO: No Floating Point Exceptions have been reported INFO: No Floating Point Exceptions have been reported - 11,503,225,226 cycles # 2.834 GHz - 21,048,377,803 instructions # 1.83 insn per cycle - 4.060868426 seconds time elapsed + 11,327,989,619 cycles # 2.830 GHz + 20,949,277,897 instructions # 1.85 insn per cycle + 4.003356859 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 1909) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum/build.avx2_d_inl0_hrd0/runTest_cpp.exe @@ -197,16 +197,16 @@ Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=0 Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+BRDHST/512y+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('512y': AVX512, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 1.889652e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 2.596697e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 2.596697e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.883433e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 2.582968e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.582968e+06 ) sec^-1 MeanMatrixElemValue = ( 1.371706e-02 +- 3.270315e-06 ) GeV^0 -TOTAL : 4.001389 sec +TOTAL : 3.982638 sec INFO: No Floating Point Exceptions have been reported INFO: No Floating Point Exceptions have been reported - 11,334,121,636 cycles # 2.824 GHz - 20,717,870,984 instructions # 1.83 insn per cycle - 4.014529771 seconds time elapsed + 11,215,050,356 cycles # 2.811 GHz + 20,617,236,409 instructions # 1.84 insn per cycle + 3.990566813 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 1647) (512y: 180) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum/build.512y_d_inl0_hrd0/runTest_cpp.exe @@ -231,16 +231,16 @@ Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=0 Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+BRDHST/512z+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('512z': AVX512, 512bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 1.585647e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 2.044820e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 2.044820e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.623833e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 2.106816e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.106816e+06 ) sec^-1 MeanMatrixElemValue = ( 1.371706e-02 +- 3.270315e-06 ) GeV^0 -TOTAL : 4.655129 sec +TOTAL : 4.513905 sec INFO: No Floating Point Exceptions have been reported INFO: No Floating Point Exceptions have been reported - 10,329,600,614 cycles # 2.214 GHz - 17,028,538,054 instructions # 1.65 insn per cycle - 4.667149794 seconds time elapsed + 9,979,044,287 cycles # 2.208 GHz + 16,913,037,388 instructions # 1.69 insn per cycle + 4.521806967 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 871) (512y: 156) (512z: 1258) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum/build.512z_d_inl0_hrd0/runTest_cpp.exe diff --git a/epochX/cudacpp/tput/logs_eemumu_mad/log_eemumu_mad_d_inl0_hrd0_common.txt b/epochX/cudacpp/tput/logs_eemumu_mad/log_eemumu_mad_d_inl0_hrd0_common.txt index a07615eec8..fb7e857ef8 100644 --- a/epochX/cudacpp/tput/logs_eemumu_mad/log_eemumu_mad_d_inl0_hrd0_common.txt +++ b/epochX/cudacpp/tput/logs_eemumu_mad/log_eemumu_mad_d_inl0_hrd0_common.txt @@ -40,7 +40,7 @@ make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum' -DATE: 2024-10-06_09:48:44 +DATE: 2024-10-06_14:07:17 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: ========================================================================= @@ -49,15 +49,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_SM_EPEM_MUPMUM_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:COMMON+RMBDEV+MESDEV/none+NAVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 6.479194e+07 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.613891e+08 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 4.774308e+08 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 6.505686e+07 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.505211e+08 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 4.753684e+08 ) sec^-1 MeanMatrixElemValue = ( 1.371879e-02 +- 3.270020e-06 ) GeV^0 -TOTAL : 1.350685 sec +TOTAL : 1.368747 sec INFO: No Floating Point Exceptions have been reported - 4,619,154,070 cycles # 2.910 GHz - 7,244,933,472 instructions # 1.57 insn per cycle - 1.645096659 seconds time elapsed + 4,641,560,808 cycles # 2.899 GHz + 7,202,532,844 instructions # 1.55 insn per cycle + 1.659890968 seconds time elapsed runNcu /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum/build.cuda_d_inl0_hrd0/check_cuda.exe -p 2048 256 1 --common ==PROF== Profiling "sigmaKin": launch__registers_per_thread 166 ==PROF== Profiling "sigmaKin": sm__sass_average_branch_targets_threads_uniform.pct 100% @@ -85,15 +85,15 @@ Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=0 Workflow summary = CPP:DBL+CXS:COMMON+RMBHST+MESHST/none+NAVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = SCALAR ('none': ~vector[1], no SIMD) -EvtsPerSec[Rmb+ME] (23) = ( 1.031231e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.202853e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.202853e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.034105e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.204854e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.204854e+06 ) sec^-1 MeanMatrixElemValue = ( 1.371879e-02 +- 3.270020e-06 ) GeV^0 -TOTAL : 6.892564 sec +TOTAL : 6.869371 sec INFO: No Floating Point Exceptions have been reported - 20,216,212,113 cycles # 2.933 GHz - 46,211,289,901 instructions # 2.29 insn per cycle - 6.898049528 seconds time elapsed + 20,127,706,225 cycles # 2.928 GHz + 46,193,046,397 instructions # 2.29 insn per cycle + 6.874895441 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 463) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum/build.none_d_inl0_hrd0/runTest_cpp.exe @@ -117,15 +117,15 @@ Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=0 Workflow summary = CPP:DBL+CXS:COMMON+RMBHST+MESHST/sse4+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[2] ('sse4': SSE4.2, 128bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 1.575355e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 2.054940e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 2.054940e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.582839e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 2.062761e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.062761e+06 ) sec^-1 MeanMatrixElemValue = ( 1.371879e-02 +- 3.270020e-06 ) GeV^0 -TOTAL : 4.814476 sec +TOTAL : 4.757321 sec INFO: No Floating Point Exceptions have been reported - 14,161,512,947 cycles # 2.938 GHz - 31,718,115,030 instructions # 2.24 insn per cycle - 4.820285845 seconds time elapsed + 14,026,082,707 cycles # 2.946 GHz + 31,624,933,435 instructions # 2.25 insn per cycle + 4.762912415 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 1662) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum/build.sse4_d_inl0_hrd0/runTest_cpp.exe @@ -149,15 +149,15 @@ Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=0 Workflow summary = CPP:DBL+CXS:COMMON+RMBHST+MESHST/avx2+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('avx2': AVX2, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 1.990481e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 2.780031e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 2.780031e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.994818e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 2.779315e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.779315e+06 ) sec^-1 MeanMatrixElemValue = ( 1.371879e-02 +- 3.270020e-06 ) GeV^0 -TOTAL : 3.993697 sec +TOTAL : 3.934254 sec INFO: No Floating Point Exceptions have been reported - 11,344,220,574 cycles # 2.837 GHz - 19,628,934,109 instructions # 1.73 insn per cycle - 3.999571252 seconds time elapsed + 11,142,822,139 cycles # 2.829 GHz + 19,489,738,039 instructions # 1.75 insn per cycle + 3.939850039 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 1909) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum/build.avx2_d_inl0_hrd0/runTest_cpp.exe @@ -181,15 +181,15 @@ Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=0 Workflow summary = CPP:DBL+CXS:COMMON+RMBHST+MESHST/512y+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('512y': AVX512, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 2.024448e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 2.841239e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 2.841239e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 2.038921e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 2.855389e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.855389e+06 ) sec^-1 MeanMatrixElemValue = ( 1.371879e-02 +- 3.270020e-06 ) GeV^0 -TOTAL : 3.944999 sec +TOTAL : 3.868760 sec INFO: No Floating Point Exceptions have been reported - 11,153,243,188 cycles # 2.824 GHz - 19,098,861,484 instructions # 1.71 insn per cycle - 3.950731996 seconds time elapsed + 10,971,960,427 cycles # 2.833 GHz + 18,960,885,602 instructions # 1.73 insn per cycle + 3.874470540 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 1647) (512y: 180) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum/build.512y_d_inl0_hrd0/runTest_cpp.exe @@ -213,15 +213,15 @@ Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=0 Workflow summary = CPP:DBL+CXS:COMMON+RMBHST+MESHST/512z+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('512z': AVX512, 512bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 1.731970e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 2.289397e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 2.289397e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.738006e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 2.296583e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.296583e+06 ) sec^-1 MeanMatrixElemValue = ( 1.371879e-02 +- 3.270020e-06 ) GeV^0 -TOTAL : 4.507668 sec +TOTAL : 4.410007 sec INFO: No Floating Point Exceptions have been reported - 9,996,448,485 cycles # 2.215 GHz - 15,693,646,767 instructions # 1.57 insn per cycle - 4.513790217 seconds time elapsed + 9,736,817,654 cycles # 2.206 GHz + 15,456,466,481 instructions # 1.59 insn per cycle + 4.415691287 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 871) (512y: 156) (512z: 1258) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum/build.512z_d_inl0_hrd0/runTest_cpp.exe diff --git a/epochX/cudacpp/tput/logs_eemumu_mad/log_eemumu_mad_d_inl0_hrd0_curhst.txt b/epochX/cudacpp/tput/logs_eemumu_mad/log_eemumu_mad_d_inl0_hrd0_curhst.txt index cf4e1a1e41..45fcae4a58 100644 --- a/epochX/cudacpp/tput/logs_eemumu_mad/log_eemumu_mad_d_inl0_hrd0_curhst.txt +++ b/epochX/cudacpp/tput/logs_eemumu_mad/log_eemumu_mad_d_inl0_hrd0_curhst.txt @@ -40,7 +40,7 @@ make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum' -DATE: 2024-10-06_09:45:58 +DATE: 2024-10-06_14:04:31 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: ========================================================================= @@ -49,15 +49,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_SM_EPEM_MUPMUM_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURHST+RMBDEV+MESDEV/none+NAVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 6.516686e+07 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.553796e+08 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 4.802555e+08 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 6.628543e+07 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.639190e+08 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 4.818050e+08 ) sec^-1 MeanMatrixElemValue = ( 1.371706e-02 +- 3.270315e-06 ) GeV^0 -TOTAL : 0.994565 sec +TOTAL : 0.986431 sec INFO: No Floating Point Exceptions have been reported - 3,557,200,491 cycles # 2.898 GHz - 7,056,373,361 instructions # 1.98 insn per cycle - 1.285636058 seconds time elapsed + 3,538,698,244 cycles # 2.900 GHz + 7,045,875,726 instructions # 1.99 insn per cycle + 1.277805273 seconds time elapsed runNcu /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum/build.cuda_d_inl0_hrd0/check_cuda.exe -p 2048 256 1 --curhst ==PROF== Profiling "sigmaKin": launch__registers_per_thread 166 ==PROF== Profiling "sigmaKin": sm__sass_average_branch_targets_threads_uniform.pct 100% @@ -85,15 +85,15 @@ Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=0 Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/none+NAVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = SCALAR ('none': ~vector[1], no SIMD) -EvtsPerSec[Rmb+ME] (23) = ( 1.036397e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.208868e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.208868e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.034433e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.204886e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.204886e+06 ) sec^-1 MeanMatrixElemValue = ( 1.371706e-02 +- 3.270315e-06 ) GeV^0 -TOTAL : 6.480716 sec +TOTAL : 6.492494 sec INFO: No Floating Point Exceptions have been reported - 19,050,518,676 cycles # 2.938 GHz - 46,087,808,907 instructions # 2.42 insn per cycle - 6.486425223 seconds time elapsed + 19,144,598,774 cycles # 2.947 GHz + 46,091,610,477 instructions # 2.41 insn per cycle + 6.498030497 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 463) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum/build.none_d_inl0_hrd0/runTest_cpp.exe @@ -117,15 +117,15 @@ Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=0 Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[2] ('sse4': SSE4.2, 128bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 1.562645e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 2.044042e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 2.044042e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.584131e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 2.064167e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.064167e+06 ) sec^-1 MeanMatrixElemValue = ( 1.371706e-02 +- 3.270315e-06 ) GeV^0 -TOTAL : 4.441615 sec +TOTAL : 4.386842 sec INFO: No Floating Point Exceptions have been reported - 13,100,732,544 cycles # 2.946 GHz - 31,624,731,275 instructions # 2.41 insn per cycle - 4.447190414 seconds time elapsed + 12,925,405,286 cycles # 2.943 GHz + 31,622,252,380 instructions # 2.45 insn per cycle + 4.392692449 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 1662) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum/build.sse4_d_inl0_hrd0/runTest_cpp.exe @@ -149,15 +149,15 @@ Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=0 Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('avx2': AVX2, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 1.962342e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 2.741135e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 2.741135e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.985844e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 2.772856e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.772856e+06 ) sec^-1 MeanMatrixElemValue = ( 1.371706e-02 +- 3.270315e-06 ) GeV^0 -TOTAL : 3.617520 sec +TOTAL : 3.576214 sec INFO: No Floating Point Exceptions have been reported - 10,105,971,200 cycles # 2.790 GHz - 19,587,417,861 instructions # 1.94 insn per cycle - 3.623303854 seconds time elapsed + 10,081,945,851 cycles # 2.815 GHz + 19,589,332,111 instructions # 1.94 insn per cycle + 3.582087720 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 1909) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum/build.avx2_d_inl0_hrd0/runTest_cpp.exe @@ -181,15 +181,15 @@ Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=0 Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('512y': AVX512, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 2.035108e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 2.854302e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 2.854302e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 2.043792e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 2.857786e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.857786e+06 ) sec^-1 MeanMatrixElemValue = ( 1.371706e-02 +- 3.270315e-06 ) GeV^0 -TOTAL : 3.498153 sec +TOTAL : 3.484012 sec INFO: No Floating Point Exceptions have been reported - 9,879,352,969 cycles # 2.820 GHz - 19,249,039,766 instructions # 1.95 insn per cycle - 3.504047287 seconds time elapsed + 9,837,675,738 cycles # 2.820 GHz + 19,249,053,256 instructions # 1.96 insn per cycle + 3.489855131 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 1647) (512y: 180) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum/build.512y_d_inl0_hrd0/runTest_cpp.exe @@ -213,15 +213,15 @@ Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=0 Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('512z': AVX512, 512bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 1.738426e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 2.300548e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 2.300548e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.734129e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 2.292307e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.292307e+06 ) sec^-1 MeanMatrixElemValue = ( 1.371706e-02 +- 3.270315e-06 ) GeV^0 -TOTAL : 4.029997 sec +TOTAL : 4.039271 sec INFO: No Floating Point Exceptions have been reported - 8,617,786,478 cycles # 2.136 GHz - 15,755,373,979 instructions # 1.83 insn per cycle - 4.035885525 seconds time elapsed + 8,606,100,886 cycles # 2.128 GHz + 15,755,354,926 instructions # 1.83 insn per cycle + 4.045171632 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 871) (512y: 156) (512z: 1258) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum/build.512z_d_inl0_hrd0/runTest_cpp.exe diff --git a/epochX/cudacpp/tput/logs_eemumu_mad/log_eemumu_mad_d_inl0_hrd0_rmbhst.txt b/epochX/cudacpp/tput/logs_eemumu_mad/log_eemumu_mad_d_inl0_hrd0_rmbhst.txt index 23a95e9b43..e26e9c6c4b 100644 --- a/epochX/cudacpp/tput/logs_eemumu_mad/log_eemumu_mad_d_inl0_hrd0_rmbhst.txt +++ b/epochX/cudacpp/tput/logs_eemumu_mad/log_eemumu_mad_d_inl0_hrd0_rmbhst.txt @@ -40,7 +40,7 @@ make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum' -DATE: 2024-10-06_09:43:10 +DATE: 2024-10-06_14:01:43 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: ========================================================================= @@ -50,15 +50,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_SM_EPEM_MUPMUM_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURHST+RMBHST+MESDEV/none+NAVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 6.035607e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.566958e+08 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 4.715605e+08 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 6.065314e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.602842e+08 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 4.762256e+08 ) sec^-1 MeanMatrixElemValue = ( 1.371706e-02 +- 3.270315e-06 ) GeV^0 -TOTAL : 1.900996 sec +TOTAL : 1.886819 sec INFO: No Floating Point Exceptions have been reported - 6,141,367,935 cycles # 2.877 GHz - 11,470,611,621 instructions # 1.87 insn per cycle - 2.190401749 seconds time elapsed + 6,171,943,223 cycles # 2.910 GHz + 11,432,322,940 instructions # 1.85 insn per cycle + 2.177198169 seconds time elapsed runNcu /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum/build.cuda_d_inl0_hrd0/check_cuda.exe -p 2048 256 1 --rmbhst WARNING! RamboHost selected: cannot use CurandDevice, will use CurandHost ==PROF== Profiling "sigmaKin": launch__registers_per_thread 166 @@ -87,15 +87,15 @@ Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=0 Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/none+NAVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = SCALAR ('none': ~vector[1], no SIMD) -EvtsPerSec[Rmb+ME] (23) = ( 1.040250e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.212161e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.212161e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.035565e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.206133e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.206133e+06 ) sec^-1 MeanMatrixElemValue = ( 1.371706e-02 +- 3.270315e-06 ) GeV^0 -TOTAL : 6.458941 sec +TOTAL : 6.487538 sec INFO: No Floating Point Exceptions have been reported - 19,062,791,283 cycles # 2.949 GHz - 46,091,693,422 instructions # 2.42 insn per cycle - 6.464859061 seconds time elapsed + 19,012,234,848 cycles # 2.928 GHz + 46,088,936,024 instructions # 2.42 insn per cycle + 6.493473738 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 463) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum/build.none_d_inl0_hrd0/runTest_cpp.exe @@ -119,15 +119,15 @@ Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=0 Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[2] ('sse4': SSE4.2, 128bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 1.576646e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 2.057103e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 2.057103e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.576994e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 2.051607e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.051607e+06 ) sec^-1 MeanMatrixElemValue = ( 1.371706e-02 +- 3.270315e-06 ) GeV^0 -TOTAL : 4.406292 sec +TOTAL : 4.403707 sec INFO: No Floating Point Exceptions have been reported - 12,965,800,121 cycles # 2.939 GHz - 31,623,980,844 instructions # 2.44 insn per cycle - 4.412202935 seconds time elapsed + 12,934,925,116 cycles # 2.934 GHz + 31,624,346,831 instructions # 2.44 insn per cycle + 4.409708711 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 1662) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum/build.sse4_d_inl0_hrd0/runTest_cpp.exe @@ -151,15 +151,15 @@ Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=0 Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('avx2': AVX2, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 1.982815e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 2.782156e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 2.782156e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.979531e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 2.775756e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.775756e+06 ) sec^-1 MeanMatrixElemValue = ( 1.371706e-02 +- 3.270315e-06 ) GeV^0 -TOTAL : 3.583602 sec +TOTAL : 3.590712 sec INFO: No Floating Point Exceptions have been reported - 10,107,254,042 cycles # 2.816 GHz - 19,587,412,579 instructions # 1.94 insn per cycle - 3.589639966 seconds time elapsed + 10,108,115,481 cycles # 2.811 GHz + 19,587,472,518 instructions # 1.94 insn per cycle + 3.596489331 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 1909) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum/build.avx2_d_inl0_hrd0/runTest_cpp.exe @@ -183,15 +183,15 @@ Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=0 Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('512y': AVX512, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 2.036151e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 2.856576e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 2.856576e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 2.034669e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 2.849230e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.849230e+06 ) sec^-1 MeanMatrixElemValue = ( 1.371706e-02 +- 3.270315e-06 ) GeV^0 -TOTAL : 3.497964 sec +TOTAL : 3.497578 sec INFO: No Floating Point Exceptions have been reported - 9,879,922,849 cycles # 2.820 GHz - 19,260,007,955 instructions # 1.95 insn per cycle - 3.503929332 seconds time elapsed + 9,864,909,299 cycles # 2.817 GHz + 19,249,043,993 instructions # 1.95 insn per cycle + 3.503493277 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 1647) (512y: 180) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum/build.512y_d_inl0_hrd0/runTest_cpp.exe @@ -215,15 +215,15 @@ Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=0 Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('512z': AVX512, 512bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 1.741980e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 2.303561e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 2.303561e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.738531e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 2.296428e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.296428e+06 ) sec^-1 MeanMatrixElemValue = ( 1.371706e-02 +- 3.270315e-06 ) GeV^0 -TOTAL : 4.020543 sec +TOTAL : 4.029592 sec INFO: No Floating Point Exceptions have been reported - 8,613,807,526 cycles # 2.140 GHz - 15,755,294,312 instructions # 1.83 insn per cycle - 4.026429840 seconds time elapsed + 8,626,430,569 cycles # 2.138 GHz + 15,755,232,603 instructions # 1.83 insn per cycle + 4.035417303 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 871) (512y: 156) (512z: 1258) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum/build.512z_d_inl0_hrd0/runTest_cpp.exe diff --git a/epochX/cudacpp/tput/logs_eemumu_mad/log_eemumu_mad_d_inl0_hrd1.txt b/epochX/cudacpp/tput/logs_eemumu_mad/log_eemumu_mad_d_inl0_hrd1.txt index 25ac5b33ed..d4dcd2f4ae 100644 --- a/epochX/cudacpp/tput/logs_eemumu_mad/log_eemumu_mad_d_inl0_hrd1.txt +++ b/epochX/cudacpp/tput/logs_eemumu_mad/log_eemumu_mad_d_inl0_hrd1.txt @@ -40,7 +40,7 @@ make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum' -DATE: 2024-10-06_08:55:23 +DATE: 2024-10-06_13:24:40 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: ========================================================================= @@ -49,15 +49,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_SM_EPEM_MUPMUM_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=1] Workflow summary = CUD:DBL+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 6.275982e+07 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.504846e+08 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 4.746692e+08 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 7.352653e+07 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.741622e+08 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 4.920325e+08 ) sec^-1 MeanMatrixElemValue = ( 1.371706e-02 +- 3.270315e-06 ) GeV^0 -TOTAL : 0.698739 sec +TOTAL : 0.661485 sec INFO: No Floating Point Exceptions have been reported - 2,671,543,996 cycles # 2.868 GHz - 4,201,680,962 instructions # 1.57 insn per cycle - 1.042000131 seconds time elapsed + 2,594,433,164 cycles # 2.897 GHz + 4,059,498,246 instructions # 1.56 insn per cycle + 0.954939632 seconds time elapsed runNcu /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum/build.cuda_d_inl0_hrd1/check_cuda.exe -p 2048 256 1 ==PROF== Profiling "sigmaKin": launch__registers_per_thread 154 ==PROF== Profiling "sigmaKin": sm__sass_average_branch_targets_threads_uniform.pct 100% @@ -85,15 +85,15 @@ Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=0 Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/none+NAVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = SCALAR ('none': ~vector[1], no SIMD) -EvtsPerSec[Rmb+ME] (23) = ( 1.030289e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.210430e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.210430e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.040431e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.211629e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.211629e+06 ) sec^-1 MeanMatrixElemValue = ( 1.371706e-02 +- 3.270315e-06 ) GeV^0 -TOTAL : 6.580678 sec +TOTAL : 6.455258 sec INFO: No Floating Point Exceptions have been reported - 19,388,414,039 cycles # 2.942 GHz - 46,168,116,276 instructions # 2.38 insn per cycle - 6.592554583 seconds time elapsed + 19,003,040,422 cycles # 2.942 GHz + 46,049,878,370 instructions # 2.42 insn per cycle + 6.460324703 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 452) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum/build.none_d_inl0_hrd1/runTest_cpp.exe @@ -117,15 +117,15 @@ Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=0 Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[2] ('sse4': SSE4.2, 128bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 1.571872e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 2.069657e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 2.069657e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.595987e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 2.081272e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.081272e+06 ) sec^-1 MeanMatrixElemValue = ( 1.371706e-02 +- 3.270315e-06 ) GeV^0 -TOTAL : 4.454497 sec +TOTAL : 4.353623 sec INFO: No Floating Point Exceptions have been reported - 13,123,917,893 cycles # 2.941 GHz - 31,665,954,915 instructions # 2.41 insn per cycle - 4.468095413 seconds time elapsed + 12,857,961,145 cycles # 2.951 GHz + 31,561,142,783 instructions # 2.45 insn per cycle + 4.358763684 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 1648) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum/build.sse4_d_inl0_hrd1/runTest_cpp.exe @@ -149,15 +149,15 @@ Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=0 Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('avx2': AVX2, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 1.982748e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 2.777393e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 2.777393e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.988714e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 2.766116e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.766116e+06 ) sec^-1 MeanMatrixElemValue = ( 1.371706e-02 +- 3.270315e-06 ) GeV^0 -TOTAL : 3.618164 sec +TOTAL : 3.570288 sec INFO: No Floating Point Exceptions have been reported - 10,210,665,805 cycles # 2.814 GHz - 19,682,748,403 instructions # 1.93 insn per cycle - 3.629801888 seconds time elapsed + 10,014,402,148 cycles # 2.802 GHz + 19,579,742,812 instructions # 1.96 insn per cycle + 3.575414332 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 1894) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum/build.avx2_d_inl0_hrd1/runTest_cpp.exe @@ -181,15 +181,15 @@ Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=0 Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('512y': AVX512, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 2.010638e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 2.831487e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 2.831487e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 2.036513e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 2.848755e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.848755e+06 ) sec^-1 MeanMatrixElemValue = ( 1.371706e-02 +- 3.270315e-06 ) GeV^0 -TOTAL : 3.575425 sec +TOTAL : 3.495305 sec INFO: No Floating Point Exceptions have been reported - 10,055,677,244 cycles # 2.805 GHz - 19,379,411,405 instructions # 1.93 insn per cycle - 3.588891240 seconds time elapsed + 9,835,731,184 cycles # 2.811 GHz + 19,274,394,474 instructions # 1.96 insn per cycle + 3.500394903 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 1636) (512y: 178) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum/build.512y_d_inl0_hrd1/runTest_cpp.exe @@ -213,15 +213,15 @@ Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=0 Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('512z': AVX512, 512bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 1.768631e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 2.372427e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 2.372427e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.787716e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 2.382515e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.382515e+06 ) sec^-1 MeanMatrixElemValue = ( 1.371706e-02 +- 3.270315e-06 ) GeV^0 -TOTAL : 4.003265 sec +TOTAL : 3.926568 sec INFO: No Floating Point Exceptions have been reported - 8,643,505,927 cycles # 2.154 GHz - 15,697,303,734 instructions # 1.82 insn per cycle - 4.017112338 seconds time elapsed + 8,427,171,909 cycles # 2.144 GHz + 15,589,889,161 instructions # 1.85 insn per cycle + 3.932043759 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 833) (512y: 153) (512z: 1240) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum/build.512z_d_inl0_hrd1/runTest_cpp.exe diff --git a/epochX/cudacpp/tput/logs_eemumu_mad/log_eemumu_mad_d_inl1_hrd0.txt b/epochX/cudacpp/tput/logs_eemumu_mad/log_eemumu_mad_d_inl1_hrd0.txt index 9d9181639f..0cba288727 100644 --- a/epochX/cudacpp/tput/logs_eemumu_mad/log_eemumu_mad_d_inl1_hrd0.txt +++ b/epochX/cudacpp/tput/logs_eemumu_mad/log_eemumu_mad_d_inl1_hrd0.txt @@ -40,7 +40,7 @@ make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum' -DATE: 2024-10-06_09:26:55 +DATE: 2024-10-06_13:45:28 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: ========================================================================= @@ -49,15 +49,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_SM_EPEM_MUPMUM_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=1] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 6.029061e+07 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.569612e+08 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 4.860356e+08 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 7.207263e+07 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.685012e+08 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 4.837960e+08 ) sec^-1 MeanMatrixElemValue = ( 1.371706e-02 +- 3.270315e-06 ) GeV^0 -TOTAL : 0.695056 sec +TOTAL : 0.669145 sec INFO: No Floating Point Exceptions have been reported - 2,704,879,803 cycles # 2.897 GHz - 4,231,460,596 instructions # 1.56 insn per cycle - 0.994220648 seconds time elapsed + 2,552,232,844 cycles # 2.814 GHz + 4,013,232,418 instructions # 1.57 insn per cycle + 0.966583157 seconds time elapsed runNcu /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum/build.cuda_d_inl1_hrd0/check_cuda.exe -p 2048 256 1 ==PROF== Profiling "sigmaKin": launch__registers_per_thread 166 ==PROF== Profiling "sigmaKin": sm__sass_average_branch_targets_threads_uniform.pct 100% @@ -85,15 +85,15 @@ Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=1 Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/none+NAVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = SCALAR ('none': ~vector[1], no SIMD) -EvtsPerSec[Rmb+ME] (23) = ( 1.606609e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 2.069672e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 2.069672e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.616848e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 2.071528e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.071528e+06 ) sec^-1 MeanMatrixElemValue = ( 1.371706e-02 +- 3.270315e-06 ) GeV^0 -TOTAL : 4.367451 sec +TOTAL : 4.301335 sec INFO: No Floating Point Exceptions have been reported - 12,912,062,009 cycles # 2.950 GHz - 32,678,927,799 instructions # 2.53 insn per cycle - 4.379017229 seconds time elapsed + 12,669,480,430 cycles # 2.942 GHz + 32,573,491,108 instructions # 2.57 insn per cycle + 4.306970379 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 281) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum/build.none_d_inl1_hrd0/runTest_cpp.exe @@ -117,15 +117,15 @@ Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=1 Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[2] ('sse4': SSE4.2, 128bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 1.977635e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 2.819919e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 2.819919e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 2.014355e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 2.859327e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.859327e+06 ) sec^-1 MeanMatrixElemValue = ( 1.371706e-02 +- 3.270315e-06 ) GeV^0 -TOTAL : 3.639596 sec +TOTAL : 3.536663 sec INFO: No Floating Point Exceptions have been reported - 10,716,876,159 cycles # 2.936 GHz - 25,005,426,831 instructions # 2.33 insn per cycle - 3.651343591 seconds time elapsed + 10,425,756,626 cycles # 2.944 GHz + 24,900,221,139 instructions # 2.39 insn per cycle + 3.541937497 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 1246) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum/build.sse4_d_inl1_hrd0/runTest_cpp.exe @@ -149,15 +149,15 @@ Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=1 Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('avx2': AVX2, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 2.209379e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.259757e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.259757e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 2.233979e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.268060e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.268060e+06 ) sec^-1 MeanMatrixElemValue = ( 1.371706e-02 +- 3.270315e-06 ) GeV^0 -TOTAL : 3.300079 sec +TOTAL : 3.222129 sec INFO: No Floating Point Exceptions have been reported - 9,398,178,742 cycles # 2.838 GHz - 16,938,114,674 instructions # 1.80 insn per cycle - 3.311853262 seconds time elapsed + 9,107,125,134 cycles # 2.823 GHz + 16,833,993,832 instructions # 1.85 insn per cycle + 3.227472591 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 1599) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum/build.avx2_d_inl1_hrd0/runTest_cpp.exe @@ -181,15 +181,15 @@ Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=1 Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('512y': AVX512, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 2.277311e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.397001e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.397001e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 2.277247e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.351012e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.351012e+06 ) sec^-1 MeanMatrixElemValue = ( 1.371706e-02 +- 3.270315e-06 ) GeV^0 -TOTAL : 3.212263 sec +TOTAL : 3.170697 sec INFO: No Floating Point Exceptions have been reported - 9,139,009,296 cycles # 2.835 GHz - 16,502,297,129 instructions # 1.81 insn per cycle - 3.223908096 seconds time elapsed + 8,901,852,698 cycles # 2.804 GHz + 16,399,026,897 instructions # 1.84 insn per cycle + 3.176742632 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 1355) (512y: 139) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum/build.512y_d_inl1_hrd0/runTest_cpp.exe @@ -213,15 +213,15 @@ Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=1 Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('512z': AVX512, 512bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 1.921368e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 2.661482e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 2.661482e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.956943e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 2.695210e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.695210e+06 ) sec^-1 MeanMatrixElemValue = ( 1.371706e-02 +- 3.270315e-06 ) GeV^0 -TOTAL : 3.727052 sec +TOTAL : 3.623105 sec INFO: No Floating Point Exceptions have been reported - 8,146,634,535 cycles # 2.180 GHz - 14,661,732,896 instructions # 1.80 insn per cycle - 3.738643291 seconds time elapsed + 7,884,772,110 cycles # 2.174 GHz + 14,559,326,319 instructions # 1.85 insn per cycle + 3.628599399 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 1003) (512y: 158) (512z: 946) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum/build.512z_d_inl1_hrd0/runTest_cpp.exe diff --git a/epochX/cudacpp/tput/logs_eemumu_mad/log_eemumu_mad_d_inl1_hrd1.txt b/epochX/cudacpp/tput/logs_eemumu_mad/log_eemumu_mad_d_inl1_hrd1.txt index abe54e8953..faf82b58c5 100644 --- a/epochX/cudacpp/tput/logs_eemumu_mad/log_eemumu_mad_d_inl1_hrd1.txt +++ b/epochX/cudacpp/tput/logs_eemumu_mad/log_eemumu_mad_d_inl1_hrd1.txt @@ -40,7 +40,7 @@ make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum' -DATE: 2024-10-06_09:27:22 +DATE: 2024-10-06_13:45:54 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: ========================================================================= @@ -49,15 +49,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_SM_EPEM_MUPMUM_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=1] [hardcodePARAM=1] Workflow summary = CUD:DBL+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 6.921706e+07 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.715910e+08 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 4.877358e+08 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 7.184031e+07 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.686531e+08 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 4.839847e+08 ) sec^-1 MeanMatrixElemValue = ( 1.371706e-02 +- 3.270315e-06 ) GeV^0 -TOTAL : 0.678632 sec +TOTAL : 0.671206 sec INFO: No Floating Point Exceptions have been reported - 2,636,898,249 cycles # 2.884 GHz - 4,067,260,892 instructions # 1.54 insn per cycle - 0.973352356 seconds time elapsed + 2,620,037,674 cycles # 2.893 GHz + 4,082,631,138 instructions # 1.56 insn per cycle + 0.967670753 seconds time elapsed runNcu /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum/build.cuda_d_inl1_hrd1/check_cuda.exe -p 2048 256 1 ==PROF== Profiling "sigmaKin": launch__registers_per_thread 154 ==PROF== Profiling "sigmaKin": sm__sass_average_branch_targets_threads_uniform.pct 100% @@ -85,15 +85,15 @@ Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=1 Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/none+NAVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = SCALAR ('none': ~vector[1], no SIMD) -EvtsPerSec[Rmb+ME] (23) = ( 2.084164e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 2.941928e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 2.941928e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 2.090670e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 2.931602e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.931602e+06 ) sec^-1 MeanMatrixElemValue = ( 1.371706e-02 +- 3.270315e-06 ) GeV^0 -TOTAL : 3.469758 sec +TOTAL : 3.421272 sec INFO: No Floating Point Exceptions have been reported - 10,217,900,291 cycles # 2.936 GHz - 25,614,437,724 instructions # 2.51 insn per cycle - 3.480862891 seconds time elapsed + 10,000,792,824 cycles # 2.920 GHz + 25,510,981,237 instructions # 2.55 insn per cycle + 3.426545380 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 236) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum/build.none_d_inl1_hrd1/runTest_cpp.exe @@ -117,15 +117,15 @@ Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=1 Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[2] ('sse4': SSE4.2, 128bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 2.313032e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.558172e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.558172e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 2.321116e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.548747e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.548747e+06 ) sec^-1 MeanMatrixElemValue = ( 1.371706e-02 +- 3.270315e-06 ) GeV^0 -TOTAL : 3.173341 sec +TOTAL : 3.127808 sec INFO: No Floating Point Exceptions have been reported - 9,354,473,123 cycles # 2.939 GHz - 21,650,720,885 instructions # 2.31 insn per cycle - 3.184272296 seconds time elapsed + 9,074,162,171 cycles # 2.897 GHz + 21,544,244,995 instructions # 2.37 insn per cycle + 3.133236084 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 1112) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum/build.sse4_d_inl1_hrd1/runTest_cpp.exe @@ -149,15 +149,15 @@ Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=1 Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('avx2': AVX2, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 2.358550e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.604458e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.604458e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 2.393375e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.623848e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.623848e+06 ) sec^-1 MeanMatrixElemValue = ( 1.371706e-02 +- 3.270315e-06 ) GeV^0 -TOTAL : 3.115986 sec +TOTAL : 3.036686 sec INFO: No Floating Point Exceptions have been reported - 8,850,186,465 cycles # 2.831 GHz - 16,062,849,181 instructions # 1.81 insn per cycle - 3.126797345 seconds time elapsed + 8,579,757,723 cycles # 2.822 GHz + 15,960,183,986 instructions # 1.86 insn per cycle + 3.042036675 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 1497) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum/build.avx2_d_inl1_hrd1/runTest_cpp.exe @@ -181,15 +181,15 @@ Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=1 Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('512y': AVX512, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 2.422935e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.724037e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.724037e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 2.467079e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.760040e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.760040e+06 ) sec^-1 MeanMatrixElemValue = ( 1.371706e-02 +- 3.270315e-06 ) GeV^0 -TOTAL : 3.043301 sec +TOTAL : 2.954362 sec INFO: No Floating Point Exceptions have been reported - 8,651,791,606 cycles # 2.834 GHz - 15,666,461,627 instructions # 1.81 insn per cycle - 3.054177777 seconds time elapsed + 8,375,901,213 cycles # 2.831 GHz + 15,563,126,307 instructions # 1.86 insn per cycle + 2.959574048 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 1264) (512y: 141) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum/build.512y_d_inl1_hrd1/runTest_cpp.exe @@ -213,15 +213,15 @@ Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=1 Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('512z': AVX512, 512bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 2.052275e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 2.908416e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 2.908416e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 2.065843e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 2.902912e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.902912e+06 ) sec^-1 MeanMatrixElemValue = ( 1.371706e-02 +- 3.270315e-06 ) GeV^0 -TOTAL : 3.514529 sec +TOTAL : 3.452572 sec INFO: No Floating Point Exceptions have been reported - 7,791,531,975 cycles # 2.211 GHz - 14,393,714,103 instructions # 1.85 insn per cycle - 3.525649878 seconds time elapsed + 7,599,329,079 cycles # 2.198 GHz + 14,288,734,983 instructions # 1.88 insn per cycle + 3.457812872 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 1031) (512y: 164) (512z: 876) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum/build.512z_d_inl1_hrd1/runTest_cpp.exe diff --git a/epochX/cudacpp/tput/logs_eemumu_mad/log_eemumu_mad_f_inl0_hrd0.txt b/epochX/cudacpp/tput/logs_eemumu_mad/log_eemumu_mad_f_inl0_hrd0.txt index fa697401ba..de1e63cf32 100644 --- a/epochX/cudacpp/tput/logs_eemumu_mad/log_eemumu_mad_f_inl0_hrd0.txt +++ b/epochX/cudacpp/tput/logs_eemumu_mad/log_eemumu_mad_f_inl0_hrd0.txt @@ -40,7 +40,7 @@ make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum' -DATE: 2024-10-06_08:56:56 +DATE: 2024-10-06_13:26:11 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: ========================================================================= @@ -49,15 +49,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_SM_EPEM_MUPMUM_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 1.318402e+08 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 6.547340e+08 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 8.573294e+08 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.419757e+08 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 6.859299e+08 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 8.848437e+08 ) sec^-1 MeanMatrixElemValue = ( 1.371687e-02 +- 3.270220e-06 ) GeV^0 -TOTAL : 0.580539 sec +TOTAL : 0.571418 sec INFO: No Floating Point Exceptions have been reported - 2,318,735,379 cycles # 2.865 GHz - 3,612,120,055 instructions # 1.56 insn per cycle - 0.879357898 seconds time elapsed + 2,300,992,217 cycles # 2.886 GHz + 3,625,442,222 instructions # 1.58 insn per cycle + 0.854852985 seconds time elapsed runNcu /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum/build.cuda_f_inl0_hrd0/check_cuda.exe -p 2048 256 1 ==PROF== Profiling "sigmaKin": launch__registers_per_thread 109 ==PROF== Profiling "sigmaKin": sm__sass_average_branch_targets_threads_uniform.pct 100% @@ -85,15 +85,15 @@ Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=0 Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/none+NAVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = SCALAR ('none': ~vector[1], no SIMD) -EvtsPerSec[Rmb+ME] (23) = ( 1.072197e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.275533e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.275533e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.076590e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.270875e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.270875e+06 ) sec^-1 MeanMatrixElemValue = ( 1.371707e-02 +- 3.270376e-06 ) GeV^0 -TOTAL : 6.274259 sec +TOTAL : 6.212065 sec INFO: No Floating Point Exceptions have been reported - 18,464,131,410 cycles # 2.940 GHz - 45,058,020,075 instructions # 2.44 insn per cycle - 6.281329583 seconds time elapsed + 18,196,256,071 cycles # 2.927 GHz + 45,006,924,286 instructions # 2.47 insn per cycle + 6.217006136 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 411) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum/build.none_f_inl0_hrd0/runTest_cpp.exe @@ -117,15 +117,15 @@ Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=0 Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('sse4': SSE4.2, 128bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 2.257463e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.446957e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.446957e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 2.262279e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.451327e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.451327e+06 ) sec^-1 MeanMatrixElemValue = ( 1.371707e-02 +- 3.270376e-06 ) GeV^0 -TOTAL : 3.180369 sec +TOTAL : 3.155082 sec INFO: No Floating Point Exceptions have been reported - 9,372,467,471 cycles # 2.941 GHz - 22,319,965,268 instructions # 2.38 insn per cycle - 3.189536232 seconds time elapsed + 9,263,231,347 cycles # 2.932 GHz + 22,275,004,889 instructions # 2.40 insn per cycle + 3.159976719 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 1954) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum/build.sse4_f_inl0_hrd0/runTest_cpp.exe @@ -149,15 +149,15 @@ Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=0 Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('avx2': AVX2, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 2.408379e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.710073e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.710073e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 2.422148e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.700397e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.700397e+06 ) sec^-1 MeanMatrixElemValue = ( 1.371706e-02 +- 3.270341e-06 ) GeV^0 -TOTAL : 3.001033 sec +TOTAL : 2.963023 sec INFO: No Floating Point Exceptions have been reported - 8,493,792,111 cycles # 2.825 GHz - 15,797,222,111 instructions # 1.86 insn per cycle - 3.010052254 seconds time elapsed + 8,379,935,544 cycles # 2.824 GHz + 15,753,700,438 instructions # 1.88 insn per cycle + 2.967976624 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 2565) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum/build.avx2_f_inl0_hrd0/runTest_cpp.exe @@ -181,15 +181,15 @@ Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=0 Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('512y': AVX512, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 2.426130e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.768067e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.768067e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 2.457419e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.782563e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.782563e+06 ) sec^-1 MeanMatrixElemValue = ( 1.371706e-02 +- 3.270341e-06 ) GeV^0 -TOTAL : 2.984448 sec +TOTAL : 2.927734 sec INFO: No Floating Point Exceptions have been reported - 8,427,466,763 cycles # 2.818 GHz - 15,640,000,146 instructions # 1.86 insn per cycle - 2.993491493 seconds time elapsed + 8,279,567,219 cycles # 2.824 GHz + 15,594,458,753 instructions # 1.88 insn per cycle + 2.932642748 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 2462) (512y: 12) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum/build.512y_f_inl0_hrd0/runTest_cpp.exe @@ -213,15 +213,15 @@ Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=0 Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[16] ('512z': AVX512, 512bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 2.427110e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.709739e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.709739e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 2.464455e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.763027e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.763027e+06 ) sec^-1 MeanMatrixElemValue = ( 1.371706e-02 +- 3.270342e-06 ) GeV^0 -TOTAL : 2.984139 sec +TOTAL : 2.925238 sec INFO: No Floating Point Exceptions have been reported - 6,725,622,216 cycles # 2.249 GHz - 12,910,486,373 instructions # 1.92 insn per cycle - 2.994013668 seconds time elapsed + 6,621,698,711 cycles # 2.261 GHz + 12,864,557,328 instructions # 1.94 insn per cycle + 2.930515499 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 1698) (512y: 16) (512z: 1440) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum/build.512z_f_inl0_hrd0/runTest_cpp.exe diff --git a/epochX/cudacpp/tput/logs_eemumu_mad/log_eemumu_mad_f_inl0_hrd0_bridge.txt b/epochX/cudacpp/tput/logs_eemumu_mad/log_eemumu_mad_f_inl0_hrd0_bridge.txt index 9136826931..fb31941e68 100644 --- a/epochX/cudacpp/tput/logs_eemumu_mad/log_eemumu_mad_f_inl0_hrd0_bridge.txt +++ b/epochX/cudacpp/tput/logs_eemumu_mad/log_eemumu_mad_f_inl0_hrd0_bridge.txt @@ -40,7 +40,7 @@ make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum' -DATE: 2024-10-06_09:37:07 +DATE: 2024-10-06_13:55:40 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: ========================================================================= @@ -53,16 +53,16 @@ WARNING! Set grid in Bridge (nevt=524288, gpublocks=2048, gputhreads=256, gpublo Process = SIGMA_SM_EPEM_MUPMUM_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 7.256593e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 7.121486e+07 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 7.121486e+07 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 7.097584e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 7.099576e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 7.099576e+07 ) sec^-1 MeanMatrixElemValue = ( 1.371710e-02 +- 3.270389e-06 ) GeV^0 -TOTAL : 1.691319 sec +TOTAL : 1.718198 sec INFO: No Floating Point Exceptions have been reported INFO: No Floating Point Exceptions have been reported - 5,610,006,933 cycles # 2.911 GHz - 10,218,919,767 instructions # 1.82 insn per cycle - 1.984436466 seconds time elapsed + 5,626,784,721 cycles # 2.887 GHz + 10,174,365,691 instructions # 1.81 insn per cycle + 2.008018816 seconds time elapsed runNcu /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum/build.cuda_f_inl0_hrd0/check_cuda.exe -p 2048 256 1 --bridge WARNING! Bridge selected: cannot use RamboDevice, will use RamboHost WARNING! RamboHost selected: cannot use CurandDevice, will use CurandHost @@ -95,16 +95,16 @@ Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=0 Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+BRDHST/none+NAVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = SCALAR ('none': ~vector[1], no SIMD) -EvtsPerSec[Rmb+ME] (23) = ( 1.060836e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.248384e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.248384e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.058680e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.250441e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.250441e+06 ) sec^-1 MeanMatrixElemValue = ( 1.371707e-02 +- 3.270376e-06 ) GeV^0 -TOTAL : 6.418392 sec +TOTAL : 6.426029 sec INFO: No Floating Point Exceptions have been reported INFO: No Floating Point Exceptions have been reported - 18,916,088,440 cycles # 2.945 GHz - 45,156,650,630 instructions # 2.39 insn per cycle - 6.425565221 seconds time elapsed + 18,943,893,209 cycles # 2.945 GHz + 45,156,764,654 instructions # 2.38 insn per cycle + 6.432870502 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 411) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum/build.none_f_inl0_hrd0/runTest_cpp.exe @@ -129,16 +129,16 @@ Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=0 Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+BRDHST/sse4+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('sse4': SSE4.2, 128bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 2.163234e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.223206e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.223206e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 2.155767e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.205831e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.205831e+06 ) sec^-1 MeanMatrixElemValue = ( 1.371707e-02 +- 3.270376e-06 ) GeV^0 -TOTAL : 3.414716 sec +TOTAL : 3.423667 sec INFO: No Floating Point Exceptions have been reported INFO: No Floating Point Exceptions have been reported - 10,073,193,872 cycles # 2.945 GHz - 23,610,645,909 instructions # 2.34 insn per cycle - 3.421707000 seconds time elapsed + 10,106,259,185 cycles # 2.947 GHz + 23,615,083,291 instructions # 2.34 insn per cycle + 3.430439821 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 1954) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum/build.sse4_f_inl0_hrd0/runTest_cpp.exe @@ -163,16 +163,16 @@ Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=0 Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+BRDHST/avx2+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('avx2': AVX2, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 2.302389e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.467769e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.467769e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 2.317024e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.476940e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.476940e+06 ) sec^-1 MeanMatrixElemValue = ( 1.371706e-02 +- 3.270341e-06 ) GeV^0 -TOTAL : 3.241454 sec +TOTAL : 3.216595 sec INFO: No Floating Point Exceptions have been reported INFO: No Floating Point Exceptions have been reported - 9,215,037,610 cycles # 2.837 GHz - 16,874,646,512 instructions # 1.83 insn per cycle - 3.248598680 seconds time elapsed + 9,141,895,834 cycles # 2.837 GHz + 16,874,531,478 instructions # 1.85 insn per cycle + 3.223659537 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 2565) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum/build.avx2_f_inl0_hrd0/runTest_cpp.exe @@ -197,16 +197,16 @@ Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=0 Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+BRDHST/512y+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('512y': AVX512, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 2.316990e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.533576e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.533576e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 2.340532e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.535196e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.535196e+06 ) sec^-1 MeanMatrixElemValue = ( 1.371706e-02 +- 3.270341e-06 ) GeV^0 -TOTAL : 3.224710 sec +TOTAL : 3.189329 sec INFO: No Floating Point Exceptions have been reported INFO: No Floating Point Exceptions have been reported - 9,166,209,661 cycles # 2.837 GHz - 16,710,284,997 instructions # 1.82 insn per cycle - 3.231713030 seconds time elapsed + 9,069,527,459 cycles # 2.838 GHz + 16,710,237,140 instructions # 1.84 insn per cycle + 3.196643307 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 2462) (512y: 12) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum/build.512y_f_inl0_hrd0/runTest_cpp.exe @@ -231,16 +231,16 @@ Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=0 Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+BRDHST/512z+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[16] ('512z': AVX512, 512bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 2.333210e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.469405e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.469405e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 2.327550e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.458993e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.458993e+06 ) sec^-1 MeanMatrixElemValue = ( 1.371706e-02 +- 3.270342e-06 ) GeV^0 -TOTAL : 3.205451 sec +TOTAL : 3.208255 sec INFO: No Floating Point Exceptions have been reported INFO: No Floating Point Exceptions have been reported - 7,432,634,328 cycles # 2.315 GHz - 14,074,642,515 instructions # 1.89 insn per cycle - 3.212353581 seconds time elapsed + 7,425,695,874 cycles # 2.310 GHz + 14,072,561,642 instructions # 1.90 insn per cycle + 3.215282834 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 1698) (512y: 16) (512z: 1440) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum/build.512z_f_inl0_hrd0/runTest_cpp.exe diff --git a/epochX/cudacpp/tput/logs_eemumu_mad/log_eemumu_mad_f_inl0_hrd0_common.txt b/epochX/cudacpp/tput/logs_eemumu_mad/log_eemumu_mad_f_inl0_hrd0_common.txt index 3c8228d85b..ab0cc9d689 100644 --- a/epochX/cudacpp/tput/logs_eemumu_mad/log_eemumu_mad_f_inl0_hrd0_common.txt +++ b/epochX/cudacpp/tput/logs_eemumu_mad/log_eemumu_mad_f_inl0_hrd0_common.txt @@ -40,7 +40,7 @@ make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum' -DATE: 2024-10-06_09:49:18 +DATE: 2024-10-06_14:07:50 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: ========================================================================= @@ -49,15 +49,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_SM_EPEM_MUPMUM_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:COMMON+RMBDEV+MESDEV/none+NAVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 1.233592e+08 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 6.244967e+08 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 8.184868e+08 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.282865e+08 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 6.346710e+08 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 8.351156e+08 ) sec^-1 MeanMatrixElemValue = ( 1.371863e-02 +- 3.269951e-06 ) GeV^0 -TOTAL : 1.220966 sec +TOTAL : 1.191185 sec INFO: No Floating Point Exceptions have been reported - 4,183,681,416 cycles # 2.867 GHz - 6,662,508,205 instructions # 1.59 insn per cycle - 1.516447212 seconds time elapsed + 4,126,006,057 cycles # 2.912 GHz + 6,651,338,537 instructions # 1.61 insn per cycle + 1.473702727 seconds time elapsed runNcu /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum/build.cuda_f_inl0_hrd0/check_cuda.exe -p 2048 256 1 --common ==PROF== Profiling "sigmaKin": launch__registers_per_thread 109 ==PROF== Profiling "sigmaKin": sm__sass_average_branch_targets_threads_uniform.pct 100% @@ -85,15 +85,15 @@ Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=0 Workflow summary = CPP:FLT+CXS:COMMON+RMBHST+MESHST/none+NAVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = SCALAR ('none': ~vector[1], no SIMD) -EvtsPerSec[Rmb+ME] (23) = ( 1.080178e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.275874e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.275874e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.079774e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.274680e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.274680e+06 ) sec^-1 MeanMatrixElemValue = ( 1.371887e-02 +- 3.270267e-06 ) GeV^0 -TOTAL : 6.537773 sec +TOTAL : 6.538906 sec INFO: No Floating Point Exceptions have been reported - 19,269,764,932 cycles # 2.946 GHz - 45,190,617,795 instructions # 2.35 insn per cycle - 6.543013626 seconds time elapsed + 19,264,590,146 cycles # 2.944 GHz + 45,188,683,576 instructions # 2.35 insn per cycle + 6.544345921 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 411) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum/build.none_f_inl0_hrd0/runTest_cpp.exe @@ -117,15 +117,15 @@ Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=0 Workflow summary = CPP:FLT+CXS:COMMON+RMBHST+MESHST/sse4+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('sse4': SSE4.2, 128bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 2.263942e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.453881e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.453881e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 2.271968e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.465796e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.465796e+06 ) sec^-1 MeanMatrixElemValue = ( 1.371887e-02 +- 3.270266e-06 ) GeV^0 -TOTAL : 3.487545 sec +TOTAL : 3.479464 sec INFO: No Floating Point Exceptions have been reported - 10,298,424,695 cycles # 2.949 GHz - 22,355,388,978 instructions # 2.17 insn per cycle - 3.493059791 seconds time elapsed + 10,279,075,927 cycles # 2.950 GHz + 22,355,322,708 instructions # 2.17 insn per cycle + 3.484861410 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 1954) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum/build.sse4_f_inl0_hrd0/runTest_cpp.exe @@ -149,15 +149,15 @@ Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=0 Workflow summary = CPP:FLT+CXS:COMMON+RMBHST+MESHST/avx2+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('avx2': AVX2, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 2.406924e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.701531e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.701531e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 2.414715e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.704184e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.704184e+06 ) sec^-1 MeanMatrixElemValue = ( 1.371885e-02 +- 3.270112e-06 ) GeV^0 -TOTAL : 3.322767 sec +TOTAL : 3.320341 sec INFO: No Floating Point Exceptions have been reported - 9,443,809,325 cycles # 2.838 GHz - 15,664,102,195 instructions # 1.66 insn per cycle - 3.328357008 seconds time elapsed + 9,427,218,822 cycles # 2.836 GHz + 15,664,192,369 instructions # 1.66 insn per cycle + 3.325866465 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 2565) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum/build.avx2_f_inl0_hrd0/runTest_cpp.exe @@ -181,15 +181,15 @@ Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=0 Workflow summary = CPP:FLT+CXS:COMMON+RMBHST+MESHST/512y+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('512y': AVX512, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 2.446360e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.803645e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.803645e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 2.458838e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.799495e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.799495e+06 ) sec^-1 MeanMatrixElemValue = ( 1.371885e-02 +- 3.270112e-06 ) GeV^0 -TOTAL : 3.287328 sec +TOTAL : 3.277724 sec INFO: No Floating Point Exceptions have been reported - 9,371,124,961 cycles # 2.847 GHz - 15,299,944,141 instructions # 1.63 insn per cycle - 3.292839828 seconds time elapsed + 9,319,976,438 cycles # 2.841 GHz + 15,305,707,646 instructions # 1.64 insn per cycle + 3.283285505 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 2462) (512y: 12) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum/build.512y_f_inl0_hrd0/runTest_cpp.exe @@ -213,15 +213,15 @@ Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=0 Workflow summary = CPP:FLT+CXS:COMMON+RMBHST+MESHST/512z+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[16] ('512z': AVX512, 512bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 2.466708e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.777222e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.777222e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 2.453878e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.750711e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.750711e+06 ) sec^-1 MeanMatrixElemValue = ( 1.371885e-02 +- 3.270112e-06 ) GeV^0 -TOTAL : 3.269312 sec +TOTAL : 3.286948 sec INFO: No Floating Point Exceptions have been reported - 7,659,274,117 cycles # 2.340 GHz - 12,573,895,764 instructions # 1.64 insn per cycle - 3.274843213 seconds time elapsed + 7,678,046,171 cycles # 2.333 GHz + 12,575,191,159 instructions # 1.64 insn per cycle + 3.292461237 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 1698) (512y: 16) (512z: 1440) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum/build.512z_f_inl0_hrd0/runTest_cpp.exe diff --git a/epochX/cudacpp/tput/logs_eemumu_mad/log_eemumu_mad_f_inl0_hrd0_curhst.txt b/epochX/cudacpp/tput/logs_eemumu_mad/log_eemumu_mad_f_inl0_hrd0_curhst.txt index 7f30dafdfd..f7ae3eef6d 100644 --- a/epochX/cudacpp/tput/logs_eemumu_mad/log_eemumu_mad_f_inl0_hrd0_curhst.txt +++ b/epochX/cudacpp/tput/logs_eemumu_mad/log_eemumu_mad_f_inl0_hrd0_curhst.txt @@ -40,7 +40,7 @@ make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum' -DATE: 2024-10-06_09:46:29 +DATE: 2024-10-06_14:05:02 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: ========================================================================= @@ -49,15 +49,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_SM_EPEM_MUPMUM_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:CURHST+RMBDEV+MESDEV/none+NAVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 1.282321e+08 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 6.333955e+08 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 8.369324e+08 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.303644e+08 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 6.434361e+08 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 8.418232e+08 ) sec^-1 MeanMatrixElemValue = ( 1.371687e-02 +- 3.270220e-06 ) GeV^0 -TOTAL : 0.867789 sec +TOTAL : 0.866254 sec INFO: No Floating Point Exceptions have been reported - 3,167,199,789 cycles # 2.899 GHz - 6,506,216,930 instructions # 2.05 insn per cycle - 1.149942283 seconds time elapsed + 3,152,800,861 cycles # 2.890 GHz + 6,465,345,394 instructions # 2.05 insn per cycle + 1.147660465 seconds time elapsed runNcu /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum/build.cuda_f_inl0_hrd0/check_cuda.exe -p 2048 256 1 --curhst ==PROF== Profiling "sigmaKin": launch__registers_per_thread 109 ==PROF== Profiling "sigmaKin": sm__sass_average_branch_targets_threads_uniform.pct 100% @@ -85,15 +85,15 @@ Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=0 Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/none+NAVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = SCALAR ('none': ~vector[1], no SIMD) -EvtsPerSec[Rmb+ME] (23) = ( 1.085219e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.281583e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.281583e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.080571e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.278486e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.278486e+06 ) sec^-1 MeanMatrixElemValue = ( 1.371707e-02 +- 3.270376e-06 ) GeV^0 -TOTAL : 6.166056 sec +TOTAL : 6.191465 sec INFO: No Floating Point Exceptions have been reported - 18,234,644,828 cycles # 2.955 GHz - 45,008,398,832 instructions # 2.47 insn per cycle - 6.171760600 seconds time elapsed + 18,259,169,052 cycles # 2.948 GHz + 45,011,584,326 instructions # 2.47 insn per cycle + 6.196701590 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 411) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum/build.none_f_inl0_hrd0/runTest_cpp.exe @@ -117,15 +117,15 @@ Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=0 Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('sse4': SSE4.2, 128bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 2.256894e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.462086e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.462086e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 2.274532e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.468156e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.468156e+06 ) sec^-1 MeanMatrixElemValue = ( 1.371707e-02 +- 3.270376e-06 ) GeV^0 -TOTAL : 3.159870 sec +TOTAL : 3.140643 sec INFO: No Floating Point Exceptions have been reported - 9,347,982,513 cycles # 2.954 GHz - 22,275,896,372 instructions # 2.38 insn per cycle - 3.165402193 seconds time elapsed + 9,278,450,051 cycles # 2.950 GHz + 22,275,298,589 instructions # 2.40 insn per cycle + 3.146145270 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 1954) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum/build.sse4_f_inl0_hrd0/runTest_cpp.exe @@ -149,15 +149,15 @@ Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=0 Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('avx2': AVX2, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 2.410366e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.712636e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.712636e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 2.408818e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.681338e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.681338e+06 ) sec^-1 MeanMatrixElemValue = ( 1.371706e-02 +- 3.270341e-06 ) GeV^0 -TOTAL : 2.983322 sec +TOTAL : 2.981008 sec INFO: No Floating Point Exceptions have been reported - 8,463,194,185 cycles # 2.833 GHz - 15,755,395,679 instructions # 1.86 insn per cycle - 2.988746216 seconds time elapsed + 8,418,448,374 cycles # 2.820 GHz + 15,754,050,232 instructions # 1.87 insn per cycle + 2.986467555 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 2565) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum/build.avx2_f_inl0_hrd0/runTest_cpp.exe @@ -181,15 +181,15 @@ Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=0 Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('512y': AVX512, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 2.454105e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.801490e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.801490e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 2.450602e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.770644e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.770644e+06 ) sec^-1 MeanMatrixElemValue = ( 1.371706e-02 +- 3.270341e-06 ) GeV^0 -TOTAL : 2.933599 sec +TOTAL : 2.940330 sec INFO: No Floating Point Exceptions have been reported - 8,319,397,972 cycles # 2.832 GHz - 15,593,973,322 instructions # 1.87 insn per cycle - 2.939101584 seconds time elapsed + 8,276,018,757 cycles # 2.810 GHz + 15,593,784,301 instructions # 1.88 insn per cycle + 2.945891422 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 2462) (512y: 12) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum/build.512y_f_inl0_hrd0/runTest_cpp.exe @@ -213,15 +213,15 @@ Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=0 Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[16] ('512z': AVX512, 512bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 2.469652e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.768397e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.768397e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 2.450470e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.766365e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.766365e+06 ) sec^-1 MeanMatrixElemValue = ( 1.371706e-02 +- 3.270342e-06 ) GeV^0 -TOTAL : 2.922384 sec +TOTAL : 2.943039 sec INFO: No Floating Point Exceptions have been reported - 6,636,368,959 cycles # 2.267 GHz - 12,865,256,567 instructions # 1.94 insn per cycle - 2.927905791 seconds time elapsed + 6,672,009,456 cycles # 2.264 GHz + 12,863,845,260 instructions # 1.93 insn per cycle + 2.948518548 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 1698) (512y: 16) (512z: 1440) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum/build.512z_f_inl0_hrd0/runTest_cpp.exe diff --git a/epochX/cudacpp/tput/logs_eemumu_mad/log_eemumu_mad_f_inl0_hrd0_rmbhst.txt b/epochX/cudacpp/tput/logs_eemumu_mad/log_eemumu_mad_f_inl0_hrd0_rmbhst.txt index e2ecb9b5fd..d4a5c63657 100644 --- a/epochX/cudacpp/tput/logs_eemumu_mad/log_eemumu_mad_f_inl0_hrd0_rmbhst.txt +++ b/epochX/cudacpp/tput/logs_eemumu_mad/log_eemumu_mad_f_inl0_hrd0_rmbhst.txt @@ -40,7 +40,7 @@ make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum' -DATE: 2024-10-06_09:43:42 +DATE: 2024-10-06_14:02:15 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: ========================================================================= @@ -50,15 +50,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_SM_EPEM_MUPMUM_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:CURHST+RMBHST+MESDEV/none+NAVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 8.979354e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 6.311142e+08 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 8.251832e+08 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 9.059137e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 6.288702e+08 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 8.239489e+08 ) sec^-1 MeanMatrixElemValue = ( 1.371710e-02 +- 3.270389e-06 ) GeV^0 -TOTAL : 1.493081 sec +TOTAL : 1.494849 sec INFO: No Floating Point Exceptions have been reported - 5,009,051,141 cycles # 2.916 GHz - 9,204,393,500 instructions # 1.84 insn per cycle - 1.774548277 seconds time elapsed + 5,006,150,613 cycles # 2.913 GHz + 9,170,133,537 instructions # 1.83 insn per cycle + 1.777433062 seconds time elapsed runNcu /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum/build.cuda_f_inl0_hrd0/check_cuda.exe -p 2048 256 1 --rmbhst WARNING! RamboHost selected: cannot use CurandDevice, will use CurandHost ==PROF== Profiling "sigmaKin": launch__registers_per_thread 109 @@ -87,15 +87,15 @@ Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=0 Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/none+NAVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = SCALAR ('none': ~vector[1], no SIMD) -EvtsPerSec[Rmb+ME] (23) = ( 1.077151e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.276926e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.276926e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.081289e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.278924e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.278924e+06 ) sec^-1 MeanMatrixElemValue = ( 1.371707e-02 +- 3.270376e-06 ) GeV^0 -TOTAL : 6.211513 sec +TOTAL : 6.187343 sec INFO: No Floating Point Exceptions have been reported - 18,299,232,198 cycles # 2.944 GHz - 45,005,768,829 instructions # 2.46 insn per cycle - 6.217115880 seconds time elapsed + 18,275,542,697 cycles # 2.952 GHz + 45,011,276,231 instructions # 2.46 insn per cycle + 6.192571804 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 411) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum/build.none_f_inl0_hrd0/runTest_cpp.exe @@ -119,15 +119,15 @@ Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=0 Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('sse4': SSE4.2, 128bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 2.268380e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.460029e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.460029e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 2.263693e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.451948e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.451948e+06 ) sec^-1 MeanMatrixElemValue = ( 1.371707e-02 +- 3.270376e-06 ) GeV^0 -TOTAL : 3.148224 sec +TOTAL : 3.155673 sec INFO: No Floating Point Exceptions have been reported - 9,293,240,022 cycles # 2.948 GHz - 22,275,553,802 instructions # 2.40 insn per cycle - 3.153857529 seconds time elapsed + 9,280,168,842 cycles # 2.936 GHz + 22,274,050,046 instructions # 2.40 insn per cycle + 3.161240450 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 1954) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum/build.sse4_f_inl0_hrd0/runTest_cpp.exe @@ -151,15 +151,15 @@ Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=0 Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('avx2': AVX2, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 2.395770e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.675698e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.675698e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 2.434611e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.722626e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.722626e+06 ) sec^-1 MeanMatrixElemValue = ( 1.371706e-02 +- 3.270341e-06 ) GeV^0 -TOTAL : 2.994942 sec +TOTAL : 2.952658 sec INFO: No Floating Point Exceptions have been reported - 8,447,981,393 cycles # 2.817 GHz - 15,754,576,494 instructions # 1.86 insn per cycle - 3.000419944 seconds time elapsed + 8,392,678,825 cycles # 2.838 GHz + 15,752,611,842 instructions # 1.88 insn per cycle + 2.958157845 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 2565) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum/build.avx2_f_inl0_hrd0/runTest_cpp.exe @@ -183,15 +183,15 @@ Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=0 Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('512y': AVX512, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 2.419912e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.751119e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.751119e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 2.469173e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.804638e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.804638e+06 ) sec^-1 MeanMatrixElemValue = ( 1.371706e-02 +- 3.270341e-06 ) GeV^0 -TOTAL : 2.971435 sec +TOTAL : 2.918079 sec INFO: No Floating Point Exceptions have been reported - 8,357,800,499 cycles # 2.808 GHz - 15,594,139,449 instructions # 1.87 insn per cycle - 2.977163262 seconds time elapsed + 8,274,026,636 cycles # 2.831 GHz + 15,588,381,345 instructions # 1.88 insn per cycle + 2.923788121 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 2462) (512y: 12) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum/build.512y_f_inl0_hrd0/runTest_cpp.exe @@ -215,15 +215,15 @@ Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=0 Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[16] ('512z': AVX512, 512bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 2.455367e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.730952e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.730952e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 2.464699e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.779174e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.779174e+06 ) sec^-1 MeanMatrixElemValue = ( 1.371706e-02 +- 3.270342e-06 ) GeV^0 -TOTAL : 2.933639 sec +TOTAL : 2.925348 sec INFO: No Floating Point Exceptions have been reported - 6,669,997,057 cycles # 2.271 GHz - 12,867,351,511 instructions # 1.93 insn per cycle - 2.938851588 seconds time elapsed + 6,658,128,291 cycles # 2.273 GHz + 12,863,781,851 instructions # 1.93 insn per cycle + 2.930985147 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 1698) (512y: 16) (512z: 1440) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum/build.512z_f_inl0_hrd0/runTest_cpp.exe diff --git a/epochX/cudacpp/tput/logs_eemumu_mad/log_eemumu_mad_f_inl0_hrd1.txt b/epochX/cudacpp/tput/logs_eemumu_mad/log_eemumu_mad_f_inl0_hrd1.txt index 9e915de581..875cd73043 100644 --- a/epochX/cudacpp/tput/logs_eemumu_mad/log_eemumu_mad_f_inl0_hrd1.txt +++ b/epochX/cudacpp/tput/logs_eemumu_mad/log_eemumu_mad_f_inl0_hrd1.txt @@ -40,7 +40,7 @@ make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum' -DATE: 2024-10-06_08:57:23 +DATE: 2024-10-06_13:26:37 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: ========================================================================= @@ -49,15 +49,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_SM_EPEM_MUPMUM_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=1] Workflow summary = CUD:FLT+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 1.310707e+08 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 6.890276e+08 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 9.030864e+08 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.432981e+08 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 7.178332e+08 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 9.385058e+08 ) sec^-1 MeanMatrixElemValue = ( 1.371687e-02 +- 3.270220e-06 ) GeV^0 -TOTAL : 0.577005 sec +TOTAL : 0.569251 sec INFO: No Floating Point Exceptions have been reported - 2,340,023,876 cycles # 2.880 GHz - 3,638,052,704 instructions # 1.55 insn per cycle - 0.886148283 seconds time elapsed + 2,279,293,688 cycles # 2.862 GHz + 3,545,637,406 instructions # 1.56 insn per cycle + 0.854622111 seconds time elapsed runNcu /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum/build.cuda_f_inl0_hrd1/check_cuda.exe -p 2048 256 1 ==PROF== Profiling "sigmaKin": launch__registers_per_thread 79 ==PROF== Profiling "sigmaKin": sm__sass_average_branch_targets_threads_uniform.pct 100% @@ -85,15 +85,15 @@ Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=0 Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/none+NAVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = SCALAR ('none': ~vector[1], no SIMD) -EvtsPerSec[Rmb+ME] (23) = ( 1.074456e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.269687e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.269687e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.077239e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.270955e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.270955e+06 ) sec^-1 MeanMatrixElemValue = ( 1.371707e-02 +- 3.270376e-06 ) GeV^0 -TOTAL : 6.244443 sec +TOTAL : 6.204463 sec INFO: No Floating Point Exceptions have been reported - 18,377,232,357 cycles # 2.941 GHz - 45,025,324,964 instructions # 2.45 insn per cycle - 6.253002386 seconds time elapsed + 18,245,705,968 cycles # 2.939 GHz + 44,981,074,419 instructions # 2.47 insn per cycle + 6.209565981 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 397) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum/build.none_f_inl0_hrd1/runTest_cpp.exe @@ -117,15 +117,15 @@ Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=0 Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('sse4': SSE4.2, 128bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 2.251309e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.439034e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.439034e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 2.264044e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.451202e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.451202e+06 ) sec^-1 MeanMatrixElemValue = ( 1.371707e-02 +- 3.270376e-06 ) GeV^0 -TOTAL : 3.184453 sec +TOTAL : 3.151629 sec INFO: No Floating Point Exceptions have been reported - 9,383,250,913 cycles # 2.940 GHz - 22,280,358,761 instructions # 2.37 insn per cycle - 3.194375038 seconds time elapsed + 9,261,679,023 cycles # 2.935 GHz + 22,236,314,458 instructions # 2.40 insn per cycle + 3.156687968 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 1935) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum/build.sse4_f_inl0_hrd1/runTest_cpp.exe @@ -149,15 +149,15 @@ Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=0 Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('avx2': AVX2, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 2.403334e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.700033e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.700033e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 2.399177e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.656523e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.656523e+06 ) sec^-1 MeanMatrixElemValue = ( 1.371706e-02 +- 3.270341e-06 ) GeV^0 -TOTAL : 3.004384 sec +TOTAL : 2.989463 sec INFO: No Floating Point Exceptions have been reported - 8,513,730,278 cycles # 2.827 GHz - 15,791,909,505 instructions # 1.85 insn per cycle - 3.013283160 seconds time elapsed + 8,381,938,671 cycles # 2.801 GHz + 15,748,042,457 instructions # 1.88 insn per cycle + 2.994499142 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 2540) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum/build.avx2_f_inl0_hrd1/runTest_cpp.exe @@ -181,15 +181,15 @@ Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=0 Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('512y': AVX512, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 2.444935e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.799463e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.799463e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 2.463666e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.784579e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.784579e+06 ) sec^-1 MeanMatrixElemValue = ( 1.371706e-02 +- 3.270341e-06 ) GeV^0 -TOTAL : 2.959849 sec +TOTAL : 2.919759 sec INFO: No Floating Point Exceptions have been reported - 8,395,161,248 cycles # 2.830 GHz - 15,634,676,534 instructions # 1.86 insn per cycle - 2.968734397 seconds time elapsed + 8,257,947,998 cycles # 2.825 GHz + 15,585,237,522 instructions # 1.89 insn per cycle + 2.924727630 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 2438) (512y: 10) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum/build.512y_f_inl0_hrd1/runTest_cpp.exe @@ -213,15 +213,15 @@ Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=0 Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[16] ('512z': AVX512, 512bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 2.454317e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.767111e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.767111e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 2.470287e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.780303e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.780303e+06 ) sec^-1 MeanMatrixElemValue = ( 1.371706e-02 +- 3.270342e-06 ) GeV^0 -TOTAL : 2.955150 sec +TOTAL : 2.916451 sec INFO: No Floating Point Exceptions have been reported - 6,701,822,130 cycles # 2.263 GHz - 12,886,633,037 instructions # 1.92 insn per cycle - 2.963931226 seconds time elapsed + 6,607,248,489 cycles # 2.262 GHz + 12,841,871,365 instructions # 1.94 insn per cycle + 2.921576547 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 1669) (512y: 16) (512z: 1427) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum/build.512z_f_inl0_hrd1/runTest_cpp.exe diff --git a/epochX/cudacpp/tput/logs_eemumu_mad/log_eemumu_mad_f_inl1_hrd0.txt b/epochX/cudacpp/tput/logs_eemumu_mad/log_eemumu_mad_f_inl1_hrd0.txt index 1fabc46555..e9304e6583 100644 --- a/epochX/cudacpp/tput/logs_eemumu_mad/log_eemumu_mad_f_inl1_hrd0.txt +++ b/epochX/cudacpp/tput/logs_eemumu_mad/log_eemumu_mad_f_inl1_hrd0.txt @@ -40,7 +40,7 @@ make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum' -DATE: 2024-10-06_09:27:46 +DATE: 2024-10-06_13:46:19 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: ========================================================================= @@ -49,15 +49,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_SM_EPEM_MUPMUM_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=1] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 1.309386e+08 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 6.516838e+08 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 8.621181e+08 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.371473e+08 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 6.825327e+08 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 8.842482e+08 ) sec^-1 MeanMatrixElemValue = ( 1.371687e-02 +- 3.270220e-06 ) GeV^0 -TOTAL : 0.581015 sec +TOTAL : 0.576412 sec INFO: No Floating Point Exceptions have been reported - 2,337,717,863 cycles # 2.893 GHz - 3,666,959,770 instructions # 1.57 insn per cycle - 0.866189287 seconds time elapsed + 2,338,556,779 cycles # 2.874 GHz + 3,632,749,421 instructions # 1.55 insn per cycle + 0.871907545 seconds time elapsed runNcu /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum/build.cuda_f_inl1_hrd0/check_cuda.exe -p 2048 256 1 ==PROF== Profiling "sigmaKin": launch__registers_per_thread 109 ==PROF== Profiling "sigmaKin": sm__sass_average_branch_targets_threads_uniform.pct 100% @@ -85,15 +85,15 @@ Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=1 Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/none+NAVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = SCALAR ('none': ~vector[1], no SIMD) -EvtsPerSec[Rmb+ME] (23) = ( 1.617887e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 2.109367e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 2.109367e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.638811e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 2.134861e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.134861e+06 ) sec^-1 MeanMatrixElemValue = ( 1.371707e-02 +- 3.270376e-06 ) GeV^0 -TOTAL : 4.275933 sec +TOTAL : 4.209215 sec INFO: No Floating Point Exceptions have been reported - 12,412,341,686 cycles # 2.900 GHz - 32,352,281,163 instructions # 2.61 insn per cycle - 4.283041784 seconds time elapsed + 12,185,435,109 cycles # 2.892 GHz + 32,296,383,921 instructions # 2.65 insn per cycle + 4.214228080 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 290) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum/build.none_f_inl1_hrd0/runTest_cpp.exe @@ -117,15 +117,15 @@ Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=1 Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('sse4': SSE4.2, 128bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 2.642717e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 4.471061e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 4.471061e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 2.667995e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 4.483093e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 4.483093e+06 ) sec^-1 MeanMatrixElemValue = ( 1.371707e-02 +- 3.270376e-06 ) GeV^0 -TOTAL : 2.775228 sec +TOTAL : 2.732781 sec INFO: No Floating Point Exceptions have been reported - 8,161,861,180 cycles # 2.934 GHz - 18,732,698,985 instructions # 2.30 insn per cycle - 2.782796507 seconds time elapsed + 8,025,505,165 cycles # 2.932 GHz + 18,688,593,713 instructions # 2.33 insn per cycle + 2.738024685 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 1534) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum/build.sse4_f_inl1_hrd0/runTest_cpp.exe @@ -149,15 +149,15 @@ Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=1 Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('avx2': AVX2, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 2.771950e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 4.635210e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 4.635210e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 2.784084e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 4.624458e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 4.624458e+06 ) sec^-1 MeanMatrixElemValue = ( 1.371706e-02 +- 3.270341e-06 ) GeV^0 -TOTAL : 2.653741 sec +TOTAL : 2.629050 sec INFO: No Floating Point Exceptions have been reported - 7,565,022,779 cycles # 2.844 GHz - 14,293,093,213 instructions # 1.89 insn per cycle - 2.661141426 seconds time elapsed + 7,452,831,784 cycles # 2.830 GHz + 14,248,373,301 instructions # 1.91 insn per cycle + 2.634160421 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 2234) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum/build.avx2_f_inl1_hrd0/runTest_cpp.exe @@ -181,15 +181,15 @@ Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=1 Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('512y': AVX512, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 2.799741e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 4.762487e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 4.762487e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 2.831918e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 4.767337e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 4.767337e+06 ) sec^-1 MeanMatrixElemValue = ( 1.371706e-02 +- 3.270341e-06 ) GeV^0 -TOTAL : 2.634363 sec +TOTAL : 2.590357 sec INFO: No Floating Point Exceptions have been reported - 7,504,285,407 cycles # 2.842 GHz - 13,994,355,792 instructions # 1.86 insn per cycle - 2.641913370 seconds time elapsed + 7,373,932,933 cycles # 2.842 GHz + 13,950,418,834 instructions # 1.89 insn per cycle + 2.595444940 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 2087) (512y: 3) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum/build.512y_f_inl1_hrd0/runTest_cpp.exe @@ -213,15 +213,15 @@ Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=1 Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[16] ('512z': AVX512, 512bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 2.507958e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.890935e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.890935e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 2.508648e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.869287e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.869287e+06 ) sec^-1 MeanMatrixElemValue = ( 1.371706e-02 +- 3.270342e-06 ) GeV^0 -TOTAL : 2.900923 sec +TOTAL : 2.880327 sec INFO: No Floating Point Exceptions have been reported - 6,641,718,947 cycles # 2.284 GHz - 13,481,348,782 instructions # 2.03 insn per cycle - 2.908502130 seconds time elapsed + 6,528,674,282 cycles # 2.263 GHz + 13,437,313,423 instructions # 2.06 insn per cycle + 2.885466107 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 2073) (512y: 1) (512z: 1201) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum/build.512z_f_inl1_hrd0/runTest_cpp.exe diff --git a/epochX/cudacpp/tput/logs_eemumu_mad/log_eemumu_mad_f_inl1_hrd1.txt b/epochX/cudacpp/tput/logs_eemumu_mad/log_eemumu_mad_f_inl1_hrd1.txt index ddc690e546..783645d387 100644 --- a/epochX/cudacpp/tput/logs_eemumu_mad/log_eemumu_mad_f_inl1_hrd1.txt +++ b/epochX/cudacpp/tput/logs_eemumu_mad/log_eemumu_mad_f_inl1_hrd1.txt @@ -40,7 +40,7 @@ make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum' -DATE: 2024-10-06_09:28:10 +DATE: 2024-10-06_13:46:42 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: ========================================================================= @@ -49,15 +49,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_SM_EPEM_MUPMUM_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=1] [hardcodePARAM=1] Workflow summary = CUD:FLT+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 1.311525e+08 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 6.893939e+08 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 9.130206e+08 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.385848e+08 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 6.950367e+08 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 9.149914e+08 ) sec^-1 MeanMatrixElemValue = ( 1.371687e-02 +- 3.270220e-06 ) GeV^0 -TOTAL : 0.580982 sec +TOTAL : 0.570449 sec INFO: No Floating Point Exceptions have been reported - 2,326,498,884 cycles # 2.887 GHz - 3,595,400,053 instructions # 1.55 insn per cycle - 0.865243472 seconds time elapsed + 2,299,975,171 cycles # 2.883 GHz + 3,614,065,946 instructions # 1.57 insn per cycle + 0.854201673 seconds time elapsed runNcu /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum/build.cuda_f_inl1_hrd1/check_cuda.exe -p 2048 256 1 ==PROF== Profiling "sigmaKin": launch__registers_per_thread 79 ==PROF== Profiling "sigmaKin": sm__sass_average_branch_targets_threads_uniform.pct 100% @@ -85,15 +85,15 @@ Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=1 Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/none+NAVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = SCALAR ('none': ~vector[1], no SIMD) -EvtsPerSec[Rmb+ME] (23) = ( 2.199736e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.210916e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.210916e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 2.210248e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.213615e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.213615e+06 ) sec^-1 MeanMatrixElemValue = ( 1.371707e-02 +- 3.270376e-06 ) GeV^0 -TOTAL : 3.247253 sec +TOTAL : 3.214282 sec INFO: No Floating Point Exceptions have been reported - 9,460,485,661 cycles # 2.907 GHz - 25,749,028,052 instructions # 2.72 insn per cycle - 3.254869601 seconds time elapsed + 9,357,416,331 cycles # 2.908 GHz + 25,707,265,233 instructions # 2.75 insn per cycle + 3.219369391 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 243) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum/build.none_f_inl1_hrd1/runTest_cpp.exe @@ -117,15 +117,15 @@ Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=1 Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('sse4': SSE4.2, 128bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 2.982142e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 5.480555e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 5.480555e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 3.006331e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 5.505447e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 5.505447e+06 ) sec^-1 MeanMatrixElemValue = ( 1.371707e-02 +- 3.270376e-06 ) GeV^0 -TOTAL : 2.498717 sec +TOTAL : 2.463562 sec INFO: No Floating Point Exceptions have been reported - 7,385,528,393 cycles # 2.949 GHz - 16,812,365,380 instructions # 2.28 insn per cycle - 2.506313604 seconds time elapsed + 7,273,133,406 cycles # 2.947 GHz + 16,769,452,855 instructions # 2.31 insn per cycle + 2.468690242 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 1311) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum/build.sse4_f_inl1_hrd1/runTest_cpp.exe @@ -149,15 +149,15 @@ Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=1 Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('avx2': AVX2, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 2.917887e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 5.065921e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 5.065921e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 2.935013e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 5.090472e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 5.090472e+06 ) sec^-1 MeanMatrixElemValue = ( 1.371706e-02 +- 3.270341e-06 ) GeV^0 -TOTAL : 2.542096 sec +TOTAL : 2.517034 sec INFO: No Floating Point Exceptions have been reported - 7,260,793,625 cycles # 2.848 GHz - 13,703,433,227 instructions # 1.89 insn per cycle - 2.549878549 seconds time elapsed + 7,172,075,904 cycles # 2.844 GHz + 13,658,614,986 instructions # 1.90 insn per cycle + 2.522312080 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 2067) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum/build.avx2_f_inl1_hrd1/runTest_cpp.exe @@ -181,15 +181,15 @@ Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=1 Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('512y': AVX512, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 2.947392e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 5.166768e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 5.166768e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 2.983678e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 5.193089e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 5.193089e+06 ) sec^-1 MeanMatrixElemValue = ( 1.371706e-02 +- 3.270341e-06 ) GeV^0 -TOTAL : 2.537410 sec +TOTAL : 2.478847 sec INFO: No Floating Point Exceptions have been reported - 7,253,478,894 cycles # 2.851 GHz - 13,505,585,795 instructions # 1.86 insn per cycle - 2.545044336 seconds time elapsed + 7,034,231,544 cycles # 2.833 GHz + 13,452,657,052 instructions # 1.91 insn per cycle + 2.484081711 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 1935) (512y: 7) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum/build.512y_f_inl1_hrd1/runTest_cpp.exe @@ -213,15 +213,15 @@ Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=1 Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[16] ('512z': AVX512, 512bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 2.612725e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 4.139660e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 4.139660e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 2.612098e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 4.137964e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 4.137964e+06 ) sec^-1 MeanMatrixElemValue = ( 1.371706e-02 +- 3.270342e-06 ) GeV^0 -TOTAL : 2.798296 sec +TOTAL : 2.778012 sec INFO: No Floating Point Exceptions have been reported - 6,447,529,861 cycles # 2.298 GHz - 13,215,855,857 instructions # 2.05 insn per cycle - 2.806480502 seconds time elapsed + 6,356,134,232 cycles # 2.284 GHz + 13,171,961,736 instructions # 2.07 insn per cycle + 2.783271056 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 2039) (512y: 2) (512z: 1081) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum/build.512z_f_inl1_hrd1/runTest_cpp.exe diff --git a/epochX/cudacpp/tput/logs_eemumu_mad/log_eemumu_mad_m_inl0_hrd0.txt b/epochX/cudacpp/tput/logs_eemumu_mad/log_eemumu_mad_m_inl0_hrd0.txt index 8e00f9820d..32cc35a153 100644 --- a/epochX/cudacpp/tput/logs_eemumu_mad/log_eemumu_mad_m_inl0_hrd0.txt +++ b/epochX/cudacpp/tput/logs_eemumu_mad/log_eemumu_mad_m_inl0_hrd0.txt @@ -40,7 +40,7 @@ make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum' -DATE: 2024-10-06_08:55:54 +DATE: 2024-10-06_13:25:10 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: ========================================================================= @@ -49,15 +49,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_SM_EPEM_MUPMUM_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:MIX+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 7.055673e+07 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.658424e+08 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 4.851508e+08 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 7.345295e+07 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.739721e+08 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 4.893916e+08 ) sec^-1 MeanMatrixElemValue = ( 1.371706e-02 +- 3.270315e-06 ) GeV^0 -TOTAL : 0.678781 sec +TOTAL : 0.665808 sec INFO: No Floating Point Exceptions have been reported - 2,628,768,348 cycles # 2.876 GHz - 4,103,389,790 instructions # 1.56 insn per cycle - 1.044225431 seconds time elapsed + 2,629,212,284 cycles # 2.892 GHz + 4,064,692,900 instructions # 1.55 insn per cycle + 0.971213973 seconds time elapsed runNcu /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum/build.cuda_m_inl0_hrd0/check_cuda.exe -p 2048 256 1 ==PROF== Profiling "sigmaKin": launch__registers_per_thread 166 ==PROF== Profiling "sigmaKin": sm__sass_average_branch_targets_threads_uniform.pct 100% @@ -85,15 +85,15 @@ Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=0 Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/none+NAVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) Internal loops fptype_sv = SCALAR ('none': ~vector[1], no SIMD) -EvtsPerSec[Rmb+ME] (23) = ( 1.011376e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.175905e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.175905e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.025371e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.190772e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.190772e+06 ) sec^-1 MeanMatrixElemValue = ( 1.371706e-02 +- 3.270315e-06 ) GeV^0 -TOTAL : 6.671756 sec +TOTAL : 6.543081 sec INFO: No Floating Point Exceptions have been reported - 19,661,999,702 cycles # 2.943 GHz - 46,395,546,050 instructions # 2.36 insn per cycle - 6.683261433 seconds time elapsed + 19,311,487,761 cycles # 2.950 GHz + 46,280,874,210 instructions # 2.40 insn per cycle + 6.548297721 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 466) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum/build.none_m_inl0_hrd0/runTest_cpp.exe @@ -117,15 +117,15 @@ Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=0 Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[2] ('sse4': SSE4.2, 128bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 1.631538e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 2.161697e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 2.161697e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.615967e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 2.127241e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.127241e+06 ) sec^-1 MeanMatrixElemValue = ( 1.371706e-02 +- 3.270315e-06 ) GeV^0 -TOTAL : 4.308846 sec +TOTAL : 4.311487 sec INFO: No Floating Point Exceptions have been reported - 12,713,127,116 cycles # 2.944 GHz - 31,571,564,120 instructions # 2.48 insn per cycle - 4.322869208 seconds time elapsed + 12,528,366,142 cycles # 2.903 GHz + 31,469,067,737 instructions # 2.51 insn per cycle + 4.316652619 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 1731) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum/build.sse4_m_inl0_hrd0/runTest_cpp.exe @@ -149,15 +149,15 @@ Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=0 Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('avx2': AVX2, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 1.963768e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 2.746755e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 2.746755e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.921434e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 2.664750e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.664750e+06 ) sec^-1 MeanMatrixElemValue = ( 1.371706e-02 +- 3.270315e-06 ) GeV^0 -TOTAL : 3.649356 sec +TOTAL : 3.690256 sec INFO: No Floating Point Exceptions have been reported - 10,294,572,937 cycles # 2.814 GHz - 19,586,622,017 instructions # 1.90 insn per cycle - 3.662289672 seconds time elapsed + 10,057,656,591 cycles # 2.722 GHz + 19,481,121,617 instructions # 1.94 insn per cycle + 3.695697812 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 2045) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum/build.avx2_m_inl0_hrd0/runTest_cpp.exe @@ -181,15 +181,15 @@ Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=0 Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('512y': AVX512, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 2.001856e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 2.818080e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 2.818080e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 2.029243e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 2.835610e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.835610e+06 ) sec^-1 MeanMatrixElemValue = ( 1.371706e-02 +- 3.270315e-06 ) GeV^0 -TOTAL : 3.589952 sec +TOTAL : 3.503324 sec INFO: No Floating Point Exceptions have been reported - 10,108,826,304 cycles # 2.808 GHz - 19,396,692,714 instructions # 1.92 insn per cycle - 3.602641354 seconds time elapsed + 9,872,201,716 cycles # 2.815 GHz + 19,291,552,898 instructions # 1.95 insn per cycle + 3.508490076 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 1799) (512y: 188) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum/build.512y_m_inl0_hrd0/runTest_cpp.exe @@ -213,15 +213,15 @@ Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=0 Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('512z': AVX512, 512bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 1.801777e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 2.420597e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 2.420597e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.813752e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 2.430977e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.430977e+06 ) sec^-1 MeanMatrixElemValue = ( 1.371706e-02 +- 3.270315e-06 ) GeV^0 -TOTAL : 3.939332 sec +TOTAL : 3.876245 sec INFO: No Floating Point Exceptions have been reported - 8,555,878,739 cycles # 2.167 GHz - 15,216,666,169 instructions # 1.78 insn per cycle - 3.951287451 seconds time elapsed + 8,347,804,839 cycles # 2.151 GHz + 15,110,716,474 instructions # 1.81 insn per cycle + 3.881494639 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 966) (512y: 154) (512z: 1330) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum/build.512z_m_inl0_hrd0/runTest_cpp.exe diff --git a/epochX/cudacpp/tput/logs_eemumu_mad/log_eemumu_mad_m_inl0_hrd1.txt b/epochX/cudacpp/tput/logs_eemumu_mad/log_eemumu_mad_m_inl0_hrd1.txt index 0283d4438d..31be23f692 100644 --- a/epochX/cudacpp/tput/logs_eemumu_mad/log_eemumu_mad_m_inl0_hrd1.txt +++ b/epochX/cudacpp/tput/logs_eemumu_mad/log_eemumu_mad_m_inl0_hrd1.txt @@ -40,7 +40,7 @@ make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum' -DATE: 2024-10-06_08:56:25 +DATE: 2024-10-06_13:25:40 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: ========================================================================= @@ -49,15 +49,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_SM_EPEM_MUPMUM_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=1] Workflow summary = CUD:MIX+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 7.048170e+07 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.671940e+08 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 4.867900e+08 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 7.332368e+07 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.723319e+08 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 4.881052e+08 ) sec^-1 MeanMatrixElemValue = ( 1.371706e-02 +- 3.270315e-06 ) GeV^0 -TOTAL : 0.677955 sec +TOTAL : 0.665957 sec INFO: No Floating Point Exceptions have been reported - 2,610,429,449 cycles # 2.847 GHz - 4,074,904,816 instructions # 1.56 insn per cycle - 1.028610198 seconds time elapsed + 2,631,132,413 cycles # 2.884 GHz + 4,099,466,879 instructions # 1.56 insn per cycle + 0.974241988 seconds time elapsed runNcu /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum/build.cuda_m_inl0_hrd1/check_cuda.exe -p 2048 256 1 ==PROF== Profiling "sigmaKin": launch__registers_per_thread 154 ==PROF== Profiling "sigmaKin": sm__sass_average_branch_targets_threads_uniform.pct 100% @@ -85,15 +85,15 @@ Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=0 Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/none+NAVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) Internal loops fptype_sv = SCALAR ('none': ~vector[1], no SIMD) -EvtsPerSec[Rmb+ME] (23) = ( 1.012794e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.178467e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.178467e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.025834e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.192151e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.192151e+06 ) sec^-1 MeanMatrixElemValue = ( 1.371706e-02 +- 3.270315e-06 ) GeV^0 -TOTAL : 6.662178 sec +TOTAL : 6.540920 sec INFO: No Floating Point Exceptions have been reported - 19,608,707,308 cycles # 2.939 GHz - 46,331,953,932 instructions # 2.36 insn per cycle - 6.674225175 seconds time elapsed + 19,237,229,772 cycles # 2.939 GHz + 46,214,397,774 instructions # 2.40 insn per cycle + 6.545970727 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 453) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum/build.none_m_inl0_hrd1/runTest_cpp.exe @@ -117,15 +117,15 @@ Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=0 Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[2] ('sse4': SSE4.2, 128bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 1.631371e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 2.156116e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 2.156116e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.637767e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 2.157269e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.157269e+06 ) sec^-1 MeanMatrixElemValue = ( 1.371706e-02 +- 3.270315e-06 ) GeV^0 -TOTAL : 4.305744 sec +TOTAL : 4.254024 sec INFO: No Floating Point Exceptions have been reported - 12,687,194,497 cycles # 2.940 GHz - 31,570,654,619 instructions # 2.49 insn per cycle - 4.317357131 seconds time elapsed + 12,519,482,151 cycles # 2.940 GHz + 31,465,791,050 instructions # 2.51 insn per cycle + 4.259162695 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 1724) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum/build.sse4_m_inl0_hrd1/runTest_cpp.exe @@ -149,15 +149,15 @@ Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=0 Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('avx2': AVX2, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 1.951503e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 2.723168e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 2.723168e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.965336e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 2.725187e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.725187e+06 ) sec^-1 MeanMatrixElemValue = ( 1.371706e-02 +- 3.270315e-06 ) GeV^0 -TOTAL : 3.669508 sec +TOTAL : 3.607649 sec INFO: No Floating Point Exceptions have been reported - 10,337,023,986 cycles # 2.809 GHz - 19,600,398,756 instructions # 1.90 insn per cycle - 3.680210311 seconds time elapsed + 10,127,554,699 cycles # 2.804 GHz + 19,495,603,367 instructions # 1.93 insn per cycle + 3.612892520 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 2036) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum/build.avx2_m_inl0_hrd1/runTest_cpp.exe @@ -181,15 +181,15 @@ Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=0 Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('512y': AVX512, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 2.000628e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 2.813640e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 2.813640e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 2.025105e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 2.838706e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.838706e+06 ) sec^-1 MeanMatrixElemValue = ( 1.371706e-02 +- 3.270315e-06 ) GeV^0 -TOTAL : 3.591164 sec +TOTAL : 3.512131 sec INFO: No Floating Point Exceptions have been reported - 10,093,463,938 cycles # 2.804 GHz - 19,298,137,282 instructions # 1.91 insn per cycle - 3.601580555 seconds time elapsed + 9,896,251,811 cycles # 2.814 GHz + 19,193,402,920 instructions # 1.94 insn per cycle + 3.517305185 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 1766) (512y: 191) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum/build.512y_m_inl0_hrd1/runTest_cpp.exe @@ -213,15 +213,15 @@ Process = SIGMA_SM_EPEM_MUPMUM_CPP [gcc 11.3.1] [inlineHel=0 Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('512z': AVX512, 512bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 1.833398e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 2.483164e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 2.483164e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.851396e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 2.505359e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.505359e+06 ) sec^-1 MeanMatrixElemValue = ( 1.371706e-02 +- 3.270315e-06 ) GeV^0 -TOTAL : 3.878021 sec +TOTAL : 3.805556 sec INFO: No Floating Point Exceptions have been reported - 8,399,559,009 cycles # 2.161 GHz - 15,073,176,103 instructions # 1.79 insn per cycle - 3.888708235 seconds time elapsed + 8,214,377,670 cycles # 2.156 GHz + 14,969,754,766 instructions # 1.82 insn per cycle + 3.810755884 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 959) (512y: 155) (512z: 1296) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/ee_mumu.mad/SubProcesses/P1_epem_mupmum/build.512z_m_inl0_hrd1/runTest_cpp.exe diff --git a/epochX/cudacpp/tput/logs_ggtt_mad/log_ggtt_mad_d_inl0_hrd0.txt b/epochX/cudacpp/tput/logs_ggtt_mad/log_ggtt_mad_d_inl0_hrd0.txt index 0abecbd859..12c9b60b75 100644 --- a/epochX/cudacpp/tput/logs_ggtt_mad/log_ggtt_mad_d_inl0_hrd0.txt +++ b/epochX/cudacpp/tput/logs_ggtt_mad/log_ggtt_mad_d_inl0_hrd0.txt @@ -40,7 +40,7 @@ make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx' -DATE: 2024-10-06_08:57:50 +DATE: 2024-10-06_13:27:04 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: ========================================================================= @@ -49,15 +49,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_SM_GG_TTX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 4.424562e+07 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 9.378226e+07 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.000814e+08 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 4.644538e+07 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 9.484518e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.009869e+08 ) sec^-1 MeanMatrixElemValue = ( 2.086689e+00 +- 3.413217e-03 ) GeV^0 -TOTAL : 0.532719 sec +TOTAL : 0.525542 sec INFO: No Floating Point Exceptions have been reported - 2,198,564,055 cycles # 2.860 GHz - 3,137,529,593 instructions # 1.43 insn per cycle - 0.850854779 seconds time elapsed + 2,193,016,812 cycles # 2.886 GHz + 3,178,032,178 instructions # 1.45 insn per cycle + 0.817155477 seconds time elapsed runNcu /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx/build.cuda_d_inl0_hrd0/check_cuda.exe -p 2048 256 1 ==PROF== Profiling "sigmaKin": launch__registers_per_thread 214 ==PROF== Profiling "sigmaKin": sm__sass_average_branch_targets_threads_uniform.pct 100% @@ -85,15 +85,15 @@ Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [ha Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/none+NAVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = SCALAR ('none': ~vector[1], no SIMD) -EvtsPerSec[Rmb+ME] (23) = ( 1.821542e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.869016e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.869016e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.830181e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.876958e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.876958e+05 ) sec^-1 MeanMatrixElemValue = ( 2.086689e+00 +- 3.413217e-03 ) GeV^0 -TOTAL : 5.902388 sec +TOTAL : 5.835754 sec INFO: No Floating Point Exceptions have been reported - 17,373,663,633 cycles # 2.939 GHz - 46,051,346,456 instructions # 2.65 insn per cycle - 5.916149203 seconds time elapsed + 17,230,624,871 cycles # 2.951 GHz + 45,935,120,994 instructions # 2.67 insn per cycle + 5.840975990 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 617) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx/build.none_d_inl0_hrd0/runTest_cpp.exe @@ -117,15 +117,15 @@ Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [ha Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[2] ('sse4': SSE4.2, 128bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 3.199984e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.364044e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.364044e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 3.223853e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.383682e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.383682e+05 ) sec^-1 MeanMatrixElemValue = ( 2.086689e+00 +- 3.413217e-03 ) GeV^0 -TOTAL : 3.425369 sec +TOTAL : 3.360328 sec INFO: No Floating Point Exceptions have been reported - 10,116,123,100 cycles # 2.945 GHz - 27,968,506,728 instructions # 2.76 insn per cycle - 3.436971917 seconds time elapsed + 9,930,373,364 cycles # 2.951 GHz + 27,847,584,501 instructions # 2.80 insn per cycle + 3.365679770 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 2536) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx/build.sse4_d_inl0_hrd0/runTest_cpp.exe @@ -149,15 +149,15 @@ Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [ha Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('avx2': AVX2, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 5.021241e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 5.422127e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 5.422127e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 5.035999e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 5.429988e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 5.429988e+05 ) sec^-1 MeanMatrixElemValue = ( 2.086689e+00 +- 3.413217e-03 ) GeV^0 -TOTAL : 2.236686 sec +TOTAL : 2.190300 sec INFO: No Floating Point Exceptions have been reported - 6,226,726,050 cycles # 2.773 GHz - 12,700,169,832 instructions # 2.04 insn per cycle - 2.249020906 seconds time elapsed + 6,070,860,290 cycles # 2.766 GHz + 12,581,395,688 instructions # 2.07 insn per cycle + 2.195703346 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 2619) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx/build.avx2_d_inl0_hrd0/runTest_cpp.exe @@ -181,15 +181,15 @@ Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [ha Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('512y': AVX512, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 5.518459e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 5.996461e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 5.996461e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 5.498159e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 5.959301e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 5.959301e+05 ) sec^-1 MeanMatrixElemValue = ( 2.086689e+00 +- 3.413217e-03 ) GeV^0 -TOTAL : 2.046768 sec +TOTAL : 2.014060 sec INFO: No Floating Point Exceptions have been reported - 5,709,909,658 cycles # 2.777 GHz - 12,140,194,379 instructions # 2.13 insn per cycle - 2.059786524 seconds time elapsed + 5,591,658,235 cycles # 2.770 GHz + 12,023,108,897 instructions # 2.15 insn per cycle + 2.019322218 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 2360) (512y: 144) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx/build.512y_d_inl0_hrd0/runTest_cpp.exe @@ -213,15 +213,15 @@ Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [ha Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('512z': AVX512, 512bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 3.403513e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.583329e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.583329e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 3.529487e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.712895e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.712895e+05 ) sec^-1 MeanMatrixElemValue = ( 2.086689e+00 +- 3.413217e-03 ) GeV^0 -TOTAL : 3.229693 sec +TOTAL : 3.077735 sec INFO: No Floating Point Exceptions have been reported - 6,051,702,488 cycles # 1.869 GHz - 8,428,750,265 instructions # 1.39 insn per cycle - 3.242969033 seconds time elapsed + 5,707,603,505 cycles # 1.852 GHz + 8,294,202,064 instructions # 1.45 insn per cycle + 3.083008121 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 1441) (512y: 122) (512z: 1802) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx/build.512z_d_inl0_hrd0/runTest_cpp.exe diff --git a/epochX/cudacpp/tput/logs_ggtt_mad/log_ggtt_mad_d_inl0_hrd0_bridge.txt b/epochX/cudacpp/tput/logs_ggtt_mad/log_ggtt_mad_d_inl0_hrd0_bridge.txt index 0a62f31f21..7cfbf90f0b 100644 --- a/epochX/cudacpp/tput/logs_ggtt_mad/log_ggtt_mad_d_inl0_hrd0_bridge.txt +++ b/epochX/cudacpp/tput/logs_ggtt_mad/log_ggtt_mad_d_inl0_hrd0_bridge.txt @@ -40,7 +40,7 @@ make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx' -DATE: 2024-10-06_09:37:36 +DATE: 2024-10-06_13:56:09 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: ========================================================================= @@ -53,16 +53,16 @@ WARNING! Set grid in Bridge (nevt=524288, gpublocks=2048, gputhreads=256, gpublo Process = SIGMA_SM_GG_TTX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 4.523249e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.008578e+07 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.008578e+07 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 4.552348e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 2.996738e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.996738e+07 ) sec^-1 MeanMatrixElemValue = ( 2.086689e+00 +- 3.413217e-03 ) GeV^0 -TOTAL : 0.943118 sec +TOTAL : 0.816271 sec INFO: No Floating Point Exceptions have been reported INFO: No Floating Point Exceptions have been reported - 3,438,006,415 cycles # 2.887 GHz - 4,812,518,572 instructions # 1.40 insn per cycle - 1.248014993 seconds time elapsed + 3,088,602,552 cycles # 2.895 GHz + 4,790,382,106 instructions # 1.55 insn per cycle + 1.122951640 seconds time elapsed runNcu /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx/build.cuda_d_inl0_hrd0/check_cuda.exe -p 2048 256 1 --bridge WARNING! Bridge selected: cannot use RamboDevice, will use RamboHost WARNING! RamboHost selected: cannot use CurandDevice, will use CurandHost @@ -95,16 +95,16 @@ Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [ha Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+BRDHST/none+NAVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = SCALAR ('none': ~vector[1], no SIMD) -EvtsPerSec[Rmb+ME] (23) = ( 1.806787e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.852935e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.852935e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.814121e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.860438e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.860438e+05 ) sec^-1 MeanMatrixElemValue = ( 2.086689e+00 +- 3.413217e-03 ) GeV^0 -TOTAL : 6.028463 sec +TOTAL : 5.973351 sec INFO: No Floating Point Exceptions have been reported INFO: No Floating Point Exceptions have been reported - 17,701,660,691 cycles # 2.931 GHz - 46,100,592,443 instructions # 2.60 insn per cycle - 6.041454793 seconds time elapsed + 17,581,001,714 cycles # 2.940 GHz + 45,999,293,150 instructions # 2.62 insn per cycle + 5.981438491 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 617) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx/build.none_d_inl0_hrd0/runTest_cpp.exe @@ -129,16 +129,16 @@ Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [ha Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+BRDHST/sse4+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[2] ('sse4': SSE4.2, 128bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 3.171570e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.328412e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.328412e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 3.166053e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.321151e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.321151e+05 ) sec^-1 MeanMatrixElemValue = ( 2.086689e+00 +- 3.413217e-03 ) GeV^0 -TOTAL : 3.537488 sec +TOTAL : 3.511645 sec INFO: No Floating Point Exceptions have been reported INFO: No Floating Point Exceptions have been reported - 10,436,410,766 cycles # 2.940 GHz - 28,150,415,987 instructions # 2.70 insn per cycle - 3.550700440 seconds time elapsed + 10,363,151,257 cycles # 2.945 GHz + 28,050,059,892 instructions # 2.71 insn per cycle + 3.520337976 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 2536) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx/build.sse4_d_inl0_hrd0/runTest_cpp.exe @@ -163,16 +163,16 @@ Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [ha Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+BRDHST/avx2+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('avx2': AVX2, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 4.940586e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 5.316252e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 5.316252e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 4.881672e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 5.246097e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 5.246097e+05 ) sec^-1 MeanMatrixElemValue = ( 2.086689e+00 +- 3.413217e-03 ) GeV^0 -TOTAL : 2.355700 sec +TOTAL : 2.348334 sec INFO: No Floating Point Exceptions have been reported INFO: No Floating Point Exceptions have been reported - 6,586,554,223 cycles # 2.781 GHz - 12,999,619,553 instructions # 1.97 insn per cycle - 2.369192751 seconds time elapsed + 6,535,974,015 cycles # 2.775 GHz + 12,898,573,826 instructions # 1.97 insn per cycle + 2.356393590 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 2619) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx/build.avx2_d_inl0_hrd0/runTest_cpp.exe @@ -197,16 +197,16 @@ Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [ha Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+BRDHST/512y+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('512y': AVX512, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 5.425137e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 5.877080e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 5.877080e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 5.392341e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 5.833954e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 5.833954e+05 ) sec^-1 MeanMatrixElemValue = ( 2.086689e+00 +- 3.413217e-03 ) GeV^0 -TOTAL : 2.160954 sec +TOTAL : 2.144095 sec INFO: No Floating Point Exceptions have been reported INFO: No Floating Point Exceptions have been reported - 6,058,497,746 cycles # 2.788 GHz - 12,422,408,910 instructions # 2.05 insn per cycle - 2.174009213 seconds time elapsed + 5,953,598,449 cycles # 2.766 GHz + 12,332,620,381 instructions # 2.07 insn per cycle + 2.153089195 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 2360) (512y: 144) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx/build.512y_d_inl0_hrd0/runTest_cpp.exe @@ -231,16 +231,16 @@ Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [ha Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+BRDHST/512z+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('512z': AVX512, 512bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 3.454260e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.633384e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.633384e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 3.508594e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.690233e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.690233e+05 ) sec^-1 MeanMatrixElemValue = ( 2.086689e+00 +- 3.413217e-03 ) GeV^0 -TOTAL : 3.271770 sec +TOTAL : 3.187897 sec INFO: No Floating Point Exceptions have been reported INFO: No Floating Point Exceptions have been reported - 6,220,081,356 cycles # 1.894 GHz - 8,655,636,644 instructions # 1.39 insn per cycle - 3.285127387 seconds time elapsed + 6,085,167,751 cycles # 1.904 GHz + 8,559,046,906 instructions # 1.41 insn per cycle + 3.196732997 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 1441) (512y: 122) (512z: 1802) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx/build.512z_d_inl0_hrd0/runTest_cpp.exe diff --git a/epochX/cudacpp/tput/logs_ggtt_mad/log_ggtt_mad_d_inl0_hrd0_common.txt b/epochX/cudacpp/tput/logs_ggtt_mad/log_ggtt_mad_d_inl0_hrd0_common.txt index 70d02af695..58de26a64f 100644 --- a/epochX/cudacpp/tput/logs_ggtt_mad/log_ggtt_mad_d_inl0_hrd0_common.txt +++ b/epochX/cudacpp/tput/logs_ggtt_mad/log_ggtt_mad_d_inl0_hrd0_common.txt @@ -40,7 +40,7 @@ make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx' -DATE: 2024-10-06_09:49:47 +DATE: 2024-10-06_14:08:19 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: ========================================================================= @@ -49,15 +49,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_SM_GG_TTX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:COMMON+RMBDEV+MESDEV/none+NAVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 4.202403e+07 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 9.187841e+07 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 9.877468e+07 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 4.283212e+07 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 9.181246e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 9.893593e+07 ) sec^-1 MeanMatrixElemValue = ( 2.079401e+00 +- 3.402993e-03 ) GeV^0 -TOTAL : 0.637594 sec +TOTAL : 0.635108 sec INFO: No Floating Point Exceptions have been reported - 2,481,390,363 cycles # 2.852 GHz - 3,619,998,982 instructions # 1.46 insn per cycle - 0.928734017 seconds time elapsed + 2,504,959,971 cycles # 2.883 GHz + 3,643,714,828 instructions # 1.45 insn per cycle + 0.926493456 seconds time elapsed runNcu /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx/build.cuda_d_inl0_hrd0/check_cuda.exe -p 2048 256 1 --common ==PROF== Profiling "sigmaKin": launch__registers_per_thread 214 ==PROF== Profiling "sigmaKin": sm__sass_average_branch_targets_threads_uniform.pct 100% @@ -85,15 +85,15 @@ Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [ha Workflow summary = CPP:DBL+CXS:COMMON+RMBHST+MESHST/none+NAVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = SCALAR ('none': ~vector[1], no SIMD) -EvtsPerSec[Rmb+ME] (23) = ( 1.808108e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.854363e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.854363e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.812653e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.859290e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.859290e+05 ) sec^-1 MeanMatrixElemValue = ( 2.079401e+00 +- 3.402993e-03 ) GeV^0 -TOTAL : 5.981515 sec +TOTAL : 5.956244 sec INFO: No Floating Point Exceptions have been reported - 17,441,882,337 cycles # 2.914 GHz - 45,980,812,555 instructions # 2.64 insn per cycle - 5.987317462 seconds time elapsed + 17,424,073,232 cycles # 2.923 GHz + 45,949,745,901 instructions # 2.64 insn per cycle + 5.962042260 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 617) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx/build.none_d_inl0_hrd0/runTest_cpp.exe @@ -117,15 +117,15 @@ Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [ha Workflow summary = CPP:DBL+CXS:COMMON+RMBHST+MESHST/sse4+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[2] ('sse4': SSE4.2, 128bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 3.173867e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.332553e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.332553e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 3.222728e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.383637e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.383637e+05 ) sec^-1 MeanMatrixElemValue = ( 2.079401e+00 +- 3.402993e-03 ) GeV^0 -TOTAL : 3.490197 sec +TOTAL : 3.422897 sec INFO: No Floating Point Exceptions have been reported - 10,215,611,800 cycles # 2.923 GHz - 27,889,324,001 instructions # 2.73 insn per cycle - 3.495993800 seconds time elapsed + 10,123,199,950 cycles # 2.953 GHz + 27,846,742,787 instructions # 2.75 insn per cycle + 3.428686932 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 2536) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx/build.sse4_d_inl0_hrd0/runTest_cpp.exe @@ -149,15 +149,15 @@ Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [ha Workflow summary = CPP:DBL+CXS:COMMON+RMBHST+MESHST/avx2+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('avx2': AVX2, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 4.999819e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 5.389873e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 5.389873e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 5.042365e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 5.432470e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 5.432470e+05 ) sec^-1 MeanMatrixElemValue = ( 2.079401e+00 +- 3.402993e-03 ) GeV^0 -TOTAL : 2.281339 sec +TOTAL : 2.250353 sec INFO: No Floating Point Exceptions have been reported - 6,287,168,374 cycles # 2.750 GHz - 12,602,929,813 instructions # 2.00 insn per cycle - 2.287435325 seconds time elapsed + 6,263,824,127 cycles # 2.777 GHz + 12,563,166,336 instructions # 2.01 insn per cycle + 2.256310175 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 2619) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx/build.avx2_d_inl0_hrd0/runTest_cpp.exe @@ -181,15 +181,15 @@ Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [ha Workflow summary = CPP:DBL+CXS:COMMON+RMBHST+MESHST/512y+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('512y': AVX512, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 5.471434e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 5.936245e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 5.936245e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 5.516126e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 5.978849e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 5.978849e+05 ) sec^-1 MeanMatrixElemValue = ( 2.079401e+00 +- 3.402993e-03 ) GeV^0 -TOTAL : 2.097327 sec +TOTAL : 2.072142 sec INFO: No Floating Point Exceptions have been reported - 5,814,420,150 cycles # 2.765 GHz - 11,994,829,914 instructions # 2.06 insn per cycle - 2.103345298 seconds time elapsed + 5,758,641,252 cycles # 2.772 GHz + 11,971,487,426 instructions # 2.08 insn per cycle + 2.078030943 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 2360) (512y: 144) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx/build.512y_d_inl0_hrd0/runTest_cpp.exe @@ -213,15 +213,15 @@ Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [ha Workflow summary = CPP:DBL+CXS:COMMON+RMBHST+MESHST/512z+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('512z': AVX512, 512bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 3.462865e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.641783e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.641783e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 3.543812e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.731795e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.731795e+05 ) sec^-1 MeanMatrixElemValue = ( 2.079401e+00 +- 3.402993e-03 ) GeV^0 -TOTAL : 3.218108 sec +TOTAL : 3.130676 sec INFO: No Floating Point Exceptions have been reported - 5,937,437,503 cycles # 1.843 GHz - 8,290,568,638 instructions # 1.40 insn per cycle - 3.224462086 seconds time elapsed + 5,897,192,075 cycles # 1.881 GHz + 8,242,646,552 instructions # 1.40 insn per cycle + 3.136649808 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 1441) (512y: 122) (512z: 1802) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx/build.512z_d_inl0_hrd0/runTest_cpp.exe diff --git a/epochX/cudacpp/tput/logs_ggtt_mad/log_ggtt_mad_d_inl0_hrd0_curhst.txt b/epochX/cudacpp/tput/logs_ggtt_mad/log_ggtt_mad_d_inl0_hrd0_curhst.txt index 794a3c9310..85646834cb 100644 --- a/epochX/cudacpp/tput/logs_ggtt_mad/log_ggtt_mad_d_inl0_hrd0_curhst.txt +++ b/epochX/cudacpp/tput/logs_ggtt_mad/log_ggtt_mad_d_inl0_hrd0_curhst.txt @@ -40,7 +40,7 @@ make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx' -DATE: 2024-10-06_09:46:56 +DATE: 2024-10-06_14:05:29 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: ========================================================================= @@ -49,15 +49,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_SM_GG_TTX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURHST+RMBDEV+MESDEV/none+NAVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 4.311257e+07 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 9.342288e+07 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.004457e+08 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 4.299237e+07 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 9.310667e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 9.958799e+07 ) sec^-1 MeanMatrixElemValue = ( 2.086689e+00 +- 3.413217e-03 ) GeV^0 -TOTAL : 0.566605 sec +TOTAL : 0.569268 sec INFO: No Floating Point Exceptions have been reported - 2,313,605,054 cycles # 2.893 GHz - 3,600,350,267 instructions # 1.56 insn per cycle - 0.856648834 seconds time elapsed + 2,337,058,501 cycles # 2.886 GHz + 3,608,061,808 instructions # 1.54 insn per cycle + 0.867038360 seconds time elapsed runNcu /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx/build.cuda_d_inl0_hrd0/check_cuda.exe -p 2048 256 1 --curhst ==PROF== Profiling "sigmaKin": launch__registers_per_thread 214 ==PROF== Profiling "sigmaKin": sm__sass_average_branch_targets_threads_uniform.pct 100% @@ -85,15 +85,15 @@ Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [ha Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/none+NAVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = SCALAR ('none': ~vector[1], no SIMD) -EvtsPerSec[Rmb+ME] (23) = ( 1.824387e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.871256e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.871256e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.834250e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.881280e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.881280e+05 ) sec^-1 MeanMatrixElemValue = ( 2.086689e+00 +- 3.413217e-03 ) GeV^0 -TOTAL : 5.855309 sec +TOTAL : 5.824646 sec INFO: No Floating Point Exceptions have been reported - 17,230,682,954 cycles # 2.940 GHz - 45,932,528,772 instructions # 2.67 insn per cycle - 5.861424268 seconds time elapsed + 17,234,324,997 cycles # 2.957 GHz + 45,935,141,298 instructions # 2.67 insn per cycle + 5.830609901 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 617) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx/build.none_d_inl0_hrd0/runTest_cpp.exe @@ -117,15 +117,15 @@ Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [ha Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[2] ('sse4': SSE4.2, 128bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 3.215073e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.378302e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.378302e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 3.214396e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.374725e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.374725e+05 ) sec^-1 MeanMatrixElemValue = ( 2.086689e+00 +- 3.413217e-03 ) GeV^0 -TOTAL : 3.372268 sec +TOTAL : 3.372715 sec INFO: No Floating Point Exceptions have been reported - 9,959,367,668 cycles # 2.949 GHz - 27,848,270,798 instructions # 2.80 insn per cycle - 3.378265573 seconds time elapsed + 9,969,297,069 cycles # 2.952 GHz + 27,849,504,873 instructions # 2.79 insn per cycle + 3.378357330 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 2536) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx/build.sse4_d_inl0_hrd0/runTest_cpp.exe @@ -149,15 +149,15 @@ Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [ha Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('avx2': AVX2, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 4.999546e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 5.391220e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 5.391220e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 5.010793e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 5.395143e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 5.395143e+05 ) sec^-1 MeanMatrixElemValue = ( 2.086689e+00 +- 3.413217e-03 ) GeV^0 -TOTAL : 2.206484 sec +TOTAL : 2.201941 sec INFO: No Floating Point Exceptions have been reported - 6,113,930,208 cycles # 2.765 GHz - 12,581,849,902 instructions # 2.06 insn per cycle - 2.212402360 seconds time elapsed + 6,099,118,232 cycles # 2.764 GHz + 12,580,684,063 instructions # 2.06 insn per cycle + 2.207897324 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 2619) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx/build.avx2_d_inl0_hrd0/runTest_cpp.exe @@ -181,15 +181,15 @@ Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [ha Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('512y': AVX512, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 5.516180e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 5.984165e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 5.984165e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 5.490274e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 5.950543e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 5.950543e+05 ) sec^-1 MeanMatrixElemValue = ( 2.086689e+00 +- 3.413217e-03 ) GeV^0 -TOTAL : 2.007710 sec +TOTAL : 2.016981 sec INFO: No Floating Point Exceptions have been reported - 5,576,628,773 cycles # 2.771 GHz - 12,020,299,868 instructions # 2.16 insn per cycle - 2.013581558 seconds time elapsed + 5,586,835,450 cycles # 2.763 GHz + 12,023,518,273 instructions # 2.15 insn per cycle + 2.022798369 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 2360) (512y: 144) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx/build.512y_d_inl0_hrd0/runTest_cpp.exe @@ -213,15 +213,15 @@ Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [ha Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('512z': AVX512, 512bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 3.502286e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.687963e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.687963e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 3.549194e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.734136e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.734136e+05 ) sec^-1 MeanMatrixElemValue = ( 2.086689e+00 +- 3.413217e-03 ) GeV^0 -TOTAL : 3.102138 sec +TOTAL : 3.062333 sec INFO: No Floating Point Exceptions have been reported - 5,751,986,200 cycles # 1.852 GHz - 8,297,969,466 instructions # 1.44 insn per cycle - 3.107697215 seconds time elapsed + 5,715,296,170 cycles # 1.863 GHz + 8,294,887,685 instructions # 1.45 insn per cycle + 3.068307420 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 1441) (512y: 122) (512z: 1802) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx/build.512z_d_inl0_hrd0/runTest_cpp.exe diff --git a/epochX/cudacpp/tput/logs_ggtt_mad/log_ggtt_mad_d_inl0_hrd0_rmbhst.txt b/epochX/cudacpp/tput/logs_ggtt_mad/log_ggtt_mad_d_inl0_hrd0_rmbhst.txt index 70a45db399..0fb673ea20 100644 --- a/epochX/cudacpp/tput/logs_ggtt_mad/log_ggtt_mad_d_inl0_hrd0_rmbhst.txt +++ b/epochX/cudacpp/tput/logs_ggtt_mad/log_ggtt_mad_d_inl0_hrd0_rmbhst.txt @@ -40,7 +40,7 @@ make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx' -DATE: 2024-10-06_09:44:10 +DATE: 2024-10-06_14:02:42 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: ========================================================================= @@ -50,15 +50,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_SM_GG_TTX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURHST+RMBHST+MESDEV/none+NAVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 5.785807e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 9.291280e+07 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 9.973584e+07 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 5.751315e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 9.363313e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.001177e+08 ) sec^-1 MeanMatrixElemValue = ( 2.086689e+00 +- 3.413217e-03 ) GeV^0 -TOTAL : 0.717756 sec +TOTAL : 0.723715 sec INFO: No Floating Point Exceptions have been reported - 2,755,914,027 cycles # 2.900 GHz - 4,368,405,962 instructions # 1.59 insn per cycle - 1.007006361 seconds time elapsed + 2,764,217,383 cycles # 2.893 GHz + 4,350,237,710 instructions # 1.57 insn per cycle + 1.014396218 seconds time elapsed runNcu /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx/build.cuda_d_inl0_hrd0/check_cuda.exe -p 2048 256 1 --rmbhst WARNING! RamboHost selected: cannot use CurandDevice, will use CurandHost ==PROF== Profiling "sigmaKin": launch__registers_per_thread 214 @@ -87,15 +87,15 @@ Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [ha Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/none+NAVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = SCALAR ('none': ~vector[1], no SIMD) -EvtsPerSec[Rmb+ME] (23) = ( 1.829948e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.877608e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.877608e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.829121e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.875520e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.875520e+05 ) sec^-1 MeanMatrixElemValue = ( 2.086689e+00 +- 3.413217e-03 ) GeV^0 -TOTAL : 5.839744 sec +TOTAL : 5.840517 sec INFO: No Floating Point Exceptions have been reported - 17,231,514,699 cycles # 2.948 GHz - 45,931,758,909 instructions # 2.67 insn per cycle - 5.845651027 seconds time elapsed + 17,254,916,316 cycles # 2.952 GHz + 45,934,778,300 instructions # 2.66 insn per cycle + 5.846056356 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 617) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx/build.none_d_inl0_hrd0/runTest_cpp.exe @@ -119,15 +119,15 @@ Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [ha Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[2] ('sse4': SSE4.2, 128bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 3.215717e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.376174e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.376174e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 3.231460e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.391640e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.391640e+05 ) sec^-1 MeanMatrixElemValue = ( 2.086689e+00 +- 3.413217e-03 ) GeV^0 -TOTAL : 3.370523 sec +TOTAL : 3.352860 sec INFO: No Floating Point Exceptions have been reported - 9,939,666,586 cycles # 2.945 GHz - 27,847,302,489 instructions # 2.80 insn per cycle - 3.376515027 seconds time elapsed + 9,929,219,846 cycles # 2.957 GHz + 27,847,201,970 instructions # 2.80 insn per cycle + 3.358767883 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 2536) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx/build.sse4_d_inl0_hrd0/runTest_cpp.exe @@ -151,15 +151,15 @@ Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [ha Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('avx2': AVX2, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 5.058902e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 5.451650e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 5.451650e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 4.991769e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 5.373698e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 5.373698e+05 ) sec^-1 MeanMatrixElemValue = ( 2.086689e+00 +- 3.413217e-03 ) GeV^0 -TOTAL : 2.181386 sec +TOTAL : 2.208309 sec INFO: No Floating Point Exceptions have been reported - 6,074,037,919 cycles # 2.778 GHz - 12,580,567,087 instructions # 2.07 insn per cycle - 2.187203017 seconds time elapsed + 6,105,955,691 cycles # 2.759 GHz + 12,581,330,349 instructions # 2.06 insn per cycle + 2.214174663 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 2619) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx/build.avx2_d_inl0_hrd0/runTest_cpp.exe @@ -183,15 +183,15 @@ Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [ha Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('512y': AVX512, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 5.484469e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 5.947491e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 5.947491e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 5.505150e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 5.971050e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 5.971050e+05 ) sec^-1 MeanMatrixElemValue = ( 2.086689e+00 +- 3.413217e-03 ) GeV^0 -TOTAL : 2.020942 sec +TOTAL : 2.013361 sec INFO: No Floating Point Exceptions have been reported - 5,589,694,694 cycles # 2.759 GHz - 12,020,772,424 instructions # 2.15 insn per cycle - 2.026934215 seconds time elapsed + 5,597,932,254 cycles # 2.773 GHz + 12,020,430,111 instructions # 2.15 insn per cycle + 2.019378341 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 2360) (512y: 144) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx/build.512y_d_inl0_hrd0/runTest_cpp.exe @@ -215,15 +215,15 @@ Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [ha Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('512z': AVX512, 512bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 3.541083e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.728456e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.728456e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 3.535343e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.718319e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.718319e+05 ) sec^-1 MeanMatrixElemValue = ( 2.086689e+00 +- 3.413217e-03 ) GeV^0 -TOTAL : 3.072814 sec +TOTAL : 3.074719 sec INFO: No Floating Point Exceptions have been reported - 5,724,538,871 cycles # 1.860 GHz - 8,297,304,281 instructions # 1.45 insn per cycle - 3.079169559 seconds time elapsed + 5,719,476,896 cycles # 1.857 GHz + 8,293,481,357 instructions # 1.45 insn per cycle + 3.080737283 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 1441) (512y: 122) (512z: 1802) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx/build.512z_d_inl0_hrd0/runTest_cpp.exe diff --git a/epochX/cudacpp/tput/logs_ggtt_mad/log_ggtt_mad_d_inl0_hrd1.txt b/epochX/cudacpp/tput/logs_ggtt_mad/log_ggtt_mad_d_inl0_hrd1.txt index 03be4a726d..81cb07b5a4 100644 --- a/epochX/cudacpp/tput/logs_ggtt_mad/log_ggtt_mad_d_inl0_hrd1.txt +++ b/epochX/cudacpp/tput/logs_ggtt_mad/log_ggtt_mad_d_inl0_hrd1.txt @@ -40,7 +40,7 @@ make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx' -DATE: 2024-10-06_08:58:15 +DATE: 2024-10-06_13:27:28 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: ========================================================================= @@ -49,15 +49,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_SM_GG_TTX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=1] Workflow summary = CUD:DBL+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 4.508928e+07 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 9.321752e+07 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.002344e+08 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 4.650820e+07 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 9.493239e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.010953e+08 ) sec^-1 MeanMatrixElemValue = ( 2.086689e+00 +- 3.413217e-03 ) GeV^0 -TOTAL : 0.536365 sec +TOTAL : 0.522855 sec INFO: No Floating Point Exceptions have been reported - 2,214,194,265 cycles # 2.876 GHz - 3,152,115,430 instructions # 1.42 insn per cycle - 0.834564895 seconds time elapsed + 2,180,312,648 cycles # 2.885 GHz + 3,165,830,903 instructions # 1.45 insn per cycle + 0.812662327 seconds time elapsed runNcu /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx/build.cuda_d_inl0_hrd1/check_cuda.exe -p 2048 256 1 ==PROF== Profiling "sigmaKin": launch__registers_per_thread 212 ==PROF== Profiling "sigmaKin": sm__sass_average_branch_targets_threads_uniform.pct 100% @@ -85,15 +85,15 @@ Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [ha Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/none+NAVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = SCALAR ('none': ~vector[1], no SIMD) -EvtsPerSec[Rmb+ME] (23) = ( 1.855453e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.904405e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.904405e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.882478e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.931887e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.931887e+05 ) sec^-1 MeanMatrixElemValue = ( 2.086689e+00 +- 3.413217e-03 ) GeV^0 -TOTAL : 5.800560 sec +TOTAL : 5.675800 sec INFO: No Floating Point Exceptions have been reported - 16,903,949,090 cycles # 2.909 GHz - 45,043,853,273 instructions # 2.66 insn per cycle - 5.813534817 seconds time elapsed + 16,756,436,616 cycles # 2.950 GHz + 44,927,253,087 instructions # 2.68 insn per cycle + 5.681059418 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 567) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx/build.none_d_inl0_hrd1/runTest_cpp.exe @@ -117,15 +117,15 @@ Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [ha Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[2] ('sse4': SSE4.2, 128bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 3.339712e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.518637e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.518637e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 3.385691e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.563707e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.563707e+05 ) sec^-1 MeanMatrixElemValue = ( 2.086689e+00 +- 3.413217e-03 ) GeV^0 -TOTAL : 3.288467 sec +TOTAL : 3.203366 sec INFO: No Floating Point Exceptions have been reported - 9,645,043,566 cycles # 2.925 GHz - 26,807,862,552 instructions # 2.78 insn per cycle - 3.301069690 seconds time elapsed + 9,482,663,797 cycles # 2.957 GHz + 26,690,087,332 instructions # 2.81 insn per cycle + 3.208512270 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 2327) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx/build.sse4_d_inl0_hrd1/runTest_cpp.exe @@ -149,15 +149,15 @@ Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [ha Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('avx2': AVX2, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 4.590385e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 4.923511e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 4.923511e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 4.597783e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 4.922355e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 4.922355e+05 ) sec^-1 MeanMatrixElemValue = ( 2.086689e+00 +- 3.413217e-03 ) GeV^0 -TOTAL : 2.431911 sec +TOTAL : 2.387462 sec INFO: No Floating Point Exceptions have been reported - 6,762,097,168 cycles # 2.769 GHz - 14,239,182,198 instructions # 2.11 insn per cycle - 2.443454156 seconds time elapsed + 6,602,269,100 cycles # 2.761 GHz + 14,119,210,714 instructions # 2.14 insn per cycle + 2.392617106 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 2711) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx/build.avx2_d_inl0_hrd1/runTest_cpp.exe @@ -181,15 +181,15 @@ Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [ha Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('512y': AVX512, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 4.784038e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 5.137564e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 5.137564e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 4.808431e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 5.153993e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 5.153993e+05 ) sec^-1 MeanMatrixElemValue = ( 2.086689e+00 +- 3.413217e-03 ) GeV^0 -TOTAL : 2.339078 sec +TOTAL : 2.287719 sec INFO: No Floating Point Exceptions have been reported - 6,493,835,738 cycles # 2.765 GHz - 13,835,177,964 instructions # 2.13 insn per cycle - 2.350490634 seconds time elapsed + 6,345,565,904 cycles # 2.768 GHz + 13,718,128,101 instructions # 2.16 insn per cycle + 2.293020845 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 2360) (512y: 298) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx/build.512y_d_inl0_hrd1/runTest_cpp.exe @@ -213,15 +213,15 @@ Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [ha Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('512z': AVX512, 512bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 3.400894e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.576119e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.576119e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 3.400114e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.569934e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.569934e+05 ) sec^-1 MeanMatrixElemValue = ( 2.086689e+00 +- 3.413217e-03 ) GeV^0 -TOTAL : 3.231977 sec +TOTAL : 3.190819 sec INFO: No Floating Point Exceptions have been reported - 6,054,126,925 cycles # 1.868 GHz - 10,181,313,288 instructions # 1.68 insn per cycle - 3.245420113 seconds time elapsed + 5,905,889,175 cycles # 1.848 GHz + 10,061,383,006 instructions # 1.70 insn per cycle + 3.196142650 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 1273) (512y: 208) (512z: 1988) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx/build.512z_d_inl0_hrd1/runTest_cpp.exe diff --git a/epochX/cudacpp/tput/logs_ggtt_mad/log_ggtt_mad_d_inl1_hrd0.txt b/epochX/cudacpp/tput/logs_ggtt_mad/log_ggtt_mad_d_inl1_hrd0.txt index f94c1448dd..3694ffda9b 100644 --- a/epochX/cudacpp/tput/logs_ggtt_mad/log_ggtt_mad_d_inl1_hrd0.txt +++ b/epochX/cudacpp/tput/logs_ggtt_mad/log_ggtt_mad_d_inl1_hrd0.txt @@ -40,7 +40,7 @@ make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx' -DATE: 2024-10-06_09:28:32 +DATE: 2024-10-06_13:47:04 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: ========================================================================= @@ -49,15 +49,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_SM_GG_TTX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=1] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 4.445619e+07 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 9.389644e+07 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 9.998797e+07 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 4.542132e+07 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 9.394266e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.002071e+08 ) sec^-1 MeanMatrixElemValue = ( 2.086689e+00 +- 3.413217e-03 ) GeV^0 -TOTAL : 0.532166 sec +TOTAL : 0.530719 sec INFO: No Floating Point Exceptions have been reported - 2,223,705,741 cycles # 2.888 GHz - 3,137,862,648 instructions # 1.41 insn per cycle - 0.826622030 seconds time elapsed + 2,196,900,750 cycles # 2.844 GHz + 3,180,900,744 instructions # 1.45 insn per cycle + 0.830118820 seconds time elapsed runNcu /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx/build.cuda_d_inl1_hrd0/check_cuda.exe -p 2048 256 1 ==PROF== Profiling "sigmaKin": launch__registers_per_thread 214 ==PROF== Profiling "sigmaKin": sm__sass_average_branch_targets_threads_uniform.pct 100% @@ -85,15 +85,15 @@ Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=1] [ha Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/none+NAVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = SCALAR ('none': ~vector[1], no SIMD) -EvtsPerSec[Rmb+ME] (23) = ( 2.243473e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 2.316891e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 2.316891e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 2.441280e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 2.524720e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.524720e+05 ) sec^-1 MeanMatrixElemValue = ( 2.086689e+00 +- 3.413217e-03 ) GeV^0 -TOTAL : 4.822701 sec +TOTAL : 4.401566 sec INFO: No Floating Point Exceptions have been reported - 14,262,425,677 cycles # 2.951 GHz - 34,462,229,045 instructions # 2.42 insn per cycle - 4.834685593 seconds time elapsed + 13,000,821,055 cycles # 2.951 GHz + 34,343,199,734 instructions # 2.64 insn per cycle + 4.407038450 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 665) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx/build.none_d_inl1_hrd0/runTest_cpp.exe @@ -117,15 +117,15 @@ Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=1] [ha Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[2] ('sse4': SSE4.2, 128bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 2.991823e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.134338e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.134338e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 2.992615e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.129735e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.129735e+05 ) sec^-1 MeanMatrixElemValue = ( 2.086689e+00 +- 3.413217e-03 ) GeV^0 -TOTAL : 3.653458 sec +TOTAL : 3.611931 sec INFO: No Floating Point Exceptions have been reported - 10,828,452,798 cycles # 2.955 GHz - 24,364,594,695 instructions # 2.25 insn per cycle - 3.665357624 seconds time elapsed + 10,665,717,031 cycles # 2.949 GHz + 24,246,374,021 instructions # 2.27 insn per cycle + 3.617383876 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 2610) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx/build.sse4_d_inl1_hrd0/runTest_cpp.exe @@ -149,15 +149,15 @@ Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=1] [ha Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('avx2': AVX2, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 4.588361e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 4.923011e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 4.923011e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 4.546324e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 4.864769e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 4.864769e+05 ) sec^-1 MeanMatrixElemValue = ( 2.086689e+00 +- 3.413217e-03 ) GeV^0 -TOTAL : 2.432860 sec +TOTAL : 2.414769 sec INFO: No Floating Point Exceptions have been reported - 6,763,126,248 cycles # 2.768 GHz - 12,520,790,366 instructions # 1.85 insn per cycle - 2.444836798 seconds time elapsed + 6,641,570,512 cycles # 2.745 GHz + 12,403,751,593 instructions # 1.87 insn per cycle + 2.420285980 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 3115) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx/build.avx2_d_inl1_hrd0/runTest_cpp.exe @@ -181,15 +181,15 @@ Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=1] [ha Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('512y': AVX512, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 4.983949e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 5.371900e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 5.371900e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 4.972695e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 5.347832e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 5.347832e+05 ) sec^-1 MeanMatrixElemValue = ( 2.086689e+00 +- 3.413217e-03 ) GeV^0 -TOTAL : 2.251146 sec +TOTAL : 2.217215 sec INFO: No Floating Point Exceptions have been reported - 6,291,656,449 cycles # 2.782 GHz - 11,662,894,163 instructions # 1.85 insn per cycle - 2.263135736 seconds time elapsed + 6,146,192,793 cycles # 2.766 GHz + 11,546,052,028 instructions # 1.88 insn per cycle + 2.223172793 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 2644) (512y: 239) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx/build.512y_d_inl1_hrd0/runTest_cpp.exe @@ -213,15 +213,15 @@ Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=1] [ha Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('512z': AVX512, 512bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 3.728872e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.941749e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.941749e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 3.744764e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.951527e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.951527e+05 ) sec^-1 MeanMatrixElemValue = ( 2.086689e+00 +- 3.413217e-03 ) GeV^0 -TOTAL : 2.960781 sec +TOTAL : 2.909333 sec INFO: No Floating Point Exceptions have been reported - 5,563,913,804 cycles # 1.872 GHz - 9,412,295,126 instructions # 1.69 insn per cycle - 2.972906161 seconds time elapsed + 5,406,333,025 cycles # 1.855 GHz + 9,292,248,954 instructions # 1.72 insn per cycle + 2.915145064 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 2099) (512y: 282) (512z: 1958) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx/build.512z_d_inl1_hrd0/runTest_cpp.exe diff --git a/epochX/cudacpp/tput/logs_ggtt_mad/log_ggtt_mad_d_inl1_hrd1.txt b/epochX/cudacpp/tput/logs_ggtt_mad/log_ggtt_mad_d_inl1_hrd1.txt index 3c1647789f..9c0b68360e 100644 --- a/epochX/cudacpp/tput/logs_ggtt_mad/log_ggtt_mad_d_inl1_hrd1.txt +++ b/epochX/cudacpp/tput/logs_ggtt_mad/log_ggtt_mad_d_inl1_hrd1.txt @@ -40,7 +40,7 @@ make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx' -DATE: 2024-10-06_09:28:57 +DATE: 2024-10-06_13:47:28 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: ========================================================================= @@ -49,15 +49,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_SM_GG_TTX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=1] [hardcodePARAM=1] Workflow summary = CUD:DBL+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 4.391002e+07 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 9.323919e+07 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 9.976474e+07 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 4.572904e+07 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 9.381633e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 9.985203e+07 ) sec^-1 MeanMatrixElemValue = ( 2.086689e+00 +- 3.413217e-03 ) GeV^0 -TOTAL : 0.534058 sec +TOTAL : 0.530862 sec INFO: No Floating Point Exceptions have been reported - 2,225,875,951 cycles # 2.883 GHz - 3,143,824,990 instructions # 1.41 insn per cycle - 0.828954123 seconds time elapsed + 2,208,280,848 cycles # 2.877 GHz + 3,180,669,575 instructions # 1.44 insn per cycle + 0.825078964 seconds time elapsed runNcu /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx/build.cuda_d_inl1_hrd1/check_cuda.exe -p 2048 256 1 ==PROF== Profiling "sigmaKin": launch__registers_per_thread 212 ==PROF== Profiling "sigmaKin": sm__sass_average_branch_targets_threads_uniform.pct 100% @@ -85,15 +85,15 @@ Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=1] [ha Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/none+NAVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = SCALAR ('none': ~vector[1], no SIMD) -EvtsPerSec[Rmb+ME] (23) = ( 2.586147e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 2.682611e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 2.682611e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 2.539377e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 2.629831e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.629831e+05 ) sec^-1 MeanMatrixElemValue = ( 2.086689e+00 +- 3.413217e-03 ) GeV^0 -TOTAL : 4.200138 sec +TOTAL : 4.235831 sec INFO: No Floating Point Exceptions have been reported - 12,457,576,414 cycles # 2.958 GHz - 35,030,140,380 instructions # 2.81 insn per cycle - 4.211834896 seconds time elapsed + 12,319,969,560 cycles # 2.905 GHz + 34,914,305,112 instructions # 2.83 insn per cycle + 4.241550200 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 430) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx/build.none_d_inl1_hrd1/runTest_cpp.exe @@ -117,15 +117,15 @@ Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=1] [ha Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[2] ('sse4': SSE4.2, 128bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 3.003695e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.145378e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.145378e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 3.009417e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.148137e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.148137e+05 ) sec^-1 MeanMatrixElemValue = ( 2.086689e+00 +- 3.413217e-03 ) GeV^0 -TOTAL : 3.637171 sec +TOTAL : 3.592072 sec INFO: No Floating Point Exceptions have been reported - 10,771,658,335 cycles # 2.953 GHz - 23,459,809,146 instructions # 2.18 insn per cycle - 3.648522280 seconds time elapsed + 10,622,202,409 cycles # 2.953 GHz + 23,340,208,769 instructions # 2.20 insn per cycle + 3.597876603 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 2378) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx/build.sse4_d_inl1_hrd1/runTest_cpp.exe @@ -149,15 +149,15 @@ Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=1] [ha Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('avx2': AVX2, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 5.029039e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 5.423785e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 5.423785e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 5.085474e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 5.485638e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 5.485638e+05 ) sec^-1 MeanMatrixElemValue = ( 2.086689e+00 +- 3.413217e-03 ) GeV^0 -TOTAL : 2.231082 sec +TOTAL : 2.170896 sec INFO: No Floating Point Exceptions have been reported - 6,224,358,348 cycles # 2.777 GHz - 11,980,138,777 instructions # 1.92 insn per cycle - 2.242426635 seconds time elapsed + 6,039,004,645 cycles # 2.775 GHz + 11,860,765,125 instructions # 1.96 insn per cycle + 2.176789286 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 2468) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx/build.avx2_d_inl1_hrd1/runTest_cpp.exe @@ -181,15 +181,15 @@ Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=1] [ha Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('512y': AVX512, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 5.044695e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 5.439218e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 5.439218e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 5.026086e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 5.416845e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 5.416845e+05 ) sec^-1 MeanMatrixElemValue = ( 2.086689e+00 +- 3.413217e-03 ) GeV^0 -TOTAL : 2.224952 sec +TOTAL : 2.194582 sec INFO: No Floating Point Exceptions have been reported - 6,216,689,838 cycles # 2.781 GHz - 11,219,235,507 instructions # 1.80 insn per cycle - 2.236216110 seconds time elapsed + 6,066,869,712 cycles # 2.759 GHz + 11,099,028,693 instructions # 1.83 insn per cycle + 2.200285606 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 2098) (512y: 174) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx/build.512y_d_inl1_hrd1/runTest_cpp.exe @@ -213,15 +213,15 @@ Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=1] [ha Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('512z': AVX512, 512bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 3.888626e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 4.118349e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 4.118349e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 3.867368e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 4.087327e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 4.087327e+05 ) sec^-1 MeanMatrixElemValue = ( 2.086689e+00 +- 3.413217e-03 ) GeV^0 -TOTAL : 2.842958 sec +TOTAL : 2.818594 sec INFO: No Floating Point Exceptions have been reported - 5,376,391,405 cycles # 1.885 GHz - 9,136,626,879 instructions # 1.70 insn per cycle - 2.854254782 seconds time elapsed + 5,253,289,224 cycles # 1.861 GHz + 9,016,309,852 instructions # 1.72 insn per cycle + 2.824586528 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 1632) (512y: 208) (512z: 1567) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx/build.512z_d_inl1_hrd1/runTest_cpp.exe diff --git a/epochX/cudacpp/tput/logs_ggtt_mad/log_ggtt_mad_f_inl0_hrd0.txt b/epochX/cudacpp/tput/logs_ggtt_mad/log_ggtt_mad_f_inl0_hrd0.txt index eed598e900..6bc06fabba 100644 --- a/epochX/cudacpp/tput/logs_ggtt_mad/log_ggtt_mad_f_inl0_hrd0.txt +++ b/epochX/cudacpp/tput/logs_ggtt_mad/log_ggtt_mad_f_inl0_hrd0.txt @@ -40,7 +40,7 @@ make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx' -DATE: 2024-10-06_08:59:31 +DATE: 2024-10-06_13:28:43 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: ========================================================================= @@ -49,15 +49,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_SM_GG_TTX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 8.348925e+07 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.730429e+08 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.847126e+08 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 8.988202e+07 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.810380e+08 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.929298e+08 ) sec^-1 MeanMatrixElemValue = ( 2.086719e+00 +- 3.413389e-03 ) GeV^0 -TOTAL : 0.489368 sec +TOTAL : 0.482560 sec INFO: No Floating Point Exceptions have been reported - 2,066,464,716 cycles # 2.888 GHz - 2,966,218,976 instructions # 1.44 insn per cycle - 0.775358949 seconds time elapsed + 2,030,553,628 cycles # 2.873 GHz + 2,925,743,931 instructions # 1.44 insn per cycle + 0.763834290 seconds time elapsed runNcu /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx/build.cuda_f_inl0_hrd0/check_cuda.exe -p 2048 256 1 ==PROF== Profiling "sigmaKin": launch__registers_per_thread 124 ==PROF== Profiling "sigmaKin": sm__sass_average_branch_targets_threads_uniform.pct 100% @@ -85,15 +85,15 @@ Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [ha Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/none+NAVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = SCALAR ('none': ~vector[1], no SIMD) -EvtsPerSec[Rmb+ME] (23) = ( 1.920704e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.976809e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.976809e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.930475e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.984782e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.984782e+05 ) sec^-1 MeanMatrixElemValue = ( 2.086780e+00 +- 3.413794e-03 ) GeV^0 -TOTAL : 5.564608 sec +TOTAL : 5.517063 sec INFO: No Floating Point Exceptions have been reported - 16,407,008,301 cycles # 2.946 GHz - 45,390,324,197 instructions # 2.77 insn per cycle - 5.572247633 seconds time elapsed + 16,238,980,959 cycles # 2.941 GHz + 45,332,785,627 instructions # 2.79 insn per cycle + 5.522172195 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 591) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx/build.none_f_inl0_hrd0/runTest_cpp.exe @@ -117,15 +117,15 @@ Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [ha Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('sse4': SSE4.2, 128bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 4.527362e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 4.867119e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 4.867119e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 4.539962e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 4.876067e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 4.876067e+05 ) sec^-1 MeanMatrixElemValue = ( 2.086779e+00 +- 3.413793e-03 ) GeV^0 -TOTAL : 2.418751 sec +TOTAL : 2.394351 sec INFO: No Floating Point Exceptions have been reported - 7,148,582,676 cycles # 2.947 GHz - 17,841,430,692 instructions # 2.50 insn per cycle - 2.426747092 seconds time elapsed + 7,081,255,610 cycles # 2.953 GHz + 17,793,019,962 instructions # 2.51 insn per cycle + 2.399317581 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 3144) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx/build.sse4_f_inl0_hrd0/runTest_cpp.exe @@ -149,15 +149,15 @@ Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [ha Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('avx2': AVX2, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 8.351940e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 9.517580e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 9.517580e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 8.272004e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 9.392729e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 9.392729e+05 ) sec^-1 MeanMatrixElemValue = ( 2.086810e+00 +- 3.414231e-03 ) GeV^0 -TOTAL : 1.358362 sec +TOTAL : 1.351770 sec INFO: No Floating Point Exceptions have been reported - 3,812,563,399 cycles # 2.792 GHz - 8,312,155,726 instructions # 2.18 insn per cycle - 1.366469053 seconds time elapsed + 3,729,927,650 cycles # 2.751 GHz + 8,262,373,488 instructions # 2.22 insn per cycle + 1.356751601 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 3367) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx/build.avx2_f_inl0_hrd0/runTest_cpp.exe @@ -181,15 +181,15 @@ Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [ha Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('512y': AVX512, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 8.799220e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.010674e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.010674e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 8.845956e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.013123e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.013123e+06 ) sec^-1 MeanMatrixElemValue = ( 2.086810e+00 +- 3.414231e-03 ) GeV^0 -TOTAL : 1.295301 sec +TOTAL : 1.269392 sec INFO: No Floating Point Exceptions have been reported - 3,622,174,398 cycles # 2.781 GHz - 7,961,498,247 instructions # 2.20 insn per cycle - 1.303182368 seconds time elapsed + 3,546,289,531 cycles # 2.784 GHz + 7,913,381,925 instructions # 2.23 insn per cycle + 1.274701941 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 3209) (512y: 20) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx/build.512y_f_inl0_hrd0/runTest_cpp.exe @@ -213,15 +213,15 @@ Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [ha Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[16] ('512z': AVX512, 512bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 6.500324e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 7.161825e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 7.161825e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 6.504044e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 7.162443e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 7.162443e+05 ) sec^-1 MeanMatrixElemValue = ( 2.086810e+00 +- 3.414231e-03 ) GeV^0 -TOTAL : 1.717843 sec +TOTAL : 1.697906 sec INFO: No Floating Point Exceptions have been reported - 3,332,199,340 cycles # 1.933 GHz - 6,146,454,565 instructions # 1.84 insn per cycle - 1.725889754 seconds time elapsed + 3,259,243,145 cycles # 1.915 GHz + 6,096,671,875 instructions # 1.87 insn per cycle + 1.703001110 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 2251) (512y: 22) (512z: 2155) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx/build.512z_f_inl0_hrd0/runTest_cpp.exe diff --git a/epochX/cudacpp/tput/logs_ggtt_mad/log_ggtt_mad_f_inl0_hrd0_bridge.txt b/epochX/cudacpp/tput/logs_ggtt_mad/log_ggtt_mad_f_inl0_hrd0_bridge.txt index ba391daf9b..0701579e00 100644 --- a/epochX/cudacpp/tput/logs_ggtt_mad/log_ggtt_mad_f_inl0_hrd0_bridge.txt +++ b/epochX/cudacpp/tput/logs_ggtt_mad/log_ggtt_mad_f_inl0_hrd0_bridge.txt @@ -40,7 +40,7 @@ make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx' -DATE: 2024-10-06_09:38:02 +DATE: 2024-10-06_13:56:35 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: ========================================================================= @@ -53,16 +53,16 @@ WARNING! Set grid in Bridge (nevt=524288, gpublocks=2048, gputhreads=256, gpublo Process = SIGMA_SM_GG_TTX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 6.962971e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 5.366502e+07 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 5.366502e+07 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 7.001138e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 5.393743e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 5.393743e+07 ) sec^-1 MeanMatrixElemValue = ( 2.086805e+00 +- 3.414078e-03 ) GeV^0 -TOTAL : 0.683449 sec +TOTAL : 0.681655 sec INFO: No Floating Point Exceptions have been reported INFO: No Floating Point Exceptions have been reported - 2,639,955,466 cycles # 2.881 GHz - 4,089,465,491 instructions # 1.55 insn per cycle - 0.973820402 seconds time elapsed + 2,642,599,060 cycles # 2.889 GHz + 4,094,934,466 instructions # 1.55 insn per cycle + 0.971584363 seconds time elapsed runNcu /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx/build.cuda_f_inl0_hrd0/check_cuda.exe -p 2048 256 1 --bridge WARNING! Bridge selected: cannot use RamboDevice, will use RamboHost WARNING! RamboHost selected: cannot use CurandDevice, will use CurandHost @@ -95,16 +95,16 @@ Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [ha Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+BRDHST/none+NAVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = SCALAR ('none': ~vector[1], no SIMD) -EvtsPerSec[Rmb+ME] (23) = ( 1.921187e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.975107e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.975107e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.928735e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.983213e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.983213e+05 ) sec^-1 MeanMatrixElemValue = ( 2.086780e+00 +- 3.413794e-03 ) GeV^0 -TOTAL : 5.590872 sec +TOTAL : 5.567760 sec INFO: No Floating Point Exceptions have been reported INFO: No Floating Point Exceptions have been reported - 16,505,827,538 cycles # 2.949 GHz - 45,383,324,587 instructions # 2.75 insn per cycle - 5.597525299 seconds time elapsed + 16,438,711,316 cycles # 2.950 GHz + 45,376,931,052 instructions # 2.76 insn per cycle + 5.575010601 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 591) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx/build.none_f_inl0_hrd0/runTest_cpp.exe @@ -129,16 +129,16 @@ Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [ha Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+BRDHST/sse4+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('sse4': SSE4.2, 128bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 4.503675e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 4.835801e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 4.835801e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 4.488400e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 4.818362e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 4.818362e+05 ) sec^-1 MeanMatrixElemValue = ( 2.086779e+00 +- 3.413793e-03 ) GeV^0 -TOTAL : 2.463825 sec +TOTAL : 2.472017 sec INFO: No Floating Point Exceptions have been reported INFO: No Floating Point Exceptions have been reported - 7,301,370,898 cycles # 2.956 GHz - 18,072,803,019 instructions # 2.48 insn per cycle - 2.471007950 seconds time elapsed + 7,300,900,664 cycles # 2.946 GHz + 18,074,316,483 instructions # 2.48 insn per cycle + 2.479144655 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 3144) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx/build.sse4_f_inl0_hrd0/runTest_cpp.exe @@ -163,16 +163,16 @@ Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [ha Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+BRDHST/avx2+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('avx2': AVX2, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 8.228346e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 9.356902e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 9.356902e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 8.201402e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 9.304884e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 9.304884e+05 ) sec^-1 MeanMatrixElemValue = ( 2.086810e+00 +- 3.414231e-03 ) GeV^0 -TOTAL : 1.409585 sec +TOTAL : 1.415527 sec INFO: No Floating Point Exceptions have been reported INFO: No Floating Point Exceptions have been reported - 3,950,274,134 cycles # 2.790 GHz - 8,500,615,795 instructions # 2.15 insn per cycle - 1.416669722 seconds time elapsed + 3,958,284,133 cycles # 2.783 GHz + 8,500,667,388 instructions # 2.15 insn per cycle + 1.422811127 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 3367) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx/build.avx2_f_inl0_hrd0/runTest_cpp.exe @@ -197,16 +197,16 @@ Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [ha Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+BRDHST/512y+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('512y': AVX512, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 8.630316e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 9.908478e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 9.908478e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 8.667599e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 9.898069e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 9.898069e+05 ) sec^-1 MeanMatrixElemValue = ( 2.086810e+00 +- 3.414231e-03 ) GeV^0 -TOTAL : 1.350838 sec +TOTAL : 1.346602 sec INFO: No Floating Point Exceptions have been reported INFO: No Floating Point Exceptions have been reported - 3,776,468,219 cycles # 2.783 GHz - 8,150,432,975 instructions # 2.16 insn per cycle - 1.357973048 seconds time elapsed + 3,766,986,307 cycles # 2.784 GHz + 8,150,402,103 instructions # 2.16 insn per cycle + 1.353852071 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 3209) (512y: 20) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx/build.512y_f_inl0_hrd0/runTest_cpp.exe @@ -231,16 +231,16 @@ Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [ha Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+BRDHST/512z+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[16] ('512z': AVX512, 512bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 6.446924e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 7.088794e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 7.088794e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 6.423848e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 7.070941e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 7.070941e+05 ) sec^-1 MeanMatrixElemValue = ( 2.086810e+00 +- 3.414231e-03 ) GeV^0 -TOTAL : 1.766906 sec +TOTAL : 1.770998 sec INFO: No Floating Point Exceptions have been reported INFO: No Floating Point Exceptions have been reported - 3,483,580,907 cycles # 1.964 GHz - 6,352,443,418 instructions # 1.82 insn per cycle - 1.774118995 seconds time elapsed + 3,490,297,909 cycles # 1.964 GHz + 6,352,446,136 instructions # 1.82 insn per cycle + 1.778150945 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 2251) (512y: 22) (512z: 2155) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx/build.512z_f_inl0_hrd0/runTest_cpp.exe diff --git a/epochX/cudacpp/tput/logs_ggtt_mad/log_ggtt_mad_f_inl0_hrd0_common.txt b/epochX/cudacpp/tput/logs_ggtt_mad/log_ggtt_mad_f_inl0_hrd0_common.txt index eaf1557b5a..a374e3d3ca 100644 --- a/epochX/cudacpp/tput/logs_ggtt_mad/log_ggtt_mad_f_inl0_hrd0_common.txt +++ b/epochX/cudacpp/tput/logs_ggtt_mad/log_ggtt_mad_f_inl0_hrd0_common.txt @@ -40,7 +40,7 @@ make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx' -DATE: 2024-10-06_09:50:12 +DATE: 2024-10-06_14:08:44 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: ========================================================================= @@ -49,15 +49,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_SM_GG_TTX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:COMMON+RMBDEV+MESDEV/none+NAVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 8.125576e+07 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.707303e+08 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.828418e+08 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 8.124818e+07 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.682549e+08 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.800122e+08 ) sec^-1 MeanMatrixElemValue = ( 2.079446e+00 +- 3.403306e-03 ) GeV^0 -TOTAL : 0.579716 sec +TOTAL : 0.580656 sec INFO: No Floating Point Exceptions have been reported - 2,336,853,883 cycles # 2.860 GHz - 3,355,823,518 instructions # 1.44 insn per cycle - 0.873538557 seconds time elapsed + 2,335,242,170 cycles # 2.874 GHz + 3,427,367,764 instructions # 1.47 insn per cycle + 0.869365004 seconds time elapsed runNcu /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx/build.cuda_f_inl0_hrd0/check_cuda.exe -p 2048 256 1 --common ==PROF== Profiling "sigmaKin": launch__registers_per_thread 124 ==PROF== Profiling "sigmaKin": sm__sass_average_branch_targets_threads_uniform.pct 100% @@ -85,15 +85,15 @@ Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [ha Workflow summary = CPP:FLT+CXS:COMMON+RMBHST+MESHST/none+NAVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = SCALAR ('none': ~vector[1], no SIMD) -EvtsPerSec[Rmb+ME] (23) = ( 1.929027e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.983438e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.983438e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.936385e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.990913e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.990913e+05 ) sec^-1 MeanMatrixElemValue = ( 2.079573e+00 +- 3.404712e-03 ) GeV^0 -TOTAL : 5.578220 sec +TOTAL : 5.558383 sec INFO: No Floating Point Exceptions have been reported - 16,412,792,219 cycles # 2.940 GHz - 45,364,108,775 instructions # 2.76 insn per cycle - 5.583854256 seconds time elapsed + 16,414,985,229 cycles # 2.951 GHz + 45,360,917,490 instructions # 2.76 insn per cycle + 5.564145023 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 591) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx/build.none_f_inl0_hrd0/runTest_cpp.exe @@ -117,15 +117,15 @@ Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [ha Workflow summary = CPP:FLT+CXS:COMMON+RMBHST+MESHST/sse4+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('sse4': SSE4.2, 128bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 4.528116e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 4.863028e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 4.863028e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 4.533600e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 4.870640e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 4.870640e+05 ) sec^-1 MeanMatrixElemValue = ( 2.079572e+00 +- 3.404712e-03 ) GeV^0 -TOTAL : 2.458830 sec +TOTAL : 2.455012 sec INFO: No Floating Point Exceptions have been reported - 7,256,357,914 cycles # 2.945 GHz - 17,803,442,746 instructions # 2.45 insn per cycle - 2.464565338 seconds time elapsed + 7,257,807,303 cycles # 2.951 GHz + 17,804,877,687 instructions # 2.45 insn per cycle + 2.460688762 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 3144) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx/build.sse4_f_inl0_hrd0/runTest_cpp.exe @@ -149,15 +149,15 @@ Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [ha Workflow summary = CPP:FLT+CXS:COMMON+RMBHST+MESHST/avx2+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('avx2': AVX2, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 8.321630e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 9.466483e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 9.466483e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 8.320583e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 9.475500e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 9.475500e+05 ) sec^-1 MeanMatrixElemValue = ( 2.079550e+00 +- 3.404207e-03 ) GeV^0 -TOTAL : 1.402951 sec +TOTAL : 1.405622 sec INFO: No Floating Point Exceptions have been reported - 3,915,341,003 cycles # 2.781 GHz - 8,245,891,296 instructions # 2.11 insn per cycle - 1.408611815 seconds time elapsed + 3,899,204,152 cycles # 2.765 GHz + 8,246,395,453 instructions # 2.11 insn per cycle + 1.411314887 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 3367) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx/build.avx2_f_inl0_hrd0/runTest_cpp.exe @@ -181,15 +181,15 @@ Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [ha Workflow summary = CPP:FLT+CXS:COMMON+RMBHST+MESHST/512y+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('512y': AVX512, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 8.769699e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.005525e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.005525e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 8.773484e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.004081e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.004081e+06 ) sec^-1 MeanMatrixElemValue = ( 2.079550e+00 +- 3.404207e-03 ) GeV^0 -TOTAL : 1.339268 sec +TOTAL : 1.340341 sec INFO: No Floating Point Exceptions have been reported - 3,730,447,512 cycles # 2.775 GHz - 7,861,984,465 instructions # 2.11 insn per cycle - 1.344998375 seconds time elapsed + 3,745,798,011 cycles # 2.785 GHz + 7,864,457,535 instructions # 2.10 insn per cycle + 1.345928092 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 3209) (512y: 20) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx/build.512y_f_inl0_hrd0/runTest_cpp.exe @@ -213,15 +213,15 @@ Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [ha Workflow summary = CPP:FLT+CXS:COMMON+RMBHST+MESHST/512z+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[16] ('512z': AVX512, 512bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 6.517692e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 7.188107e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 7.188107e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 6.516726e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 7.174104e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 7.174104e+05 ) sec^-1 MeanMatrixElemValue = ( 2.079550e+00 +- 3.404208e-03 ) GeV^0 -TOTAL : 1.753383 sec +TOTAL : 1.753681 sec INFO: No Floating Point Exceptions have been reported - 3,445,483,739 cycles # 1.959 GHz - 6,046,658,237 instructions # 1.75 insn per cycle - 1.759146158 seconds time elapsed + 3,433,552,639 cycles # 1.953 GHz + 6,046,953,497 instructions # 1.76 insn per cycle + 1.759227573 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 2251) (512y: 22) (512z: 2155) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx/build.512z_f_inl0_hrd0/runTest_cpp.exe diff --git a/epochX/cudacpp/tput/logs_ggtt_mad/log_ggtt_mad_f_inl0_hrd0_curhst.txt b/epochX/cudacpp/tput/logs_ggtt_mad/log_ggtt_mad_f_inl0_hrd0_curhst.txt index 0132142a7f..f867f1fda9 100644 --- a/epochX/cudacpp/tput/logs_ggtt_mad/log_ggtt_mad_f_inl0_hrd0_curhst.txt +++ b/epochX/cudacpp/tput/logs_ggtt_mad/log_ggtt_mad_f_inl0_hrd0_curhst.txt @@ -40,7 +40,7 @@ make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx' -DATE: 2024-10-06_09:47:21 +DATE: 2024-10-06_14:05:54 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: ========================================================================= @@ -49,15 +49,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_SM_GG_TTX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:CURHST+RMBDEV+MESDEV/none+NAVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 8.231900e+07 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.718618e+08 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.843172e+08 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 8.358541e+07 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.734628e+08 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.855238e+08 ) sec^-1 MeanMatrixElemValue = ( 2.086719e+00 +- 3.413389e-03 ) GeV^0 -TOTAL : 0.520956 sec +TOTAL : 0.521013 sec INFO: No Floating Point Exceptions have been reported - 2,145,908,279 cycles # 2.880 GHz - 3,342,720,192 instructions # 1.56 insn per cycle - 0.802555619 seconds time elapsed + 2,150,474,630 cycles # 2.885 GHz + 3,378,484,178 instructions # 1.57 insn per cycle + 0.802200013 seconds time elapsed runNcu /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx/build.cuda_f_inl0_hrd0/check_cuda.exe -p 2048 256 1 --curhst ==PROF== Profiling "sigmaKin": launch__registers_per_thread 124 ==PROF== Profiling "sigmaKin": sm__sass_average_branch_targets_threads_uniform.pct 100% @@ -85,15 +85,15 @@ Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [ha Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/none+NAVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = SCALAR ('none': ~vector[1], no SIMD) -EvtsPerSec[Rmb+ME] (23) = ( 1.929666e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.983661e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.983661e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.936419e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.991718e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.991718e+05 ) sec^-1 MeanMatrixElemValue = ( 2.086780e+00 +- 3.413794e-03 ) GeV^0 -TOTAL : 5.518695 sec +TOTAL : 5.500669 sec INFO: No Floating Point Exceptions have been reported - 16,237,309,072 cycles # 2.940 GHz - 45,332,194,999 instructions # 2.79 insn per cycle - 5.524338903 seconds time elapsed + 16,243,456,976 cycles # 2.950 GHz + 45,331,864,982 instructions # 2.79 insn per cycle + 5.506347589 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 591) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx/build.none_f_inl0_hrd0/runTest_cpp.exe @@ -117,15 +117,15 @@ Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [ha Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('sse4': SSE4.2, 128bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 4.531812e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 4.871745e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 4.871745e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 4.539074e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 4.875984e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 4.875984e+05 ) sec^-1 MeanMatrixElemValue = ( 2.086779e+00 +- 3.413793e-03 ) GeV^0 -TOTAL : 2.400101 sec +TOTAL : 2.396296 sec INFO: No Floating Point Exceptions have been reported - 7,092,917,063 cycles # 2.949 GHz - 17,790,950,300 instructions # 2.51 insn per cycle - 2.405895056 seconds time elapsed + 7,089,075,019 cycles # 2.952 GHz + 17,790,748,794 instructions # 2.51 insn per cycle + 2.402102973 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 3144) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx/build.sse4_f_inl0_hrd0/runTest_cpp.exe @@ -149,15 +149,15 @@ Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [ha Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('avx2': AVX2, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 8.364764e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 9.520513e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 9.520513e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 8.334242e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 9.482347e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 9.482347e+05 ) sec^-1 MeanMatrixElemValue = ( 2.086810e+00 +- 3.414231e-03 ) GeV^0 -TOTAL : 1.339869 sec +TOTAL : 1.345471 sec INFO: No Floating Point Exceptions have been reported - 3,746,789,760 cycles # 2.786 GHz - 8,261,610,745 instructions # 2.20 insn per cycle - 1.345882215 seconds time elapsed + 3,742,344,939 cycles # 2.771 GHz + 8,261,651,919 instructions # 2.21 insn per cycle + 1.351112654 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 3367) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx/build.avx2_f_inl0_hrd0/runTest_cpp.exe @@ -181,15 +181,15 @@ Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [ha Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('512y': AVX512, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 8.818621e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.013746e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.013746e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 8.830982e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.010123e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.010123e+06 ) sec^-1 MeanMatrixElemValue = ( 2.086810e+00 +- 3.414231e-03 ) GeV^0 -TOTAL : 1.275913 sec +TOTAL : 1.273479 sec INFO: No Floating Point Exceptions have been reported - 3,561,649,230 cycles # 2.781 GHz - 7,911,264,889 instructions # 2.22 insn per cycle - 1.281614236 seconds time elapsed + 3,557,820,802 cycles # 2.787 GHz + 7,912,431,339 instructions # 2.22 insn per cycle + 1.279203364 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 3209) (512y: 20) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx/build.512y_f_inl0_hrd0/runTest_cpp.exe @@ -213,15 +213,15 @@ Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [ha Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[16] ('512z': AVX512, 512bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 6.490214e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 7.139560e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 7.139560e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 6.479757e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 7.138944e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 7.138944e+05 ) sec^-1 MeanMatrixElemValue = ( 2.086810e+00 +- 3.414231e-03 ) GeV^0 -TOTAL : 1.701983 sec +TOTAL : 1.705449 sec INFO: No Floating Point Exceptions have been reported - 3,270,370,699 cycles # 1.916 GHz - 6,096,029,839 instructions # 1.86 insn per cycle - 1.707817189 seconds time elapsed + 3,281,260,592 cycles # 1.918 GHz + 6,096,727,410 instructions # 1.86 insn per cycle + 1.711340491 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 2251) (512y: 22) (512z: 2155) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx/build.512z_f_inl0_hrd0/runTest_cpp.exe diff --git a/epochX/cudacpp/tput/logs_ggtt_mad/log_ggtt_mad_f_inl0_hrd0_rmbhst.txt b/epochX/cudacpp/tput/logs_ggtt_mad/log_ggtt_mad_f_inl0_hrd0_rmbhst.txt index 55c92f68ec..b94fa5c767 100644 --- a/epochX/cudacpp/tput/logs_ggtt_mad/log_ggtt_mad_f_inl0_hrd0_rmbhst.txt +++ b/epochX/cudacpp/tput/logs_ggtt_mad/log_ggtt_mad_f_inl0_hrd0_rmbhst.txt @@ -40,7 +40,7 @@ make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx' -DATE: 2024-10-06_09:44:35 +DATE: 2024-10-06_14:03:08 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: ========================================================================= @@ -50,15 +50,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_SM_GG_TTX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:CURHST+RMBHST+MESDEV/none+NAVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 8.418560e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.722658e+08 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.839243e+08 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 8.632708e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.731576e+08 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.852598e+08 ) sec^-1 MeanMatrixElemValue = ( 2.086805e+00 +- 3.414078e-03 ) GeV^0 -TOTAL : 0.630221 sec +TOTAL : 0.628418 sec INFO: No Floating Point Exceptions have been reported - 2,475,236,721 cycles # 2.897 GHz - 3,823,734,565 instructions # 1.54 insn per cycle - 0.911361538 seconds time elapsed + 2,452,122,761 cycles # 2.875 GHz + 3,823,221,566 instructions # 1.56 insn per cycle + 0.909820522 seconds time elapsed runNcu /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx/build.cuda_f_inl0_hrd0/check_cuda.exe -p 2048 256 1 --rmbhst WARNING! RamboHost selected: cannot use CurandDevice, will use CurandHost ==PROF== Profiling "sigmaKin": launch__registers_per_thread 124 @@ -87,15 +87,15 @@ Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [ha Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/none+NAVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = SCALAR ('none': ~vector[1], no SIMD) -EvtsPerSec[Rmb+ME] (23) = ( 1.933112e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.987540e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.987540e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.935490e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.989949e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.989949e+05 ) sec^-1 MeanMatrixElemValue = ( 2.086780e+00 +- 3.413794e-03 ) GeV^0 -TOTAL : 5.510619 sec +TOTAL : 5.504188 sec INFO: No Floating Point Exceptions have been reported - 16,239,692,933 cycles # 2.945 GHz - 45,332,021,728 instructions # 2.79 insn per cycle - 5.516250908 seconds time elapsed + 16,247,801,226 cycles # 2.950 GHz + 45,332,171,156 instructions # 2.79 insn per cycle + 5.509791740 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 591) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx/build.none_f_inl0_hrd0/runTest_cpp.exe @@ -119,15 +119,15 @@ Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [ha Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('sse4': SSE4.2, 128bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 4.528380e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 4.868469e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 4.868469e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 4.500324e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 4.833675e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 4.833675e+05 ) sec^-1 MeanMatrixElemValue = ( 2.086779e+00 +- 3.413793e-03 ) GeV^0 -TOTAL : 2.402631 sec +TOTAL : 2.417381 sec INFO: No Floating Point Exceptions have been reported - 7,087,618,340 cycles # 2.944 GHz - 17,790,727,043 instructions # 2.51 insn per cycle - 2.408346877 seconds time elapsed + 7,093,589,030 cycles # 2.929 GHz + 17,792,565,402 instructions # 2.51 insn per cycle + 2.423150920 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 3144) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx/build.sse4_f_inl0_hrd0/runTest_cpp.exe @@ -151,15 +151,15 @@ Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [ha Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('avx2': AVX2, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 8.367783e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 9.536121e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 9.536121e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 8.369797e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 9.529589e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 9.529589e+05 ) sec^-1 MeanMatrixElemValue = ( 2.086810e+00 +- 3.414231e-03 ) GeV^0 -TOTAL : 1.339197 sec +TOTAL : 1.338775 sec INFO: No Floating Point Exceptions have been reported - 3,748,433,186 cycles # 2.789 GHz - 8,262,218,774 instructions # 2.20 insn per cycle - 1.344812605 seconds time elapsed + 3,745,924,449 cycles # 2.788 GHz + 8,261,498,207 instructions # 2.21 insn per cycle + 1.344519433 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 3367) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx/build.avx2_f_inl0_hrd0/runTest_cpp.exe @@ -183,15 +183,15 @@ Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [ha Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('512y': AVX512, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 8.816225e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.011910e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.011910e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 8.780761e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.006934e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.006934e+06 ) sec^-1 MeanMatrixElemValue = ( 2.086810e+00 +- 3.414231e-03 ) GeV^0 -TOTAL : 1.274973 sec +TOTAL : 1.280884 sec INFO: No Floating Point Exceptions have been reported - 3,561,414,995 cycles # 2.782 GHz - 7,912,015,045 instructions # 2.22 insn per cycle - 1.280637958 seconds time elapsed + 3,563,008,513 cycles # 2.771 GHz + 7,911,475,094 instructions # 2.22 insn per cycle + 1.286796749 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 3209) (512y: 20) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx/build.512y_f_inl0_hrd0/runTest_cpp.exe @@ -215,15 +215,15 @@ Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [ha Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[16] ('512z': AVX512, 512bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 6.504790e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 7.157762e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 7.157762e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 6.481023e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 7.194083e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 7.194083e+05 ) sec^-1 MeanMatrixElemValue = ( 2.086810e+00 +- 3.414231e-03 ) GeV^0 -TOTAL : 1.699252 sec +TOTAL : 1.705916 sec INFO: No Floating Point Exceptions have been reported - 3,270,672,138 cycles # 1.919 GHz - 6,095,863,693 instructions # 1.86 insn per cycle - 1.704973507 seconds time elapsed + 3,307,680,225 cycles # 1.934 GHz + 6,098,822,728 instructions # 1.84 insn per cycle + 1.711342648 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 2251) (512y: 22) (512z: 2155) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx/build.512z_f_inl0_hrd0/runTest_cpp.exe diff --git a/epochX/cudacpp/tput/logs_ggtt_mad/log_ggtt_mad_f_inl0_hrd1.txt b/epochX/cudacpp/tput/logs_ggtt_mad/log_ggtt_mad_f_inl0_hrd1.txt index 5e80ecf473..695b4c25da 100644 --- a/epochX/cudacpp/tput/logs_ggtt_mad/log_ggtt_mad_f_inl0_hrd1.txt +++ b/epochX/cudacpp/tput/logs_ggtt_mad/log_ggtt_mad_f_inl0_hrd1.txt @@ -40,7 +40,7 @@ make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx' -DATE: 2024-10-06_08:59:52 +DATE: 2024-10-06_13:29:04 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: ========================================================================= @@ -49,15 +49,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_SM_GG_TTX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=1] Workflow summary = CUD:FLT+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 8.326131e+07 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.746336e+08 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.856838e+08 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 8.940406e+07 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.793424e+08 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.906276e+08 ) sec^-1 MeanMatrixElemValue = ( 2.086719e+00 +- 3.413389e-03 ) GeV^0 -TOTAL : 0.493394 sec +TOTAL : 0.483167 sec INFO: No Floating Point Exceptions have been reported - 2,062,281,894 cycles # 2.861 GHz - 2,938,913,241 instructions # 1.43 insn per cycle - 0.784913836 seconds time elapsed + 2,034,803,981 cycles # 2.871 GHz + 2,916,122,334 instructions # 1.43 insn per cycle + 0.765729955 seconds time elapsed runNcu /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx/build.cuda_f_inl0_hrd1/check_cuda.exe -p 2048 256 1 ==PROF== Profiling "sigmaKin": launch__registers_per_thread 126 ==PROF== Profiling "sigmaKin": sm__sass_average_branch_targets_threads_uniform.pct 100% @@ -85,15 +85,15 @@ Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [ha Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/none+NAVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = SCALAR ('none': ~vector[1], no SIMD) -EvtsPerSec[Rmb+ME] (23) = ( 1.953822e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 2.011638e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 2.011638e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.973078e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 2.030916e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.030916e+05 ) sec^-1 MeanMatrixElemValue = ( 2.086780e+00 +- 3.413794e-03 ) GeV^0 -TOTAL : 5.471139 sec +TOTAL : 5.399943 sec INFO: No Floating Point Exceptions have been reported - 16,020,529,034 cycles # 2.925 GHz - 44,492,038,074 instructions # 2.78 insn per cycle - 5.480388445 seconds time elapsed + 15,930,356,234 cycles # 2.948 GHz + 44,442,633,706 instructions # 2.79 insn per cycle + 5.404884606 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 536) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx/build.none_f_inl0_hrd1/runTest_cpp.exe @@ -117,15 +117,15 @@ Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [ha Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('sse4': SSE4.2, 128bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 5.317220e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 5.788673e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 5.788673e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 5.338927e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 5.809254e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 5.809254e+05 ) sec^-1 MeanMatrixElemValue = ( 2.086779e+00 +- 3.413793e-03 ) GeV^0 -TOTAL : 2.075008 sec +TOTAL : 2.048756 sec INFO: No Floating Point Exceptions have been reported - 6,135,177,420 cycles # 2.947 GHz - 17,131,917,948 instructions # 2.79 insn per cycle - 2.082995277 seconds time elapsed + 6,055,036,918 cycles # 2.949 GHz + 17,080,783,834 instructions # 2.82 insn per cycle + 2.053837921 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 2863) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx/build.sse4_f_inl0_hrd1/runTest_cpp.exe @@ -149,15 +149,15 @@ Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [ha Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('avx2': AVX2, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 6.077036e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 6.672972e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 6.672972e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 6.062011e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 6.643980e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 6.643980e+05 ) sec^-1 MeanMatrixElemValue = ( 2.086810e+00 +- 3.414231e-03 ) GeV^0 -TOTAL : 1.827961 sec +TOTAL : 1.815047 sec INFO: No Floating Point Exceptions have been reported - 5,098,745,585 cycles # 2.778 GHz - 10,277,927,063 instructions # 2.02 insn per cycle - 1.836088116 seconds time elapsed + 5,022,230,249 cycles # 2.762 GHz + 10,226,910,883 instructions # 2.04 insn per cycle + 1.820096481 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 3907) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx/build.avx2_f_inl0_hrd1/runTest_cpp.exe @@ -181,15 +181,15 @@ Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [ha Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('512y': AVX512, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 6.138089e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 6.753320e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 6.753320e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 6.161457e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 6.755109e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 6.755109e+05 ) sec^-1 MeanMatrixElemValue = ( 2.086810e+00 +- 3.414231e-03 ) GeV^0 -TOTAL : 1.811229 sec +TOTAL : 1.786162 sec INFO: No Floating Point Exceptions have been reported - 5,047,478,028 cycles # 2.778 GHz - 10,048,355,032 instructions # 1.99 insn per cycle - 1.819572790 seconds time elapsed + 4,964,461,316 cycles # 2.773 GHz + 9,996,638,840 instructions # 2.01 insn per cycle + 1.791311843 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 3806) (512y: 2) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx/build.512y_f_inl0_hrd1/runTest_cpp.exe @@ -213,15 +213,15 @@ Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [ha Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[16] ('512z': AVX512, 512bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 4.690006e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 5.022722e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 5.022722e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 4.679256e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 5.008473e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 5.008473e+05 ) sec^-1 MeanMatrixElemValue = ( 2.086810e+00 +- 3.414231e-03 ) GeV^0 -TOTAL : 2.339710 sec +TOTAL : 2.325471 sec INFO: No Floating Point Exceptions have been reported - 4,430,484,038 cycles # 1.888 GHz - 8,494,687,635 instructions # 1.92 insn per cycle - 2.347901015 seconds time elapsed + 4,363,514,008 cycles # 1.873 GHz + 8,446,324,538 instructions # 1.94 insn per cycle + 2.330460622 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 2746) (512y: 4) (512z: 2754) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx/build.512z_f_inl0_hrd1/runTest_cpp.exe diff --git a/epochX/cudacpp/tput/logs_ggtt_mad/log_ggtt_mad_f_inl1_hrd0.txt b/epochX/cudacpp/tput/logs_ggtt_mad/log_ggtt_mad_f_inl1_hrd0.txt index 8666f655aa..efb6be0553 100644 --- a/epochX/cudacpp/tput/logs_ggtt_mad/log_ggtt_mad_f_inl1_hrd0.txt +++ b/epochX/cudacpp/tput/logs_ggtt_mad/log_ggtt_mad_f_inl1_hrd0.txt @@ -40,7 +40,7 @@ make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx' -DATE: 2024-10-06_09:29:20 +DATE: 2024-10-06_13:47:51 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: ========================================================================= @@ -49,15 +49,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_SM_GG_TTX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=1] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 8.502979e+07 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.757241e+08 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.878370e+08 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 8.559091e+07 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.763554e+08 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.884434e+08 ) sec^-1 MeanMatrixElemValue = ( 2.086719e+00 +- 3.413389e-03 ) GeV^0 -TOTAL : 0.488684 sec +TOTAL : 0.490618 sec INFO: No Floating Point Exceptions have been reported - 2,072,092,086 cycles # 2.888 GHz - 2,980,809,123 instructions # 1.44 insn per cycle - 0.774128701 seconds time elapsed + 2,057,097,157 cycles # 2.860 GHz + 2,945,304,271 instructions # 1.43 insn per cycle + 0.777399478 seconds time elapsed runNcu /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx/build.cuda_f_inl1_hrd0/check_cuda.exe -p 2048 256 1 ==PROF== Profiling "sigmaKin": launch__registers_per_thread 124 ==PROF== Profiling "sigmaKin": sm__sass_average_branch_targets_threads_uniform.pct 100% @@ -85,15 +85,15 @@ Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=1] [ha Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/none+NAVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = SCALAR ('none': ~vector[1], no SIMD) -EvtsPerSec[Rmb+ME] (23) = ( 2.497944e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 2.591831e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 2.591831e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 2.474404e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 2.565269e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.565269e+05 ) sec^-1 MeanMatrixElemValue = ( 2.086780e+00 +- 3.413794e-03 ) GeV^0 -TOTAL : 4.301034 sec +TOTAL : 4.322994 sec INFO: No Floating Point Exceptions have been reported - 12,652,758,977 cycles # 2.937 GHz - 34,660,886,060 instructions # 2.74 insn per cycle - 4.309086604 seconds time elapsed + 12,582,674,952 cycles # 2.908 GHz + 34,612,368,274 instructions # 2.75 insn per cycle + 4.328495589 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 683) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx/build.none_f_inl1_hrd0/runTest_cpp.exe @@ -117,15 +117,15 @@ Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=1] [ha Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('sse4': SSE4.2, 128bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 5.170038e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 5.622090e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 5.622090e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 5.275656e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 5.735409e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 5.735409e+05 ) sec^-1 MeanMatrixElemValue = ( 2.086779e+00 +- 3.413793e-03 ) GeV^0 -TOTAL : 2.133248 sec +TOTAL : 2.073924 sec INFO: No Floating Point Exceptions have been reported - 6,307,478,134 cycles # 2.947 GHz - 14,873,781,997 instructions # 2.36 insn per cycle - 2.140857047 seconds time elapsed + 6,122,254,676 cycles # 2.945 GHz + 14,813,765,530 instructions # 2.42 insn per cycle + 2.079513663 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 2975) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx/build.sse4_f_inl1_hrd0/runTest_cpp.exe @@ -149,15 +149,15 @@ Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=1] [ha Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('avx2': AVX2, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 7.248492e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 8.104502e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 8.104502e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 7.145291e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 7.968665e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 7.968665e+05 ) sec^-1 MeanMatrixElemValue = ( 2.086810e+00 +- 3.414231e-03 ) GeV^0 -TOTAL : 1.548692 sec +TOTAL : 1.552929 sec INFO: No Floating Point Exceptions have been reported - 4,331,332,767 cycles # 2.784 GHz - 9,119,017,787 instructions # 2.11 insn per cycle - 1.556682967 seconds time elapsed + 4,268,472,137 cycles # 2.740 GHz + 9,070,134,074 instructions # 2.12 insn per cycle + 1.558519696 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 4456) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx/build.avx2_f_inl1_hrd0/runTest_cpp.exe @@ -181,15 +181,15 @@ Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=1] [ha Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('512y': AVX512, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 7.353371e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 8.251881e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 8.251881e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 7.378793e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 8.256613e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 8.256613e+05 ) sec^-1 MeanMatrixElemValue = ( 2.086810e+00 +- 3.414231e-03 ) GeV^0 -TOTAL : 1.529166 sec +TOTAL : 1.507164 sec INFO: No Floating Point Exceptions have been reported - 4,288,032,705 cycles # 2.791 GHz - 8,709,611,506 instructions # 2.03 insn per cycle - 1.537124060 seconds time elapsed + 4,197,224,058 cycles # 2.776 GHz + 8,659,612,617 instructions # 2.06 insn per cycle + 1.512779224 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 4233) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx/build.512y_f_inl1_hrd0/runTest_cpp.exe @@ -213,15 +213,15 @@ Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=1] [ha Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[16] ('512z': AVX512, 512bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 5.411255e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 5.862053e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 5.862053e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 5.420614e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 5.867583e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 5.867583e+05 ) sec^-1 MeanMatrixElemValue = ( 2.086810e+00 +- 3.414231e-03 ) GeV^0 -TOTAL : 2.041395 sec +TOTAL : 2.020943 sec INFO: No Floating Point Exceptions have been reported - 3,904,121,018 cycles # 1.906 GHz - 7,856,412,999 instructions # 2.01 insn per cycle - 2.049301951 seconds time elapsed + 3,841,936,815 cycles # 1.897 GHz + 7,806,407,423 instructions # 2.03 insn per cycle + 2.026447043 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 4273) (512y: 0) (512z: 2558) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx/build.512z_f_inl1_hrd0/runTest_cpp.exe diff --git a/epochX/cudacpp/tput/logs_ggtt_mad/log_ggtt_mad_f_inl1_hrd1.txt b/epochX/cudacpp/tput/logs_ggtt_mad/log_ggtt_mad_f_inl1_hrd1.txt index 74b1cf75ec..e590e6f1ae 100644 --- a/epochX/cudacpp/tput/logs_ggtt_mad/log_ggtt_mad_f_inl1_hrd1.txt +++ b/epochX/cudacpp/tput/logs_ggtt_mad/log_ggtt_mad_f_inl1_hrd1.txt @@ -40,7 +40,7 @@ make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx' -DATE: 2024-10-06_09:29:40 +DATE: 2024-10-06_13:48:11 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: ========================================================================= @@ -49,15 +49,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_SM_GG_TTX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=1] [hardcodePARAM=1] Workflow summary = CUD:FLT+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 8.573239e+07 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.755917e+08 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.881516e+08 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 8.525484e+07 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.756900e+08 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.872161e+08 ) sec^-1 MeanMatrixElemValue = ( 2.086719e+00 +- 3.413389e-03 ) GeV^0 -TOTAL : 0.487451 sec +TOTAL : 0.491682 sec INFO: No Floating Point Exceptions have been reported - 2,067,657,057 cycles # 2.894 GHz - 2,969,147,079 instructions # 1.44 insn per cycle - 0.771604792 seconds time elapsed + 2,060,842,300 cycles # 2.870 GHz + 2,943,515,957 instructions # 1.43 insn per cycle + 0.777326007 seconds time elapsed runNcu /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx/build.cuda_f_inl1_hrd1/check_cuda.exe -p 2048 256 1 ==PROF== Profiling "sigmaKin": launch__registers_per_thread 126 ==PROF== Profiling "sigmaKin": sm__sass_average_branch_targets_threads_uniform.pct 100% @@ -85,15 +85,15 @@ Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=1] [ha Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/none+NAVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = SCALAR ('none': ~vector[1], no SIMD) -EvtsPerSec[Rmb+ME] (23) = ( 2.674902e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 2.781976e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 2.781976e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 2.613601e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 2.714385e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.714385e+05 ) sec^-1 MeanMatrixElemValue = ( 2.086780e+00 +- 3.413794e-03 ) GeV^0 -TOTAL : 4.022349 sec +TOTAL : 4.097369 sec INFO: No Floating Point Exceptions have been reported - 11,884,847,246 cycles # 2.950 GHz - 35,128,022,846 instructions # 2.96 insn per cycle - 4.030241157 seconds time elapsed + 11,821,624,462 cycles # 2.882 GHz + 35,078,217,079 instructions # 2.97 insn per cycle + 4.102919050 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 453) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx/build.none_f_inl1_hrd1/runTest_cpp.exe @@ -117,15 +117,15 @@ Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=1] [ha Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('sse4': SSE4.2, 128bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 5.473588e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 5.982990e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 5.982990e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 5.483240e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 5.982428e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 5.982428e+05 ) sec^-1 MeanMatrixElemValue = ( 2.086779e+00 +- 3.413793e-03 ) GeV^0 -TOTAL : 2.018275 sec +TOTAL : 1.998177 sec INFO: No Floating Point Exceptions have been reported - 5,977,087,994 cycles # 2.951 GHz - 14,582,659,278 instructions # 2.44 insn per cycle - 2.026172081 seconds time elapsed + 5,918,363,329 cycles # 2.955 GHz + 14,533,402,874 instructions # 2.46 insn per cycle + 2.003607094 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 2569) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx/build.sse4_f_inl1_hrd1/runTest_cpp.exe @@ -149,15 +149,15 @@ Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=1] [ha Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('avx2': AVX2, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 7.377553e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 8.279187e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 8.279187e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 7.416173e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 8.298448e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 8.298448e+05 ) sec^-1 MeanMatrixElemValue = ( 2.086810e+00 +- 3.414231e-03 ) GeV^0 -TOTAL : 1.524553 sec +TOTAL : 1.499537 sec INFO: No Floating Point Exceptions have been reported - 4,234,763,555 cycles # 2.764 GHz - 8,897,798,804 instructions # 2.10 insn per cycle - 1.532761317 seconds time elapsed + 4,161,931,800 cycles # 2.767 GHz + 8,847,825,596 instructions # 2.13 insn per cycle + 1.504922803 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 3552) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx/build.avx2_f_inl1_hrd1/runTest_cpp.exe @@ -181,15 +181,15 @@ Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=1] [ha Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('512y': AVX512, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 7.495273e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 8.420338e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 8.420338e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 7.496792e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 8.392974e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 8.392974e+05 ) sec^-1 MeanMatrixElemValue = ( 2.086810e+00 +- 3.414231e-03 ) GeV^0 -TOTAL : 1.502506 sec +TOTAL : 1.483974 sec INFO: No Floating Point Exceptions have been reported - 4,214,392,060 cycles # 2.792 GHz - 8,461,762,117 instructions # 2.01 insn per cycle - 1.510417354 seconds time elapsed + 4,137,375,501 cycles # 2.779 GHz + 8,411,314,354 instructions # 2.03 insn per cycle + 1.489567560 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 3296) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx/build.512y_f_inl1_hrd1/runTest_cpp.exe @@ -213,15 +213,15 @@ Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=1] [ha Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[16] ('512z': AVX512, 512bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 5.487070e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 5.949626e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 5.949626e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 5.473494e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 5.932901e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 5.932901e+05 ) sec^-1 MeanMatrixElemValue = ( 2.086810e+00 +- 3.414231e-03 ) GeV^0 -TOTAL : 2.014420 sec +TOTAL : 2.001653 sec INFO: No Floating Point Exceptions have been reported - 3,856,759,695 cycles # 1.908 GHz - 7,749,847,516 instructions # 2.01 insn per cycle - 2.022398856 seconds time elapsed + 3,784,116,903 cycles # 1.886 GHz + 7,699,555,675 instructions # 2.03 insn per cycle + 2.007145599 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 3289) (512y: 0) (512z: 2110) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx/build.512z_f_inl1_hrd1/runTest_cpp.exe diff --git a/epochX/cudacpp/tput/logs_ggtt_mad/log_ggtt_mad_m_inl0_hrd0.txt b/epochX/cudacpp/tput/logs_ggtt_mad/log_ggtt_mad_m_inl0_hrd0.txt index 46bc87b45e..ed9350aa1f 100644 --- a/epochX/cudacpp/tput/logs_ggtt_mad/log_ggtt_mad_m_inl0_hrd0.txt +++ b/epochX/cudacpp/tput/logs_ggtt_mad/log_ggtt_mad_m_inl0_hrd0.txt @@ -40,7 +40,7 @@ make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx' -DATE: 2024-10-06_08:58:41 +DATE: 2024-10-06_13:27:53 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: ========================================================================= @@ -49,15 +49,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_SM_GG_TTX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:MIX+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 4.456560e+07 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 9.379988e+07 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.000705e+08 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 4.644823e+07 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 9.479488e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.010556e+08 ) sec^-1 MeanMatrixElemValue = ( 2.086689e+00 +- 3.413217e-03 ) GeV^0 -TOTAL : 0.534501 sec +TOTAL : 0.524977 sec INFO: No Floating Point Exceptions have been reported - 2,219,584,721 cycles # 2.878 GHz - 3,138,987,562 instructions # 1.41 insn per cycle - 0.829330920 seconds time elapsed + 2,183,989,947 cycles # 2.885 GHz + 3,174,833,913 instructions # 1.45 insn per cycle + 0.815368524 seconds time elapsed runNcu /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx/build.cuda_m_inl0_hrd0/check_cuda.exe -p 2048 256 1 ==PROF== Profiling "sigmaKin": launch__registers_per_thread 214 ==PROF== Profiling "sigmaKin": sm__sass_average_branch_targets_threads_uniform.pct 100% @@ -85,15 +85,15 @@ Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [ha Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/none+NAVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) Internal loops fptype_sv = SCALAR ('none': ~vector[1], no SIMD) -EvtsPerSec[Rmb+ME] (23) = ( 1.813220e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.859845e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.859845e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.812292e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.858591e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.858591e+05 ) sec^-1 MeanMatrixElemValue = ( 2.086689e+00 +- 3.413217e-03 ) GeV^0 -TOTAL : 5.928691 sec +TOTAL : 5.893688 sec INFO: No Floating Point Exceptions have been reported - 17,514,594,449 cycles # 2.949 GHz - 46,201,641,620 instructions # 2.64 insn per cycle - 5.940965337 seconds time elapsed + 17,362,803,402 cycles # 2.944 GHz + 46,084,017,452 instructions # 2.65 insn per cycle + 5.898903693 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 617) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx/build.none_m_inl0_hrd0/runTest_cpp.exe @@ -117,15 +117,15 @@ Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [ha Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[2] ('sse4': SSE4.2, 128bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 3.229159e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.395479e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.395479e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 3.235473e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.396927e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.396927e+05 ) sec^-1 MeanMatrixElemValue = ( 2.086689e+00 +- 3.413217e-03 ) GeV^0 -TOTAL : 3.394540 sec +TOTAL : 3.347757 sec INFO: No Floating Point Exceptions have been reported - 10,052,901,757 cycles # 2.953 GHz - 27,702,324,481 instructions # 2.76 insn per cycle - 3.406321535 seconds time elapsed + 9,897,975,136 cycles # 2.953 GHz + 27,584,117,693 instructions # 2.79 insn per cycle + 3.353003266 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 2581) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx/build.sse4_m_inl0_hrd0/runTest_cpp.exe @@ -149,15 +149,15 @@ Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [ha Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('avx2': AVX2, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 5.062332e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 5.465524e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 5.465524e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 5.098160e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 5.496708e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 5.496708e+05 ) sec^-1 MeanMatrixElemValue = ( 2.086689e+00 +- 3.413217e-03 ) GeV^0 -TOTAL : 2.217135 sec +TOTAL : 2.163523 sec INFO: No Floating Point Exceptions have been reported - 6,171,509,914 cycles # 2.770 GHz - 12,603,170,569 instructions # 2.04 insn per cycle - 2.229995554 seconds time elapsed + 6,013,212,358 cycles # 2.774 GHz + 12,483,654,309 instructions # 2.08 insn per cycle + 2.168711908 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 2773) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx/build.avx2_m_inl0_hrd0/runTest_cpp.exe @@ -181,15 +181,15 @@ Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [ha Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('512y': AVX512, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 5.580384e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 6.068896e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 6.068896e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 5.611454e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 6.088949e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 6.088949e+05 ) sec^-1 MeanMatrixElemValue = ( 2.086689e+00 +- 3.413217e-03 ) GeV^0 -TOTAL : 2.025421 sec +TOTAL : 1.974779 sec INFO: No Floating Point Exceptions have been reported - 5,651,741,681 cycles # 2.776 GHz - 12,038,443,177 instructions # 2.13 insn per cycle - 2.038138408 seconds time elapsed + 5,503,300,837 cycles # 2.781 GHz + 11,920,665,852 instructions # 2.17 insn per cycle + 1.979903882 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 2518) (512y: 146) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx/build.512y_m_inl0_hrd0/runTest_cpp.exe @@ -213,15 +213,15 @@ Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [ha Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('512z': AVX512, 512bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 3.630973e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.831034e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.831034e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 3.612045e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.807094e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.807094e+05 ) sec^-1 MeanMatrixElemValue = ( 2.086689e+00 +- 3.413217e-03 ) GeV^0 -TOTAL : 3.034632 sec +TOTAL : 3.009558 sec INFO: No Floating Point Exceptions have been reported - 5,740,712,408 cycles # 1.885 GHz - 8,225,599,297 instructions # 1.43 insn per cycle - 3.047056631 seconds time elapsed + 5,590,700,075 cycles # 1.855 GHz + 8,108,733,626 instructions # 1.45 insn per cycle + 3.014813962 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 1671) (512y: 126) (512z: 1862) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx/build.512z_m_inl0_hrd0/runTest_cpp.exe diff --git a/epochX/cudacpp/tput/logs_ggtt_mad/log_ggtt_mad_m_inl0_hrd1.txt b/epochX/cudacpp/tput/logs_ggtt_mad/log_ggtt_mad_m_inl0_hrd1.txt index ffa5410982..411a407c80 100644 --- a/epochX/cudacpp/tput/logs_ggtt_mad/log_ggtt_mad_m_inl0_hrd1.txt +++ b/epochX/cudacpp/tput/logs_ggtt_mad/log_ggtt_mad_m_inl0_hrd1.txt @@ -40,7 +40,7 @@ make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx' -DATE: 2024-10-06_08:59:06 +DATE: 2024-10-06_13:28:18 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: ========================================================================= @@ -49,15 +49,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_SM_GG_TTX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=1] Workflow summary = CUD:MIX+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 4.422071e+07 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 9.351796e+07 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 9.985674e+07 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 4.635146e+07 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 9.428698e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.004494e+08 ) sec^-1 MeanMatrixElemValue = ( 2.086689e+00 +- 3.413217e-03 ) GeV^0 -TOTAL : 0.535641 sec +TOTAL : 0.522197 sec INFO: No Floating Point Exceptions have been reported - 2,214,747,611 cycles # 2.879 GHz - 3,172,033,471 instructions # 1.43 insn per cycle - 0.829540839 seconds time elapsed + 2,179,732,075 cycles # 2.880 GHz + 3,103,973,177 instructions # 1.42 insn per cycle + 0.813496399 seconds time elapsed runNcu /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx/build.cuda_m_inl0_hrd1/check_cuda.exe -p 2048 256 1 ==PROF== Profiling "sigmaKin": launch__registers_per_thread 212 ==PROF== Profiling "sigmaKin": sm__sass_average_branch_targets_threads_uniform.pct 100% @@ -85,15 +85,15 @@ Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [ha Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/none+NAVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) Internal loops fptype_sv = SCALAR ('none': ~vector[1], no SIMD) -EvtsPerSec[Rmb+ME] (23) = ( 1.862163e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.911340e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.911340e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.856541e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.904486e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.904486e+05 ) sec^-1 MeanMatrixElemValue = ( 2.086689e+00 +- 3.413217e-03 ) GeV^0 -TOTAL : 5.777703 sec +TOTAL : 5.753609 sec INFO: No Floating Point Exceptions have been reported - 17,097,861,095 cycles # 2.954 GHz - 45,230,787,591 instructions # 2.65 insn per cycle - 5.789414615 seconds time elapsed + 16,914,759,829 cycles # 2.938 GHz + 45,109,201,326 instructions # 2.67 insn per cycle + 5.758744777 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 568) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx/build.none_m_inl0_hrd1/runTest_cpp.exe @@ -117,15 +117,15 @@ Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [ha Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[2] ('sse4': SSE4.2, 128bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 3.356972e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.536408e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.536408e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 3.354543e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.530510e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.530510e+05 ) sec^-1 MeanMatrixElemValue = ( 2.086689e+00 +- 3.413217e-03 ) GeV^0 -TOTAL : 3.270231 sec +TOTAL : 3.233379 sec INFO: No Floating Point Exceptions have been reported - 9,665,855,757 cycles # 2.946 GHz - 26,370,377,514 instructions # 2.73 insn per cycle - 3.281726897 seconds time elapsed + 9,513,279,080 cycles # 2.938 GHz + 26,252,275,360 instructions # 2.76 insn per cycle + 3.238616994 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 2386) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx/build.sse4_m_inl0_hrd1/runTest_cpp.exe @@ -149,15 +149,15 @@ Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [ha Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('avx2': AVX2, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 4.515319e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 4.832036e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 4.832036e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 4.513969e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 4.826300e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 4.826300e+05 ) sec^-1 MeanMatrixElemValue = ( 2.086689e+00 +- 3.413217e-03 ) GeV^0 -TOTAL : 2.470441 sec +TOTAL : 2.429574 sec INFO: No Floating Point Exceptions have been reported - 6,884,599,220 cycles # 2.774 GHz - 14,150,233,239 instructions # 2.06 insn per cycle - 2.482504065 seconds time elapsed + 6,733,073,158 cycles # 2.767 GHz + 14,032,458,743 instructions # 2.08 insn per cycle + 2.434764702 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 2896) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx/build.avx2_m_inl0_hrd1/runTest_cpp.exe @@ -181,15 +181,15 @@ Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [ha Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('512y': AVX512, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 4.744762e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 5.096792e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 5.096792e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 4.752950e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 5.094299e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 5.094299e+05 ) sec^-1 MeanMatrixElemValue = ( 2.086689e+00 +- 3.413217e-03 ) GeV^0 -TOTAL : 2.356796 sec +TOTAL : 2.312119 sec INFO: No Floating Point Exceptions have been reported - 6,551,408,744 cycles # 2.767 GHz - 13,642,717,150 instructions # 2.08 insn per cycle - 2.368190066 seconds time elapsed + 6,402,182,386 cycles # 2.764 GHz + 13,523,162,780 instructions # 2.11 insn per cycle + 2.317403276 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 2535) (512y: 302) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx/build.512y_m_inl0_hrd1/runTest_cpp.exe @@ -213,15 +213,15 @@ Process = SIGMA_SM_GG_TTX_CPP [gcc 11.3.1] [inlineHel=0] [ha Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('512z': AVX512, 512bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 3.568399e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.763148e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.763148e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 3.623021e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.817335e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.817335e+05 ) sec^-1 MeanMatrixElemValue = ( 2.086689e+00 +- 3.413217e-03 ) GeV^0 -TOTAL : 3.086761 sec +TOTAL : 3.000443 sec INFO: No Floating Point Exceptions have been reported - 5,741,113,391 cycles # 1.854 GHz - 9,326,512,235 instructions # 1.62 insn per cycle - 3.098253222 seconds time elapsed + 5,574,514,801 cycles # 1.855 GHz + 9,207,921,365 instructions # 1.65 insn per cycle + 3.005820952 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 1456) (512y: 212) (512z: 2060) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_tt.mad/SubProcesses/P1_gg_ttx/build.512z_m_inl0_hrd1/runTest_cpp.exe diff --git a/epochX/cudacpp/tput/logs_ggttg_mad/log_ggttg_mad_d_inl0_hrd0.txt b/epochX/cudacpp/tput/logs_ggttg_mad/log_ggttg_mad_d_inl0_hrd0.txt index 028292e268..b449231e68 100644 --- a/epochX/cudacpp/tput/logs_ggttg_mad/log_ggttg_mad_d_inl0_hrd0.txt +++ b/epochX/cudacpp/tput/logs_ggttg_mad/log_ggttg_mad_d_inl0_hrd0.txt @@ -40,7 +40,7 @@ make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttg.mad/SubProcesses/P1_gg_ttxg' -DATE: 2024-10-06_09:00:14 +DATE: 2024-10-06_13:29:26 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: ========================================================================= @@ -49,15 +49,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_SM_GG_TTXG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 7.612194e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 8.849217e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 8.964394e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 7.886994e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 8.999184e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 9.091869e+06 ) sec^-1 MeanMatrixElemValue = ( 1.008920e+02 +- 5.001681e+01 ) GeV^-2 -TOTAL : 0.477846 sec +TOTAL : 0.470526 sec INFO: No Floating Point Exceptions have been reported - 1,998,983,760 cycles # 2.871 GHz - 2,812,176,587 instructions # 1.41 insn per cycle - 0.759674168 seconds time elapsed + 1,980,771,638 cycles # 2.881 GHz + 2,862,394,539 instructions # 1.45 insn per cycle + 0.744667505 seconds time elapsed runNcu /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttg.mad/SubProcesses/P1_gg_ttxg/build.cuda_d_inl0_hrd0/check_cuda.exe -p 64 256 1 ==PROF== Profiling "sigmaKin": launch__registers_per_thread 255 ==PROF== Profiling "sigmaKin": sm__sass_average_branch_targets_threads_uniform.pct 100% @@ -67,15 +67,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_SM_GG_TTXG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 1.042987e+07 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.232338e+07 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.242858e+07 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.054055e+07 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.229940e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.239975e+07 ) sec^-1 MeanMatrixElemValue = ( 6.734461e+02 +- 4.775415e+02 ) GeV^-2 -TOTAL : 0.621402 sec +TOTAL : 0.608373 sec INFO: No Floating Point Exceptions have been reported - 2,510,286,495 cycles # 2.883 GHz - 3,752,986,245 instructions # 1.50 insn per cycle - 0.931747637 seconds time elapsed + 2,433,856,904 cycles # 2.874 GHz + 3,720,631,061 instructions # 1.53 insn per cycle + 0.905376658 seconds time elapsed ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttg.mad/SubProcesses/P1_gg_ttxg/build.cuda_d_inl0_hrd0/runTest_cuda.exe INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: FE_DIVBYZERO, FE_INVALID, FE_OVERFLOW @@ -100,15 +100,15 @@ Process = SIGMA_SM_GG_TTXG_CPP [gcc 11.3.1] [inlineHel=0] [h Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/none+NAVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = SCALAR ('none': ~vector[1], no SIMD) -EvtsPerSec[Rmb+ME] (23) = ( 2.434605e+04 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 2.446812e+04 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 2.446812e+04 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 2.427372e+04 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 2.439351e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.439351e+04 ) sec^-1 MeanMatrixElemValue = ( 1.008920e+02 +- 5.001681e+01 ) GeV^-2 -TOTAL : 6.752117 sec +TOTAL : 6.770074 sec INFO: No Floating Point Exceptions have been reported - 19,916,103,310 cycles # 2.949 GHz - 59,916,518,373 instructions # 3.01 insn per cycle - 6.756066066 seconds time elapsed + 19,921,073,411 cycles # 2.941 GHz + 59,914,505,967 instructions # 3.01 insn per cycle + 6.774216977 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 1199) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttg.mad/SubProcesses/P1_gg_ttxg/build.none_d_inl0_hrd0/runTest_cpp.exe @@ -132,15 +132,15 @@ Process = SIGMA_SM_GG_TTXG_CPP [gcc 11.3.1] [inlineHel=0] [h Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[2] ('sse4': SSE4.2, 128bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 4.568526e+04 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 4.611480e+04 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 4.611480e+04 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 4.601586e+04 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 4.643511e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 4.643511e+04 ) sec^-1 MeanMatrixElemValue = ( 1.008920e+02 +- 5.001681e+01 ) GeV^-2 -TOTAL : 3.606956 sec +TOTAL : 3.579972 sec INFO: No Floating Point Exceptions have been reported - 10,571,212,167 cycles # 2.928 GHz - 31,086,653,440 instructions # 2.94 insn per cycle - 3.611892241 seconds time elapsed + 10,559,948,820 cycles # 2.947 GHz + 31,084,717,678 instructions # 2.94 insn per cycle + 3.584214703 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 5221) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttg.mad/SubProcesses/P1_gg_ttxg/build.sse4_d_inl0_hrd0/runTest_cpp.exe @@ -164,15 +164,15 @@ Process = SIGMA_SM_GG_TTXG_CPP [gcc 11.3.1] [inlineHel=0] [h Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('avx2': AVX2, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 9.091675e+04 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 9.256165e+04 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 9.256165e+04 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 9.136334e+04 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 9.300599e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 9.300599e+04 ) sec^-1 MeanMatrixElemValue = ( 1.008920e+02 +- 5.001681e+01 ) GeV^-2 -TOTAL : 1.823998 sec +TOTAL : 1.812838 sec INFO: No Floating Point Exceptions have been reported - 4,999,238,647 cycles # 2.738 GHz - 11,406,827,724 instructions # 2.28 insn per cycle - 1.827985092 seconds time elapsed + 4,995,392,395 cycles # 2.750 GHz + 11,405,360,468 instructions # 2.28 insn per cycle + 1.817036394 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 4642) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttg.mad/SubProcesses/P1_gg_ttxg/build.avx2_d_inl0_hrd0/runTest_cpp.exe @@ -196,15 +196,15 @@ Process = SIGMA_SM_GG_TTXG_CPP [gcc 11.3.1] [inlineHel=0] [h Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('512y': AVX512, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 1.026950e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.047965e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.047965e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.025674e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.046195e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.046195e+05 ) sec^-1 MeanMatrixElemValue = ( 1.008920e+02 +- 5.001681e+01 ) GeV^-2 -TOTAL : 1.617207 sec +TOTAL : 1.618204 sec INFO: No Floating Point Exceptions have been reported - 4,447,500,259 cycles # 2.747 GHz - 10,665,398,274 instructions # 2.40 insn per cycle - 1.621167175 seconds time elapsed + 4,442,976,773 cycles # 2.741 GHz + 10,664,510,521 instructions # 2.40 insn per cycle + 1.622470428 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 4378) (512y: 92) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttg.mad/SubProcesses/P1_gg_ttxg/build.512y_d_inl0_hrd0/runTest_cpp.exe @@ -228,15 +228,15 @@ Process = SIGMA_SM_GG_TTXG_CPP [gcc 11.3.1] [inlineHel=0] [h Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('512z': AVX512, 512bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 7.168386e+04 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 7.273905e+04 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 7.273905e+04 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 7.094451e+04 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 7.194460e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 7.194460e+04 ) sec^-1 MeanMatrixElemValue = ( 1.008920e+02 +- 5.001681e+01 ) GeV^-2 -TOTAL : 2.309115 sec +TOTAL : 2.331405 sec INFO: No Floating Point Exceptions have been reported - 4,128,751,307 cycles # 1.785 GHz - 5,972,449,468 instructions # 1.45 insn per cycle - 2.314144205 seconds time elapsed + 4,125,968,713 cycles # 1.767 GHz + 5,970,752,171 instructions # 1.45 insn per cycle + 2.335699486 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 1620) (512y: 94) (512z: 3577) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttg.mad/SubProcesses/P1_gg_ttxg/build.512z_d_inl0_hrd0/runTest_cpp.exe diff --git a/epochX/cudacpp/tput/logs_ggttg_mad/log_ggttg_mad_d_inl0_hrd0_bridge.txt b/epochX/cudacpp/tput/logs_ggttg_mad/log_ggttg_mad_d_inl0_hrd0_bridge.txt index 76636470b0..098859b1f9 100644 --- a/epochX/cudacpp/tput/logs_ggttg_mad/log_ggttg_mad_d_inl0_hrd0_bridge.txt +++ b/epochX/cudacpp/tput/logs_ggttg_mad/log_ggttg_mad_d_inl0_hrd0_bridge.txt @@ -40,7 +40,7 @@ make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttg.mad/SubProcesses/P1_gg_ttxg' -DATE: 2024-10-06_09:38:23 +DATE: 2024-10-06_13:56:56 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: ========================================================================= @@ -53,16 +53,16 @@ WARNING! Set grid in Bridge (nevt=16384, gpublocks=64, gputhreads=256, gpublocks Process = SIGMA_SM_GG_TTXG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 2.472313e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 6.180220e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 6.180220e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 2.487613e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 6.137406e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 6.137406e+06 ) sec^-1 MeanMatrixElemValue = ( 1.008920e+02 +- 5.001681e+01 ) GeV^-2 -TOTAL : 0.504857 sec +TOTAL : 0.502957 sec INFO: No Floating Point Exceptions have been reported INFO: No Floating Point Exceptions have been reported - 2,064,539,289 cycles # 2.862 GHz - 3,123,566,672 instructions # 1.51 insn per cycle - 0.778239097 seconds time elapsed + 2,066,621,912 cycles # 2.874 GHz + 3,131,649,457 instructions # 1.52 insn per cycle + 0.775712566 seconds time elapsed runNcu /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttg.mad/SubProcesses/P1_gg_ttxg/build.cuda_d_inl0_hrd0/check_cuda.exe -p 64 256 1 --bridge WARNING! Bridge selected: cannot use RamboDevice, will use RamboHost WARNING! RamboHost selected: cannot use CurandDevice, will use CurandHost @@ -80,16 +80,16 @@ WARNING! Set grid in Bridge (nevt=524288, gpublocks=2048, gputhreads=256, gpublo Process = SIGMA_SM_GG_TTXG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 2.683325e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 9.341961e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 9.341961e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 2.633276e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 9.328654e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 9.328654e+06 ) sec^-1 MeanMatrixElemValue = ( 6.734461e+02 +- 4.775415e+02 ) GeV^-2 -TOTAL : 0.833212 sec +TOTAL : 0.840105 sec INFO: No Floating Point Exceptions have been reported INFO: No Floating Point Exceptions have been reported - 3,141,452,514 cycles # 2.889 GHz - 4,965,295,428 instructions # 1.58 insn per cycle - 1.145190233 seconds time elapsed + 3,146,941,649 cycles # 2.877 GHz + 5,028,600,902 instructions # 1.60 insn per cycle + 1.155345164 seconds time elapsed ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttg.mad/SubProcesses/P1_gg_ttxg/build.cuda_d_inl0_hrd0/runTest_cuda.exe INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: FE_DIVBYZERO, FE_INVALID, FE_OVERFLOW @@ -115,16 +115,16 @@ Process = SIGMA_SM_GG_TTXG_CPP [gcc 11.3.1] [inlineHel=0] [h Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+BRDHST/none+NAVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = SCALAR ('none': ~vector[1], no SIMD) -EvtsPerSec[Rmb+ME] (23) = ( 2.439308e+04 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 2.451643e+04 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 2.451643e+04 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 2.438401e+04 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 2.450574e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.450574e+04 ) sec^-1 MeanMatrixElemValue = ( 1.008920e+02 +- 5.001681e+01 ) GeV^-2 -TOTAL : 6.745227 sec +TOTAL : 6.748082 sec INFO: No Floating Point Exceptions have been reported INFO: No Floating Point Exceptions have been reported - 19,922,326,116 cycles # 2.952 GHz - 59,921,657,661 instructions # 3.01 insn per cycle - 6.749767217 seconds time elapsed + 19,942,067,521 cycles # 2.954 GHz + 59,921,466,571 instructions # 3.00 insn per cycle + 6.752688401 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 1199) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttg.mad/SubProcesses/P1_gg_ttxg/build.none_d_inl0_hrd0/runTest_cpp.exe @@ -149,16 +149,16 @@ Process = SIGMA_SM_GG_TTXG_CPP [gcc 11.3.1] [inlineHel=0] [h Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+BRDHST/sse4+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[2] ('sse4': SSE4.2, 128bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 4.590762e+04 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 4.634359e+04 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 4.634359e+04 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 4.606355e+04 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 4.649624e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 4.649624e+04 ) sec^-1 MeanMatrixElemValue = ( 1.008920e+02 +- 5.001681e+01 ) GeV^-2 -TOTAL : 3.596308 sec +TOTAL : 3.585448 sec INFO: No Floating Point Exceptions have been reported INFO: No Floating Point Exceptions have been reported - 10,606,556,243 cycles # 2.946 GHz - 31,132,640,347 instructions # 2.94 insn per cycle - 3.600784290 seconds time elapsed + 10,611,470,097 cycles # 2.957 GHz + 31,134,511,981 instructions # 2.93 insn per cycle + 3.590096203 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 5221) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttg.mad/SubProcesses/P1_gg_ttxg/build.sse4_d_inl0_hrd0/runTest_cpp.exe @@ -183,16 +183,16 @@ Process = SIGMA_SM_GG_TTXG_CPP [gcc 11.3.1] [inlineHel=0] [h Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+BRDHST/avx2+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('avx2': AVX2, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 9.045361e+04 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 9.212711e+04 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 9.212711e+04 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 9.105176e+04 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 9.273678e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 9.273678e+04 ) sec^-1 MeanMatrixElemValue = ( 1.008920e+02 +- 5.001681e+01 ) GeV^-2 -TOTAL : 1.840181 sec +TOTAL : 1.827491 sec INFO: No Floating Point Exceptions have been reported INFO: No Floating Point Exceptions have been reported - 5,054,435,549 cycles # 2.741 GHz - 11,457,891,523 instructions # 2.27 insn per cycle - 1.844724432 seconds time elapsed + 5,037,454,984 cycles # 2.751 GHz + 11,454,279,434 instructions # 2.27 insn per cycle + 1.831952090 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 4642) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttg.mad/SubProcesses/P1_gg_ttxg/build.avx2_d_inl0_hrd0/runTest_cpp.exe @@ -217,16 +217,16 @@ Process = SIGMA_SM_GG_TTXG_CPP [gcc 11.3.1] [inlineHel=0] [h Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+BRDHST/512y+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('512y': AVX512, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 1.028589e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.049854e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.049854e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.026081e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.047095e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.047095e+05 ) sec^-1 MeanMatrixElemValue = ( 1.008920e+02 +- 5.001681e+01 ) GeV^-2 -TOTAL : 1.621206 sec +TOTAL : 1.624470 sec INFO: No Floating Point Exceptions have been reported INFO: No Floating Point Exceptions have been reported - 4,484,828,096 cycles # 2.760 GHz - 10,715,944,638 instructions # 2.39 insn per cycle - 1.625802151 seconds time elapsed + 4,482,342,799 cycles # 2.753 GHz + 10,712,617,570 instructions # 2.39 insn per cycle + 1.628968333 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 4378) (512y: 92) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttg.mad/SubProcesses/P1_gg_ttxg/build.512y_d_inl0_hrd0/runTest_cpp.exe @@ -251,16 +251,16 @@ Process = SIGMA_SM_GG_TTXG_CPP [gcc 11.3.1] [inlineHel=0] [h Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+BRDHST/512z+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('512z': AVX512, 512bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 7.165257e+04 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 7.268564e+04 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 7.268564e+04 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 7.069120e+04 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 7.172509e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 7.172509e+04 ) sec^-1 MeanMatrixElemValue = ( 1.008920e+02 +- 5.001681e+01 ) GeV^-2 -TOTAL : 2.316443 sec +TOTAL : 2.348637 sec INFO: No Floating Point Exceptions have been reported INFO: No Floating Point Exceptions have been reported - 4,162,925,866 cycles # 1.795 GHz - 6,008,954,577 instructions # 1.44 insn per cycle - 2.321140123 seconds time elapsed + 4,170,875,646 cycles # 1.773 GHz + 6,010,624,907 instructions # 1.44 insn per cycle + 2.353230021 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 1620) (512y: 94) (512z: 3577) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttg.mad/SubProcesses/P1_gg_ttxg/build.512z_d_inl0_hrd0/runTest_cpp.exe diff --git a/epochX/cudacpp/tput/logs_ggttg_mad/log_ggttg_mad_d_inl0_hrd1.txt b/epochX/cudacpp/tput/logs_ggttg_mad/log_ggttg_mad_d_inl0_hrd1.txt index 49402063e2..dd9c5510f4 100644 --- a/epochX/cudacpp/tput/logs_ggttg_mad/log_ggttg_mad_d_inl0_hrd1.txt +++ b/epochX/cudacpp/tput/logs_ggttg_mad/log_ggttg_mad_d_inl0_hrd1.txt @@ -40,7 +40,7 @@ make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttg.mad/SubProcesses/P1_gg_ttxg' -DATE: 2024-10-06_09:00:40 +DATE: 2024-10-06_13:29:51 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: ========================================================================= @@ -49,15 +49,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_SM_GG_TTXG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=1] Workflow summary = CUD:DBL+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 7.575064e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 8.921304e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 9.028957e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 7.901172e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 9.020040e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 9.114700e+06 ) sec^-1 MeanMatrixElemValue = ( 1.008920e+02 +- 5.001681e+01 ) GeV^-2 -TOTAL : 0.477703 sec +TOTAL : 0.472677 sec INFO: No Floating Point Exceptions have been reported - 1,994,590,518 cycles # 2.865 GHz - 2,848,992,929 instructions # 1.43 insn per cycle - 0.754407053 seconds time elapsed + 1,969,533,191 cycles # 2.859 GHz + 2,819,300,861 instructions # 1.43 insn per cycle + 0.746665217 seconds time elapsed runNcu /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttg.mad/SubProcesses/P1_gg_ttxg/build.cuda_d_inl0_hrd1/check_cuda.exe -p 64 256 1 ==PROF== Profiling "sigmaKin": launch__registers_per_thread 255 ==PROF== Profiling "sigmaKin": sm__sass_average_branch_targets_threads_uniform.pct 100% @@ -67,15 +67,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_SM_GG_TTXG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=1] Workflow summary = CUD:DBL+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 1.042325e+07 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.231825e+07 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.242712e+07 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.056058e+07 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.232796e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.242742e+07 ) sec^-1 MeanMatrixElemValue = ( 6.734461e+02 +- 4.775415e+02 ) GeV^-2 -TOTAL : 0.616679 sec +TOTAL : 0.606629 sec INFO: No Floating Point Exceptions have been reported - 2,463,746,118 cycles # 2.874 GHz - 3,716,874,386 instructions # 1.51 insn per cycle - 0.917442132 seconds time elapsed + 2,428,725,444 cycles # 2.874 GHz + 3,744,070,416 instructions # 1.54 insn per cycle + 0.903910002 seconds time elapsed ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttg.mad/SubProcesses/P1_gg_ttxg/build.cuda_d_inl0_hrd1/runTest_cuda.exe INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: FE_DIVBYZERO, FE_INVALID, FE_OVERFLOW @@ -100,15 +100,15 @@ Process = SIGMA_SM_GG_TTXG_CPP [gcc 11.3.1] [inlineHel=0] [h Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/none+NAVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = SCALAR ('none': ~vector[1], no SIMD) -EvtsPerSec[Rmb+ME] (23) = ( 2.437110e+04 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 2.449363e+04 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 2.449363e+04 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 2.424768e+04 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 2.436840e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.436840e+04 ) sec^-1 MeanMatrixElemValue = ( 1.008920e+02 +- 5.001681e+01 ) GeV^-2 -TOTAL : 6.744189 sec +TOTAL : 6.777440 sec INFO: No Floating Point Exceptions have been reported - 19,899,963,729 cycles # 2.950 GHz - 60,130,622,589 instructions # 3.02 insn per cycle - 6.748077481 seconds time elapsed + 19,901,719,688 cycles # 2.936 GHz + 60,132,173,903 instructions # 3.02 insn per cycle + 6.781450785 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 1322) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttg.mad/SubProcesses/P1_gg_ttxg/build.none_d_inl0_hrd1/runTest_cpp.exe @@ -132,15 +132,15 @@ Process = SIGMA_SM_GG_TTXG_CPP [gcc 11.3.1] [inlineHel=0] [h Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[2] ('sse4': SSE4.2, 128bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 4.632122e+04 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 4.676125e+04 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 4.676125e+04 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 4.637476e+04 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 4.680658e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 4.680658e+04 ) sec^-1 MeanMatrixElemValue = ( 1.008920e+02 +- 5.001681e+01 ) GeV^-2 -TOTAL : 3.557509 sec +TOTAL : 3.552815 sec INFO: No Floating Point Exceptions have been reported - 10,482,296,489 cycles # 2.944 GHz - 30,686,942,862 instructions # 2.93 insn per cycle - 3.561419011 seconds time elapsed + 10,466,773,742 cycles # 2.943 GHz + 30,685,309,780 instructions # 2.93 insn per cycle + 3.556946287 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 5047) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttg.mad/SubProcesses/P1_gg_ttxg/build.sse4_d_inl0_hrd1/runTest_cpp.exe @@ -164,15 +164,15 @@ Process = SIGMA_SM_GG_TTXG_CPP [gcc 11.3.1] [inlineHel=0] [h Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('avx2': AVX2, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 8.842314e+04 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 8.999775e+04 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 8.999775e+04 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 8.843947e+04 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 8.997853e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 8.997853e+04 ) sec^-1 MeanMatrixElemValue = ( 1.008920e+02 +- 5.001681e+01 ) GeV^-2 -TOTAL : 1.874706 sec +TOTAL : 1.872398 sec INFO: No Floating Point Exceptions have been reported - 5,138,957,277 cycles # 2.738 GHz - 11,840,408,683 instructions # 2.30 insn per cycle - 1.878700358 seconds time elapsed + 5,149,050,685 cycles # 2.746 GHz + 11,838,766,868 instructions # 2.30 insn per cycle + 1.876512510 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 4746) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttg.mad/SubProcesses/P1_gg_ttxg/build.avx2_d_inl0_hrd1/runTest_cpp.exe @@ -196,15 +196,15 @@ Process = SIGMA_SM_GG_TTXG_CPP [gcc 11.3.1] [inlineHel=0] [h Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('512y': AVX512, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 9.602387e+04 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 9.789550e+04 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 9.789550e+04 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 9.435013e+04 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 9.608771e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 9.608771e+04 ) sec^-1 MeanMatrixElemValue = ( 1.008920e+02 +- 5.001681e+01 ) GeV^-2 -TOTAL : 1.727070 sec +TOTAL : 1.756135 sec INFO: No Floating Point Exceptions have been reported - 4,726,480,466 cycles # 2.731 GHz - 11,165,052,550 instructions # 2.36 insn per cycle - 1.731070886 seconds time elapsed + 4,832,076,257 cycles # 2.746 GHz + 11,163,330,358 instructions # 2.31 insn per cycle + 1.760329651 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 4403) (512y: 246) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttg.mad/SubProcesses/P1_gg_ttxg/build.512y_d_inl0_hrd1/runTest_cpp.exe @@ -228,15 +228,15 @@ Process = SIGMA_SM_GG_TTXG_CPP [gcc 11.3.1] [inlineHel=0] [h Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('512z': AVX512, 512bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 7.101185e+04 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 7.203049e+04 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 7.203049e+04 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 7.074917e+04 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 7.175754e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 7.175754e+04 ) sec^-1 MeanMatrixElemValue = ( 1.008920e+02 +- 5.001681e+01 ) GeV^-2 -TOTAL : 2.329881 sec +TOTAL : 2.337569 sec INFO: No Floating Point Exceptions have been reported - 4,155,200,887 cycles # 1.781 GHz - 6,223,800,996 instructions # 1.50 insn per cycle - 2.334090572 seconds time elapsed + 4,153,292,886 cycles # 1.774 GHz + 6,222,221,062 instructions # 1.50 insn per cycle + 2.341780189 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 1516) (512y: 139) (512z: 3679) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttg.mad/SubProcesses/P1_gg_ttxg/build.512z_d_inl0_hrd1/runTest_cpp.exe diff --git a/epochX/cudacpp/tput/logs_ggttg_mad/log_ggttg_mad_f_inl0_hrd0.txt b/epochX/cudacpp/tput/logs_ggttg_mad/log_ggttg_mad_f_inl0_hrd0.txt index b4d9344f80..03f33d4dc9 100644 --- a/epochX/cudacpp/tput/logs_ggttg_mad/log_ggttg_mad_f_inl0_hrd0.txt +++ b/epochX/cudacpp/tput/logs_ggttg_mad/log_ggttg_mad_f_inl0_hrd0.txt @@ -40,7 +40,7 @@ make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttg.mad/SubProcesses/P1_gg_ttxg' -DATE: 2024-10-06_09:01:57 +DATE: 2024-10-06_13:31:09 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: ========================================================================= @@ -49,15 +49,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_SM_GG_TTXG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 1.641235e+07 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 2.015793e+07 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 2.057654e+07 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.746438e+07 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 2.070563e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.106052e+07 ) sec^-1 MeanMatrixElemValue = ( 1.008472e+02 +- 5.002447e+01 ) GeV^-2 -TOTAL : 0.457409 sec +TOTAL : 0.456368 sec INFO: No Floating Point Exceptions have been reported - 1,937,244,275 cycles # 2.867 GHz - 2,710,892,637 instructions # 1.40 insn per cycle - 0.733854811 seconds time elapsed + 1,933,093,566 cycles # 2.874 GHz + 2,756,060,171 instructions # 1.43 insn per cycle + 0.730006638 seconds time elapsed runNcu /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttg.mad/SubProcesses/P1_gg_ttxg/build.cuda_f_inl0_hrd0/check_cuda.exe -p 64 256 1 ==PROF== Profiling "sigmaKin": launch__registers_per_thread 226 ==PROF== Profiling "sigmaKin": sm__sass_average_branch_targets_threads_uniform.pct 100% @@ -67,15 +67,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_SM_GG_TTXG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 2.672412e+07 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.384843e+07 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.427387e+07 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 2.759195e+07 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.408962e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.449374e+07 ) sec^-1 MeanMatrixElemValue = ( 6.630097e+02 +- 4.770717e+02 ) GeV^-2 -TOTAL : 0.509900 sec +TOTAL : 0.505692 sec INFO: No Floating Point Exceptions have been reported - 2,162,696,786 cycles # 2.871 GHz - 3,100,226,347 instructions # 1.43 insn per cycle - 0.811215095 seconds time elapsed + 2,099,612,654 cycles # 2.869 GHz + 2,998,560,797 instructions # 1.43 insn per cycle + 0.790734913 seconds time elapsed ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttg.mad/SubProcesses/P1_gg_ttxg/build.cuda_f_inl0_hrd0/runTest_cuda.exe INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: FE_DIVBYZERO, FE_INVALID, FE_OVERFLOW @@ -100,15 +100,15 @@ Process = SIGMA_SM_GG_TTXG_CPP [gcc 11.3.1] [inlineHel=0] [h Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/none+NAVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = SCALAR ('none': ~vector[1], no SIMD) -EvtsPerSec[Rmb+ME] (23) = ( 2.513642e+04 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 2.526564e+04 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 2.526564e+04 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 2.518717e+04 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 2.532278e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.532278e+04 ) sec^-1 MeanMatrixElemValue = ( 1.009236e+02 +- 5.002643e+01 ) GeV^-2 -TOTAL : 6.537186 sec +TOTAL : 6.523364 sec INFO: No Floating Point Exceptions have been reported - 19,278,711,706 cycles # 2.948 GHz - 59,616,757,005 instructions # 3.09 insn per cycle - 6.541004954 seconds time elapsed + 19,245,836,019 cycles # 2.949 GHz + 59,614,242,393 instructions # 3.10 insn per cycle + 6.527451938 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 959) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttg.mad/SubProcesses/P1_gg_ttxg/build.none_f_inl0_hrd0/runTest_cpp.exe @@ -132,15 +132,15 @@ Process = SIGMA_SM_GG_TTXG_CPP [gcc 11.3.1] [inlineHel=0] [h Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('sse4': SSE4.2, 128bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 8.120315e+04 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 8.259615e+04 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 8.259615e+04 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 8.098694e+04 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 8.234725e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 8.234725e+04 ) sec^-1 MeanMatrixElemValue = ( 1.009236e+02 +- 5.002643e+01 ) GeV^-2 -TOTAL : 2.035209 sec +TOTAL : 2.040545 sec INFO: No Floating Point Exceptions have been reported - 6,010,527,138 cycles # 2.949 GHz - 17,061,942,080 instructions # 2.84 insn per cycle - 2.038918474 seconds time elapsed + 6,011,562,637 cycles # 2.942 GHz + 17,063,620,334 instructions # 2.84 insn per cycle + 2.044540810 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 5856) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttg.mad/SubProcesses/P1_gg_ttxg/build.sse4_f_inl0_hrd0/runTest_cpp.exe @@ -164,15 +164,15 @@ Process = SIGMA_SM_GG_TTXG_CPP [gcc 11.3.1] [inlineHel=0] [h Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('avx2': AVX2, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 1.748972e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.811746e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.811746e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.734172e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.795567e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.795567e+05 ) sec^-1 MeanMatrixElemValue = ( 1.008857e+02 +- 5.002468e+01 ) GeV^-2 -TOTAL : 0.954915 sec +TOTAL : 0.962548 sec INFO: No Floating Point Exceptions have been reported - 2,640,169,352 cycles # 2.756 GHz - 6,187,458,591 instructions # 2.34 insn per cycle - 0.958678404 seconds time elapsed + 2,639,165,766 cycles # 2.732 GHz + 6,187,855,401 instructions # 2.34 insn per cycle + 0.966575337 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 5105) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttg.mad/SubProcesses/P1_gg_ttxg/build.avx2_f_inl0_hrd0/runTest_cpp.exe @@ -196,15 +196,15 @@ Process = SIGMA_SM_GG_TTXG_CPP [gcc 11.3.1] [inlineHel=0] [h Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('512y': AVX512, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 1.923079e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.998771e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.998771e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.924864e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.999116e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.999116e+05 ) sec^-1 MeanMatrixElemValue = ( 1.008857e+02 +- 5.002468e+01 ) GeV^-2 -TOTAL : 0.870044 sec +TOTAL : 0.868874 sec INFO: No Floating Point Exceptions have been reported - 2,402,321,989 cycles # 2.751 GHz - 5,790,080,813 instructions # 2.41 insn per cycle - 0.873863245 seconds time elapsed + 2,401,156,334 cycles # 2.753 GHz + 5,790,098,479 instructions # 2.41 insn per cycle + 0.872995682 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 4906) (512y: 37) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttg.mad/SubProcesses/P1_gg_ttxg/build.512y_f_inl0_hrd0/runTest_cpp.exe @@ -228,15 +228,15 @@ Process = SIGMA_SM_GG_TTXG_CPP [gcc 11.3.1] [inlineHel=0] [h Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[16] ('512z': AVX512, 512bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 1.455132e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.498332e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.498332e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.446132e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.489145e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.489145e+05 ) sec^-1 MeanMatrixElemValue = ( 1.008856e+02 +- 5.002468e+01 ) GeV^-2 -TOTAL : 1.146003 sec +TOTAL : 1.154804 sec INFO: No Floating Point Exceptions have been reported - 2,072,911,951 cycles # 1.804 GHz - 3,391,607,808 instructions # 1.64 insn per cycle - 1.149850121 seconds time elapsed + 2,074,608,657 cycles # 1.791 GHz + 3,391,460,898 instructions # 1.63 insn per cycle + 1.158896259 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 2237) (512y: 37) (512z: 3789) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttg.mad/SubProcesses/P1_gg_ttxg/build.512z_f_inl0_hrd0/runTest_cpp.exe diff --git a/epochX/cudacpp/tput/logs_ggttg_mad/log_ggttg_mad_f_inl0_hrd0_bridge.txt b/epochX/cudacpp/tput/logs_ggttg_mad/log_ggttg_mad_f_inl0_hrd0_bridge.txt index 89f1af02c0..1ed69c98cc 100644 --- a/epochX/cudacpp/tput/logs_ggttg_mad/log_ggttg_mad_f_inl0_hrd0_bridge.txt +++ b/epochX/cudacpp/tput/logs_ggttg_mad/log_ggttg_mad_f_inl0_hrd0_bridge.txt @@ -40,7 +40,7 @@ make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttg.mad/SubProcesses/P1_gg_ttxg' -DATE: 2024-10-06_09:38:49 +DATE: 2024-10-06_13:57:22 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: ========================================================================= @@ -53,16 +53,16 @@ WARNING! Set grid in Bridge (nevt=16384, gpublocks=64, gputhreads=256, gpublocks Process = SIGMA_SM_GG_TTXG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 4.430077e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.496267e+07 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.496267e+07 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 4.435332e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.509622e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.509622e+07 ) sec^-1 MeanMatrixElemValue = ( 1.009071e+02 +- 5.002295e+01 ) GeV^-2 -TOTAL : 0.468595 sec +TOTAL : 0.471204 sec INFO: No Floating Point Exceptions have been reported INFO: No Floating Point Exceptions have been reported - 1,993,529,576 cycles # 2.878 GHz - 2,894,144,626 instructions # 1.45 insn per cycle - 0.749153323 seconds time elapsed + 1,970,875,021 cycles # 2.882 GHz + 2,878,082,597 instructions # 1.46 insn per cycle + 0.742702928 seconds time elapsed runNcu /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttg.mad/SubProcesses/P1_gg_ttxg/build.cuda_f_inl0_hrd0/check_cuda.exe -p 64 256 1 --bridge WARNING! Bridge selected: cannot use RamboDevice, will use RamboHost WARNING! RamboHost selected: cannot use CurandDevice, will use CurandHost @@ -80,16 +80,16 @@ WARNING! Set grid in Bridge (nevt=524288, gpublocks=2048, gputhreads=256, gpublo Process = SIGMA_SM_GG_TTXG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 4.508973e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 2.254431e+07 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 2.254431e+07 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 4.530545e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 2.261839e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.261839e+07 ) sec^-1 MeanMatrixElemValue = ( 6.737499e+02 +- 4.776369e+02 ) GeV^-2 -TOTAL : 0.658615 sec +TOTAL : 0.655296 sec INFO: No Floating Point Exceptions have been reported INFO: No Floating Point Exceptions have been reported - 2,580,648,777 cycles # 2.888 GHz - 3,894,936,658 instructions # 1.51 insn per cycle - 0.952346890 seconds time elapsed + 2,569,276,650 cycles # 2.888 GHz + 3,913,635,588 instructions # 1.52 insn per cycle + 0.947925925 seconds time elapsed ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttg.mad/SubProcesses/P1_gg_ttxg/build.cuda_f_inl0_hrd0/runTest_cuda.exe INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: FE_DIVBYZERO, FE_INVALID, FE_OVERFLOW @@ -115,16 +115,16 @@ Process = SIGMA_SM_GG_TTXG_CPP [gcc 11.3.1] [inlineHel=0] [h Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+BRDHST/none+NAVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = SCALAR ('none': ~vector[1], no SIMD) -EvtsPerSec[Rmb+ME] (23) = ( 2.506368e+04 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 2.519408e+04 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 2.519408e+04 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 2.514256e+04 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 2.527740e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.527740e+04 ) sec^-1 MeanMatrixElemValue = ( 1.009236e+02 +- 5.002643e+01 ) GeV^-2 -TOTAL : 6.559975 sec +TOTAL : 6.541306 sec INFO: No Floating Point Exceptions have been reported INFO: No Floating Point Exceptions have been reported - 19,280,512,048 cycles # 2.938 GHz - 59,619,141,119 instructions # 3.09 insn per cycle - 6.564243260 seconds time elapsed + 19,277,830,708 cycles # 2.946 GHz + 59,619,246,897 instructions # 3.09 insn per cycle + 6.545709514 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 959) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttg.mad/SubProcesses/P1_gg_ttxg/build.none_f_inl0_hrd0/runTest_cpp.exe @@ -149,16 +149,16 @@ Process = SIGMA_SM_GG_TTXG_CPP [gcc 11.3.1] [inlineHel=0] [h Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+BRDHST/sse4+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('sse4': SSE4.2, 128bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 8.092271e+04 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 8.230160e+04 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 8.230160e+04 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 8.110374e+04 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 8.247879e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 8.247879e+04 ) sec^-1 MeanMatrixElemValue = ( 1.009236e+02 +- 5.002643e+01 ) GeV^-2 -TOTAL : 2.047307 sec +TOTAL : 2.042598 sec INFO: No Floating Point Exceptions have been reported INFO: No Floating Point Exceptions have been reported - 6,043,775,465 cycles # 2.947 GHz - 17,111,089,922 instructions # 2.83 insn per cycle - 2.051614364 seconds time elapsed + 6,035,950,953 cycles # 2.950 GHz + 17,110,696,820 instructions # 2.83 insn per cycle + 2.046759841 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 5856) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttg.mad/SubProcesses/P1_gg_ttxg/build.sse4_f_inl0_hrd0/runTest_cpp.exe @@ -183,16 +183,16 @@ Process = SIGMA_SM_GG_TTXG_CPP [gcc 11.3.1] [inlineHel=0] [h Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+BRDHST/avx2+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('avx2': AVX2, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 1.748354e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.809701e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.809701e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.740709e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.802541e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.802541e+05 ) sec^-1 MeanMatrixElemValue = ( 1.008857e+02 +- 5.002468e+01 ) GeV^-2 -TOTAL : 0.959425 sec +TOTAL : 0.964079 sec INFO: No Floating Point Exceptions have been reported INFO: No Floating Point Exceptions have been reported - 2,659,679,748 cycles # 2.761 GHz - 6,224,393,438 instructions # 2.34 insn per cycle - 0.963869172 seconds time elapsed + 2,661,800,157 cycles # 2.751 GHz + 6,224,579,044 instructions # 2.34 insn per cycle + 0.968361421 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 5105) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttg.mad/SubProcesses/P1_gg_ttxg/build.avx2_f_inl0_hrd0/runTest_cpp.exe @@ -217,16 +217,16 @@ Process = SIGMA_SM_GG_TTXG_CPP [gcc 11.3.1] [inlineHel=0] [h Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+BRDHST/512y+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('512y': AVX512, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 1.927524e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 2.002486e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 2.002486e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.906247e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.980857e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.980857e+05 ) sec^-1 MeanMatrixElemValue = ( 1.008857e+02 +- 5.002468e+01 ) GeV^-2 -TOTAL : 0.872058 sec +TOTAL : 0.882109 sec INFO: No Floating Point Exceptions have been reported INFO: No Floating Point Exceptions have been reported - 2,421,094,163 cycles # 2.765 GHz - 5,826,830,021 instructions # 2.41 insn per cycle - 0.876372578 seconds time elapsed + 2,424,428,424 cycles # 2.737 GHz + 5,826,936,286 instructions # 2.40 insn per cycle + 0.886424442 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 4906) (512y: 37) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttg.mad/SubProcesses/P1_gg_ttxg/build.512y_f_inl0_hrd0/runTest_cpp.exe @@ -251,16 +251,16 @@ Process = SIGMA_SM_GG_TTXG_CPP [gcc 11.3.1] [inlineHel=0] [h Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+BRDHST/512z+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[16] ('512z': AVX512, 512bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 1.443486e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.486864e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.486864e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.454316e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.496981e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.496981e+05 ) sec^-1 MeanMatrixElemValue = ( 1.008856e+02 +- 5.002468e+01 ) GeV^-2 -TOTAL : 1.160150 sec +TOTAL : 1.151045 sec INFO: No Floating Point Exceptions have been reported INFO: No Floating Point Exceptions have been reported - 2,098,432,349 cycles # 1.804 GHz - 3,433,067,927 instructions # 1.64 insn per cycle - 1.164579445 seconds time elapsed + 2,095,449,942 cycles # 1.815 GHz + 3,432,948,266 instructions # 1.64 insn per cycle + 1.155382155 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 2237) (512y: 37) (512z: 3789) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttg.mad/SubProcesses/P1_gg_ttxg/build.512z_f_inl0_hrd0/runTest_cpp.exe diff --git a/epochX/cudacpp/tput/logs_ggttg_mad/log_ggttg_mad_f_inl0_hrd1.txt b/epochX/cudacpp/tput/logs_ggttg_mad/log_ggttg_mad_f_inl0_hrd1.txt index 7537d3c84d..cd1e4fad5a 100644 --- a/epochX/cudacpp/tput/logs_ggttg_mad/log_ggttg_mad_f_inl0_hrd1.txt +++ b/epochX/cudacpp/tput/logs_ggttg_mad/log_ggttg_mad_f_inl0_hrd1.txt @@ -40,7 +40,7 @@ make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttg.mad/SubProcesses/P1_gg_ttxg' -DATE: 2024-10-06_09:02:18 +DATE: 2024-10-06_13:31:30 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: ========================================================================= @@ -49,15 +49,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_SM_GG_TTXG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=1] Workflow summary = CUD:FLT+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 1.658659e+07 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 2.027503e+07 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 2.066373e+07 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.759352e+07 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 2.085168e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.119725e+07 ) sec^-1 MeanMatrixElemValue = ( 1.008472e+02 +- 5.002447e+01 ) GeV^-2 -TOTAL : 0.462988 sec +TOTAL : 0.456098 sec INFO: No Floating Point Exceptions have been reported - 1,956,715,427 cycles # 2.872 GHz - 2,757,694,861 instructions # 1.41 insn per cycle - 0.742544959 seconds time elapsed + 1,932,195,880 cycles # 2.871 GHz + 2,721,814,729 instructions # 1.41 insn per cycle + 0.730106904 seconds time elapsed runNcu /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttg.mad/SubProcesses/P1_gg_ttxg/build.cuda_f_inl0_hrd1/check_cuda.exe -p 64 256 1 ==PROF== Profiling "sigmaKin": launch__registers_per_thread 226 ==PROF== Profiling "sigmaKin": sm__sass_average_branch_targets_threads_uniform.pct 100% @@ -67,15 +67,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_SM_GG_TTXG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=1] Workflow summary = CUD:FLT+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 2.669827e+07 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.371215e+07 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.415741e+07 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 2.748743e+07 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.394614e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.433963e+07 ) sec^-1 MeanMatrixElemValue = ( 6.630097e+02 +- 4.770717e+02 ) GeV^-2 -TOTAL : 0.505811 sec +TOTAL : 0.501651 sec INFO: No Floating Point Exceptions have been reported - 2,123,611,289 cycles # 2.883 GHz - 3,083,974,467 instructions # 1.45 insn per cycle - 0.793454464 seconds time elapsed + 2,090,057,094 cycles # 2.865 GHz + 3,037,467,879 instructions # 1.45 insn per cycle + 0.786765239 seconds time elapsed ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttg.mad/SubProcesses/P1_gg_ttxg/build.cuda_f_inl0_hrd1/runTest_cuda.exe INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: FE_DIVBYZERO, FE_INVALID, FE_OVERFLOW @@ -100,15 +100,15 @@ Process = SIGMA_SM_GG_TTXG_CPP [gcc 11.3.1] [inlineHel=0] [h Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/none+NAVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = SCALAR ('none': ~vector[1], no SIMD) -EvtsPerSec[Rmb+ME] (23) = ( 2.488365e+04 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 2.501255e+04 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 2.501255e+04 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 2.492713e+04 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 2.505485e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.505485e+04 ) sec^-1 MeanMatrixElemValue = ( 1.009236e+02 +- 5.002643e+01 ) GeV^-2 -TOTAL : 6.602997 sec +TOTAL : 6.591207 sec INFO: No Floating Point Exceptions have been reported - 19,409,400,884 cycles # 2.938 GHz - 59,351,848,666 instructions # 3.06 insn per cycle - 6.606759387 seconds time elapsed + 19,393,904,369 cycles # 2.941 GHz + 59,352,124,048 instructions # 3.06 insn per cycle + 6.595267893 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 1027) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttg.mad/SubProcesses/P1_gg_ttxg/build.none_f_inl0_hrd1/runTest_cpp.exe @@ -132,15 +132,15 @@ Process = SIGMA_SM_GG_TTXG_CPP [gcc 11.3.1] [inlineHel=0] [h Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('sse4': SSE4.2, 128bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 8.484090e+04 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 8.633368e+04 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 8.633368e+04 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 8.467962e+04 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 8.615642e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 8.615642e+04 ) sec^-1 MeanMatrixElemValue = ( 1.009236e+02 +- 5.002643e+01 ) GeV^-2 -TOTAL : 1.948837 sec +TOTAL : 1.952305 sec INFO: No Floating Point Exceptions have been reported - 5,764,162,956 cycles # 2.953 GHz - 16,849,716,772 instructions # 2.92 insn per cycle - 1.952678468 seconds time elapsed + 5,763,798,203 cycles # 2.947 GHz + 16,849,875,225 instructions # 2.92 insn per cycle + 1.956395411 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 5610) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttg.mad/SubProcesses/P1_gg_ttxg/build.sse4_f_inl0_hrd1/runTest_cpp.exe @@ -164,15 +164,15 @@ Process = SIGMA_SM_GG_TTXG_CPP [gcc 11.3.1] [inlineHel=0] [h Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('avx2': AVX2, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 1.522405e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.569181e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.569181e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.519067e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.566183e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.566183e+05 ) sec^-1 MeanMatrixElemValue = ( 1.008857e+02 +- 5.002468e+01 ) GeV^-2 -TOTAL : 1.094041 sec +TOTAL : 1.096895 sec INFO: No Floating Point Exceptions have been reported - 3,018,102,108 cycles # 2.750 GHz - 6,848,568,360 instructions # 2.27 insn per cycle - 1.098202042 seconds time elapsed + 3,019,040,088 cycles # 2.751 GHz + 6,849,189,231 instructions # 2.27 insn per cycle + 1.101114433 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 5735) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttg.mad/SubProcesses/P1_gg_ttxg/build.avx2_f_inl0_hrd1/runTest_cpp.exe @@ -196,15 +196,15 @@ Process = SIGMA_SM_GG_TTXG_CPP [gcc 11.3.1] [inlineHel=0] [h Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('512y': AVX512, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 1.654265e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.710055e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.710055e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.645174e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.699505e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.699505e+05 ) sec^-1 MeanMatrixElemValue = ( 1.008857e+02 +- 5.002468e+01 ) GeV^-2 -TOTAL : 1.008735 sec +TOTAL : 1.013611 sec INFO: No Floating Point Exceptions have been reported - 2,794,533,058 cycles # 2.762 GHz - 6,437,695,564 instructions # 2.30 insn per cycle - 1.012558685 seconds time elapsed + 2,792,117,172 cycles # 2.745 GHz + 6,437,435,410 instructions # 2.31 insn per cycle + 1.017640888 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 5509) (512y: 23) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttg.mad/SubProcesses/P1_gg_ttxg/build.512y_f_inl0_hrd1/runTest_cpp.exe @@ -228,15 +228,15 @@ Process = SIGMA_SM_GG_TTXG_CPP [gcc 11.3.1] [inlineHel=0] [h Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[16] ('512z': AVX512, 512bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 1.323435e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.360072e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.360072e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.320484e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.356410e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.356410e+05 ) sec^-1 MeanMatrixElemValue = ( 1.008856e+02 +- 5.002468e+01 ) GeV^-2 -TOTAL : 1.258302 sec +TOTAL : 1.260518 sec INFO: No Floating Point Exceptions have been reported - 2,251,923,496 cycles # 1.787 GHz - 3,755,291,572 instructions # 1.67 insn per cycle - 1.262174564 seconds time elapsed + 2,253,184,751 cycles # 1.783 GHz + 3,756,058,079 instructions # 1.67 insn per cycle + 1.264586489 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 2467) (512y: 28) (512z: 4084) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttg.mad/SubProcesses/P1_gg_ttxg/build.512z_f_inl0_hrd1/runTest_cpp.exe diff --git a/epochX/cudacpp/tput/logs_ggttg_mad/log_ggttg_mad_m_inl0_hrd0.txt b/epochX/cudacpp/tput/logs_ggttg_mad/log_ggttg_mad_m_inl0_hrd0.txt index 6b4617ba56..d1821e3fae 100644 --- a/epochX/cudacpp/tput/logs_ggttg_mad/log_ggttg_mad_m_inl0_hrd0.txt +++ b/epochX/cudacpp/tput/logs_ggttg_mad/log_ggttg_mad_m_inl0_hrd0.txt @@ -40,7 +40,7 @@ make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttg.mad/SubProcesses/P1_gg_ttxg' -DATE: 2024-10-06_09:01:06 +DATE: 2024-10-06_13:30:17 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: ========================================================================= @@ -49,15 +49,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_SM_GG_TTXG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:MIX+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 7.531107e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 8.896113e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 9.014318e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 7.794134e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 8.905230e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 8.999204e+06 ) sec^-1 MeanMatrixElemValue = ( 1.008920e+02 +- 5.001681e+01 ) GeV^-2 -TOTAL : 0.482291 sec +TOTAL : 0.475249 sec INFO: No Floating Point Exceptions have been reported - 1,996,726,100 cycles # 2.869 GHz - 2,875,927,393 instructions # 1.44 insn per cycle - 0.757518934 seconds time elapsed + 1,969,996,926 cycles # 2.852 GHz + 2,821,925,580 instructions # 1.43 insn per cycle + 0.749818298 seconds time elapsed runNcu /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttg.mad/SubProcesses/P1_gg_ttxg/build.cuda_m_inl0_hrd0/check_cuda.exe -p 64 256 1 ==PROF== Profiling "sigmaKin": launch__registers_per_thread 255 ==PROF== Profiling "sigmaKin": sm__sass_average_branch_targets_threads_uniform.pct 100% @@ -67,15 +67,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_SM_GG_TTXG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:MIX+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 1.039985e+07 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.227093e+07 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.238483e+07 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.052872e+07 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.229112e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.239124e+07 ) sec^-1 MeanMatrixElemValue = ( 6.734461e+02 +- 4.775415e+02 ) GeV^-2 -TOTAL : 0.618226 sec +TOTAL : 0.612866 sec INFO: No Floating Point Exceptions have been reported - 2,476,524,825 cycles # 2.883 GHz - 3,787,822,568 instructions # 1.53 insn per cycle - 0.918414719 seconds time elapsed + 2,444,224,412 cycles # 2.882 GHz + 3,707,217,935 instructions # 1.52 insn per cycle + 0.910631075 seconds time elapsed ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttg.mad/SubProcesses/P1_gg_ttxg/build.cuda_m_inl0_hrd0/runTest_cuda.exe INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: FE_DIVBYZERO, FE_INVALID, FE_OVERFLOW @@ -100,15 +100,15 @@ Process = SIGMA_SM_GG_TTXG_CPP [gcc 11.3.1] [inlineHel=0] [h Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/none+NAVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) Internal loops fptype_sv = SCALAR ('none': ~vector[1], no SIMD) -EvtsPerSec[Rmb+ME] (23) = ( 2.396101e+04 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 2.408087e+04 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 2.408087e+04 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 2.390456e+04 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 2.402142e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.402142e+04 ) sec^-1 MeanMatrixElemValue = ( 1.008920e+02 +- 5.001681e+01 ) GeV^-2 -TOTAL : 6.859626 sec +TOTAL : 6.874398 sec INFO: No Floating Point Exceptions have been reported - 20,206,369,377 cycles # 2.945 GHz - 60,950,595,896 instructions # 3.02 insn per cycle - 6.863727850 seconds time elapsed + 20,186,349,250 cycles # 2.936 GHz + 60,947,677,285 instructions # 3.02 insn per cycle + 6.878560477 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 1220) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttg.mad/SubProcesses/P1_gg_ttxg/build.none_m_inl0_hrd0/runTest_cpp.exe @@ -132,15 +132,15 @@ Process = SIGMA_SM_GG_TTXG_CPP [gcc 11.3.1] [inlineHel=0] [h Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[2] ('sse4': SSE4.2, 128bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 4.651759e+04 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 4.695029e+04 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 4.695029e+04 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 4.645117e+04 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 4.688033e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 4.688033e+04 ) sec^-1 MeanMatrixElemValue = ( 1.008920e+02 +- 5.001681e+01 ) GeV^-2 -TOTAL : 3.542669 sec +TOTAL : 3.546690 sec INFO: No Floating Point Exceptions have been reported - 10,470,195,857 cycles # 2.953 GHz - 30,822,635,750 instructions # 2.94 insn per cycle - 3.546724112 seconds time elapsed + 10,463,978,818 cycles # 2.947 GHz + 30,820,961,481 instructions # 2.95 insn per cycle + 3.551091582 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 5351) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttg.mad/SubProcesses/P1_gg_ttxg/build.sse4_m_inl0_hrd0/runTest_cpp.exe @@ -164,15 +164,15 @@ Process = SIGMA_SM_GG_TTXG_CPP [gcc 11.3.1] [inlineHel=0] [h Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('avx2': AVX2, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 9.177717e+04 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 9.345070e+04 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 9.345070e+04 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 9.226232e+04 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 9.393448e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 9.393448e+04 ) sec^-1 MeanMatrixElemValue = ( 1.008920e+02 +- 5.001681e+01 ) GeV^-2 -TOTAL : 1.805877 sec +TOTAL : 1.795592 sec INFO: No Floating Point Exceptions have been reported - 4,960,900,655 cycles # 2.742 GHz - 11,360,293,322 instructions # 2.29 insn per cycle - 1.809915904 seconds time elapsed + 4,958,822,845 cycles # 2.757 GHz + 11,359,118,807 instructions # 2.29 insn per cycle + 1.799734262 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 4776) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttg.mad/SubProcesses/P1_gg_ttxg/build.avx2_m_inl0_hrd0/runTest_cpp.exe @@ -196,15 +196,15 @@ Process = SIGMA_SM_GG_TTXG_CPP [gcc 11.3.1] [inlineHel=0] [h Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('512y': AVX512, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 1.047166e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.068679e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.068679e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.043285e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.064651e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.064651e+05 ) sec^-1 MeanMatrixElemValue = ( 1.008920e+02 +- 5.001681e+01 ) GeV^-2 -TOTAL : 1.585052 sec +TOTAL : 1.589981 sec INFO: No Floating Point Exceptions have been reported - 4,379,448,731 cycles # 2.757 GHz - 10,610,063,505 instructions # 2.42 insn per cycle - 1.588995755 seconds time elapsed + 4,374,564,542 cycles # 2.745 GHz + 10,608,419,194 instructions # 2.43 insn per cycle + 1.594129009 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 4503) (512y: 84) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttg.mad/SubProcesses/P1_gg_ttxg/build.512y_m_inl0_hrd0/runTest_cpp.exe @@ -228,15 +228,15 @@ Process = SIGMA_SM_GG_TTXG_CPP [gcc 11.3.1] [inlineHel=0] [h Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('512z': AVX512, 512bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 6.890582e+04 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 6.987179e+04 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 6.987179e+04 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 6.970118e+04 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 7.066362e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 7.066362e+04 ) sec^-1 MeanMatrixElemValue = ( 1.008920e+02 +- 5.001681e+01 ) GeV^-2 -TOTAL : 2.401138 sec +TOTAL : 2.372653 sec INFO: No Floating Point Exceptions have been reported - 4,243,505,288 cycles # 1.765 GHz - 6,171,567,257 instructions # 1.45 insn per cycle - 2.405218093 seconds time elapsed + 4,232,460,771 cycles # 1.781 GHz + 6,168,515,459 instructions # 1.46 insn per cycle + 2.376867299 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 2143) (512y: 116) (512z: 3653) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttg.mad/SubProcesses/P1_gg_ttxg/build.512z_m_inl0_hrd0/runTest_cpp.exe diff --git a/epochX/cudacpp/tput/logs_ggttg_mad/log_ggttg_mad_m_inl0_hrd1.txt b/epochX/cudacpp/tput/logs_ggttg_mad/log_ggttg_mad_m_inl0_hrd1.txt index 1a268fb0a6..382581c123 100644 --- a/epochX/cudacpp/tput/logs_ggttg_mad/log_ggttg_mad_m_inl0_hrd1.txt +++ b/epochX/cudacpp/tput/logs_ggttg_mad/log_ggttg_mad_m_inl0_hrd1.txt @@ -40,7 +40,7 @@ make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttg.mad/SubProcesses/P1_gg_ttxg' -DATE: 2024-10-06_09:01:31 +DATE: 2024-10-06_13:30:43 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: ========================================================================= @@ -49,15 +49,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_SM_GG_TTXG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=1] Workflow summary = CUD:MIX+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 7.506525e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 8.876419e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 8.986419e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 7.865056e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 8.995398e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 9.088088e+06 ) sec^-1 MeanMatrixElemValue = ( 1.008920e+02 +- 5.001681e+01 ) GeV^-2 -TOTAL : 0.475723 sec +TOTAL : 0.471878 sec INFO: No Floating Point Exceptions have been reported - 1,989,777,196 cycles # 2.876 GHz - 2,865,221,599 instructions # 1.44 insn per cycle - 0.750464789 seconds time elapsed + 1,982,576,237 cycles # 2.882 GHz + 2,838,129,895 instructions # 1.43 insn per cycle + 0.746041915 seconds time elapsed runNcu /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttg.mad/SubProcesses/P1_gg_ttxg/build.cuda_m_inl0_hrd1/check_cuda.exe -p 64 256 1 ==PROF== Profiling "sigmaKin": launch__registers_per_thread 255 ==PROF== Profiling "sigmaKin": sm__sass_average_branch_targets_threads_uniform.pct 100% @@ -67,15 +67,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_SM_GG_TTXG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=1] Workflow summary = CUD:MIX+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 1.040967e+07 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.229706e+07 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.240646e+07 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.057755e+07 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.235252e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.245312e+07 ) sec^-1 MeanMatrixElemValue = ( 6.734461e+02 +- 4.775415e+02 ) GeV^-2 -TOTAL : 0.612359 sec +TOTAL : 0.604221 sec INFO: No Floating Point Exceptions have been reported - 2,465,408,367 cycles # 2.885 GHz - 3,759,784,229 instructions # 1.53 insn per cycle - 0.914073870 seconds time elapsed + 2,451,997,755 cycles # 2.883 GHz + 3,733,977,708 instructions # 1.52 insn per cycle + 0.909520068 seconds time elapsed ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttg.mad/SubProcesses/P1_gg_ttxg/build.cuda_m_inl0_hrd1/runTest_cuda.exe INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: FE_DIVBYZERO, FE_INVALID, FE_OVERFLOW @@ -100,15 +100,15 @@ Process = SIGMA_SM_GG_TTXG_CPP [gcc 11.3.1] [inlineHel=0] [h Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/none+NAVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) Internal loops fptype_sv = SCALAR ('none': ~vector[1], no SIMD) -EvtsPerSec[Rmb+ME] (23) = ( 2.395973e+04 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 2.407808e+04 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 2.407808e+04 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 2.375078e+04 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 2.386702e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.386702e+04 ) sec^-1 MeanMatrixElemValue = ( 1.008920e+02 +- 5.001681e+01 ) GeV^-2 -TOTAL : 6.859771 sec +TOTAL : 6.918678 sec INFO: No Floating Point Exceptions have been reported - 20,239,178,144 cycles # 2.949 GHz - 61,173,779,461 instructions # 3.02 insn per cycle - 6.863706451 seconds time elapsed + 20,256,030,558 cycles # 2.926 GHz + 61,172,693,155 instructions # 3.02 insn per cycle + 6.922962926 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 1272) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttg.mad/SubProcesses/P1_gg_ttxg/build.none_m_inl0_hrd1/runTest_cpp.exe @@ -132,15 +132,15 @@ Process = SIGMA_SM_GG_TTXG_CPP [gcc 11.3.1] [inlineHel=0] [h Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[2] ('sse4': SSE4.2, 128bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 4.702334e+04 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 4.747762e+04 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 4.747762e+04 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 4.684164e+04 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 4.728785e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 4.728785e+04 ) sec^-1 MeanMatrixElemValue = ( 1.008920e+02 +- 5.001681e+01 ) GeV^-2 -TOTAL : 3.505938 sec +TOTAL : 3.517048 sec INFO: No Floating Point Exceptions have been reported - 10,333,154,234 cycles # 2.946 GHz - 30,534,348,115 instructions # 2.95 insn per cycle - 3.510016853 seconds time elapsed + 10,329,368,027 cycles # 2.934 GHz + 30,533,631,418 instructions # 2.96 insn per cycle + 3.521220340 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 5155) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttg.mad/SubProcesses/P1_gg_ttxg/build.sse4_m_inl0_hrd1/runTest_cpp.exe @@ -164,15 +164,15 @@ Process = SIGMA_SM_GG_TTXG_CPP [gcc 11.3.1] [inlineHel=0] [h Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('avx2': AVX2, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 8.861323e+04 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 9.018375e+04 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 9.018375e+04 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 8.824536e+04 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 8.978126e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 8.978126e+04 ) sec^-1 MeanMatrixElemValue = ( 1.008920e+02 +- 5.001681e+01 ) GeV^-2 -TOTAL : 1.870783 sec +TOTAL : 1.876258 sec INFO: No Floating Point Exceptions have been reported - 5,160,894,050 cycles # 2.755 GHz - 11,875,310,688 instructions # 2.30 insn per cycle - 1.874839635 seconds time elapsed + 5,141,779,065 cycles # 2.735 GHz + 11,872,506,644 instructions # 2.31 insn per cycle + 1.880419461 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 4887) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttg.mad/SubProcesses/P1_gg_ttxg/build.avx2_m_inl0_hrd1/runTest_cpp.exe @@ -196,15 +196,15 @@ Process = SIGMA_SM_GG_TTXG_CPP [gcc 11.3.1] [inlineHel=0] [h Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('512y': AVX512, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 9.768245e+04 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 9.957717e+04 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 9.957717e+04 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 9.775388e+04 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 9.963725e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 9.963725e+04 ) sec^-1 MeanMatrixElemValue = ( 1.008920e+02 +- 5.001681e+01 ) GeV^-2 -TOTAL : 1.697611 sec +TOTAL : 1.695576 sec INFO: No Floating Point Exceptions have been reported - 4,679,050,155 cycles # 2.751 GHz - 11,168,862,734 instructions # 2.39 insn per cycle - 1.701628470 seconds time elapsed + 4,673,717,265 cycles # 2.751 GHz + 11,165,709,024 instructions # 2.39 insn per cycle + 1.699766528 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 4508) (512y: 239) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttg.mad/SubProcesses/P1_gg_ttxg/build.512y_m_inl0_hrd1/runTest_cpp.exe @@ -228,15 +228,15 @@ Process = SIGMA_SM_GG_TTXG_CPP [gcc 11.3.1] [inlineHel=0] [h Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('512z': AVX512, 512bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 6.922687e+04 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 7.020028e+04 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 7.020028e+04 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 6.930942e+04 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 7.026180e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 7.026180e+04 ) sec^-1 MeanMatrixElemValue = ( 1.008920e+02 +- 5.001681e+01 ) GeV^-2 -TOTAL : 2.390116 sec +TOTAL : 2.385664 sec INFO: No Floating Point Exceptions have been reported - 4,256,907,095 cycles # 1.778 GHz - 6,411,350,564 instructions # 1.51 insn per cycle - 2.394737171 seconds time elapsed + 4,245,101,293 cycles # 1.777 GHz + 6,409,294,991 instructions # 1.51 insn per cycle + 2.389907643 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 2039) (512y: 162) (512z: 3731) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttg.mad/SubProcesses/P1_gg_ttxg/build.512z_m_inl0_hrd1/runTest_cpp.exe diff --git a/epochX/cudacpp/tput/logs_ggttgg_mad/log_ggttgg_mad_d_inl0_hrd0.txt b/epochX/cudacpp/tput/logs_ggttgg_mad/log_ggttgg_mad_d_inl0_hrd0.txt index fe9e9669c6..71b08d3969 100644 --- a/epochX/cudacpp/tput/logs_ggttgg_mad/log_ggttgg_mad_d_inl0_hrd0.txt +++ b/epochX/cudacpp/tput/logs_ggttgg_mad/log_ggttgg_mad_d_inl0_hrd0.txt @@ -40,7 +40,7 @@ make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg' -DATE: 2024-10-06_09:02:40 +DATE: 2024-10-06_13:31:51 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: ========================================================================= @@ -49,15 +49,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_SM_GG_TTXGG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 3.331120e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.359202e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.361250e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 3.317296e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.339130e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.340859e+05 ) sec^-1 MeanMatrixElemValue = ( 4.063123e+00 +- 2.368970e+00 ) GeV^-4 -TOTAL : 0.543918 sec +TOTAL : 0.532404 sec INFO: No Floating Point Exceptions have been reported - 2,225,694,406 cycles # 2.884 GHz - 3,483,451,829 instructions # 1.57 insn per cycle - 0.837015502 seconds time elapsed + 2,199,379,460 cycles # 2.866 GHz + 3,447,036,091 instructions # 1.57 insn per cycle + 0.824650276 seconds time elapsed runNcu /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.cuda_d_inl0_hrd0/check_cuda.exe -p 64 256 1 ==PROF== Profiling "sigmaKin": launch__registers_per_thread 255 ==PROF== Profiling "sigmaKin": sm__sass_average_branch_targets_threads_uniform.pct 100% @@ -67,15 +67,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_SM_GG_TTXGG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 4.134422e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 4.164730e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 4.165914e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 4.142399e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 4.170404e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 4.171653e+05 ) sec^-1 MeanMatrixElemValue = ( 6.665112e+00 +- 5.002651e+00 ) GeV^-4 -TOTAL : 3.052190 sec +TOTAL : 3.033705 sec INFO: No Floating Point Exceptions have been reported - 9,689,726,748 cycles # 2.928 GHz - 22,118,867,491 instructions # 2.28 insn per cycle - 3.368998161 seconds time elapsed + 9,656,011,710 cycles # 2.927 GHz + 20,474,711,303 instructions # 2.12 insn per cycle + 3.355152313 seconds time elapsed ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.cuda_d_inl0_hrd0/runTest_cuda.exe INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: FE_DIVBYZERO, FE_INVALID, FE_OVERFLOW @@ -100,15 +100,15 @@ Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=0] [ Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/none+NAVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = SCALAR ('none': ~vector[1], no SIMD) -EvtsPerSec[Rmb+ME] (23) = ( 1.884002e+03 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.884932e+03 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.884932e+03 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.884781e+03 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.885684e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.885684e+03 ) sec^-1 MeanMatrixElemValue = ( 4.063123e+00 +- 2.368970e+00 ) GeV^-4 -TOTAL : 8.713540 sec +TOTAL : 8.708002 sec INFO: No Floating Point Exceptions have been reported - 25,683,805,881 cycles # 2.947 GHz - 78,963,253,936 instructions # 3.07 insn per cycle - 8.717598721 seconds time elapsed + 25,645,009,137 cycles # 2.944 GHz + 78,955,874,590 instructions # 3.08 insn per cycle + 8.712243433 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 4842) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.none_d_inl0_hrd0/runTest_cpp.exe @@ -132,15 +132,15 @@ Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=0] [ Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[2] ('sse4': SSE4.2, 128bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 3.540501e+03 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.543820e+03 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.543820e+03 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 3.541329e+03 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.544511e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.544511e+03 ) sec^-1 MeanMatrixElemValue = ( 4.063123e+00 +- 2.368970e+00 ) GeV^-4 -TOTAL : 4.639000 sec +TOTAL : 4.637339 sec INFO: No Floating Point Exceptions have been reported - 13,090,618,968 cycles # 2.820 GHz - 39,561,040,325 instructions # 3.02 insn per cycle - 4.644193645 seconds time elapsed + 13,073,829,228 cycles # 2.817 GHz + 39,561,226,768 instructions # 3.03 insn per cycle + 4.641846645 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4:13192) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.sse4_d_inl0_hrd0/runTest_cpp.exe @@ -164,15 +164,15 @@ Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=0] [ Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('avx2': AVX2, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 8.087246e+03 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 8.103223e+03 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 8.103223e+03 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 8.049714e+03 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 8.065612e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 8.065612e+03 ) sec^-1 MeanMatrixElemValue = ( 4.063123e+00 +- 2.368970e+00 ) GeV^-4 -TOTAL : 2.035017 sec +TOTAL : 2.043838 sec INFO: No Floating Point Exceptions have been reported - 5,608,597,608 cycles # 2.752 GHz - 13,825,354,537 instructions # 2.47 insn per cycle - 2.039075619 seconds time elapsed + 5,608,774,799 cycles # 2.740 GHz + 13,823,654,205 instructions # 2.46 insn per cycle + 2.048457884 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2:11520) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.avx2_d_inl0_hrd0/runTest_cpp.exe @@ -196,15 +196,15 @@ Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=0] [ Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('512y': AVX512, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 9.190120e+03 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 9.211201e+03 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 9.211201e+03 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 9.203409e+03 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 9.225011e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 9.225011e+03 ) sec^-1 MeanMatrixElemValue = ( 4.063123e+00 +- 2.368970e+00 ) GeV^-4 -TOTAL : 1.791765 sec +TOTAL : 1.788627 sec INFO: No Floating Point Exceptions have been reported - 4,921,067,926 cycles # 2.743 GHz - 12,507,200,724 instructions # 2.54 insn per cycle - 1.798123347 seconds time elapsed + 4,914,019,688 cycles # 2.742 GHz + 12,505,199,387 instructions # 2.54 insn per cycle + 1.793117412 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2:10439) (512y: 89) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.512y_d_inl0_hrd0/runTest_cpp.exe @@ -228,15 +228,15 @@ Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=0] [ Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('512z': AVX512, 512bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 7.012553e+03 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 7.024911e+03 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 7.024911e+03 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 6.991432e+03 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 7.003964e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 7.003964e+03 ) sec^-1 MeanMatrixElemValue = ( 4.063123e+00 +- 2.368970e+00 ) GeV^-4 -TOTAL : 2.347251 sec +TOTAL : 2.352089 sec INFO: No Floating Point Exceptions have been reported - 4,147,263,675 cycles # 1.765 GHz - 6,394,266,077 instructions # 1.54 insn per cycle - 2.352573303 seconds time elapsed + 4,140,389,696 cycles # 1.758 GHz + 6,392,460,964 instructions # 1.54 insn per cycle + 2.356471090 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 1978) (512y: 101) (512z: 9386) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.512z_d_inl0_hrd0/runTest_cpp.exe diff --git a/epochX/cudacpp/tput/logs_ggttgg_mad/log_ggttgg_mad_d_inl0_hrd0_bridge.txt b/epochX/cudacpp/tput/logs_ggttgg_mad/log_ggttgg_mad_d_inl0_hrd0_bridge.txt index bc0987eea5..382729ebe0 100644 --- a/epochX/cudacpp/tput/logs_ggttgg_mad/log_ggttgg_mad_d_inl0_hrd0_bridge.txt +++ b/epochX/cudacpp/tput/logs_ggttgg_mad/log_ggttgg_mad_d_inl0_hrd0_bridge.txt @@ -40,7 +40,7 @@ make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg' -DATE: 2024-10-06_09:39:37 +DATE: 2024-10-06_13:58:09 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: ========================================================================= @@ -53,16 +53,16 @@ WARNING! Set grid in Bridge (nevt=16384, gpublocks=64, gputhreads=256, gpublocks Process = SIGMA_SM_GG_TTXGG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 2.954093e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.263620e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.263620e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 2.969810e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.274652e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.274652e+05 ) sec^-1 MeanMatrixElemValue = ( 4.063123e+00 +- 2.368970e+00 ) GeV^-4 -TOTAL : 0.526732 sec +TOTAL : 0.528129 sec INFO: No Floating Point Exceptions have been reported INFO: No Floating Point Exceptions have been reported - 2,227,837,882 cycles # 2.883 GHz - 3,476,505,124 instructions # 1.56 insn per cycle - 0.832118305 seconds time elapsed + 2,217,669,368 cycles # 2.887 GHz + 3,446,589,857 instructions # 1.55 insn per cycle + 0.829571226 seconds time elapsed runNcu /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.cuda_d_inl0_hrd0/check_cuda.exe -p 64 256 1 --bridge WARNING! Bridge selected: cannot use RamboDevice, will use RamboHost WARNING! RamboHost selected: cannot use CurandDevice, will use CurandHost @@ -80,16 +80,16 @@ WARNING! Set grid in Bridge (nevt=524288, gpublocks=2048, gputhreads=256, gpublo Process = SIGMA_SM_GG_TTXGG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 3.643761e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 4.124122e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 4.124122e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 3.639238e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 4.123938e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 4.123938e+05 ) sec^-1 MeanMatrixElemValue = ( 6.665112e+00 +- 5.002651e+00 ) GeV^-4 -TOTAL : 3.301805 sec +TOTAL : 3.307492 sec INFO: No Floating Point Exceptions have been reported INFO: No Floating Point Exceptions have been reported - 10,501,615,955 cycles # 2.935 GHz - 23,489,948,913 instructions # 2.24 insn per cycle - 3.634545913 seconds time elapsed + 10,419,982,735 cycles # 2.912 GHz + 22,532,459,235 instructions # 2.16 insn per cycle + 3.638269314 seconds time elapsed ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.cuda_d_inl0_hrd0/runTest_cuda.exe INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: FE_DIVBYZERO, FE_INVALID, FE_OVERFLOW @@ -115,16 +115,16 @@ Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=0] [ Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+BRDHST/none+NAVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = SCALAR ('none': ~vector[1], no SIMD) -EvtsPerSec[Rmb+ME] (23) = ( 1.879294e+03 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.880182e+03 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.880182e+03 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.886173e+03 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.887135e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.887135e+03 ) sec^-1 MeanMatrixElemValue = ( 4.063123e+00 +- 2.368970e+00 ) GeV^-4 -TOTAL : 8.737845 sec +TOTAL : 8.706036 sec INFO: No Floating Point Exceptions have been reported INFO: No Floating Point Exceptions have been reported - 25,658,913,414 cycles # 2.936 GHz - 78,963,594,343 instructions # 3.08 insn per cycle - 8.742435740 seconds time elapsed + 25,672,868,709 cycles # 2.948 GHz + 78,967,719,116 instructions # 3.08 insn per cycle + 8.710738056 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 4842) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.none_d_inl0_hrd0/runTest_cpp.exe @@ -149,16 +149,16 @@ Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=0] [ Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+BRDHST/sse4+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[2] ('sse4': SSE4.2, 128bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 3.518464e+03 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.521735e+03 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.521735e+03 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 3.552148e+03 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.555428e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.555428e+03 ) sec^-1 MeanMatrixElemValue = ( 4.063123e+00 +- 2.368970e+00 ) GeV^-4 -TOTAL : 4.671849 sec +TOTAL : 4.627684 sec INFO: No Floating Point Exceptions have been reported INFO: No Floating Point Exceptions have been reported - 13,102,544,659 cycles # 2.802 GHz - 39,572,381,519 instructions # 3.02 insn per cycle - 4.676455621 seconds time elapsed + 13,068,501,146 cycles # 2.822 GHz + 39,571,959,778 instructions # 3.03 insn per cycle + 4.632273504 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4:13192) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.sse4_d_inl0_hrd0/runTest_cpp.exe @@ -183,16 +183,16 @@ Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=0] [ Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+BRDHST/avx2+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('avx2': AVX2, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 8.057114e+03 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 8.073561e+03 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 8.073561e+03 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 8.048761e+03 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 8.065809e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 8.065809e+03 ) sec^-1 MeanMatrixElemValue = ( 4.063123e+00 +- 2.368970e+00 ) GeV^-4 -TOTAL : 2.046600 sec +TOTAL : 2.048497 sec INFO: No Floating Point Exceptions have been reported INFO: No Floating Point Exceptions have been reported - 5,627,314,455 cycles # 2.744 GHz - 13,834,298,777 instructions # 2.46 insn per cycle - 2.051219882 seconds time elapsed + 5,627,660,910 cycles # 2.742 GHz + 13,834,355,033 instructions # 2.46 insn per cycle + 2.053207635 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2:11520) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.avx2_d_inl0_hrd0/runTest_cpp.exe @@ -217,16 +217,16 @@ Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=0] [ Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+BRDHST/512y+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('512y': AVX512, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 9.239341e+03 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 9.261385e+03 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 9.261385e+03 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 9.125680e+03 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 9.147272e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 9.147272e+03 ) sec^-1 MeanMatrixElemValue = ( 4.063123e+00 +- 2.368970e+00 ) GeV^-4 -TOTAL : 1.786219 sec +TOTAL : 1.808173 sec INFO: No Floating Point Exceptions have been reported INFO: No Floating Point Exceptions have been reported - 4,931,565,389 cycles # 2.756 GHz - 12,515,991,121 instructions # 2.54 insn per cycle - 1.790909503 seconds time elapsed + 4,963,577,928 cycles # 2.739 GHz + 12,516,060,761 instructions # 2.52 insn per cycle + 1.812767934 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2:10439) (512y: 89) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.512y_d_inl0_hrd0/runTest_cpp.exe @@ -251,16 +251,16 @@ Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=0] [ Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+BRDHST/512z+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('512z': AVX512, 512bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 7.038188e+03 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 7.051446e+03 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 7.051446e+03 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 7.003148e+03 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 7.016489e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 7.016489e+03 ) sec^-1 MeanMatrixElemValue = ( 4.063123e+00 +- 2.368970e+00 ) GeV^-4 -TOTAL : 2.341272 sec +TOTAL : 2.353407 sec INFO: No Floating Point Exceptions have been reported INFO: No Floating Point Exceptions have been reported - 4,150,945,217 cycles # 1.770 GHz - 6,403,675,117 instructions # 1.54 insn per cycle - 2.345955468 seconds time elapsed + 4,152,986,591 cycles # 1.763 GHz + 6,403,901,885 instructions # 1.54 insn per cycle + 2.358083297 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 1978) (512y: 101) (512z: 9386) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.512z_d_inl0_hrd0/runTest_cpp.exe diff --git a/epochX/cudacpp/tput/logs_ggttgg_mad/log_ggttgg_mad_d_inl0_hrd0_common.txt b/epochX/cudacpp/tput/logs_ggttgg_mad/log_ggttgg_mad_d_inl0_hrd0_common.txt index be31042fc1..848eedc051 100644 --- a/epochX/cudacpp/tput/logs_ggttgg_mad/log_ggttgg_mad_d_inl0_hrd0_common.txt +++ b/epochX/cudacpp/tput/logs_ggttgg_mad/log_ggttgg_mad_d_inl0_hrd0_common.txt @@ -40,7 +40,7 @@ make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg' -DATE: 2024-10-06_09:50:33 +DATE: 2024-10-06_14:09:05 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: ========================================================================= @@ -49,15 +49,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_SM_GG_TTXGG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:COMMON+RMBDEV+MESDEV/none+NAVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 3.314159e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.339458e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.341417e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 3.295842e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.323377e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.325402e+05 ) sec^-1 MeanMatrixElemValue = ( 4.197467e-01 +- 3.250467e-01 ) GeV^-4 -TOTAL : 0.518126 sec +TOTAL : 0.521241 sec INFO: No Floating Point Exceptions have been reported - 2,164,802,026 cycles # 2.881 GHz - 3,409,915,390 instructions # 1.58 insn per cycle - 0.811338657 seconds time elapsed + 2,169,446,983 cycles # 2.882 GHz + 3,300,027,635 instructions # 1.52 insn per cycle + 0.813489760 seconds time elapsed runNcu /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.cuda_d_inl0_hrd0/check_cuda.exe -p 64 256 1 --common ==PROF== Profiling "sigmaKin": launch__registers_per_thread 255 ==PROF== Profiling "sigmaKin": sm__sass_average_branch_targets_threads_uniform.pct 100% @@ -67,15 +67,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_SM_GG_TTXGG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:COMMON+RMBDEV+MESDEV/none+NAVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 4.134613e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 4.165487e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 4.166746e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 4.131685e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 4.162971e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 4.164228e+05 ) sec^-1 MeanMatrixElemValue = ( 1.252232e+02 +- 1.234346e+02 ) GeV^-4 -TOTAL : 3.140406 sec +TOTAL : 3.146191 sec INFO: No Floating Point Exceptions have been reported - 9,973,053,404 cycles # 2.934 GHz - 20,986,544,572 instructions # 2.10 insn per cycle - 3.455765313 seconds time elapsed + 9,988,761,971 cycles # 2.934 GHz + 20,928,451,976 instructions # 2.10 insn per cycle + 3.460675886 seconds time elapsed ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.cuda_d_inl0_hrd0/runTest_cuda.exe INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: FE_DIVBYZERO, FE_INVALID, FE_OVERFLOW @@ -100,15 +100,15 @@ Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=0] [ Workflow summary = CPP:DBL+CXS:COMMON+RMBHST+MESHST/none+NAVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = SCALAR ('none': ~vector[1], no SIMD) -EvtsPerSec[Rmb+ME] (23) = ( 1.884135e+03 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.885033e+03 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.885033e+03 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.886941e+03 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.887829e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.887829e+03 ) sec^-1 MeanMatrixElemValue = ( 4.197467e-01 +- 3.250467e-01 ) GeV^-4 -TOTAL : 8.712529 sec +TOTAL : 8.699641 sec INFO: No Floating Point Exceptions have been reported - 25,691,717,185 cycles # 2.948 GHz - 78,960,325,856 instructions # 3.07 insn per cycle - 8.716734440 seconds time elapsed + 25,666,417,950 cycles # 2.949 GHz + 78,955,658,367 instructions # 3.08 insn per cycle + 8.703834790 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 4842) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.none_d_inl0_hrd0/runTest_cpp.exe @@ -132,15 +132,15 @@ Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=0] [ Workflow summary = CPP:DBL+CXS:COMMON+RMBHST+MESHST/sse4+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[2] ('sse4': SSE4.2, 128bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 3.543458e+03 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.546697e+03 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.546697e+03 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 3.510639e+03 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.513904e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.513904e+03 ) sec^-1 MeanMatrixElemValue = ( 4.197467e-01 +- 3.250467e-01 ) GeV^-4 -TOTAL : 4.636367 sec +TOTAL : 4.679389 sec INFO: No Floating Point Exceptions have been reported - 13,067,183,546 cycles # 2.816 GHz - 39,558,454,763 instructions # 3.03 insn per cycle - 4.640590687 seconds time elapsed + 13,056,476,216 cycles # 2.788 GHz + 39,559,065,466 instructions # 3.03 insn per cycle + 4.683623585 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4:13192) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.sse4_d_inl0_hrd0/runTest_cpp.exe @@ -164,15 +164,15 @@ Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=0] [ Workflow summary = CPP:DBL+CXS:COMMON+RMBHST+MESHST/avx2+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('avx2': AVX2, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 8.084806e+03 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 8.101064e+03 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 8.101064e+03 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 8.053459e+03 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 8.070554e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 8.070554e+03 ) sec^-1 MeanMatrixElemValue = ( 4.197467e-01 +- 3.250467e-01 ) GeV^-4 -TOTAL : 2.036679 sec +TOTAL : 2.044484 sec INFO: No Floating Point Exceptions have been reported - 5,613,470,524 cycles # 2.752 GHz - 13,823,796,455 instructions # 2.46 insn per cycle - 2.040900437 seconds time elapsed + 5,613,396,251 cycles # 2.741 GHz + 13,822,782,179 instructions # 2.46 insn per cycle + 2.048768248 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2:11520) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.avx2_d_inl0_hrd0/runTest_cpp.exe @@ -196,15 +196,15 @@ Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=0] [ Workflow summary = CPP:DBL+CXS:COMMON+RMBHST+MESHST/512y+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('512y': AVX512, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 9.198723e+03 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 9.219905e+03 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 9.219905e+03 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 9.215245e+03 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 9.237589e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 9.237589e+03 ) sec^-1 MeanMatrixElemValue = ( 4.197467e-01 +- 3.250467e-01 ) GeV^-4 -TOTAL : 1.791160 sec +TOTAL : 1.787793 sec INFO: No Floating Point Exceptions have been reported - 4,922,288,820 cycles # 2.743 GHz - 12,503,388,745 instructions # 2.54 insn per cycle - 1.795321275 seconds time elapsed + 4,924,768,938 cycles # 2.749 GHz + 12,503,183,314 instructions # 2.54 insn per cycle + 1.792244763 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2:10439) (512y: 89) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.512y_d_inl0_hrd0/runTest_cpp.exe @@ -228,15 +228,15 @@ Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=0] [ Workflow summary = CPP:DBL+CXS:COMMON+RMBHST+MESHST/512z+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('512z': AVX512, 512bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 6.975365e+03 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 6.987686e+03 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 6.987686e+03 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 6.982861e+03 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 6.995773e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 6.995773e+03 ) sec^-1 MeanMatrixElemValue = ( 4.197467e-01 +- 3.250467e-01 ) GeV^-4 -TOTAL : 2.359532 sec +TOTAL : 2.356930 sec INFO: No Floating Point Exceptions have been reported - 4,155,009,705 cycles # 1.759 GHz - 6,390,945,346 instructions # 1.54 insn per cycle - 2.363732897 seconds time elapsed + 4,141,627,443 cycles # 1.755 GHz + 6,390,437,057 instructions # 1.54 insn per cycle + 2.361150566 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 1978) (512y: 101) (512z: 9386) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.512z_d_inl0_hrd0/runTest_cpp.exe diff --git a/epochX/cudacpp/tput/logs_ggttgg_mad/log_ggttgg_mad_d_inl0_hrd0_curhst.txt b/epochX/cudacpp/tput/logs_ggttgg_mad/log_ggttgg_mad_d_inl0_hrd0_curhst.txt index 437b6b7cbd..ec0085ca9b 100644 --- a/epochX/cudacpp/tput/logs_ggttgg_mad/log_ggttgg_mad_d_inl0_hrd0_curhst.txt +++ b/epochX/cudacpp/tput/logs_ggttgg_mad/log_ggttgg_mad_d_inl0_hrd0_curhst.txt @@ -40,7 +40,7 @@ make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg' -DATE: 2024-10-06_09:47:41 +DATE: 2024-10-06_14:06:14 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: ========================================================================= @@ -49,15 +49,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_SM_GG_TTXGG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURHST+RMBDEV+MESDEV/none+NAVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 3.310053e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.334627e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.336677e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 3.309236e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.335080e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.336806e+05 ) sec^-1 MeanMatrixElemValue = ( 4.063123e+00 +- 2.368970e+00 ) GeV^-4 -TOTAL : 0.518612 sec +TOTAL : 0.517025 sec INFO: No Floating Point Exceptions have been reported - 2,156,837,380 cycles # 2.875 GHz - 3,433,389,555 instructions # 1.59 insn per cycle - 0.811650542 seconds time elapsed + 2,167,534,276 cycles # 2.891 GHz + 3,417,640,132 instructions # 1.58 insn per cycle + 0.810003775 seconds time elapsed runNcu /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.cuda_d_inl0_hrd0/check_cuda.exe -p 64 256 1 --curhst ==PROF== Profiling "sigmaKin": launch__registers_per_thread 255 ==PROF== Profiling "sigmaKin": sm__sass_average_branch_targets_threads_uniform.pct 100% @@ -67,15 +67,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_SM_GG_TTXGG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURHST+RMBDEV+MESDEV/none+NAVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 4.128944e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 4.159258e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 4.160487e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 4.133971e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 4.164055e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 4.165319e+05 ) sec^-1 MeanMatrixElemValue = ( 6.665112e+00 +- 5.002651e+00 ) GeV^-4 -TOTAL : 3.091523 sec +TOTAL : 3.086230 sec INFO: No Floating Point Exceptions have been reported - 9,825,563,648 cycles # 2.933 GHz - 22,802,776,931 instructions # 2.32 insn per cycle - 3.405923259 seconds time elapsed + 9,804,284,274 cycles # 2.932 GHz + 22,348,157,560 instructions # 2.28 insn per cycle + 3.399717524 seconds time elapsed ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.cuda_d_inl0_hrd0/runTest_cuda.exe INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: FE_DIVBYZERO, FE_INVALID, FE_OVERFLOW @@ -100,15 +100,15 @@ Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=0] [ Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/none+NAVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = SCALAR ('none': ~vector[1], no SIMD) -EvtsPerSec[Rmb+ME] (23) = ( 1.890035e+03 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.890938e+03 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.890938e+03 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.886300e+03 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.887246e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.887246e+03 ) sec^-1 MeanMatrixElemValue = ( 4.063123e+00 +- 2.368970e+00 ) GeV^-4 -TOTAL : 8.683864 sec +TOTAL : 8.700943 sec INFO: No Floating Point Exceptions have been reported - 25,635,022,031 cycles # 2.951 GHz - 78,960,809,140 instructions # 3.08 insn per cycle - 8.688143049 seconds time elapsed + 25,644,603,592 cycles # 2.946 GHz + 78,955,601,424 instructions # 3.08 insn per cycle + 8.705235584 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 4842) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.none_d_inl0_hrd0/runTest_cpp.exe @@ -132,15 +132,15 @@ Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=0] [ Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[2] ('sse4': SSE4.2, 128bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 3.535619e+03 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.538805e+03 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.538805e+03 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 3.545819e+03 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.549127e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.549127e+03 ) sec^-1 MeanMatrixElemValue = ( 4.063123e+00 +- 2.368970e+00 ) GeV^-4 -TOTAL : 4.644682 sec +TOTAL : 4.631284 sec INFO: No Floating Point Exceptions have been reported - 13,070,212,228 cycles # 2.812 GHz - 39,558,910,913 instructions # 3.03 insn per cycle - 4.648863484 seconds time elapsed + 13,056,543,557 cycles # 2.817 GHz + 39,558,712,514 instructions # 3.03 insn per cycle + 4.635447592 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4:13192) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.sse4_d_inl0_hrd0/runTest_cpp.exe @@ -164,15 +164,15 @@ Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=0] [ Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('avx2': AVX2, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 7.974136e+03 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 7.989764e+03 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 7.989764e+03 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 8.092118e+03 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 8.108059e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 8.108059e+03 ) sec^-1 MeanMatrixElemValue = ( 4.063123e+00 +- 2.368970e+00 ) GeV^-4 -TOTAL : 2.062978 sec +TOTAL : 2.033326 sec INFO: No Floating Point Exceptions have been reported - 5,609,565,523 cycles # 2.715 GHz - 13,823,736,601 instructions # 2.46 insn per cycle - 2.067208066 seconds time elapsed + 5,607,844,649 cycles # 2.753 GHz + 13,823,687,936 instructions # 2.47 insn per cycle + 2.037582485 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2:11520) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.avx2_d_inl0_hrd0/runTest_cpp.exe @@ -196,15 +196,15 @@ Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=0] [ Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('512y': AVX512, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 9.256862e+03 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 9.278276e+03 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 9.278276e+03 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 9.232566e+03 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 9.253794e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 9.253794e+03 ) sec^-1 MeanMatrixElemValue = ( 4.063123e+00 +- 2.368970e+00 ) GeV^-4 -TOTAL : 1.778135 sec +TOTAL : 1.782861 sec INFO: No Floating Point Exceptions have been reported - 4,913,104,520 cycles # 2.758 GHz - 12,505,156,898 instructions # 2.55 insn per cycle - 1.782374042 seconds time elapsed + 4,917,381,373 cycles # 2.753 GHz + 12,505,310,983 instructions # 2.54 insn per cycle + 1.787008671 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2:10439) (512y: 89) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.512y_d_inl0_hrd0/runTest_cpp.exe @@ -228,15 +228,15 @@ Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=0] [ Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('512z': AVX512, 512bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 7.040533e+03 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 7.053211e+03 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 7.053211e+03 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 6.996576e+03 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 7.008750e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 7.008750e+03 ) sec^-1 MeanMatrixElemValue = ( 4.063123e+00 +- 2.368970e+00 ) GeV^-4 -TOTAL : 2.335968 sec +TOTAL : 2.350739 sec INFO: No Floating Point Exceptions have been reported - 4,137,289,106 cycles # 1.769 GHz - 6,392,511,975 instructions # 1.55 insn per cycle - 2.340416062 seconds time elapsed + 4,139,775,312 cycles # 1.759 GHz + 6,392,598,672 instructions # 1.54 insn per cycle + 2.355091216 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 1978) (512y: 101) (512z: 9386) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.512z_d_inl0_hrd0/runTest_cpp.exe diff --git a/epochX/cudacpp/tput/logs_ggttgg_mad/log_ggttgg_mad_d_inl0_hrd0_rmbhst.txt b/epochX/cudacpp/tput/logs_ggttgg_mad/log_ggttgg_mad_d_inl0_hrd0_rmbhst.txt index f2b15e4b6f..b67da1a0f7 100644 --- a/epochX/cudacpp/tput/logs_ggttgg_mad/log_ggttgg_mad_d_inl0_hrd0_rmbhst.txt +++ b/epochX/cudacpp/tput/logs_ggttgg_mad/log_ggttgg_mad_d_inl0_hrd0_rmbhst.txt @@ -40,7 +40,7 @@ make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg' -DATE: 2024-10-06_09:44:55 +DATE: 2024-10-06_14:03:28 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: ========================================================================= @@ -50,15 +50,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_SM_GG_TTXGG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURHST+RMBHST+MESDEV/none+NAVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 3.041462e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.325366e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.327398e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 3.053632e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.327712e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.329530e+05 ) sec^-1 MeanMatrixElemValue = ( 4.063123e+00 +- 2.368970e+00 ) GeV^-4 -TOTAL : 0.520118 sec +TOTAL : 0.521962 sec INFO: No Floating Point Exceptions have been reported - 2,177,158,293 cycles # 2.891 GHz - 3,464,316,990 instructions # 1.59 insn per cycle - 0.812097316 seconds time elapsed + 2,171,071,442 cycles # 2.875 GHz + 3,455,195,557 instructions # 1.59 insn per cycle + 0.814743016 seconds time elapsed runNcu /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.cuda_d_inl0_hrd0/check_cuda.exe -p 64 256 1 --rmbhst WARNING! RamboHost selected: cannot use CurandDevice, will use CurandHost ==PROF== Profiling "sigmaKin": launch__registers_per_thread 255 @@ -70,15 +70,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_SM_GG_TTXGG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURHST+RMBHST+MESDEV/none+NAVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 3.734798e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 4.174453e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 4.175668e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 3.729651e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 4.166908e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 4.168126e+05 ) sec^-1 MeanMatrixElemValue = ( 6.665112e+00 +- 5.002651e+00 ) GeV^-4 -TOTAL : 3.213650 sec +TOTAL : 3.218762 sec INFO: No Floating Point Exceptions have been reported - 10,150,922,529 cycles # 2.918 GHz - 23,231,659,490 instructions # 2.29 insn per cycle - 3.538737264 seconds time elapsed + 10,193,756,505 cycles # 2.934 GHz + 23,404,941,253 instructions # 2.30 insn per cycle + 3.533628132 seconds time elapsed ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.cuda_d_inl0_hrd0/runTest_cuda.exe INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: FE_DIVBYZERO, FE_INVALID, FE_OVERFLOW @@ -103,15 +103,15 @@ Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=0] [ Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/none+NAVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = SCALAR ('none': ~vector[1], no SIMD) -EvtsPerSec[Rmb+ME] (23) = ( 1.885407e+03 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.886309e+03 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.886309e+03 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.887063e+03 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.887974e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.887974e+03 ) sec^-1 MeanMatrixElemValue = ( 4.063123e+00 +- 2.368970e+00 ) GeV^-4 -TOTAL : 8.705137 sec +TOTAL : 8.697508 sec INFO: No Floating Point Exceptions have been reported - 25,650,530,800 cycles # 2.946 GHz - 78,960,008,246 instructions # 3.08 insn per cycle - 8.709419634 seconds time elapsed + 25,667,274,636 cycles # 2.950 GHz + 78,960,080,977 instructions # 3.08 insn per cycle + 8.701718616 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 4842) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.none_d_inl0_hrd0/runTest_cpp.exe @@ -135,15 +135,15 @@ Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=0] [ Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[2] ('sse4': SSE4.2, 128bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 3.551750e+03 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.554937e+03 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.554937e+03 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 3.536997e+03 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.540210e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.540210e+03 ) sec^-1 MeanMatrixElemValue = ( 4.063123e+00 +- 2.368970e+00 ) GeV^-4 -TOTAL : 4.623453 sec +TOTAL : 4.643051 sec INFO: No Floating Point Exceptions have been reported - 13,056,946,389 cycles # 2.822 GHz - 39,559,090,760 instructions # 3.03 insn per cycle - 4.627712527 seconds time elapsed + 13,084,389,837 cycles # 2.816 GHz + 39,561,543,889 instructions # 3.02 insn per cycle + 4.647450943 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4:13192) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.sse4_d_inl0_hrd0/runTest_cpp.exe @@ -167,15 +167,15 @@ Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=0] [ Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('avx2': AVX2, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 8.090893e+03 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 8.106933e+03 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 8.106933e+03 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 8.101047e+03 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 8.117126e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 8.117126e+03 ) sec^-1 MeanMatrixElemValue = ( 4.063123e+00 +- 2.368970e+00 ) GeV^-4 -TOTAL : 2.033338 sec +TOTAL : 2.031056 sec INFO: No Floating Point Exceptions have been reported - 5,609,780,879 cycles # 2.754 GHz - 13,824,722,765 instructions # 2.46 insn per cycle - 2.037509617 seconds time elapsed + 5,611,644,593 cycles # 2.758 GHz + 13,823,538,186 instructions # 2.46 insn per cycle + 2.035370310 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2:11520) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.avx2_d_inl0_hrd0/runTest_cpp.exe @@ -199,15 +199,15 @@ Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=0] [ Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('512y': AVX512, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 9.188897e+03 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 9.209893e+03 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 9.209893e+03 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 9.204059e+03 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 9.225405e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 9.225405e+03 ) sec^-1 MeanMatrixElemValue = ( 4.063123e+00 +- 2.368970e+00 ) GeV^-4 -TOTAL : 1.791081 sec +TOTAL : 1.788146 sec INFO: No Floating Point Exceptions have been reported - 4,916,057,270 cycles # 2.740 GHz - 12,505,186,935 instructions # 2.54 insn per cycle - 1.795355106 seconds time elapsed + 4,921,309,454 cycles # 2.747 GHz + 12,505,223,055 instructions # 2.54 insn per cycle + 1.792497523 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2:10439) (512y: 89) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.512y_d_inl0_hrd0/runTest_cpp.exe @@ -231,15 +231,15 @@ Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=0] [ Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('512z': AVX512, 512bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 7.019116e+03 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 7.031683e+03 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 7.031683e+03 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 6.986413e+03 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 6.998530e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 6.998530e+03 ) sec^-1 MeanMatrixElemValue = ( 4.063123e+00 +- 2.368970e+00 ) GeV^-4 -TOTAL : 2.343107 sec +TOTAL : 2.354190 sec INFO: No Floating Point Exceptions have been reported - 4,136,898,273 cycles # 1.763 GHz - 6,392,336,539 instructions # 1.55 insn per cycle - 2.347534329 seconds time elapsed + 4,140,634,558 cycles # 1.756 GHz + 6,393,482,085 instructions # 1.54 insn per cycle + 2.358557003 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 1978) (512y: 101) (512z: 9386) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.512z_d_inl0_hrd0/runTest_cpp.exe diff --git a/epochX/cudacpp/tput/logs_ggttgg_mad/log_ggttgg_mad_d_inl0_hrd1.txt b/epochX/cudacpp/tput/logs_ggttgg_mad/log_ggttgg_mad_d_inl0_hrd1.txt index 99e413a8a3..982ffe1b2b 100644 --- a/epochX/cudacpp/tput/logs_ggttgg_mad/log_ggttgg_mad_d_inl0_hrd1.txt +++ b/epochX/cudacpp/tput/logs_ggttgg_mad/log_ggttgg_mad_d_inl0_hrd1.txt @@ -40,7 +40,7 @@ make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg' -DATE: 2024-10-06_09:03:14 +DATE: 2024-10-06_13:32:25 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: ========================================================================= @@ -49,15 +49,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_SM_GG_TTXGG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=1] Workflow summary = CUD:DBL+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 3.332738e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.357821e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.359802e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 3.327807e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.349266e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.350964e+05 ) sec^-1 MeanMatrixElemValue = ( 4.063123e+00 +- 2.368970e+00 ) GeV^-4 -TOTAL : 0.542209 sec +TOTAL : 0.529277 sec INFO: No Floating Point Exceptions have been reported - 2,220,139,727 cycles # 2.875 GHz - 3,465,138,857 instructions # 1.56 insn per cycle - 0.835706398 seconds time elapsed + 2,200,344,284 cycles # 2.877 GHz + 3,486,749,682 instructions # 1.58 insn per cycle + 0.822133970 seconds time elapsed runNcu /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.cuda_d_inl0_hrd1/check_cuda.exe -p 64 256 1 ==PROF== Profiling "sigmaKin": launch__registers_per_thread 255 ==PROF== Profiling "sigmaKin": sm__sass_average_branch_targets_threads_uniform.pct 100% @@ -67,15 +67,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_SM_GG_TTXGG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=1] Workflow summary = CUD:DBL+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 4.145716e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 4.176488e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 4.177708e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 4.150063e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 4.178151e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 4.179340e+05 ) sec^-1 MeanMatrixElemValue = ( 6.665112e+00 +- 5.002651e+00 ) GeV^-4 -TOTAL : 3.039240 sec +TOTAL : 3.039653 sec INFO: No Floating Point Exceptions have been reported - 9,630,090,535 cycles # 2.918 GHz - 21,945,170,652 instructions # 2.28 insn per cycle - 3.356721463 seconds time elapsed + 9,648,174,573 cycles # 2.918 GHz + 21,910,579,255 instructions # 2.27 insn per cycle + 3.364182887 seconds time elapsed ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.cuda_d_inl0_hrd1/runTest_cuda.exe INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: FE_DIVBYZERO, FE_INVALID, FE_OVERFLOW @@ -100,15 +100,15 @@ Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=0] [ Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/none+NAVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = SCALAR ('none': ~vector[1], no SIMD) -EvtsPerSec[Rmb+ME] (23) = ( 1.881580e+03 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.882499e+03 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.882499e+03 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.889919e+03 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.890821e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.890821e+03 ) sec^-1 MeanMatrixElemValue = ( 4.063123e+00 +- 2.368970e+00 ) GeV^-4 -TOTAL : 8.723377 sec +TOTAL : 8.683861 sec INFO: No Floating Point Exceptions have been reported - 25,611,709,249 cycles # 2.935 GHz - 78,703,444,126 instructions # 3.07 insn per cycle - 8.727502935 seconds time elapsed + 25,598,543,536 cycles # 2.947 GHz + 78,701,379,367 instructions # 3.07 insn per cycle + 8.688058225 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 4191) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.none_d_inl0_hrd1/runTest_cpp.exe @@ -132,15 +132,15 @@ Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=0] [ Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[2] ('sse4': SSE4.2, 128bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 3.593581e+03 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.596889e+03 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.596889e+03 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 3.558251e+03 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.561555e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.561555e+03 ) sec^-1 MeanMatrixElemValue = ( 4.063123e+00 +- 2.368970e+00 ) GeV^-4 -TOTAL : 4.571814 sec +TOTAL : 4.614915 sec INFO: No Floating Point Exceptions have been reported - 13,039,592,628 cycles # 2.851 GHz - 39,453,086,877 instructions # 3.03 insn per cycle - 4.575893049 seconds time elapsed + 13,024,005,332 cycles # 2.820 GHz + 39,449,715,983 instructions # 3.03 insn per cycle + 4.619074440 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4:12966) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.sse4_d_inl0_hrd1/runTest_cpp.exe @@ -164,15 +164,15 @@ Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=0] [ Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('avx2': AVX2, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 7.986878e+03 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 8.003760e+03 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 8.003760e+03 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 7.985523e+03 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 8.001224e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 8.001224e+03 ) sec^-1 MeanMatrixElemValue = ( 4.063123e+00 +- 2.368970e+00 ) GeV^-4 -TOTAL : 2.061484 sec +TOTAL : 2.060022 sec INFO: No Floating Point Exceptions have been reported - 5,673,128,561 cycles # 2.749 GHz - 13,911,820,426 instructions # 2.45 insn per cycle - 2.066505881 seconds time elapsed + 5,683,179,619 cycles # 2.754 GHz + 13,909,934,258 instructions # 2.45 insn per cycle + 2.064355920 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2:11582) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.avx2_d_inl0_hrd1/runTest_cpp.exe @@ -196,15 +196,15 @@ Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=0] [ Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('512y': AVX512, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 9.098916e+03 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 9.119150e+03 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 9.119150e+03 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 8.875636e+03 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 8.896102e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 8.896102e+03 ) sec^-1 MeanMatrixElemValue = ( 4.063123e+00 +- 2.368970e+00 ) GeV^-4 -TOTAL : 1.809563 sec +TOTAL : 1.853744 sec INFO: No Floating Point Exceptions have been reported - 4,990,015,585 cycles # 2.753 GHz - 12,604,471,256 instructions # 2.53 insn per cycle - 1.813650628 seconds time elapsed + 4,995,716,178 cycles # 2.690 GHz + 12,603,883,841 instructions # 2.52 insn per cycle + 1.858023648 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2:10423) (512y: 241) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.512y_d_inl0_hrd1/runTest_cpp.exe @@ -228,15 +228,15 @@ Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=0] [ Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('512z': AVX512, 512bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 6.910207e+03 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 6.922434e+03 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 6.922434e+03 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 6.954217e+03 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 6.966257e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 6.966257e+03 ) sec^-1 MeanMatrixElemValue = ( 4.063123e+00 +- 2.368970e+00 ) GeV^-4 -TOTAL : 2.380650 sec +TOTAL : 2.365078 sec INFO: No Floating Point Exceptions have been reported - 4,192,440,259 cycles # 1.759 GHz - 6,502,191,985 instructions # 1.55 insn per cycle - 2.384674618 seconds time elapsed + 4,167,405,055 cycles # 1.759 GHz + 6,500,256,546 instructions # 1.56 insn per cycle + 2.369538287 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 1754) (512y: 193) (512z: 9382) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.512z_d_inl0_hrd1/runTest_cpp.exe diff --git a/epochX/cudacpp/tput/logs_ggttgg_mad/log_ggttgg_mad_d_inl1_hrd0.txt b/epochX/cudacpp/tput/logs_ggttgg_mad/log_ggttgg_mad_d_inl1_hrd0.txt index 76362e2777..0672db26e5 100644 --- a/epochX/cudacpp/tput/logs_ggttgg_mad/log_ggttgg_mad_d_inl1_hrd0.txt +++ b/epochX/cudacpp/tput/logs_ggttgg_mad/log_ggttgg_mad_d_inl1_hrd0.txt @@ -40,7 +40,7 @@ make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg' -DATE: 2024-10-06_09:30:00 +DATE: 2024-10-06_13:48:30 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: ========================================================================= @@ -49,15 +49,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_SM_GG_TTXGG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=1] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 3.108959e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.129301e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.130870e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 3.109700e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.130557e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.132153e+05 ) sec^-1 MeanMatrixElemValue = ( 4.063123e+00 +- 2.368970e+00 ) GeV^-4 -TOTAL : 0.545749 sec +TOTAL : 0.540942 sec INFO: No Floating Point Exceptions have been reported - 2,205,865,001 cycles # 2.840 GHz - 3,412,138,367 instructions # 1.55 insn per cycle - 0.835130533 seconds time elapsed + 2,232,421,251 cycles # 2.885 GHz + 3,517,425,515 instructions # 1.58 insn per cycle + 0.830972815 seconds time elapsed runNcu /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.cuda_d_inl1_hrd0/check_cuda.exe -p 64 256 1 ==PROF== Profiling "sigmaKin": launch__registers_per_thread 255 ==PROF== Profiling "sigmaKin": sm__sass_average_branch_targets_threads_uniform.pct 100% @@ -67,15 +67,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_SM_GG_TTXGG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=1] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 3.747537e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.771352e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.772362e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 3.750926e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.774382e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.775350e+05 ) sec^-1 MeanMatrixElemValue = ( 6.665112e+00 +- 5.002651e+00 ) GeV^-4 -TOTAL : 3.317305 sec +TOTAL : 3.314015 sec INFO: No Floating Point Exceptions have been reported - 10,470,225,400 cycles # 2.928 GHz - 22,893,642,046 instructions # 2.19 insn per cycle - 3.632348979 seconds time elapsed + 10,492,307,516 cycles # 2.930 GHz + 24,086,185,836 instructions # 2.30 insn per cycle + 3.639042501 seconds time elapsed ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.cuda_d_inl1_hrd0/runTest_cuda.exe INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: FE_DIVBYZERO, FE_INVALID, FE_OVERFLOW @@ -100,15 +100,15 @@ Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=1] [ Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/none+NAVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = SCALAR ('none': ~vector[1], no SIMD) -EvtsPerSec[Rmb+ME] (23) = ( 4.279433e+02 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 4.279917e+02 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 4.279917e+02 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 4.272567e+02 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 4.273033e+02 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 4.273033e+02 ) sec^-1 MeanMatrixElemValue = ( 4.063123e+00 +- 2.368970e+00 ) GeV^-4 -TOTAL : 38.330200 sec +TOTAL : 38.390869 sec INFO: No Floating Point Exceptions have been reported - 112,786,835,820 cycles # 2.943 GHz - 144,812,254,859 instructions # 1.28 insn per cycle - 38.334547107 seconds time elapsed + 112,741,390,818 cycles # 2.937 GHz + 144,808,337,392 instructions # 1.28 insn per cycle + 38.395163410 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4:21273) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.none_d_inl1_hrd0/runTest_cpp.exe @@ -132,15 +132,15 @@ Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=1] [ Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[2] ('sse4': SSE4.2, 128bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 3.132336e+03 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.134792e+03 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.134792e+03 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 3.091053e+03 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.093529e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.093529e+03 ) sec^-1 MeanMatrixElemValue = ( 4.063123e+00 +- 2.368970e+00 ) GeV^-4 -TOTAL : 5.242571 sec +TOTAL : 5.311932 sec INFO: No Floating Point Exceptions have been reported - 14,761,048,074 cycles # 2.814 GHz - 37,609,615,991 instructions # 2.55 insn per cycle - 5.246531710 seconds time elapsed + 14,721,237,699 cycles # 2.770 GHz + 37,605,065,083 instructions # 2.55 insn per cycle + 5.316369398 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4:68172) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.sse4_d_inl1_hrd0/runTest_cpp.exe @@ -164,15 +164,15 @@ Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=1] [ Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('avx2': AVX2, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 7.367426e+03 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 7.381363e+03 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 7.381363e+03 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 7.397775e+03 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 7.411492e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 7.411492e+03 ) sec^-1 MeanMatrixElemValue = ( 4.063123e+00 +- 2.368970e+00 ) GeV^-4 -TOTAL : 2.233268 sec +TOTAL : 2.223441 sec INFO: No Floating Point Exceptions have been reported - 6,121,196,467 cycles # 2.737 GHz - 13,054,881,187 instructions # 2.13 insn per cycle - 2.237420808 seconds time elapsed + 6,117,147,586 cycles # 2.747 GHz + 13,053,000,272 instructions # 2.13 insn per cycle + 2.227901327 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2:46946) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.avx2_d_inl1_hrd0/runTest_cpp.exe @@ -196,15 +196,15 @@ Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=1] [ Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('512y': AVX512, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 8.964974e+03 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 8.985321e+03 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 8.985321e+03 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 8.937767e+03 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 8.958388e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 8.958388e+03 ) sec^-1 MeanMatrixElemValue = ( 4.063123e+00 +- 2.368970e+00 ) GeV^-4 -TOTAL : 1.836637 sec +TOTAL : 1.841322 sec INFO: No Floating Point Exceptions have been reported - 5,064,709,437 cycles # 2.753 GHz - 11,452,008,336 instructions # 2.26 insn per cycle - 1.840705951 seconds time elapsed + 5,072,280,397 cycles # 2.750 GHz + 11,450,236,584 instructions # 2.26 insn per cycle + 1.845652473 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2:40486) (512y: 285) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.512y_d_inl1_hrd0/runTest_cpp.exe @@ -228,15 +228,15 @@ Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=1] [ Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('512z': AVX512, 512bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 7.358991e+03 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 7.372760e+03 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 7.372760e+03 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 7.333122e+03 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 7.346477e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 7.346477e+03 ) sec^-1 MeanMatrixElemValue = ( 4.063123e+00 +- 2.368970e+00 ) GeV^-4 -TOTAL : 2.235964 sec +TOTAL : 2.243345 sec INFO: No Floating Point Exceptions have been reported - 3,956,538,826 cycles # 1.767 GHz - 5,928,749,634 instructions # 1.50 insn per cycle - 2.240037452 seconds time elapsed + 3,956,279,866 cycles # 1.761 GHz + 5,926,801,436 instructions # 1.50 insn per cycle + 2.247692965 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 2444) (512y: 337) (512z:39338) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.512z_d_inl1_hrd0/runTest_cpp.exe diff --git a/epochX/cudacpp/tput/logs_ggttgg_mad/log_ggttgg_mad_d_inl1_hrd1.txt b/epochX/cudacpp/tput/logs_ggttgg_mad/log_ggttgg_mad_d_inl1_hrd1.txt index 5040f4b335..d5f96b8c8e 100644 --- a/epochX/cudacpp/tput/logs_ggttgg_mad/log_ggttgg_mad_d_inl1_hrd1.txt +++ b/epochX/cudacpp/tput/logs_ggttgg_mad/log_ggttgg_mad_d_inl1_hrd1.txt @@ -40,7 +40,7 @@ make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg' -DATE: 2024-10-06_09:31:09 +DATE: 2024-10-06_13:49:39 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: ========================================================================= @@ -49,15 +49,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_SM_GG_TTXGG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=1] [hardcodePARAM=1] Workflow summary = CUD:DBL+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 3.107076e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.130192e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.131670e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 3.101940e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.122575e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.124001e+05 ) sec^-1 MeanMatrixElemValue = ( 4.063123e+00 +- 2.368970e+00 ) GeV^-4 -TOTAL : 0.539226 sec +TOTAL : 0.540166 sec INFO: No Floating Point Exceptions have been reported - 2,240,615,938 cycles # 2.902 GHz - 3,467,491,001 instructions # 1.55 insn per cycle - 0.828466018 seconds time elapsed + 2,231,518,048 cycles # 2.888 GHz + 3,509,842,963 instructions # 1.57 insn per cycle + 0.829879936 seconds time elapsed runNcu /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.cuda_d_inl1_hrd1/check_cuda.exe -p 64 256 1 ==PROF== Profiling "sigmaKin": launch__registers_per_thread 255 ==PROF== Profiling "sigmaKin": sm__sass_average_branch_targets_threads_uniform.pct 100% @@ -67,15 +67,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_SM_GG_TTXGG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=1] [hardcodePARAM=1] Workflow summary = CUD:DBL+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 3.751881e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.775679e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.776668e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 3.754320e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.777904e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.778900e+05 ) sec^-1 MeanMatrixElemValue = ( 6.665112e+00 +- 5.002651e+00 ) GeV^-4 -TOTAL : 3.303070 sec +TOTAL : 3.302218 sec INFO: No Floating Point Exceptions have been reported - 10,434,569,638 cycles # 2.930 GHz - 24,118,235,140 instructions # 2.31 insn per cycle - 3.617886016 seconds time elapsed + 10,397,495,221 cycles # 2.921 GHz + 23,399,129,234 instructions # 2.25 insn per cycle + 3.615578185 seconds time elapsed ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.cuda_d_inl1_hrd1/runTest_cuda.exe INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: FE_DIVBYZERO, FE_INVALID, FE_OVERFLOW @@ -100,15 +100,15 @@ Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=1] [ Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/none+NAVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = SCALAR ('none': ~vector[1], no SIMD) -EvtsPerSec[Rmb+ME] (23) = ( 4.241409e+02 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 4.241886e+02 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 4.241886e+02 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 4.224464e+02 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 4.224922e+02 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 4.224922e+02 ) sec^-1 MeanMatrixElemValue = ( 4.063123e+00 +- 2.368970e+00 ) GeV^-4 -TOTAL : 38.674103 sec +TOTAL : 38.828259 sec INFO: No Floating Point Exceptions have been reported - 113,958,477,984 cycles # 2.947 GHz - 144,286,195,418 instructions # 1.27 insn per cycle - 38.678088373 seconds time elapsed + 113,821,244,168 cycles # 2.931 GHz + 144,301,310,273 instructions # 1.27 insn per cycle + 38.832492297 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4:21024) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.none_d_inl1_hrd1/runTest_cpp.exe @@ -132,15 +132,15 @@ Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=1] [ Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[2] ('sse4': SSE4.2, 128bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 3.007169e+03 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.009483e+03 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.009483e+03 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 3.003479e+03 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.005773e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.005773e+03 ) sec^-1 MeanMatrixElemValue = ( 4.063123e+00 +- 2.368970e+00 ) GeV^-4 -TOTAL : 5.460584 sec +TOTAL : 5.466566 sec INFO: No Floating Point Exceptions have been reported - 15,281,187,875 cycles # 2.797 GHz - 37,839,169,102 instructions # 2.48 insn per cycle - 5.464853538 seconds time elapsed + 15,293,977,457 cycles # 2.796 GHz + 37,837,140,466 instructions # 2.47 insn per cycle + 5.470916861 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4:68594) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.sse4_d_inl1_hrd1/runTest_cpp.exe @@ -164,15 +164,15 @@ Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=1] [ Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('avx2': AVX2, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 7.567317e+03 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 7.582163e+03 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 7.582163e+03 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 7.502858e+03 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 7.516857e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 7.516857e+03 ) sec^-1 MeanMatrixElemValue = ( 4.063123e+00 +- 2.368970e+00 ) GeV^-4 -TOTAL : 2.174218 sec +TOTAL : 2.192482 sec INFO: No Floating Point Exceptions have been reported - 6,020,206,289 cycles # 2.765 GHz - 12,923,983,464 instructions # 2.15 insn per cycle - 2.178219828 seconds time elapsed + 6,010,537,528 cycles # 2.737 GHz + 12,921,629,729 instructions # 2.15 insn per cycle + 2.196820897 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2:46048) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.avx2_d_inl1_hrd1/runTest_cpp.exe @@ -196,15 +196,15 @@ Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=1] [ Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('512y': AVX512, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 8.900478e+03 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 8.920792e+03 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 8.920792e+03 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 8.863333e+03 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 8.882926e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 8.882926e+03 ) sec^-1 MeanMatrixElemValue = ( 4.063123e+00 +- 2.368970e+00 ) GeV^-4 -TOTAL : 1.849478 sec +TOTAL : 1.856786 sec INFO: No Floating Point Exceptions have been reported - 5,102,330,026 cycles # 2.754 GHz - 11,453,366,172 instructions # 2.24 insn per cycle - 1.853513717 seconds time elapsed + 5,096,301,664 cycles # 2.739 GHz + 11,451,048,704 instructions # 2.25 insn per cycle + 1.861090397 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2:40151) (512y: 219) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.512y_d_inl1_hrd1/runTest_cpp.exe @@ -228,15 +228,15 @@ Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=1] [ Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('512z': AVX512, 512bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 7.368242e+03 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 7.382314e+03 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 7.382314e+03 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 7.326356e+03 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 7.339645e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 7.339645e+03 ) sec^-1 MeanMatrixElemValue = ( 4.063123e+00 +- 2.368970e+00 ) GeV^-4 -TOTAL : 2.232876 sec +TOTAL : 2.245157 sec INFO: No Floating Point Exceptions have been reported - 3,951,515,189 cycles # 1.767 GHz - 5,896,746,544 instructions # 1.49 insn per cycle - 2.236852257 seconds time elapsed + 3,952,645,484 cycles # 1.758 GHz + 5,894,393,037 instructions # 1.49 insn per cycle + 2.249770359 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 1959) (512y: 259) (512z:38977) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.512z_d_inl1_hrd1/runTest_cpp.exe diff --git a/epochX/cudacpp/tput/logs_ggttgg_mad/log_ggttgg_mad_f_inl0_hrd0.txt b/epochX/cudacpp/tput/logs_ggttgg_mad/log_ggttgg_mad_f_inl0_hrd0.txt index c4676334b0..4ebb0cf87d 100644 --- a/epochX/cudacpp/tput/logs_ggttgg_mad/log_ggttgg_mad_f_inl0_hrd0.txt +++ b/epochX/cudacpp/tput/logs_ggttgg_mad/log_ggttgg_mad_f_inl0_hrd0.txt @@ -40,7 +40,7 @@ make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg' -DATE: 2024-10-06_09:04:57 +DATE: 2024-10-06_13:34:08 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: ========================================================================= @@ -49,15 +49,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_SM_GG_TTXGG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 5.476973e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 5.519601e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 5.523500e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 5.536878e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 5.570398e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 5.573979e+05 ) sec^-1 MeanMatrixElemValue = ( 4.059596e+00 +- 2.368053e+00 ) GeV^-4 -TOTAL : 0.498075 sec +TOTAL : 0.490230 sec INFO: No Floating Point Exceptions have been reported - 2,049,620,143 cycles # 2.856 GHz - 3,058,097,989 instructions # 1.49 insn per cycle - 0.977244524 seconds time elapsed + 2,041,780,354 cycles # 2.883 GHz + 3,033,645,851 instructions # 1.49 insn per cycle + 0.767724951 seconds time elapsed runNcu /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.cuda_f_inl0_hrd0/check_cuda.exe -p 64 256 1 ==PROF== Profiling "sigmaKin": launch__registers_per_thread 255 ==PROF== Profiling "sigmaKin": sm__sass_average_branch_targets_threads_uniform.pct 100% @@ -67,15 +67,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_SM_GG_TTXGG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 8.124860e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 8.187008e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 8.189727e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 8.131124e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 8.185658e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 8.188064e+05 ) sec^-1 MeanMatrixElemValue = ( 6.664703e+00 +- 5.072736e+00 ) GeV^-4 -TOTAL : 1.797790 sec +TOTAL : 1.791623 sec INFO: No Floating Point Exceptions have been reported - 5,916,497,978 cycles # 2.910 GHz - 12,115,730,956 instructions # 2.05 insn per cycle - 2.090370837 seconds time elapsed + 5,917,581,929 cycles # 2.918 GHz + 12,821,644,892 instructions # 2.17 insn per cycle + 2.084652141 seconds time elapsed ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.cuda_f_inl0_hrd0/runTest_cuda.exe INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: FE_DIVBYZERO, FE_INVALID, FE_OVERFLOW @@ -100,15 +100,15 @@ Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=0] [ Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/none+NAVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = SCALAR ('none': ~vector[1], no SIMD) -EvtsPerSec[Rmb+ME] (23) = ( 1.932981e+03 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.933931e+03 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.933931e+03 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.935874e+03 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.936844e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.936844e+03 ) sec^-1 MeanMatrixElemValue = ( 4.060121e+00 +- 2.367902e+00 ) GeV^-4 -TOTAL : 8.490769 sec +TOTAL : 8.476637 sec INFO: No Floating Point Exceptions have been reported - 24,922,868,630 cycles # 2.935 GHz - 79,110,265,707 instructions # 3.17 insn per cycle - 8.496015758 seconds time elapsed + 24,985,094,401 cycles # 2.947 GHz + 79,109,998,236 instructions # 3.17 insn per cycle + 8.480748547 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 3572) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.none_f_inl0_hrd0/runTest_cpp.exe @@ -132,15 +132,15 @@ Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=0] [ Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('sse4': SSE4.2, 128bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 6.975543e+03 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 6.988298e+03 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 6.988298e+03 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 7.005261e+03 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 7.017873e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 7.017873e+03 ) sec^-1 MeanMatrixElemValue = ( 4.060119e+00 +- 2.367901e+00 ) GeV^-4 -TOTAL : 2.356100 sec +TOTAL : 2.346120 sec INFO: No Floating Point Exceptions have been reported - 6,536,263,436 cycles # 2.771 GHz - 20,271,266,485 instructions # 3.10 insn per cycle - 2.362378155 seconds time elapsed + 6,538,387,237 cycles # 2.783 GHz + 20,270,803,093 instructions # 3.10 insn per cycle + 2.350422395 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4:13779) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.sse4_f_inl0_hrd0/runTest_cpp.exe @@ -164,15 +164,15 @@ Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=0] [ Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('avx2': AVX2, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 1.588631e+04 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.595153e+04 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.595153e+04 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.590664e+04 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.597045e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.597045e+04 ) sec^-1 MeanMatrixElemValue = ( 4.060560e+00 +- 2.367611e+00 ) GeV^-4 -TOTAL : 1.038490 sec +TOTAL : 1.035799 sec INFO: No Floating Point Exceptions have been reported - 2,837,721,779 cycles # 2.726 GHz - 7,066,858,765 instructions # 2.49 insn per cycle - 1.044464831 seconds time elapsed + 2,837,301,360 cycles # 2.731 GHz + 7,066,268,745 instructions # 2.49 insn per cycle + 1.039949773 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2:12055) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.avx2_f_inl0_hrd0/runTest_cpp.exe @@ -196,15 +196,15 @@ Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=0] [ Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('512y': AVX512, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 1.762421e+04 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.770702e+04 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.770702e+04 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.798486e+04 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.806728e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.806728e+04 ) sec^-1 MeanMatrixElemValue = ( 4.060560e+00 +- 2.367611e+00 ) GeV^-4 -TOTAL : 0.936394 sec +TOTAL : 0.916857 sec INFO: No Floating Point Exceptions have been reported - 2,577,125,275 cycles # 2.745 GHz - 6,404,206,024 instructions # 2.49 insn per cycle - 0.941322355 seconds time elapsed + 2,527,079,944 cycles # 2.746 GHz + 6,403,406,399 instructions # 2.53 insn per cycle + 0.921063658 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2:11019) (512y: 44) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.512y_f_inl0_hrd0/runTest_cpp.exe @@ -228,15 +228,15 @@ Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=0] [ Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[16] ('512z': AVX512, 512bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 1.409980e+04 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.415034e+04 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.415034e+04 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.409719e+04 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.414746e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.414746e+04 ) sec^-1 MeanMatrixElemValue = ( 4.060562e+00 +- 2.367612e+00 ) GeV^-4 -TOTAL : 1.170914 sec +TOTAL : 1.168631 sec INFO: No Floating Point Exceptions have been reported - 2,069,436,546 cycles # 1.766 GHz - 3,304,699,013 instructions # 1.60 insn per cycle - 1.174781391 seconds time elapsed + 2,066,337,797 cycles # 1.763 GHz + 3,304,628,320 instructions # 1.60 insn per cycle + 1.172706329 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 2603) (512y: 44) (512z: 9605) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.512z_f_inl0_hrd0/runTest_cpp.exe diff --git a/epochX/cudacpp/tput/logs_ggttgg_mad/log_ggttgg_mad_f_inl0_hrd0_bridge.txt b/epochX/cudacpp/tput/logs_ggttgg_mad/log_ggttgg_mad_f_inl0_hrd0_bridge.txt index dec260c3af..d08ca82f72 100644 --- a/epochX/cudacpp/tput/logs_ggttgg_mad/log_ggttgg_mad_f_inl0_hrd0_bridge.txt +++ b/epochX/cudacpp/tput/logs_ggttgg_mad/log_ggttgg_mad_f_inl0_hrd0_bridge.txt @@ -40,7 +40,7 @@ make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg' -DATE: 2024-10-06_09:40:11 +DATE: 2024-10-06_13:58:44 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: ========================================================================= @@ -53,16 +53,16 @@ WARNING! Set grid in Bridge (nevt=16384, gpublocks=64, gputhreads=256, gpublocks Process = SIGMA_SM_GG_TTXGG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 4.924368e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 5.456718e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 5.456718e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 4.910165e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 5.441564e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 5.441564e+05 ) sec^-1 MeanMatrixElemValue = ( 4.048178e+00 +- 2.364571e+00 ) GeV^-4 -TOTAL : 0.481369 sec +TOTAL : 0.483451 sec INFO: No Floating Point Exceptions have been reported INFO: No Floating Point Exceptions have been reported - 2,011,468,293 cycles # 2.883 GHz - 2,972,689,221 instructions # 1.48 insn per cycle - 0.755097926 seconds time elapsed + 2,025,778,102 cycles # 2.856 GHz + 3,014,621,726 instructions # 1.49 insn per cycle + 0.768235893 seconds time elapsed runNcu /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.cuda_f_inl0_hrd0/check_cuda.exe -p 64 256 1 --bridge WARNING! Bridge selected: cannot use RamboDevice, will use RamboHost WARNING! RamboHost selected: cannot use CurandDevice, will use CurandHost @@ -80,16 +80,16 @@ WARNING! Set grid in Bridge (nevt=524288, gpublocks=2048, gputhreads=256, gpublo Process = SIGMA_SM_GG_TTXGG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 6.978465e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 8.128974e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 8.128974e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 6.940987e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 8.085822e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 8.085822e+05 ) sec^-1 MeanMatrixElemValue = ( 6.641709e+00 +- 4.994248e+00 ) GeV^-4 -TOTAL : 1.967107 sec +TOTAL : 1.971741 sec INFO: No Floating Point Exceptions have been reported INFO: No Floating Point Exceptions have been reported - 6,502,759,539 cycles # 2.928 GHz - 13,854,302,325 instructions # 2.13 insn per cycle - 2.276466534 seconds time elapsed + 6,452,893,215 cycles # 2.917 GHz + 12,932,516,587 instructions # 2.00 insn per cycle + 2.268561899 seconds time elapsed ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.cuda_f_inl0_hrd0/runTest_cuda.exe INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: FE_DIVBYZERO, FE_INVALID, FE_OVERFLOW @@ -115,16 +115,16 @@ Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=0] [ Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+BRDHST/none+NAVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = SCALAR ('none': ~vector[1], no SIMD) -EvtsPerSec[Rmb+ME] (23) = ( 1.944212e+03 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.945160e+03 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.945160e+03 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.944197e+03 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.945119e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.945119e+03 ) sec^-1 MeanMatrixElemValue = ( 4.060121e+00 +- 2.367902e+00 ) GeV^-4 -TOTAL : 8.443358 sec +TOTAL : 8.443154 sec INFO: No Floating Point Exceptions have been reported INFO: No Floating Point Exceptions have been reported - 24,934,407,175 cycles # 2.952 GHz - 79,115,502,595 instructions # 3.17 insn per cycle - 8.447759712 seconds time elapsed + 24,931,161,134 cycles # 2.952 GHz + 79,118,780,002 instructions # 3.17 insn per cycle + 8.447511962 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 3572) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.none_f_inl0_hrd0/runTest_cpp.exe @@ -149,16 +149,16 @@ Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=0] [ Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+BRDHST/sse4+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('sse4': SSE4.2, 128bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 7.020230e+03 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 7.033459e+03 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 7.033459e+03 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 7.008519e+03 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 7.021495e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 7.021495e+03 ) sec^-1 MeanMatrixElemValue = ( 4.060119e+00 +- 2.367901e+00 ) GeV^-4 -TOTAL : 2.344217 sec +TOTAL : 2.347792 sec INFO: No Floating Point Exceptions have been reported INFO: No Floating Point Exceptions have been reported - 6,541,090,448 cycles # 2.786 GHz - 20,280,124,954 instructions # 3.10 insn per cycle - 2.348689069 seconds time elapsed + 6,549,748,515 cycles # 2.786 GHz + 20,281,316,740 instructions # 3.10 insn per cycle + 2.352098397 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4:13779) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.sse4_f_inl0_hrd0/runTest_cpp.exe @@ -183,16 +183,16 @@ Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=0] [ Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+BRDHST/avx2+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('avx2': AVX2, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 1.604920e+04 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.611581e+04 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.611581e+04 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.597931e+04 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.604592e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.604592e+04 ) sec^-1 MeanMatrixElemValue = ( 4.060560e+00 +- 2.367611e+00 ) GeV^-4 -TOTAL : 1.029784 sec +TOTAL : 1.034082 sec INFO: No Floating Point Exceptions have been reported INFO: No Floating Point Exceptions have been reported - 2,846,767,262 cycles # 2.755 GHz - 7,076,446,064 instructions # 2.49 insn per cycle - 1.034215836 seconds time elapsed + 2,847,664,561 cycles # 2.744 GHz + 7,075,785,603 instructions # 2.48 insn per cycle + 1.038468407 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2:12055) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.avx2_f_inl0_hrd0/runTest_cpp.exe @@ -217,16 +217,16 @@ Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=0] [ Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+BRDHST/512y+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('512y': AVX512, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 1.797566e+04 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.806224e+04 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.806224e+04 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.806966e+04 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.815407e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.815407e+04 ) sec^-1 MeanMatrixElemValue = ( 4.060560e+00 +- 2.367611e+00 ) GeV^-4 -TOTAL : 0.920078 sec +TOTAL : 0.915485 sec INFO: No Floating Point Exceptions have been reported INFO: No Floating Point Exceptions have been reported - 2,539,792,408 cycles # 2.749 GHz - 6,413,266,409 instructions # 2.53 insn per cycle - 0.924434981 seconds time elapsed + 2,535,040,117 cycles # 2.758 GHz + 6,413,499,393 instructions # 2.53 insn per cycle + 0.919871466 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2:11019) (512y: 44) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.512y_f_inl0_hrd0/runTest_cpp.exe @@ -251,16 +251,16 @@ Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=0] [ Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+BRDHST/512z+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[16] ('512z': AVX512, 512bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 1.411104e+04 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.416189e+04 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.416189e+04 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.399107e+04 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.404075e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.404075e+04 ) sec^-1 MeanMatrixElemValue = ( 4.060562e+00 +- 2.367612e+00 ) GeV^-4 -TOTAL : 1.170311 sec +TOTAL : 1.180128 sec INFO: No Floating Point Exceptions have been reported INFO: No Floating Point Exceptions have been reported - 2,078,956,436 cycles # 1.771 GHz - 3,314,205,136 instructions # 1.59 insn per cycle - 1.174679954 seconds time elapsed + 2,080,681,737 cycles # 1.758 GHz + 3,314,450,825 instructions # 1.59 insn per cycle + 1.184513033 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 2603) (512y: 44) (512z: 9605) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.512z_f_inl0_hrd0/runTest_cpp.exe diff --git a/epochX/cudacpp/tput/logs_ggttgg_mad/log_ggttgg_mad_f_inl0_hrd0_common.txt b/epochX/cudacpp/tput/logs_ggttgg_mad/log_ggttgg_mad_f_inl0_hrd0_common.txt index 3ebd5caeb8..350a5af8fc 100644 --- a/epochX/cudacpp/tput/logs_ggttgg_mad/log_ggttgg_mad_f_inl0_hrd0_common.txt +++ b/epochX/cudacpp/tput/logs_ggttgg_mad/log_ggttgg_mad_f_inl0_hrd0_common.txt @@ -40,7 +40,7 @@ make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg' -DATE: 2024-10-06_09:51:07 +DATE: 2024-10-06_14:09:39 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: ========================================================================= @@ -49,15 +49,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_SM_GG_TTXGG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:COMMON+RMBDEV+MESDEV/none+NAVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 5.481675e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 5.521755e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 5.525865e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 5.468521e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 5.510087e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 5.513866e+05 ) sec^-1 MeanMatrixElemValue = ( 4.159396e-01 +- 3.238803e-01 ) GeV^-4 -TOTAL : 0.477918 sec +TOTAL : 0.479274 sec INFO: No Floating Point Exceptions have been reported - 1,990,228,523 cycles # 2.864 GHz - 2,978,927,673 instructions # 1.50 insn per cycle - 0.751663902 seconds time elapsed + 1,998,214,102 cycles # 2.877 GHz + 2,979,894,690 instructions # 1.49 insn per cycle + 0.752077949 seconds time elapsed runNcu /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.cuda_f_inl0_hrd0/check_cuda.exe -p 64 256 1 --common ==PROF== Profiling "sigmaKin": launch__registers_per_thread 255 ==PROF== Profiling "sigmaKin": sm__sass_average_branch_targets_threads_uniform.pct 100% @@ -67,15 +67,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_SM_GG_TTXGG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:COMMON+RMBDEV+MESDEV/none+NAVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 8.037728e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 8.099183e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 8.101846e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 8.046951e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 8.107331e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 8.110154e+05 ) sec^-1 MeanMatrixElemValue = ( 1.094367e+02 +- 1.071509e+02 ) GeV^-4 -TOTAL : 1.886731 sec +TOTAL : 1.883937 sec INFO: No Floating Point Exceptions have been reported - 6,136,710,401 cycles # 2.909 GHz - 13,142,850,218 instructions # 2.14 insn per cycle - 2.175693489 seconds time elapsed + 6,163,820,196 cycles # 2.926 GHz + 13,104,701,324 instructions # 2.13 insn per cycle + 2.172581568 seconds time elapsed ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.cuda_f_inl0_hrd0/runTest_cuda.exe INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: FE_DIVBYZERO, FE_INVALID, FE_OVERFLOW @@ -100,15 +100,15 @@ Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=0] [ Workflow summary = CPP:FLT+CXS:COMMON+RMBHST+MESHST/none+NAVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = SCALAR ('none': ~vector[1], no SIMD) -EvtsPerSec[Rmb+ME] (23) = ( 1.941292e+03 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.942240e+03 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.942240e+03 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.915553e+03 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.916469e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.916469e+03 ) sec^-1 MeanMatrixElemValue = ( 4.208459e-01 +- 3.253446e-01 ) GeV^-4 -TOTAL : 8.455097 sec +TOTAL : 8.567710 sec INFO: No Floating Point Exceptions have been reported - 24,914,950,228 cycles # 2.946 GHz - 79,111,045,664 instructions # 3.18 insn per cycle - 8.459383915 seconds time elapsed + 25,247,962,013 cycles # 2.946 GHz + 79,109,499,779 instructions # 3.13 insn per cycle + 8.571905753 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 3572) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.none_f_inl0_hrd0/runTest_cpp.exe @@ -132,15 +132,15 @@ Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=0] [ Workflow summary = CPP:FLT+CXS:COMMON+RMBHST+MESHST/sse4+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('sse4': SSE4.2, 128bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 6.977213e+03 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 6.990041e+03 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 6.990041e+03 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 6.970500e+03 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 6.983344e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 6.983344e+03 ) sec^-1 MeanMatrixElemValue = ( 4.208457e-01 +- 3.253445e-01 ) GeV^-4 -TOTAL : 2.356205 sec +TOTAL : 2.359638 sec INFO: No Floating Point Exceptions have been reported - 6,550,546,250 cycles # 2.776 GHz - 20,269,237,886 instructions # 3.09 insn per cycle - 2.360272003 seconds time elapsed + 6,530,438,913 cycles # 2.764 GHz + 20,271,637,597 instructions # 3.10 insn per cycle + 2.363733582 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4:13779) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.sse4_f_inl0_hrd0/runTest_cpp.exe @@ -164,15 +164,15 @@ Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=0] [ Workflow summary = CPP:FLT+CXS:COMMON+RMBHST+MESHST/avx2+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('avx2': AVX2, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 1.601317e+04 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.608084e+04 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.608084e+04 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.588782e+04 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.595699e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.595699e+04 ) sec^-1 MeanMatrixElemValue = ( 4.214978e-01 +- 3.255521e-01 ) GeV^-4 -TOTAL : 1.030095 sec +TOTAL : 1.038380 sec INFO: No Floating Point Exceptions have been reported - 2,839,431,727 cycles # 2.748 GHz - 7,063,774,184 instructions # 2.49 insn per cycle - 1.034210988 seconds time elapsed + 2,841,007,073 cycles # 2.727 GHz + 7,063,916,408 instructions # 2.49 insn per cycle + 1.042733919 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2:12055) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.avx2_f_inl0_hrd0/runTest_cpp.exe @@ -196,15 +196,15 @@ Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=0] [ Workflow summary = CPP:FLT+CXS:COMMON+RMBHST+MESHST/512y+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('512y': AVX512, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 1.801735e+04 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.810193e+04 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.810193e+04 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.804889e+04 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.813296e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.813296e+04 ) sec^-1 MeanMatrixElemValue = ( 4.214978e-01 +- 3.255521e-01 ) GeV^-4 -TOTAL : 0.916264 sec +TOTAL : 0.914551 sec INFO: No Floating Point Exceptions have been reported - 2,529,614,240 cycles # 2.751 GHz - 6,399,972,746 instructions # 2.53 insn per cycle - 0.920311559 seconds time elapsed + 2,527,969,688 cycles # 2.754 GHz + 6,399,972,291 instructions # 2.53 insn per cycle + 0.918627500 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2:11019) (512y: 44) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.512y_f_inl0_hrd0/runTest_cpp.exe @@ -228,15 +228,15 @@ Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=0] [ Workflow summary = CPP:FLT+CXS:COMMON+RMBHST+MESHST/512z+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[16] ('512z': AVX512, 512bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 1.413582e+04 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.418711e+04 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.418711e+04 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.412165e+04 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.417447e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.417447e+04 ) sec^-1 MeanMatrixElemValue = ( 4.214981e-01 +- 3.255523e-01 ) GeV^-4 -TOTAL : 1.166574 sec +TOTAL : 1.168203 sec INFO: No Floating Point Exceptions have been reported - 2,070,023,042 cycles # 1.769 GHz - 3,300,470,940 instructions # 1.59 insn per cycle - 1.170621524 seconds time elapsed + 2,073,564,630 cycles # 1.769 GHz + 3,302,278,912 instructions # 1.59 insn per cycle + 1.172558407 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 2603) (512y: 44) (512z: 9605) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.512z_f_inl0_hrd0/runTest_cpp.exe diff --git a/epochX/cudacpp/tput/logs_ggttgg_mad/log_ggttgg_mad_f_inl0_hrd0_curhst.txt b/epochX/cudacpp/tput/logs_ggttgg_mad/log_ggttgg_mad_f_inl0_hrd0_curhst.txt index 8aa78a916d..519781a54c 100644 --- a/epochX/cudacpp/tput/logs_ggttgg_mad/log_ggttgg_mad_f_inl0_hrd0_curhst.txt +++ b/epochX/cudacpp/tput/logs_ggttgg_mad/log_ggttgg_mad_f_inl0_hrd0_curhst.txt @@ -40,7 +40,7 @@ make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg' -DATE: 2024-10-06_09:48:16 +DATE: 2024-10-06_14:06:48 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: ========================================================================= @@ -49,15 +49,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_SM_GG_TTXGG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:CURHST+RMBDEV+MESDEV/none+NAVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 5.460370e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 5.501314e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 5.505347e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 5.472378e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 5.513510e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 5.517095e+05 ) sec^-1 MeanMatrixElemValue = ( 4.059596e+00 +- 2.368053e+00 ) GeV^-4 -TOTAL : 0.475676 sec +TOTAL : 0.479017 sec INFO: No Floating Point Exceptions have been reported - 1,998,344,168 cycles # 2.886 GHz - 3,027,104,836 instructions # 1.51 insn per cycle - 0.748859673 seconds time elapsed + 2,002,079,791 cycles # 2.885 GHz + 3,032,109,731 instructions # 1.51 insn per cycle + 0.751322232 seconds time elapsed runNcu /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.cuda_f_inl0_hrd0/check_cuda.exe -p 64 256 1 --curhst ==PROF== Profiling "sigmaKin": launch__registers_per_thread 255 ==PROF== Profiling "sigmaKin": sm__sass_average_branch_targets_threads_uniform.pct 100% @@ -67,15 +67,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_SM_GG_TTXGG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:CURHST+RMBDEV+MESDEV/none+NAVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 8.172168e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 8.234506e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 8.237328e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 8.133605e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 8.195426e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 8.198026e+05 ) sec^-1 MeanMatrixElemValue = ( 6.664703e+00 +- 5.072736e+00 ) GeV^-4 -TOTAL : 1.821851 sec +TOTAL : 1.835260 sec INFO: No Floating Point Exceptions have been reported - 6,001,499,639 cycles # 2.924 GHz - 13,042,334,044 instructions # 2.17 insn per cycle - 2.109220847 seconds time elapsed + 5,953,245,675 cycles # 2.885 GHz + 12,120,019,760 instructions # 2.04 insn per cycle + 2.124019849 seconds time elapsed ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.cuda_f_inl0_hrd0/runTest_cuda.exe INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: FE_DIVBYZERO, FE_INVALID, FE_OVERFLOW @@ -100,15 +100,15 @@ Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=0] [ Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/none+NAVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = SCALAR ('none': ~vector[1], no SIMD) -EvtsPerSec[Rmb+ME] (23) = ( 1.941510e+03 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.942442e+03 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.942442e+03 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.945324e+03 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.946257e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.946257e+03 ) sec^-1 MeanMatrixElemValue = ( 4.060121e+00 +- 2.367902e+00 ) GeV^-4 -TOTAL : 8.452200 sec +TOTAL : 8.435674 sec INFO: No Floating Point Exceptions have been reported - 24,907,540,526 cycles # 2.946 GHz - 79,109,866,227 instructions # 3.18 insn per cycle - 8.456266423 seconds time elapsed + 24,912,556,021 cycles # 2.952 GHz + 79,109,818,221 instructions # 3.18 insn per cycle + 8.439775154 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 3572) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.none_f_inl0_hrd0/runTest_cpp.exe @@ -132,15 +132,15 @@ Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=0] [ Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('sse4': SSE4.2, 128bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 7.017369e+03 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 7.030395e+03 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 7.030395e+03 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 7.035744e+03 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 7.048318e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 7.048318e+03 ) sec^-1 MeanMatrixElemValue = ( 4.060119e+00 +- 2.367901e+00 ) GeV^-4 -TOTAL : 2.341887 sec +TOTAL : 2.336007 sec INFO: No Floating Point Exceptions have been reported - 6,533,658,672 cycles # 2.786 GHz - 20,270,788,705 instructions # 3.10 insn per cycle - 2.345994128 seconds time elapsed + 6,537,003,938 cycles # 2.794 GHz + 20,270,600,364 instructions # 3.10 insn per cycle + 2.340127016 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4:13779) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.sse4_f_inl0_hrd0/runTest_cpp.exe @@ -164,15 +164,15 @@ Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=0] [ Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('avx2': AVX2, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 1.604029e+04 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.610893e+04 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.610893e+04 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.608155e+04 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.614719e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.614719e+04 ) sec^-1 MeanMatrixElemValue = ( 4.060560e+00 +- 2.367611e+00 ) GeV^-4 -TOTAL : 1.027451 sec +TOTAL : 1.024751 sec INFO: No Floating Point Exceptions have been reported - 2,836,206,155 cycles # 2.751 GHz - 7,065,988,768 instructions # 2.49 insn per cycle - 1.031531216 seconds time elapsed + 2,834,240,254 cycles # 2.757 GHz + 7,065,872,567 instructions # 2.49 insn per cycle + 1.028806934 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2:12055) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.avx2_f_inl0_hrd0/runTest_cpp.exe @@ -196,15 +196,15 @@ Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=0] [ Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('512y': AVX512, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 1.796598e+04 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.804847e+04 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.804847e+04 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.798553e+04 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.806887e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.806887e+04 ) sec^-1 MeanMatrixElemValue = ( 4.060560e+00 +- 2.367611e+00 ) GeV^-4 -TOTAL : 0.917896 sec +TOTAL : 0.916858 sec INFO: No Floating Point Exceptions have been reported - 2,527,698,465 cycles # 2.744 GHz - 6,403,574,368 instructions # 2.53 insn per cycle - 0.921906155 seconds time elapsed + 2,528,915,077 cycles # 2.748 GHz + 6,404,017,474 instructions # 2.53 insn per cycle + 0.920921022 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2:11019) (512y: 44) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.512y_f_inl0_hrd0/runTest_cpp.exe @@ -228,15 +228,15 @@ Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=0] [ Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[16] ('512z': AVX512, 512bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 1.414079e+04 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.419125e+04 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.419125e+04 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.410434e+04 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.415649e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.415649e+04 ) sec^-1 MeanMatrixElemValue = ( 4.060562e+00 +- 2.367612e+00 ) GeV^-4 -TOTAL : 1.164994 sec +TOTAL : 1.168051 sec INFO: No Floating Point Exceptions have been reported - 2,068,678,617 cycles # 1.770 GHz - 3,304,093,166 instructions # 1.60 insn per cycle - 1.169236265 seconds time elapsed + 2,069,062,400 cycles # 1.766 GHz + 3,303,996,801 instructions # 1.60 insn per cycle + 1.172398393 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 2603) (512y: 44) (512z: 9605) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.512z_f_inl0_hrd0/runTest_cpp.exe diff --git a/epochX/cudacpp/tput/logs_ggttgg_mad/log_ggttgg_mad_f_inl0_hrd0_rmbhst.txt b/epochX/cudacpp/tput/logs_ggttgg_mad/log_ggttgg_mad_f_inl0_hrd0_rmbhst.txt index 59696ff16e..4f32506624 100644 --- a/epochX/cudacpp/tput/logs_ggttgg_mad/log_ggttgg_mad_f_inl0_hrd0_rmbhst.txt +++ b/epochX/cudacpp/tput/logs_ggttgg_mad/log_ggttgg_mad_f_inl0_hrd0_rmbhst.txt @@ -40,7 +40,7 @@ make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg' -DATE: 2024-10-06_09:45:30 +DATE: 2024-10-06_14:04:03 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: ========================================================================= @@ -50,15 +50,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_SM_GG_TTXGG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:CURHST+RMBHST+MESDEV/none+NAVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 5.026958e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 5.513975e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 5.517845e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 5.027823e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 5.526046e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 5.529647e+05 ) sec^-1 MeanMatrixElemValue = ( 4.048178e+00 +- 2.364571e+00 ) GeV^-4 -TOTAL : 0.478506 sec +TOTAL : 0.480665 sec INFO: No Floating Point Exceptions have been reported - 1,992,355,788 cycles # 2.865 GHz - 3,027,729,409 instructions # 1.52 insn per cycle - 0.751914958 seconds time elapsed + 1,998,601,792 cycles # 2.872 GHz + 3,011,866,350 instructions # 1.51 insn per cycle + 0.754198369 seconds time elapsed runNcu /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.cuda_f_inl0_hrd0/check_cuda.exe -p 64 256 1 --rmbhst WARNING! RamboHost selected: cannot use CurandDevice, will use CurandHost ==PROF== Profiling "sigmaKin": launch__registers_per_thread 255 @@ -70,15 +70,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_SM_GG_TTXGG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:CURHST+RMBHST+MESDEV/none+NAVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 7.156008e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 8.226322e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 8.229025e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 7.122434e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 8.190178e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 8.192953e+05 ) sec^-1 MeanMatrixElemValue = ( 6.641709e+00 +- 4.994248e+00 ) GeV^-4 -TOTAL : 1.900625 sec +TOTAL : 1.903238 sec INFO: No Floating Point Exceptions have been reported - 6,225,372,770 cycles # 2.919 GHz - 12,616,761,411 instructions # 2.03 insn per cycle - 2.188103626 seconds time elapsed + 6,212,943,803 cycles # 2.911 GHz + 13,208,812,426 instructions # 2.13 insn per cycle + 2.191711057 seconds time elapsed ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.cuda_f_inl0_hrd0/runTest_cuda.exe INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: FE_DIVBYZERO, FE_INVALID, FE_OVERFLOW @@ -103,15 +103,15 @@ Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=0] [ Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/none+NAVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = SCALAR ('none': ~vector[1], no SIMD) -EvtsPerSec[Rmb+ME] (23) = ( 1.942577e+03 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.943527e+03 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.943527e+03 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.944704e+03 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.945635e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.945635e+03 ) sec^-1 MeanMatrixElemValue = ( 4.060121e+00 +- 2.367902e+00 ) GeV^-4 -TOTAL : 8.447888 sec +TOTAL : 8.438513 sec INFO: No Floating Point Exceptions have been reported - 24,912,816,300 cycles # 2.948 GHz - 79,110,249,403 instructions # 3.18 insn per cycle - 8.452014602 seconds time elapsed + 24,901,911,376 cycles # 2.950 GHz + 79,109,793,455 instructions # 3.18 insn per cycle + 8.442571973 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 3572) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.none_f_inl0_hrd0/runTest_cpp.exe @@ -135,15 +135,15 @@ Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=0] [ Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('sse4': SSE4.2, 128bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 6.980733e+03 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 6.993141e+03 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 6.993141e+03 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 6.927529e+03 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 6.939783e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 6.939783e+03 ) sec^-1 MeanMatrixElemValue = ( 4.060119e+00 +- 2.367901e+00 ) GeV^-4 -TOTAL : 2.354354 sec +TOTAL : 2.372232 sec INFO: No Floating Point Exceptions have been reported - 6,535,460,807 cycles # 2.772 GHz - 20,270,869,690 instructions # 3.10 insn per cycle - 2.358646539 seconds time elapsed + 6,544,489,875 cycles # 2.755 GHz + 20,270,859,114 instructions # 3.10 insn per cycle + 2.376284397 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4:13779) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.sse4_f_inl0_hrd0/runTest_cpp.exe @@ -167,15 +167,15 @@ Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=0] [ Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('avx2': AVX2, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 1.603543e+04 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.610156e+04 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.610156e+04 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.608575e+04 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.615179e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.615179e+04 ) sec^-1 MeanMatrixElemValue = ( 4.060560e+00 +- 2.367611e+00 ) GeV^-4 -TOTAL : 1.027888 sec +TOTAL : 1.024685 sec INFO: No Floating Point Exceptions have been reported - 2,837,672,612 cycles # 2.752 GHz - 7,066,358,168 instructions # 2.49 insn per cycle - 1.031930682 seconds time elapsed + 2,836,740,404 cycles # 2.759 GHz + 7,066,728,541 instructions # 2.49 insn per cycle + 1.028900682 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2:12055) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.avx2_f_inl0_hrd0/runTest_cpp.exe @@ -199,15 +199,15 @@ Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=0] [ Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('512y': AVX512, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 1.798975e+04 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.807399e+04 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.807399e+04 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.803317e+04 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.811831e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.811831e+04 ) sec^-1 MeanMatrixElemValue = ( 4.060560e+00 +- 2.367611e+00 ) GeV^-4 -TOTAL : 0.916670 sec +TOTAL : 0.914441 sec INFO: No Floating Point Exceptions have been reported - 2,525,901,356 cycles # 2.745 GHz - 6,403,453,175 instructions # 2.54 insn per cycle - 0.920789172 seconds time elapsed + 2,527,228,616 cycles # 2.753 GHz + 6,403,420,073 instructions # 2.53 insn per cycle + 0.918642480 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2:11019) (512y: 44) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.512y_f_inl0_hrd0/runTest_cpp.exe @@ -231,15 +231,15 @@ Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=0] [ Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[16] ('512z': AVX512, 512bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 1.406582e+04 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.411589e+04 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.411589e+04 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.414480e+04 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.419579e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.419579e+04 ) sec^-1 MeanMatrixElemValue = ( 4.060562e+00 +- 2.367612e+00 ) GeV^-4 -TOTAL : 1.171278 sec +TOTAL : 1.164654 sec INFO: No Floating Point Exceptions have been reported - 2,071,908,739 cycles # 1.764 GHz - 3,303,987,486 instructions # 1.59 insn per cycle - 1.175442581 seconds time elapsed + 2,068,849,941 cycles # 1.771 GHz + 3,303,897,585 instructions # 1.60 insn per cycle + 1.168873616 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 2603) (512y: 44) (512z: 9605) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.512z_f_inl0_hrd0/runTest_cpp.exe diff --git a/epochX/cudacpp/tput/logs_ggttgg_mad/log_ggttgg_mad_f_inl0_hrd1.txt b/epochX/cudacpp/tput/logs_ggttgg_mad/log_ggttgg_mad_f_inl0_hrd1.txt index fc006f8d57..4d5f852ee9 100644 --- a/epochX/cudacpp/tput/logs_ggttgg_mad/log_ggttgg_mad_f_inl0_hrd1.txt +++ b/epochX/cudacpp/tput/logs_ggttgg_mad/log_ggttgg_mad_f_inl0_hrd1.txt @@ -40,7 +40,7 @@ make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg' -DATE: 2024-10-06_09:05:24 +DATE: 2024-10-06_13:34:34 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: ========================================================================= @@ -49,15 +49,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_SM_GG_TTXGG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=1] Workflow summary = CUD:FLT+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 5.473150e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 5.513248e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 5.516891e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 5.498464e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 5.532121e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 5.535446e+05 ) sec^-1 MeanMatrixElemValue = ( 4.059596e+00 +- 2.368053e+00 ) GeV^-4 -TOTAL : 0.500476 sec +TOTAL : 0.491592 sec INFO: No Floating Point Exceptions have been reported - 2,066,687,911 cycles # 2.859 GHz - 3,064,980,702 instructions # 1.48 insn per cycle - 0.941605450 seconds time elapsed + 2,044,391,610 cycles # 2.885 GHz + 3,086,577,499 instructions # 1.51 insn per cycle + 0.768514832 seconds time elapsed runNcu /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.cuda_f_inl0_hrd1/check_cuda.exe -p 64 256 1 ==PROF== Profiling "sigmaKin": launch__registers_per_thread 255 ==PROF== Profiling "sigmaKin": sm__sass_average_branch_targets_threads_uniform.pct 100% @@ -67,15 +67,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_SM_GG_TTXGG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=1] Workflow summary = CUD:FLT+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 8.096999e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 8.159101e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 8.161763e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 8.099359e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 8.153452e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 8.155863e+05 ) sec^-1 MeanMatrixElemValue = ( 6.664703e+00 +- 5.072736e+00 ) GeV^-4 -TOTAL : 1.803372 sec +TOTAL : 1.797533 sec INFO: No Floating Point Exceptions have been reported - 5,931,019,959 cycles # 2.909 GHz - 12,491,679,666 instructions # 2.11 insn per cycle - 2.096189929 seconds time elapsed + 5,915,980,346 cycles # 2.915 GHz + 11,835,771,041 instructions # 2.00 insn per cycle + 2.086054349 seconds time elapsed ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.cuda_f_inl0_hrd1/runTest_cuda.exe INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: FE_DIVBYZERO, FE_INVALID, FE_OVERFLOW @@ -100,15 +100,15 @@ Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=0] [ Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/none+NAVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = SCALAR ('none': ~vector[1], no SIMD) -EvtsPerSec[Rmb+ME] (23) = ( 1.927739e+03 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.928675e+03 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.928675e+03 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.937105e+03 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.938072e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.938072e+03 ) sec^-1 MeanMatrixElemValue = ( 4.060121e+00 +- 2.367902e+00 ) GeV^-4 -TOTAL : 8.512686 sec +TOTAL : 8.471126 sec INFO: No Floating Point Exceptions have been reported - 24,976,995,918 cycles # 2.933 GHz - 78,849,322,260 instructions # 3.16 insn per cycle - 8.521021644 seconds time elapsed + 24,925,339,532 cycles # 2.941 GHz + 78,844,267,323 instructions # 3.16 insn per cycle + 8.475209050 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 3092) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.none_f_inl0_hrd1/runTest_cpp.exe @@ -132,15 +132,15 @@ Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=0] [ Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('sse4': SSE4.2, 128bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 7.196617e+03 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 7.210064e+03 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 7.210064e+03 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 7.169975e+03 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 7.183220e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 7.183220e+03 ) sec^-1 MeanMatrixElemValue = ( 4.060119e+00 +- 2.367901e+00 ) GeV^-4 -TOTAL : 2.283841 sec +TOTAL : 2.292115 sec INFO: No Floating Point Exceptions have been reported - 6,462,353,077 cycles # 2.825 GHz - 20,230,287,596 instructions # 3.13 insn per cycle - 2.291660153 seconds time elapsed + 6,463,815,708 cycles # 2.816 GHz + 20,230,236,051 instructions # 3.13 insn per cycle + 2.296276707 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4:13491) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.sse4_f_inl0_hrd1/runTest_cpp.exe @@ -164,15 +164,15 @@ Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=0] [ Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('avx2': AVX2, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 1.507603e+04 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.513399e+04 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.513399e+04 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.519621e+04 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.525746e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.525746e+04 ) sec^-1 MeanMatrixElemValue = ( 4.060560e+00 +- 2.367611e+00 ) GeV^-4 -TOTAL : 1.094262 sec +TOTAL : 1.083878 sec INFO: No Floating Point Exceptions have been reported - 2,977,852,840 cycles # 2.716 GHz - 7,207,139,157 instructions # 2.42 insn per cycle - 1.100869463 seconds time elapsed + 2,979,564,806 cycles # 2.740 GHz + 7,207,603,732 instructions # 2.42 insn per cycle + 1.088052861 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2:12437) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.avx2_f_inl0_hrd1/runTest_cpp.exe @@ -196,15 +196,15 @@ Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=0] [ Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('512y': AVX512, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 1.740158e+04 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.747960e+04 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.747960e+04 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.735132e+04 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.742768e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.742768e+04 ) sec^-1 MeanMatrixElemValue = ( 4.060560e+00 +- 2.367611e+00 ) GeV^-4 -TOTAL : 0.947565 sec +TOTAL : 0.949974 sec INFO: No Floating Point Exceptions have been reported - 2,615,044,427 cycles # 2.750 GHz - 6,545,142,442 instructions # 2.50 insn per cycle - 0.954571468 seconds time elapsed + 2,613,122,224 cycles # 2.741 GHz + 6,545,365,697 instructions # 2.50 insn per cycle + 0.954239148 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2:11449) (512y: 27) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.512y_f_inl0_hrd1/runTest_cpp.exe @@ -228,15 +228,15 @@ Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=0] [ Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[16] ('512z': AVX512, 512bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 1.344321e+04 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.349023e+04 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.349023e+04 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.358265e+04 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.362953e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.362953e+04 ) sec^-1 MeanMatrixElemValue = ( 4.060562e+00 +- 2.367612e+00 ) GeV^-4 -TOTAL : 1.225060 sec +TOTAL : 1.212363 sec INFO: No Floating Point Exceptions have been reported - 2,140,395,059 cycles # 1.742 GHz - 3,462,158,546 instructions # 1.62 insn per cycle - 1.232075146 seconds time elapsed + 2,139,735,372 cycles # 1.761 GHz + 3,461,675,599 instructions # 1.62 insn per cycle + 1.216529449 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 3037) (512y: 25) (512z: 9677) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.512z_f_inl0_hrd1/runTest_cpp.exe diff --git a/epochX/cudacpp/tput/logs_ggttgg_mad/log_ggttgg_mad_f_inl1_hrd0.txt b/epochX/cudacpp/tput/logs_ggttgg_mad/log_ggttgg_mad_f_inl1_hrd0.txt index 507fa267fb..4eb31e9e23 100644 --- a/epochX/cudacpp/tput/logs_ggttgg_mad/log_ggttgg_mad_f_inl1_hrd0.txt +++ b/epochX/cudacpp/tput/logs_ggttgg_mad/log_ggttgg_mad_f_inl1_hrd0.txt @@ -40,7 +40,7 @@ make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg' -DATE: 2024-10-06_09:32:18 +DATE: 2024-10-06_13:50:49 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: ========================================================================= @@ -49,15 +49,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_SM_GG_TTXGG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=1] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 5.570913e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 5.612300e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 5.616113e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 5.534236e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 5.575724e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 5.579434e+05 ) sec^-1 MeanMatrixElemValue = ( 4.059597e+00 +- 2.368053e+00 ) GeV^-4 -TOTAL : 0.500062 sec +TOTAL : 0.498024 sec INFO: No Floating Point Exceptions have been reported - 2,077,093,809 cycles # 2.883 GHz - 3,095,482,027 instructions # 1.49 insn per cycle - 0.782648151 seconds time elapsed + 2,052,284,204 cycles # 2.869 GHz + 3,044,837,172 instructions # 1.48 insn per cycle + 0.777402103 seconds time elapsed runNcu /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.cuda_f_inl1_hrd0/check_cuda.exe -p 64 256 1 ==PROF== Profiling "sigmaKin": launch__registers_per_thread 255 ==PROF== Profiling "sigmaKin": sm__sass_average_branch_targets_threads_uniform.pct 100% @@ -67,15 +67,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_SM_GG_TTXGG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=1] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 8.624378e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 8.693284e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 8.696098e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 8.652081e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 8.717247e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 8.720111e+05 ) sec^-1 MeanMatrixElemValue = ( 6.664703e+00 +- 5.072736e+00 ) GeV^-4 -TOTAL : 1.736663 sec +TOTAL : 1.730818 sec INFO: No Floating Point Exceptions have been reported - 5,745,039,966 cycles # 2.917 GHz - 12,243,347,327 instructions # 2.13 insn per cycle - 2.029186282 seconds time elapsed + 5,764,100,640 cycles # 2.916 GHz + 12,007,166,869 instructions # 2.08 insn per cycle + 2.035341726 seconds time elapsed ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.cuda_f_inl1_hrd0/runTest_cuda.exe INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: FE_DIVBYZERO, FE_INVALID, FE_OVERFLOW @@ -100,15 +100,15 @@ Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=1] [ Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/none+NAVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = SCALAR ('none': ~vector[1], no SIMD) -EvtsPerSec[Rmb+ME] (23) = ( 5.610943e+02 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 5.611718e+02 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 5.611718e+02 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 5.597707e+02 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 5.598492e+02 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 5.598492e+02 ) sec^-1 MeanMatrixElemValue = ( 4.059969e+00 +- 2.367799e+00 ) GeV^-4 -TOTAL : 29.233986 sec +TOTAL : 29.303387 sec INFO: No Floating Point Exceptions have been reported - 86,131,386,822 cycles # 2.946 GHz - 135,652,659,903 instructions # 1.57 insn per cycle - 29.237672033 seconds time elapsed + 86,145,087,897 cycles # 2.940 GHz + 135,666,532,475 instructions # 1.57 insn per cycle + 29.307524144 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4:15856) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.none_f_inl1_hrd0/runTest_cpp.exe @@ -132,15 +132,15 @@ Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=1] [ Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('sse4': SSE4.2, 128bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 6.849906e+03 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 6.862163e+03 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 6.862163e+03 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 6.855261e+03 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 6.867301e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 6.867301e+03 ) sec^-1 MeanMatrixElemValue = ( 4.059962e+00 +- 2.367792e+00 ) GeV^-4 -TOTAL : 2.399244 sec +TOTAL : 2.397415 sec INFO: No Floating Point Exceptions have been reported - 6,757,771,203 cycles # 2.813 GHz - 19,352,943,673 instructions # 2.86 insn per cycle - 2.403059869 seconds time elapsed + 6,761,810,138 cycles # 2.816 GHz + 19,353,174,647 instructions # 2.86 insn per cycle + 2.401726417 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4:69577) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.sse4_f_inl1_hrd0/runTest_cpp.exe @@ -164,15 +164,15 @@ Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=1] [ Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('avx2': AVX2, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 1.430057e+04 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.435326e+04 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.435326e+04 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.421128e+04 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.426359e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.426359e+04 ) sec^-1 MeanMatrixElemValue = ( 4.060903e+00 +- 2.367377e+00 ) GeV^-4 -TOTAL : 1.151867 sec +TOTAL : 1.159110 sec INFO: No Floating Point Exceptions have been reported - 3,169,480,733 cycles # 2.744 GHz - 6,794,963,559 instructions # 2.14 insn per cycle - 1.155607574 seconds time elapsed + 3,193,473,644 cycles # 2.747 GHz + 6,795,779,412 instructions # 2.13 insn per cycle + 1.163488937 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2:49034) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.avx2_f_inl1_hrd0/runTest_cpp.exe @@ -196,15 +196,15 @@ Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=1] [ Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('512y': AVX512, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 1.731154e+04 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.739005e+04 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.739005e+04 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.713637e+04 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.721194e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.721194e+04 ) sec^-1 MeanMatrixElemValue = ( 4.060903e+00 +- 2.367377e+00 ) GeV^-4 -TOTAL : 0.952402 sec +TOTAL : 0.962273 sec INFO: No Floating Point Exceptions have been reported - 2,622,407,179 cycles # 2.744 GHz - 5,970,044,618 instructions # 2.28 insn per cycle - 0.956238068 seconds time elapsed + 2,654,896,093 cycles # 2.749 GHz + 5,970,528,480 instructions # 2.25 insn per cycle + 0.966563657 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2:42602) (512y: 11) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.512y_f_inl1_hrd0/runTest_cpp.exe @@ -228,15 +228,15 @@ Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=1] [ Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[16] ('512z': AVX512, 512bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 1.414435e+04 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.419474e+04 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.419474e+04 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.406116e+04 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.411353e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.411353e+04 ) sec^-1 MeanMatrixElemValue = ( 4.060905e+00 +- 2.367377e+00 ) GeV^-4 -TOTAL : 1.165045 sec +TOTAL : 1.171595 sec INFO: No Floating Point Exceptions have been reported - 2,067,228,248 cycles # 1.769 GHz - 3,495,098,954 instructions # 1.69 insn per cycle - 1.168981438 seconds time elapsed + 2,069,653,669 cycles # 1.761 GHz + 3,495,286,510 instructions # 1.69 insn per cycle + 1.175733500 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 5208) (512y: 3) (512z:44858) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.512z_f_inl1_hrd0/runTest_cpp.exe diff --git a/epochX/cudacpp/tput/logs_ggttgg_mad/log_ggttgg_mad_f_inl1_hrd1.txt b/epochX/cudacpp/tput/logs_ggttgg_mad/log_ggttgg_mad_f_inl1_hrd1.txt index 2595c32afa..399b507978 100644 --- a/epochX/cudacpp/tput/logs_ggttgg_mad/log_ggttgg_mad_f_inl1_hrd1.txt +++ b/epochX/cudacpp/tput/logs_ggttgg_mad/log_ggttgg_mad_f_inl1_hrd1.txt @@ -40,7 +40,7 @@ make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg' -DATE: 2024-10-06_09:33:09 +DATE: 2024-10-06_13:51:40 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: ========================================================================= @@ -49,15 +49,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_SM_GG_TTXGG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=1] [hardcodePARAM=1] Workflow summary = CUD:FLT+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 5.573938e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 5.613715e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 5.617455e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 5.576824e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 5.615427e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 5.619151e+05 ) sec^-1 MeanMatrixElemValue = ( 4.059597e+00 +- 2.368053e+00 ) GeV^-4 -TOTAL : 0.493227 sec +TOTAL : 0.497790 sec INFO: No Floating Point Exceptions have been reported - 2,049,677,908 cycles # 2.879 GHz - 3,032,655,926 instructions # 1.48 insn per cycle - 0.769218706 seconds time elapsed + 2,053,285,244 cycles # 2.871 GHz + 3,062,104,949 instructions # 1.49 insn per cycle + 0.776202753 seconds time elapsed runNcu /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.cuda_f_inl1_hrd1/check_cuda.exe -p 64 256 1 ==PROF== Profiling "sigmaKin": launch__registers_per_thread 255 ==PROF== Profiling "sigmaKin": sm__sass_average_branch_targets_threads_uniform.pct 100% @@ -67,15 +67,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_SM_GG_TTXGG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=1] [hardcodePARAM=1] Workflow summary = CUD:FLT+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 8.673337e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 8.742674e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 8.745488e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 8.703662e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 8.770102e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 8.773043e+05 ) sec^-1 MeanMatrixElemValue = ( 6.664703e+00 +- 5.072736e+00 ) GeV^-4 -TOTAL : 1.731870 sec +TOTAL : 1.731591 sec INFO: No Floating Point Exceptions have been reported - 5,773,880,906 cycles # 2.919 GHz - 12,286,627,464 instructions # 2.13 insn per cycle - 2.034768323 seconds time elapsed + 5,754,613,710 cycles # 2.913 GHz + 11,370,596,750 instructions # 1.98 insn per cycle + 2.034362427 seconds time elapsed ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.cuda_f_inl1_hrd1/runTest_cuda.exe INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: FE_DIVBYZERO, FE_INVALID, FE_OVERFLOW @@ -100,15 +100,15 @@ Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=1] [ Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/none+NAVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = SCALAR ('none': ~vector[1], no SIMD) -EvtsPerSec[Rmb+ME] (23) = ( 5.600277e+02 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 5.601076e+02 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 5.601076e+02 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 5.615034e+02 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 5.615830e+02 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 5.615830e+02 ) sec^-1 MeanMatrixElemValue = ( 4.059969e+00 +- 2.367799e+00 ) GeV^-4 -TOTAL : 29.289301 sec +TOTAL : 29.212729 sec INFO: No Floating Point Exceptions have been reported - 86,207,606,672 cycles # 2.943 GHz - 135,355,986,373 instructions # 1.57 insn per cycle - 29.293063672 seconds time elapsed + 85,783,261,520 cycles # 2.936 GHz + 135,355,808,788 instructions # 1.58 insn per cycle + 29.216858686 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4:15471) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.none_f_inl1_hrd1/runTest_cpp.exe @@ -132,15 +132,15 @@ Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=1] [ Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('sse4': SSE4.2, 128bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 6.848001e+03 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 6.860244e+03 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 6.860244e+03 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 6.819874e+03 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 6.831807e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 6.831807e+03 ) sec^-1 MeanMatrixElemValue = ( 4.059962e+00 +- 2.367792e+00 ) GeV^-4 -TOTAL : 2.399823 sec +TOTAL : 2.409786 sec INFO: No Floating Point Exceptions have been reported - 6,855,955,670 cycles # 2.853 GHz - 19,471,788,292 instructions # 2.84 insn per cycle - 2.403723205 seconds time elapsed + 6,854,190,543 cycles # 2.840 GHz + 19,473,264,864 instructions # 2.84 insn per cycle + 2.414044045 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4:69876) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.sse4_f_inl1_hrd1/runTest_cpp.exe @@ -164,15 +164,15 @@ Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=1] [ Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('avx2': AVX2, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 1.455129e+04 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.460639e+04 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.460639e+04 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.457232e+04 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.462783e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.462783e+04 ) sec^-1 MeanMatrixElemValue = ( 4.060903e+00 +- 2.367377e+00 ) GeV^-4 -TOTAL : 1.132031 sec +TOTAL : 1.130262 sec INFO: No Floating Point Exceptions have been reported - 3,102,391,764 cycles # 2.733 GHz - 6,715,014,781 instructions # 2.16 insn per cycle - 1.135898458 seconds time elapsed + 3,107,015,891 cycles # 2.741 GHz + 6,715,453,628 instructions # 2.16 insn per cycle + 1.134440947 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2:47692) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.avx2_f_inl1_hrd1/runTest_cpp.exe @@ -196,15 +196,15 @@ Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=1] [ Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('512y': AVX512, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 1.738588e+04 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.746518e+04 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.746518e+04 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.732835e+04 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.740736e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.740736e+04 ) sec^-1 MeanMatrixElemValue = ( 4.060903e+00 +- 2.367377e+00 ) GeV^-4 -TOTAL : 0.948137 sec +TOTAL : 0.951428 sec INFO: No Floating Point Exceptions have been reported - 2,626,199,962 cycles # 2.761 GHz - 5,966,019,567 instructions # 2.27 insn per cycle - 0.951931849 seconds time elapsed + 2,626,658,648 cycles # 2.750 GHz + 5,966,473,000 instructions # 2.27 insn per cycle + 0.955816667 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2:41858) (512y: 13) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.512y_f_inl1_hrd1/runTest_cpp.exe @@ -228,15 +228,15 @@ Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=1] [ Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[16] ('512z': AVX512, 512bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 1.414552e+04 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.419616e+04 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.419616e+04 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.399739e+04 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.405039e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.405039e+04 ) sec^-1 MeanMatrixElemValue = ( 4.060905e+00 +- 2.367377e+00 ) GeV^-4 -TOTAL : 1.164736 sec +TOTAL : 1.176739 sec INFO: No Floating Point Exceptions have been reported - 2,067,746,434 cycles # 1.771 GHz - 3,487,891,958 instructions # 1.69 insn per cycle - 1.168545250 seconds time elapsed + 2,069,290,363 cycles # 1.753 GHz + 3,487,567,338 instructions # 1.69 insn per cycle + 1.180844900 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 4171) (512y: 4) (512z:44494) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.512z_f_inl1_hrd1/runTest_cpp.exe diff --git a/epochX/cudacpp/tput/logs_ggttgg_mad/log_ggttgg_mad_m_inl0_hrd0.txt b/epochX/cudacpp/tput/logs_ggttgg_mad/log_ggttgg_mad_m_inl0_hrd0.txt index a3a2deda6e..bcebed4d6d 100644 --- a/epochX/cudacpp/tput/logs_ggttgg_mad/log_ggttgg_mad_m_inl0_hrd0.txt +++ b/epochX/cudacpp/tput/logs_ggttgg_mad/log_ggttgg_mad_m_inl0_hrd0.txt @@ -40,7 +40,7 @@ make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg' -DATE: 2024-10-06_09:03:48 +DATE: 2024-10-06_13:32:59 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: ========================================================================= @@ -49,15 +49,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_SM_GG_TTXGG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:MIX+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 3.318725e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.347238e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.349358e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 3.322648e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.344378e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.345975e+05 ) sec^-1 MeanMatrixElemValue = ( 4.063123e+00 +- 2.368970e+00 ) GeV^-4 -TOTAL : 0.539210 sec +TOTAL : 0.531755 sec INFO: No Floating Point Exceptions have been reported - 2,220,963,802 cycles # 2.880 GHz - 3,406,426,816 instructions # 1.53 insn per cycle - 0.832307462 seconds time elapsed + 2,214,351,581 cycles # 2.887 GHz + 3,440,633,222 instructions # 1.55 insn per cycle + 0.823632890 seconds time elapsed runNcu /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.cuda_m_inl0_hrd0/check_cuda.exe -p 64 256 1 ==PROF== Profiling "sigmaKin": launch__registers_per_thread 255 ==PROF== Profiling "sigmaKin": sm__sass_average_branch_targets_threads_uniform.pct 100% @@ -67,15 +67,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_SM_GG_TTXGG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:MIX+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 4.134167e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 4.164785e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 4.165985e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 4.138694e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 4.166643e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 4.167815e+05 ) sec^-1 MeanMatrixElemValue = ( 6.665112e+00 +- 5.002651e+00 ) GeV^-4 -TOTAL : 3.047029 sec +TOTAL : 3.038430 sec INFO: No Floating Point Exceptions have been reported - 9,687,290,131 cycles # 2.924 GHz - 21,862,744,253 instructions # 2.26 insn per cycle - 3.379254641 seconds time elapsed + 9,679,691,491 cycles # 2.926 GHz + 22,235,208,068 instructions # 2.30 insn per cycle + 3.364367051 seconds time elapsed ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.cuda_m_inl0_hrd0/runTest_cuda.exe INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: FE_DIVBYZERO, FE_INVALID, FE_OVERFLOW @@ -100,15 +100,15 @@ Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=0] [ Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/none+NAVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) Internal loops fptype_sv = SCALAR ('none': ~vector[1], no SIMD) -EvtsPerSec[Rmb+ME] (23) = ( 1.868179e+03 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.869079e+03 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.869079e+03 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.858884e+03 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.859756e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.859756e+03 ) sec^-1 MeanMatrixElemValue = ( 4.063123e+00 +- 2.368970e+00 ) GeV^-4 -TOTAL : 8.786228 sec +TOTAL : 8.828839 sec INFO: No Floating Point Exceptions have been reported - 25,910,148,307 cycles # 2.949 GHz - 79,427,985,275 instructions # 3.07 insn per cycle - 8.790193498 seconds time elapsed + 25,913,905,261 cycles # 2.934 GHz + 79,426,517,168 instructions # 3.07 insn per cycle + 8.833079742 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 4775) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.none_m_inl0_hrd0/runTest_cpp.exe @@ -132,15 +132,15 @@ Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=0] [ Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[2] ('sse4': SSE4.2, 128bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 3.521065e+03 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.524381e+03 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.524381e+03 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 3.522114e+03 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.525268e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.525268e+03 ) sec^-1 MeanMatrixElemValue = ( 4.063123e+00 +- 2.368970e+00 ) GeV^-4 -TOTAL : 4.666859 sec +TOTAL : 4.662712 sec INFO: No Floating Point Exceptions have been reported - 12,831,991,791 cycles # 2.749 GHz - 38,825,085,312 instructions # 3.03 insn per cycle - 4.671138327 seconds time elapsed + 12,818,468,807 cycles # 2.747 GHz + 38,823,378,269 instructions # 3.03 insn per cycle + 4.666970309 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4:13173) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.sse4_m_inl0_hrd0/runTest_cpp.exe @@ -164,15 +164,15 @@ Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=0] [ Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('avx2': AVX2, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 8.087173e+03 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 8.104021e+03 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 8.104021e+03 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 8.036672e+03 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 8.052384e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 8.052384e+03 ) sec^-1 MeanMatrixElemValue = ( 4.063123e+00 +- 2.368970e+00 ) GeV^-4 -TOTAL : 2.035173 sec +TOTAL : 2.046992 sec INFO: No Floating Point Exceptions have been reported - 5,594,158,972 cycles # 2.744 GHz - 13,617,938,147 instructions # 2.43 insn per cycle - 2.039272194 seconds time elapsed + 5,588,972,002 cycles # 2.726 GHz + 13,615,212,723 instructions # 2.44 insn per cycle + 2.051247677 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2:11427) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.avx2_m_inl0_hrd0/runTest_cpp.exe @@ -196,15 +196,15 @@ Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=0] [ Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('512y': AVX512, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 9.329915e+03 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 9.351715e+03 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 9.351715e+03 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 9.259017e+03 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 9.280879e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 9.280879e+03 ) sec^-1 MeanMatrixElemValue = ( 4.063123e+00 +- 2.368970e+00 ) GeV^-4 -TOTAL : 1.764965 sec +TOTAL : 1.777497 sec INFO: No Floating Point Exceptions have been reported - 4,865,961,098 cycles # 2.752 GHz - 12,296,280,016 instructions # 2.53 insn per cycle - 1.768959352 seconds time elapsed + 4,864,001,740 cycles # 2.730 GHz + 12,296,691,692 instructions # 2.53 insn per cycle + 1.782775231 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2:10331) (512y: 80) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.512y_m_inl0_hrd0/runTest_cpp.exe @@ -228,15 +228,15 @@ Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=0] [ Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('512z': AVX512, 512bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 6.944494e+03 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 6.956947e+03 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 6.956947e+03 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 6.952402e+03 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 6.964268e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 6.964268e+03 ) sec^-1 MeanMatrixElemValue = ( 4.063123e+00 +- 2.368970e+00 ) GeV^-4 -TOTAL : 2.368908 sec +TOTAL : 2.365608 sec INFO: No Floating Point Exceptions have been reported - 4,175,656,001 cycles # 1.761 GHz - 6,394,856,033 instructions # 1.53 insn per cycle - 2.373043514 seconds time elapsed + 4,168,817,415 cycles # 1.760 GHz + 6,392,051,112 instructions # 1.53 insn per cycle + 2.369892537 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 1983) (512y: 92) (512z: 9360) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.512z_m_inl0_hrd0/runTest_cpp.exe diff --git a/epochX/cudacpp/tput/logs_ggttgg_mad/log_ggttgg_mad_m_inl0_hrd1.txt b/epochX/cudacpp/tput/logs_ggttgg_mad/log_ggttgg_mad_m_inl0_hrd1.txt index f598011718..329ccd728d 100644 --- a/epochX/cudacpp/tput/logs_ggttgg_mad/log_ggttgg_mad_m_inl0_hrd1.txt +++ b/epochX/cudacpp/tput/logs_ggttgg_mad/log_ggttgg_mad_m_inl0_hrd1.txt @@ -40,7 +40,7 @@ make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg' -DATE: 2024-10-06_09:04:22 +DATE: 2024-10-06_13:33:34 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: ========================================================================= @@ -49,15 +49,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_SM_GG_TTXGG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=1] Workflow summary = CUD:MIX+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 3.335025e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.357927e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.359916e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 3.343282e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.365123e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.366869e+05 ) sec^-1 MeanMatrixElemValue = ( 4.063123e+00 +- 2.368970e+00 ) GeV^-4 -TOTAL : 0.537353 sec +TOTAL : 0.533882 sec INFO: No Floating Point Exceptions have been reported - 2,216,980,042 cycles # 2.869 GHz - 3,463,326,813 instructions # 1.56 insn per cycle - 0.836472238 seconds time elapsed + 2,216,202,893 cycles # 2.884 GHz + 3,431,484,043 instructions # 1.55 insn per cycle + 0.826865719 seconds time elapsed runNcu /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.cuda_m_inl0_hrd1/check_cuda.exe -p 64 256 1 ==PROF== Profiling "sigmaKin": launch__registers_per_thread 255 ==PROF== Profiling "sigmaKin": sm__sass_average_branch_targets_threads_uniform.pct 100% @@ -67,15 +67,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_SM_GG_TTXGG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=1] Workflow summary = CUD:MIX+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 4.141323e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 4.172030e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 4.173253e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 4.144428e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 4.172417e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 4.173588e+05 ) sec^-1 MeanMatrixElemValue = ( 6.665112e+00 +- 5.002651e+00 ) GeV^-4 -TOTAL : 3.034442 sec +TOTAL : 3.026211 sec INFO: No Floating Point Exceptions have been reported - 9,665,974,027 cycles # 2.922 GHz - 21,248,987,108 instructions # 2.20 insn per cycle - 3.363171619 seconds time elapsed + 9,620,836,593 cycles # 2.930 GHz + 21,758,300,843 instructions # 2.26 insn per cycle + 3.339519269 seconds time elapsed ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.cuda_m_inl0_hrd1/runTest_cuda.exe INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: FE_DIVBYZERO, FE_INVALID, FE_OVERFLOW @@ -100,15 +100,15 @@ Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=0] [ Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/none+NAVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) Internal loops fptype_sv = SCALAR ('none': ~vector[1], no SIMD) -EvtsPerSec[Rmb+ME] (23) = ( 1.862251e+03 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.863154e+03 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.863154e+03 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.863924e+03 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.864820e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.864820e+03 ) sec^-1 MeanMatrixElemValue = ( 4.063123e+00 +- 2.368970e+00 ) GeV^-4 -TOTAL : 8.813876 sec +TOTAL : 8.804619 sec INFO: No Floating Point Exceptions have been reported - 25,987,730,158 cycles # 2.948 GHz - 79,453,128,863 instructions # 3.06 insn per cycle - 8.817767368 seconds time elapsed + 25,995,436,563 cycles # 2.952 GHz + 79,451,930,959 instructions # 3.06 insn per cycle + 8.808705602 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 4431) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.none_m_inl0_hrd1/runTest_cpp.exe @@ -132,15 +132,15 @@ Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=0] [ Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[2] ('sse4': SSE4.2, 128bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 3.512571e+03 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.515785e+03 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.515785e+03 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 3.503636e+03 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.506720e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.506720e+03 ) sec^-1 MeanMatrixElemValue = ( 4.063123e+00 +- 2.368970e+00 ) GeV^-4 -TOTAL : 4.675994 sec +TOTAL : 4.686943 sec INFO: No Floating Point Exceptions have been reported - 12,822,983,844 cycles # 2.741 GHz - 38,780,874,555 instructions # 3.02 insn per cycle - 4.681038643 seconds time elapsed + 12,810,062,086 cycles # 2.731 GHz + 38,781,561,445 instructions # 3.03 insn per cycle + 4.691164608 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4:12935) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.sse4_m_inl0_hrd1/runTest_cpp.exe @@ -164,15 +164,15 @@ Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=0] [ Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('avx2': AVX2, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 8.056370e+03 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 8.072927e+03 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 8.072927e+03 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 8.104055e+03 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 8.120152e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 8.120152e+03 ) sec^-1 MeanMatrixElemValue = ( 4.063123e+00 +- 2.368970e+00 ) GeV^-4 -TOTAL : 2.042531 sec +TOTAL : 2.029657 sec INFO: No Floating Point Exceptions have been reported - 5,590,175,615 cycles # 2.733 GHz - 13,732,675,080 instructions # 2.46 insn per cycle - 2.046647326 seconds time elapsed + 5,587,308,453 cycles # 2.748 GHz + 13,733,180,312 instructions # 2.46 insn per cycle + 2.033870056 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2:11510) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.avx2_m_inl0_hrd1/runTest_cpp.exe @@ -196,15 +196,15 @@ Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=0] [ Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('512y': AVX512, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 9.148791e+03 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 9.170046e+03 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 9.170046e+03 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 9.138404e+03 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 9.159519e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 9.159519e+03 ) sec^-1 MeanMatrixElemValue = ( 4.063123e+00 +- 2.368970e+00 ) GeV^-4 -TOTAL : 1.800883 sec +TOTAL : 1.800887 sec INFO: No Floating Point Exceptions have been reported - 4,955,825,709 cycles # 2.749 GHz - 12,423,990,964 instructions # 2.51 insn per cycle - 1.804980058 seconds time elapsed + 4,957,775,767 cycles # 2.748 GHz + 12,422,283,070 instructions # 2.51 insn per cycle + 1.805210655 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2:10322) (512y: 240) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.512y_m_inl0_hrd1/runTest_cpp.exe @@ -228,15 +228,15 @@ Process = SIGMA_SM_GG_TTXGG_CPP [gcc 11.3.1] [inlineHel=0] [ Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('512z': AVX512, 512bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 6.851374e+03 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 6.863307e+03 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 6.863307e+03 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 6.874859e+03 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 6.886979e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 6.886979e+03 ) sec^-1 MeanMatrixElemValue = ( 4.063123e+00 +- 2.368970e+00 ) GeV^-4 -TOTAL : 2.400794 sec +TOTAL : 2.391767 sec INFO: No Floating Point Exceptions have been reported - 4,218,682,996 cycles # 1.755 GHz - 6,496,899,309 instructions # 1.54 insn per cycle - 2.406253121 seconds time elapsed + 4,175,961,745 cycles # 1.743 GHz + 6,495,139,197 instructions # 1.56 insn per cycle + 2.396085091 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 1806) (512y: 190) (512z: 9358) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttgg.mad/SubProcesses/P1_gg_ttxgg/build.512z_m_inl0_hrd1/runTest_cpp.exe diff --git a/epochX/cudacpp/tput/logs_ggttggg_mad/log_ggttggg_mad_d_inl0_hrd0.txt b/epochX/cudacpp/tput/logs_ggttggg_mad/log_ggttggg_mad_d_inl0_hrd0.txt index 17692fc5fb..e30b93b855 100644 --- a/epochX/cudacpp/tput/logs_ggttggg_mad/log_ggttggg_mad_d_inl0_hrd0.txt +++ b/epochX/cudacpp/tput/logs_ggttggg_mad/log_ggttggg_mad_d_inl0_hrd0.txt @@ -21,7 +21,7 @@ make: Nothing to be done for 'all'. make: Nothing to be done for 'all'. -DATE: 2024-10-06_09:07:10 +DATE: 2024-10-06_13:36:19 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: ========================================================================= @@ -30,15 +30,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_SM_GG_TTXGGG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 4.059500e+02 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 4.059934e+02 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 4.060148e+02 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 4.058345e+02 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 4.058760e+02 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 4.058904e+02 ) sec^-1 MeanMatrixElemValue = ( 1.187066e-05 +- 9.825549e-06 ) GeV^-6 -TOTAL : 2.453264 sec +TOTAL : 2.435619 sec INFO: No Floating Point Exceptions have been reported - 8,089,923,192 cycles # 2.904 GHz - 15,932,007,883 instructions # 1.97 insn per cycle - 2.843483231 seconds time elapsed + 8,055,742,313 cycles # 2.921 GHz + 18,367,948,395 instructions # 2.28 insn per cycle + 2.817152257 seconds time elapsed runNcu /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttggg.mad/SubProcesses/P1_gg_ttxggg/build.cuda_d_inl0_hrd0/check_cuda.exe -p 1 256 1 ==PROF== Profiling "sigmaKin": launch__registers_per_thread 255 ==PROF== Profiling "sigmaKin": sm__sass_average_branch_targets_threads_uniform.pct 100% @@ -48,15 +48,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_SM_GG_TTXGGG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 9.246459e+03 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 9.248360e+03 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 9.248591e+03 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 9.297833e+03 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 9.299850e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 9.300088e+03 ) sec^-1 MeanMatrixElemValue = ( 1.856249e-04 +- 8.329951e-05 ) GeV^-6 -TOTAL : 4.019480 sec +TOTAL : 3.986357 sec INFO: No Floating Point Exceptions have been reported - 12,563,980,059 cycles # 2.886 GHz - 29,860,686,581 instructions # 2.38 insn per cycle - 4.410635015 seconds time elapsed + 12,539,581,052 cycles # 2.903 GHz + 27,516,622,577 instructions # 2.19 insn per cycle + 4.378929253 seconds time elapsed ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttggg.mad/SubProcesses/P1_gg_ttxggg/build.cuda_d_inl0_hrd0/runTest_cuda.exe INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: FE_DIVBYZERO, FE_INVALID, FE_OVERFLOW @@ -81,15 +81,15 @@ Process = SIGMA_SM_GG_TTXGGG_CPP [gcc 11.3.1] [inlineHel=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/none+NAVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = SCALAR ('none': ~vector[1], no SIMD) -EvtsPerSec[Rmb+ME] (23) = ( 7.535286e+01 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 7.535490e+01 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 7.535490e+01 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 7.757344e+01 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 7.757558e+01 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 7.757558e+01 ) sec^-1 MeanMatrixElemValue = ( 1.187066e-05 +- 9.825549e-06 ) GeV^-6 -TOTAL : 7.007645 sec +TOTAL : 6.802929 sec INFO: No Floating Point Exceptions have been reported - 18,987,096,753 cycles # 2.709 GHz - 53,904,905,030 instructions # 2.84 insn per cycle - 7.011475835 seconds time elapsed + 19,399,624,992 cycles # 2.850 GHz + 53,903,045,283 instructions # 2.78 insn per cycle + 6.807081068 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4:32424) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttggg.mad/SubProcesses/P1_gg_ttxggg/build.none_d_inl0_hrd0/runTest_cpp.exe @@ -113,15 +113,15 @@ Process = SIGMA_SM_GG_TTXGGG_CPP [gcc 11.3.1] [inlineHel=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[2] ('sse4': SSE4.2, 128bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 1.576045e+02 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.576133e+02 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.576133e+02 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.577705e+02 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.577792e+02 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.577792e+02 ) sec^-1 MeanMatrixElemValue = ( 1.187066e-05 +- 9.825549e-06 ) GeV^-6 -TOTAL : 3.352060 sec +TOTAL : 3.348748 sec INFO: No Floating Point Exceptions have been reported - 9,813,557,960 cycles # 2.925 GHz - 27,153,109,398 instructions # 2.77 insn per cycle - 3.355902855 seconds time elapsed + 9,851,737,137 cycles # 2.939 GHz + 27,151,210,829 instructions # 2.76 insn per cycle + 3.353023572 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4:96492) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttggg.mad/SubProcesses/P1_gg_ttxggg/build.sse4_d_inl0_hrd0/runTest_cpp.exe @@ -145,15 +145,15 @@ Process = SIGMA_SM_GG_TTXGGG_CPP [gcc 11.3.1] [inlineHel=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('avx2': AVX2, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 3.392533e+02 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.392946e+02 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.392946e+02 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 3.390546e+02 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.390981e+02 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.390981e+02 ) sec^-1 MeanMatrixElemValue = ( 1.187066e-05 +- 9.825549e-06 ) GeV^-6 -TOTAL : 1.558312 sec +TOTAL : 1.559518 sec INFO: No Floating Point Exceptions have been reported - 4,259,121,658 cycles # 2.728 GHz - 9,591,809,021 instructions # 2.25 insn per cycle - 1.562248696 seconds time elapsed + 4,277,788,878 cycles # 2.737 GHz + 9,591,035,920 instructions # 2.24 insn per cycle + 1.563993127 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2:84961) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttggg.mad/SubProcesses/P1_gg_ttxggg/build.avx2_d_inl0_hrd0/runTest_cpp.exe @@ -177,15 +177,15 @@ Process = SIGMA_SM_GG_TTXGGG_CPP [gcc 11.3.1] [inlineHel=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('512y': AVX512, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 3.852746e+02 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.853256e+02 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.853256e+02 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 3.883922e+02 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.884521e+02 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.884521e+02 ) sec^-1 MeanMatrixElemValue = ( 1.187066e-05 +- 9.825549e-06 ) GeV^-6 -TOTAL : 1.371089 sec +TOTAL : 1.361536 sec INFO: No Floating Point Exceptions have been reported - 3,728,351,942 cycles # 2.713 GHz - 8,515,110,933 instructions # 2.28 insn per cycle - 1.374961080 seconds time elapsed + 3,730,442,775 cycles # 2.732 GHz + 8,515,212,284 instructions # 2.28 insn per cycle + 1.365939098 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2:80609) (512y: 90) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttggg.mad/SubProcesses/P1_gg_ttxggg/build.512y_d_inl0_hrd0/runTest_cpp.exe @@ -209,15 +209,15 @@ Process = SIGMA_SM_GG_TTXGGG_CPP [gcc 11.3.1] [inlineHel=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('512z': AVX512, 512bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 3.432608e+02 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.433087e+02 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.433087e+02 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 3.439785e+02 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.440316e+02 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.440316e+02 ) sec^-1 MeanMatrixElemValue = ( 1.187066e-05 +- 9.825549e-06 ) GeV^-6 -TOTAL : 1.541076 sec +TOTAL : 1.538183 sec INFO: No Floating Point Exceptions have been reported - 2,702,698,179 cycles # 1.750 GHz - 4,282,306,811 instructions # 1.58 insn per cycle - 1.545099546 seconds time elapsed + 2,696,585,466 cycles # 1.749 GHz + 4,281,827,355 instructions # 1.59 insn per cycle + 1.542666578 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 2856) (512y: 102) (512z:79114) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttggg.mad/SubProcesses/P1_gg_ttxggg/build.512z_d_inl0_hrd0/runTest_cpp.exe diff --git a/epochX/cudacpp/tput/logs_ggttggg_mad/log_ggttggg_mad_d_inl0_hrd0_bridge.txt b/epochX/cudacpp/tput/logs_ggttggg_mad/log_ggttggg_mad_d_inl0_hrd0_bridge.txt index 1cf857b709..0128600c53 100644 --- a/epochX/cudacpp/tput/logs_ggttggg_mad/log_ggttggg_mad_d_inl0_hrd0_bridge.txt +++ b/epochX/cudacpp/tput/logs_ggttggg_mad/log_ggttggg_mad_d_inl0_hrd0_bridge.txt @@ -21,7 +21,7 @@ make: Nothing to be done for 'all'. make: Nothing to be done for 'all'. -DATE: 2024-10-06_09:40:38 +DATE: 2024-10-06_13:59:10 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: ========================================================================= @@ -34,16 +34,16 @@ WARNING! Set grid in Bridge (nevt=256, gpublocks=1, gputhreads=256, gpublocks*gp Process = SIGMA_SM_GG_TTXGGG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 4.054825e+02 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 4.057209e+02 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 4.057209e+02 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 4.053079e+02 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 4.054677e+02 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 4.054677e+02 ) sec^-1 MeanMatrixElemValue = ( 1.187066e-05 +- 9.825549e-06 ) GeV^-6 -TOTAL : 2.388056 sec +TOTAL : 2.389831 sec INFO: No Floating Point Exceptions have been reported INFO: No Floating Point Exceptions have been reported - 7,931,671,790 cycles # 2.924 GHz - 17,623,602,431 instructions # 2.22 insn per cycle - 2.770306640 seconds time elapsed + 7,953,005,800 cycles # 2.929 GHz + 16,815,343,785 instructions # 2.11 insn per cycle + 2.772332655 seconds time elapsed runNcu /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttggg.mad/SubProcesses/P1_gg_ttxggg/build.cuda_d_inl0_hrd0/check_cuda.exe -p 1 256 1 --bridge WARNING! Bridge selected: cannot use RamboDevice, will use RamboHost WARNING! RamboHost selected: cannot use CurandDevice, will use CurandHost @@ -61,16 +61,16 @@ WARNING! Set grid in Bridge (nevt=16384, gpublocks=64, gputhreads=256, gpublocks Process = SIGMA_SM_GG_TTXGGG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 9.226146e+03 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 9.260909e+03 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 9.260909e+03 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 9.170034e+03 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 9.202189e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 9.202189e+03 ) sec^-1 MeanMatrixElemValue = ( 1.856249e-04 +- 8.329951e-05 ) GeV^-6 -TOTAL : 3.992337 sec +TOTAL : 4.005203 sec INFO: No Floating Point Exceptions have been reported INFO: No Floating Point Exceptions have been reported - 12,629,951,963 cycles # 2.926 GHz - 29,269,734,483 instructions # 2.32 insn per cycle - 4.375813430 seconds time elapsed + 12,762,607,389 cycles # 2.939 GHz + 29,896,342,782 instructions # 2.34 insn per cycle + 4.398297484 seconds time elapsed ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttggg.mad/SubProcesses/P1_gg_ttxggg/build.cuda_d_inl0_hrd0/runTest_cuda.exe INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: FE_DIVBYZERO, FE_INVALID, FE_OVERFLOW @@ -96,16 +96,16 @@ Process = SIGMA_SM_GG_TTXGGG_CPP [gcc 11.3.1] [inlineHel=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+BRDHST/none+NAVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = SCALAR ('none': ~vector[1], no SIMD) -EvtsPerSec[Rmb+ME] (23) = ( 7.889828e+01 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 7.890068e+01 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 7.890068e+01 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 7.610031e+01 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 7.610281e+01 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 7.610281e+01 ) sec^-1 MeanMatrixElemValue = ( 1.187066e-05 +- 9.825549e-06 ) GeV^-6 -TOTAL : 6.696425 sec +TOTAL : 6.940942 sec INFO: No Floating Point Exceptions have been reported INFO: No Floating Point Exceptions have been reported - 18,936,809,312 cycles # 2.827 GHz - 53,907,854,112 instructions # 2.85 insn per cycle - 6.700731218 seconds time elapsed + 18,985,130,324 cycles # 2.734 GHz + 53,911,984,395 instructions # 2.84 insn per cycle + 6.945312072 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4:32424) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttggg.mad/SubProcesses/P1_gg_ttxggg/build.none_d_inl0_hrd0/runTest_cpp.exe @@ -130,16 +130,16 @@ Process = SIGMA_SM_GG_TTXGGG_CPP [gcc 11.3.1] [inlineHel=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+BRDHST/sse4+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[2] ('sse4': SSE4.2, 128bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 1.586455e+02 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.586548e+02 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.586548e+02 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.572983e+02 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.573070e+02 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.573070e+02 ) sec^-1 MeanMatrixElemValue = ( 1.187066e-05 +- 9.825549e-06 ) GeV^-6 -TOTAL : 3.330534 sec +TOTAL : 3.359143 sec INFO: No Floating Point Exceptions have been reported INFO: No Floating Point Exceptions have been reported - 9,805,857,457 cycles # 2.941 GHz - 27,153,288,385 instructions # 2.77 insn per cycle - 3.335034911 seconds time elapsed + 9,837,543,704 cycles # 2.925 GHz + 27,153,515,644 instructions # 2.76 insn per cycle + 3.363547371 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4:96492) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttggg.mad/SubProcesses/P1_gg_ttxggg/build.sse4_d_inl0_hrd0/runTest_cpp.exe @@ -164,16 +164,16 @@ Process = SIGMA_SM_GG_TTXGGG_CPP [gcc 11.3.1] [inlineHel=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+BRDHST/avx2+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('avx2': AVX2, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 3.386158e+02 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.386550e+02 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.386550e+02 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 3.398355e+02 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.398782e+02 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.398782e+02 ) sec^-1 MeanMatrixElemValue = ( 1.187066e-05 +- 9.825549e-06 ) GeV^-6 -TOTAL : 1.562759 sec +TOTAL : 1.556417 sec INFO: No Floating Point Exceptions have been reported INFO: No Floating Point Exceptions have been reported - 4,284,138,212 cycles # 2.735 GHz - 9,593,930,746 instructions # 2.24 insn per cycle - 1.567182963 seconds time elapsed + 4,268,946,339 cycles # 2.736 GHz + 9,593,704,479 instructions # 2.25 insn per cycle + 1.560851267 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2:84961) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttggg.mad/SubProcesses/P1_gg_ttxggg/build.avx2_d_inl0_hrd0/runTest_cpp.exe @@ -198,16 +198,16 @@ Process = SIGMA_SM_GG_TTXGGG_CPP [gcc 11.3.1] [inlineHel=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+BRDHST/512y+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('512y': AVX512, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 3.892770e+02 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.893321e+02 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.893321e+02 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 3.822961e+02 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.823484e+02 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.823484e+02 ) sec^-1 MeanMatrixElemValue = ( 1.187066e-05 +- 9.825549e-06 ) GeV^-6 -TOTAL : 1.359134 sec +TOTAL : 1.383388 sec INFO: No Floating Point Exceptions have been reported INFO: No Floating Point Exceptions have been reported - 3,729,884,424 cycles # 2.737 GHz - 8,517,697,790 instructions # 2.28 insn per cycle - 1.363667603 seconds time elapsed + 3,782,833,998 cycles # 2.727 GHz + 8,518,016,663 instructions # 2.25 insn per cycle + 1.387835964 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2:80609) (512y: 90) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttggg.mad/SubProcesses/P1_gg_ttxggg/build.512y_d_inl0_hrd0/runTest_cpp.exe @@ -232,16 +232,16 @@ Process = SIGMA_SM_GG_TTXGGG_CPP [gcc 11.3.1] [inlineHel=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+BRDHST/512z+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('512z': AVX512, 512bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 3.423206e+02 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.423718e+02 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.423718e+02 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 3.424552e+02 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.425142e+02 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.425142e+02 ) sec^-1 MeanMatrixElemValue = ( 1.187066e-05 +- 9.825549e-06 ) GeV^-6 -TOTAL : 1.547281 sec +TOTAL : 1.543636 sec INFO: No Floating Point Exceptions have been reported INFO: No Floating Point Exceptions have been reported - 2,698,269,777 cycles # 1.739 GHz - 4,283,935,635 instructions # 1.59 insn per cycle - 1.552053679 seconds time elapsed + 2,701,518,227 cycles # 1.746 GHz + 4,284,495,680 instructions # 1.59 insn per cycle + 1.548295046 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 2856) (512y: 102) (512z:79114) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttggg.mad/SubProcesses/P1_gg_ttxggg/build.512z_d_inl0_hrd0/runTest_cpp.exe diff --git a/epochX/cudacpp/tput/logs_ggttggg_mad/log_ggttggg_mad_d_inl0_hrd1.txt b/epochX/cudacpp/tput/logs_ggttggg_mad/log_ggttggg_mad_d_inl0_hrd1.txt index bc67f5cacf..b079b632d0 100644 --- a/epochX/cudacpp/tput/logs_ggttggg_mad/log_ggttggg_mad_d_inl0_hrd1.txt +++ b/epochX/cudacpp/tput/logs_ggttggg_mad/log_ggttggg_mad_d_inl0_hrd1.txt @@ -21,7 +21,7 @@ make: Nothing to be done for 'all'. make: Nothing to be done for 'all'. -DATE: 2024-10-06_09:08:38 +DATE: 2024-10-06_13:37:48 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: ========================================================================= @@ -30,15 +30,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_SM_GG_TTXGGG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=1] Workflow summary = CUD:DBL+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 4.058591e+02 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 4.058974e+02 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 4.059077e+02 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 4.057507e+02 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 4.057894e+02 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 4.058057e+02 ) sec^-1 MeanMatrixElemValue = ( 1.187066e-05 +- 9.825549e-06 ) GeV^-6 -TOTAL : 2.451568 sec +TOTAL : 2.434662 sec INFO: No Floating Point Exceptions have been reported - 8,115,809,761 cycles # 2.919 GHz - 18,292,352,744 instructions # 2.25 insn per cycle - 2.835762935 seconds time elapsed + 8,080,434,276 cycles # 2.928 GHz + 18,433,666,131 instructions # 2.28 insn per cycle + 2.816222127 seconds time elapsed runNcu /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttggg.mad/SubProcesses/P1_gg_ttxggg/build.cuda_d_inl0_hrd1/check_cuda.exe -p 1 256 1 ==PROF== Profiling "sigmaKin": launch__registers_per_thread 255 ==PROF== Profiling "sigmaKin": sm__sass_average_branch_targets_threads_uniform.pct 100% @@ -48,15 +48,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_SM_GG_TTXGGG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=1] Workflow summary = CUD:DBL+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 9.228388e+03 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 9.230439e+03 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 9.230672e+03 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 9.253166e+03 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 9.255222e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 9.255472e+03 ) sec^-1 MeanMatrixElemValue = ( 1.856249e-04 +- 8.329951e-05 ) GeV^-6 -TOTAL : 4.019291 sec +TOTAL : 4.008140 sec INFO: No Floating Point Exceptions have been reported - 12,725,284,497 cycles # 2.922 GHz - 29,505,773,730 instructions # 2.32 insn per cycle - 4.410068917 seconds time elapsed + 12,666,534,126 cycles # 2.920 GHz + 30,194,831,941 instructions # 2.38 insn per cycle + 4.393384655 seconds time elapsed ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttggg.mad/SubProcesses/P1_gg_ttxggg/build.cuda_d_inl0_hrd1/runTest_cuda.exe INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: FE_DIVBYZERO, FE_INVALID, FE_OVERFLOW @@ -81,15 +81,15 @@ Process = SIGMA_SM_GG_TTXGGG_CPP [gcc 11.3.1] [inlineHel=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/none+NAVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = SCALAR ('none': ~vector[1], no SIMD) -EvtsPerSec[Rmb+ME] (23) = ( 7.905987e+01 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 7.906203e+01 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 7.906203e+01 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 7.886318e+01 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 7.886557e+01 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 7.886557e+01 ) sec^-1 MeanMatrixElemValue = ( 1.187066e-05 +- 9.825549e-06 ) GeV^-6 -TOTAL : 6.685741 sec +TOTAL : 6.699349 sec INFO: No Floating Point Exceptions have been reported - 18,901,791,742 cycles # 2.826 GHz - 53,936,334,501 instructions # 2.85 insn per cycle - 6.689520607 seconds time elapsed + 18,933,509,401 cycles # 2.825 GHz + 53,933,226,307 instructions # 2.85 insn per cycle + 6.703345685 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4:32022) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttggg.mad/SubProcesses/P1_gg_ttxggg/build.none_d_inl0_hrd1/runTest_cpp.exe @@ -113,15 +113,15 @@ Process = SIGMA_SM_GG_TTXGGG_CPP [gcc 11.3.1] [inlineHel=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[2] ('sse4': SSE4.2, 128bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 1.555988e+02 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.556078e+02 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.556078e+02 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.562530e+02 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.562616e+02 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.562616e+02 ) sec^-1 MeanMatrixElemValue = ( 1.187066e-05 +- 9.825549e-06 ) GeV^-6 -TOTAL : 3.395185 sec +TOTAL : 3.381822 sec INFO: No Floating Point Exceptions have been reported - 9,954,308,036 cycles # 2.929 GHz - 27,130,330,125 instructions # 2.73 insn per cycle - 3.399134205 seconds time elapsed + 9,933,647,149 cycles # 2.935 GHz + 27,131,936,820 instructions # 2.73 insn per cycle + 3.386036118 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4:96368) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttggg.mad/SubProcesses/P1_gg_ttxggg/build.sse4_d_inl0_hrd1/runTest_cpp.exe @@ -145,15 +145,15 @@ Process = SIGMA_SM_GG_TTXGGG_CPP [gcc 11.3.1] [inlineHel=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('avx2': AVX2, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 3.364235e+02 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.364649e+02 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.364649e+02 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 3.363944e+02 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.364338e+02 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.364338e+02 ) sec^-1 MeanMatrixElemValue = ( 1.187066e-05 +- 9.825549e-06 ) GeV^-6 -TOTAL : 1.571658 sec +TOTAL : 1.571305 sec INFO: No Floating Point Exceptions have been reported - 4,284,967,782 cycles # 2.721 GHz - 9,585,542,173 instructions # 2.24 insn per cycle - 1.575575323 seconds time elapsed + 4,305,710,743 cycles # 2.734 GHz + 9,585,486,941 instructions # 2.23 insn per cycle + 1.575549178 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2:84968) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttggg.mad/SubProcesses/P1_gg_ttxggg/build.avx2_d_inl0_hrd1/runTest_cpp.exe @@ -177,15 +177,15 @@ Process = SIGMA_SM_GG_TTXGGG_CPP [gcc 11.3.1] [inlineHel=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('512y': AVX512, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 3.898680e+02 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.899276e+02 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.899276e+02 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 3.896620e+02 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.897155e+02 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.897155e+02 ) sec^-1 MeanMatrixElemValue = ( 1.187066e-05 +- 9.825549e-06 ) GeV^-6 -TOTAL : 1.358371 sec +TOTAL : 1.356720 sec INFO: No Floating Point Exceptions have been reported - 3,717,774,700 cycles # 2.731 GHz - 8,507,853,536 instructions # 2.29 insn per cycle - 1.362296235 seconds time elapsed + 3,717,000,862 cycles # 2.732 GHz + 8,507,862,265 instructions # 2.29 insn per cycle + 1.361038289 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2:80632) (512y: 240) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttggg.mad/SubProcesses/P1_gg_ttxggg/build.512y_d_inl0_hrd1/runTest_cpp.exe @@ -209,15 +209,15 @@ Process = SIGMA_SM_GG_TTXGGG_CPP [gcc 11.3.1] [inlineHel=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('512z': AVX512, 512bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 3.399522e+02 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.400013e+02 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.400013e+02 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 3.420215e+02 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.420705e+02 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.420705e+02 ) sec^-1 MeanMatrixElemValue = ( 1.187066e-05 +- 9.825549e-06 ) GeV^-6 -TOTAL : 1.555521 sec +TOTAL : 1.545915 sec INFO: No Floating Point Exceptions have been reported - 2,693,302,897 cycles # 1.729 GHz - 4,281,674,096 instructions # 1.59 insn per cycle - 1.559394081 seconds time elapsed + 2,700,127,917 cycles # 1.743 GHz + 4,281,664,135 instructions # 1.59 insn per cycle + 1.550500292 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 2693) (512y: 184) (512z:79098) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttggg.mad/SubProcesses/P1_gg_ttxggg/build.512z_d_inl0_hrd1/runTest_cpp.exe diff --git a/epochX/cudacpp/tput/logs_ggttggg_mad/log_ggttggg_mad_f_inl0_hrd0.txt b/epochX/cudacpp/tput/logs_ggttggg_mad/log_ggttggg_mad_f_inl0_hrd0.txt index e477be7c61..00f664a687 100644 --- a/epochX/cudacpp/tput/logs_ggttggg_mad/log_ggttggg_mad_f_inl0_hrd0.txt +++ b/epochX/cudacpp/tput/logs_ggttggg_mad/log_ggttggg_mad_f_inl0_hrd0.txt @@ -21,7 +21,7 @@ make: Nothing to be done for 'all'. make: Nothing to be done for 'all'. -DATE: 2024-10-06_09:13:00 +DATE: 2024-10-06_13:42:07 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: ========================================================================= @@ -30,15 +30,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_SM_GG_TTXGGG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 6.207250e+02 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 6.207995e+02 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 6.208247e+02 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 6.208759e+02 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 6.209618e+02 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 6.209820e+02 ) sec^-1 MeanMatrixElemValue = ( 1.186984e-05 +- 9.824899e-06 ) GeV^-6 -TOTAL : 1.762040 sec +TOTAL : 1.743632 sec INFO: No Floating Point Exceptions have been reported - 5,937,636,063 cycles # 2.916 GHz - 12,374,083,331 instructions # 2.08 insn per cycle - 2.091996677 seconds time elapsed + 5,874,577,159 cycles # 2.920 GHz + 11,865,690,030 instructions # 2.02 insn per cycle + 2.068805444 seconds time elapsed runNcu /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttggg.mad/SubProcesses/P1_gg_ttxggg/build.cuda_f_inl0_hrd0/check_cuda.exe -p 1 256 1 ==PROF== Profiling "sigmaKin": launch__registers_per_thread 255 ==PROF== Profiling "sigmaKin": sm__sass_average_branch_targets_threads_uniform.pct 100% @@ -48,15 +48,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_SM_GG_TTXGGG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 2.149439e+04 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 2.150073e+04 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 2.150179e+04 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 2.125566e+04 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 2.126183e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.126266e+04 ) sec^-1 MeanMatrixElemValue = ( 1.856829e-04 +- 8.333437e-05 ) GeV^-6 -TOTAL : 2.066345 sec +TOTAL : 2.064937 sec INFO: No Floating Point Exceptions have been reported - 6,803,203,568 cycles # 2.918 GHz - 14,656,096,283 instructions # 2.15 insn per cycle - 2.390130877 seconds time elapsed + 6,840,656,997 cycles # 2.930 GHz + 14,802,831,433 instructions # 2.16 insn per cycle + 2.391457760 seconds time elapsed ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttggg.mad/SubProcesses/P1_gg_ttxggg/build.cuda_f_inl0_hrd0/runTest_cuda.exe INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: FE_DIVBYZERO, FE_INVALID, FE_OVERFLOW @@ -81,15 +81,15 @@ Process = SIGMA_SM_GG_TTXGGG_CPP [gcc 11.3.1] [inlineHel=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/none+NAVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = SCALAR ('none': ~vector[1], no SIMD) -EvtsPerSec[Rmb+ME] (23) = ( 8.548424e+01 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 8.548685e+01 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 8.548685e+01 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 8.566565e+01 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 8.566823e+01 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 8.566823e+01 ) sec^-1 MeanMatrixElemValue = ( 1.187013e-05 +- 9.825040e-06 ) GeV^-6 -TOTAL : 6.179003 sec +TOTAL : 6.165105 sec INFO: No Floating Point Exceptions have been reported - 18,168,840,210 cycles # 2.939 GHz - 53,911,011,794 instructions # 2.97 insn per cycle - 6.183081263 seconds time elapsed + 18,129,629,131 cycles # 2.939 GHz + 53,910,982,907 instructions # 2.97 insn per cycle + 6.169177584 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4:20141) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttggg.mad/SubProcesses/P1_gg_ttxggg/build.none_f_inl0_hrd0/runTest_cpp.exe @@ -113,15 +113,15 @@ Process = SIGMA_SM_GG_TTXGGG_CPP [gcc 11.3.1] [inlineHel=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('sse4': SSE4.2, 128bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 3.395658e+02 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.396067e+02 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.396067e+02 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 3.392968e+02 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.393373e+02 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.393373e+02 ) sec^-1 MeanMatrixElemValue = ( 1.187013e-05 +- 9.825037e-06 ) GeV^-6 -TOTAL : 1.556967 sec +TOTAL : 1.558200 sec INFO: No Floating Point Exceptions have been reported - 4,597,936,627 cycles # 2.947 GHz - 13,808,300,252 instructions # 3.00 insn per cycle - 1.560798930 seconds time elapsed + 4,594,592,424 cycles # 2.942 GHz + 13,807,607,540 instructions # 3.01 insn per cycle + 1.562387640 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4:97016) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttggg.mad/SubProcesses/P1_gg_ttxggg/build.sse4_f_inl0_hrd0/runTest_cpp.exe @@ -145,15 +145,15 @@ Process = SIGMA_SM_GG_TTXGGG_CPP [gcc 11.3.1] [inlineHel=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('avx2': AVX2, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 6.833708e+02 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 6.835461e+02 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 6.835461e+02 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 6.868426e+02 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 6.870195e+02 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 6.870195e+02 ) sec^-1 MeanMatrixElemValue = ( 1.187188e-05 +- 9.826767e-06 ) GeV^-6 -TOTAL : 0.774770 sec +TOTAL : 0.770782 sec INFO: No Floating Point Exceptions have been reported - 2,127,367,774 cycles # 2.734 GHz - 4,836,875,487 instructions # 2.27 insn per cycle - 0.778636721 seconds time elapsed + 2,124,637,907 cycles # 2.744 GHz + 4,836,919,651 instructions # 2.28 insn per cycle + 0.774932950 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2:85494) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttggg.mad/SubProcesses/P1_gg_ttxggg/build.avx2_f_inl0_hrd0/runTest_cpp.exe @@ -177,15 +177,15 @@ Process = SIGMA_SM_GG_TTXGGG_CPP [gcc 11.3.1] [inlineHel=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('512y': AVX512, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 7.729108e+02 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 7.731291e+02 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 7.731291e+02 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 7.577511e+02 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 7.579596e+02 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 7.579596e+02 ) sec^-1 MeanMatrixElemValue = ( 1.187188e-05 +- 9.826767e-06 ) GeV^-6 -TOTAL : 0.685221 sec +TOTAL : 0.698095 sec INFO: No Floating Point Exceptions have been reported - 1,884,703,570 cycles # 2.737 GHz - 4,291,263,737 instructions # 2.28 insn per cycle - 0.689203509 seconds time elapsed + 1,882,945,377 cycles # 2.683 GHz + 4,291,270,918 instructions # 2.28 insn per cycle + 0.702323588 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2:81183) (512y: 45) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttggg.mad/SubProcesses/P1_gg_ttxggg/build.512y_f_inl0_hrd0/runTest_cpp.exe @@ -209,15 +209,15 @@ Process = SIGMA_SM_GG_TTXGGG_CPP [gcc 11.3.1] [inlineHel=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[16] ('512z': AVX512, 512bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 6.870048e+02 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 6.872187e+02 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 6.872187e+02 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 6.876416e+02 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 6.878654e+02 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 6.878654e+02 ) sec^-1 MeanMatrixElemValue = ( 1.187188e-05 +- 9.826771e-06 ) GeV^-6 -TOTAL : 0.771101 sec +TOTAL : 0.770420 sec INFO: No Floating Point Exceptions have been reported - 1,354,646,750 cycles # 1.748 GHz - 2,162,779,823 instructions # 1.60 insn per cycle - 0.775438585 seconds time elapsed + 1,354,563,133 cycles # 1.750 GHz + 2,162,744,719 instructions # 1.60 insn per cycle + 0.774864328 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 3481) (512y: 45) (512z:79330) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttggg.mad/SubProcesses/P1_gg_ttxggg/build.512z_f_inl0_hrd0/runTest_cpp.exe diff --git a/epochX/cudacpp/tput/logs_ggttggg_mad/log_ggttggg_mad_f_inl0_hrd0_bridge.txt b/epochX/cudacpp/tput/logs_ggttggg_mad/log_ggttggg_mad_f_inl0_hrd0_bridge.txt index 09d523a948..56fc8272f1 100644 --- a/epochX/cudacpp/tput/logs_ggttggg_mad/log_ggttggg_mad_f_inl0_hrd0_bridge.txt +++ b/epochX/cudacpp/tput/logs_ggttggg_mad/log_ggttggg_mad_f_inl0_hrd0_bridge.txt @@ -21,7 +21,7 @@ make: Nothing to be done for 'all'. make: Nothing to be done for 'all'. -DATE: 2024-10-06_09:42:06 +DATE: 2024-10-06_14:00:39 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: ========================================================================= @@ -34,16 +34,16 @@ WARNING! Set grid in Bridge (nevt=256, gpublocks=1, gputhreads=256, gpublocks*gp Process = SIGMA_SM_GG_TTXGGG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 6.291704e+02 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 6.296560e+02 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 6.296560e+02 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 6.295103e+02 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 6.300241e+02 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 6.300241e+02 ) sec^-1 MeanMatrixElemValue = ( 1.187093e-05 +- 9.825663e-06 ) GeV^-6 -TOTAL : 1.680127 sec +TOTAL : 1.684758 sec INFO: No Floating Point Exceptions have been reported INFO: No Floating Point Exceptions have been reported - 5,675,085,647 cycles # 2.923 GHz - 11,509,492,893 instructions # 2.03 insn per cycle - 1.997903242 seconds time elapsed + 5,700,510,430 cycles # 2.911 GHz + 10,234,376,160 instructions # 1.80 insn per cycle + 2.015386162 seconds time elapsed runNcu /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttggg.mad/SubProcesses/P1_gg_ttxggg/build.cuda_f_inl0_hrd0/check_cuda.exe -p 1 256 1 --bridge WARNING! Bridge selected: cannot use RamboDevice, will use RamboHost WARNING! RamboHost selected: cannot use CurandDevice, will use CurandHost @@ -61,16 +61,16 @@ WARNING! Set grid in Bridge (nevt=16384, gpublocks=64, gputhreads=256, gpublocks Process = SIGMA_SM_GG_TTXGGG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 2.120892e+04 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 2.132073e+04 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 2.132073e+04 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 2.139854e+04 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 2.151336e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.151336e+04 ) sec^-1 MeanMatrixElemValue = ( 1.856440e-04 +- 8.331091e-05 ) GeV^-6 -TOTAL : 2.037220 sec +TOTAL : 2.034763 sec INFO: No Floating Point Exceptions have been reported INFO: No Floating Point Exceptions have been reported - 6,712,310,342 cycles # 2.924 GHz - 13,777,135,261 instructions # 2.05 insn per cycle - 2.354099539 seconds time elapsed + 6,701,866,598 cycles # 2.920 GHz + 14,655,886,015 instructions # 2.19 insn per cycle + 2.352695614 seconds time elapsed ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttggg.mad/SubProcesses/P1_gg_ttxggg/build.cuda_f_inl0_hrd0/runTest_cuda.exe INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: FE_DIVBYZERO, FE_INVALID, FE_OVERFLOW @@ -96,16 +96,16 @@ Process = SIGMA_SM_GG_TTXGGG_CPP [gcc 11.3.1] [inlineHel=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+BRDHST/none+NAVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = SCALAR ('none': ~vector[1], no SIMD) -EvtsPerSec[Rmb+ME] (23) = ( 8.574125e+01 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 8.574397e+01 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 8.574397e+01 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 8.548835e+01 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 8.549095e+01 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 8.549095e+01 ) sec^-1 MeanMatrixElemValue = ( 1.187013e-05 +- 9.825040e-06 ) GeV^-6 -TOTAL : 6.159980 sec +TOTAL : 6.179117 sec INFO: No Floating Point Exceptions have been reported INFO: No Floating Point Exceptions have been reported - 18,121,008,944 cycles # 2.940 GHz - 53,916,989,652 instructions # 2.98 insn per cycle - 6.164330765 seconds time elapsed + 18,163,498,222 cycles # 2.938 GHz + 53,915,750,690 instructions # 2.97 insn per cycle + 6.183344434 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4:20141) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttggg.mad/SubProcesses/P1_gg_ttxggg/build.none_f_inl0_hrd0/runTest_cpp.exe @@ -130,16 +130,16 @@ Process = SIGMA_SM_GG_TTXGGG_CPP [gcc 11.3.1] [inlineHel=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+BRDHST/sse4+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('sse4': SSE4.2, 128bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 3.371688e+02 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.372089e+02 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.372089e+02 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 3.391067e+02 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.391513e+02 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.391513e+02 ) sec^-1 MeanMatrixElemValue = ( 1.187013e-05 +- 9.825037e-06 ) GeV^-6 -TOTAL : 1.568419 sec +TOTAL : 1.558852 sec INFO: No Floating Point Exceptions have been reported INFO: No Floating Point Exceptions have been reported - 4,624,959,734 cycles # 2.942 GHz - 13,809,578,618 instructions # 2.99 insn per cycle - 1.572870258 seconds time elapsed + 4,602,065,463 cycles # 2.945 GHz + 13,809,542,979 instructions # 3.00 insn per cycle + 1.563316731 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4:97016) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttggg.mad/SubProcesses/P1_gg_ttxggg/build.sse4_f_inl0_hrd0/runTest_cpp.exe @@ -164,16 +164,16 @@ Process = SIGMA_SM_GG_TTXGGG_CPP [gcc 11.3.1] [inlineHel=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+BRDHST/avx2+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('avx2': AVX2, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 6.853120e+02 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 6.854860e+02 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 6.854860e+02 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 6.795177e+02 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 6.796847e+02 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 6.796847e+02 ) sec^-1 MeanMatrixElemValue = ( 1.187188e-05 +- 9.826767e-06 ) GeV^-6 -TOTAL : 0.772760 sec +TOTAL : 0.779244 sec INFO: No Floating Point Exceptions have been reported INFO: No Floating Point Exceptions have been reported - 2,127,660,282 cycles # 2.740 GHz - 4,839,303,130 instructions # 2.27 insn per cycle - 0.777110537 seconds time elapsed + 2,133,913,657 cycles # 2.726 GHz + 4,838,988,858 instructions # 2.27 insn per cycle + 0.783553929 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2:85494) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttggg.mad/SubProcesses/P1_gg_ttxggg/build.avx2_f_inl0_hrd0/runTest_cpp.exe @@ -198,16 +198,16 @@ Process = SIGMA_SM_GG_TTXGGG_CPP [gcc 11.3.1] [inlineHel=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+BRDHST/512y+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('512y': AVX512, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 7.707103e+02 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 7.709607e+02 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 7.709607e+02 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 7.688746e+02 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 7.690985e+02 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 7.690985e+02 ) sec^-1 MeanMatrixElemValue = ( 1.187188e-05 +- 9.826767e-06 ) GeV^-6 -TOTAL : 0.687680 sec +TOTAL : 0.689430 sec INFO: No Floating Point Exceptions have been reported INFO: No Floating Point Exceptions have been reported - 1,889,891,398 cycles # 2.733 GHz - 4,293,271,631 instructions # 2.27 insn per cycle - 0.692031150 seconds time elapsed + 1,899,795,909 cycles # 2.741 GHz + 4,293,253,577 instructions # 2.26 insn per cycle + 0.693818174 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2:81183) (512y: 45) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttggg.mad/SubProcesses/P1_gg_ttxggg/build.512y_f_inl0_hrd0/runTest_cpp.exe @@ -232,16 +232,16 @@ Process = SIGMA_SM_GG_TTXGGG_CPP [gcc 11.3.1] [inlineHel=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+BRDHST/512z+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[16] ('512z': AVX512, 512bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 6.738421e+02 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 6.740575e+02 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 6.740575e+02 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 6.906370e+02 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 6.908687e+02 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 6.908687e+02 ) sec^-1 MeanMatrixElemValue = ( 1.187188e-05 +- 9.826771e-06 ) GeV^-6 -TOTAL : 0.785848 sec +TOTAL : 0.767940 sec INFO: No Floating Point Exceptions have been reported INFO: No Floating Point Exceptions have been reported - 1,358,106,687 cycles # 1.720 GHz - 2,165,384,980 instructions # 1.59 insn per cycle - 0.790493646 seconds time elapsed + 1,356,687,429 cycles # 1.758 GHz + 2,164,841,068 instructions # 1.60 insn per cycle + 0.772551506 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 3481) (512y: 45) (512z:79330) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttggg.mad/SubProcesses/P1_gg_ttxggg/build.512z_f_inl0_hrd0/runTest_cpp.exe diff --git a/epochX/cudacpp/tput/logs_ggttggg_mad/log_ggttggg_mad_f_inl0_hrd1.txt b/epochX/cudacpp/tput/logs_ggttggg_mad/log_ggttggg_mad_f_inl0_hrd1.txt index 33a64296d4..86380a02ff 100644 --- a/epochX/cudacpp/tput/logs_ggttggg_mad/log_ggttggg_mad_f_inl0_hrd1.txt +++ b/epochX/cudacpp/tput/logs_ggttggg_mad/log_ggttggg_mad_f_inl0_hrd1.txt @@ -21,7 +21,7 @@ make: Nothing to be done for 'all'. make: Nothing to be done for 'all'. -DATE: 2024-10-06_09:14:03 +DATE: 2024-10-06_13:43:10 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: ========================================================================= @@ -30,15 +30,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_SM_GG_TTXGGG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=1] Workflow summary = CUD:FLT+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 6.196404e+02 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 6.197145e+02 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 6.197475e+02 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 6.193918e+02 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 6.194639e+02 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 6.194913e+02 ) sec^-1 MeanMatrixElemValue = ( 1.186984e-05 +- 9.824899e-06 ) GeV^-6 -TOTAL : 1.762965 sec +TOTAL : 1.746576 sec INFO: No Floating Point Exceptions have been reported - 5,951,937,078 cycles # 2.924 GHz - 11,910,577,864 instructions # 2.00 insn per cycle - 2.092003198 seconds time elapsed + 5,884,488,777 cycles # 2.920 GHz + 12,658,099,656 instructions # 2.15 insn per cycle + 2.071697511 seconds time elapsed runNcu /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttggg.mad/SubProcesses/P1_gg_ttxggg/build.cuda_f_inl0_hrd1/check_cuda.exe -p 1 256 1 ==PROF== Profiling "sigmaKin": launch__registers_per_thread 255 ==PROF== Profiling "sigmaKin": sm__sass_average_branch_targets_threads_uniform.pct 100% @@ -48,15 +48,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_SM_GG_TTXGGG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=1] Workflow summary = CUD:FLT+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 2.150073e+04 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 2.150749e+04 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 2.150840e+04 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 2.149700e+04 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 2.150325e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.150420e+04 ) sec^-1 MeanMatrixElemValue = ( 1.856829e-04 +- 8.333437e-05 ) GeV^-6 -TOTAL : 2.074025 sec +TOTAL : 2.054180 sec INFO: No Floating Point Exceptions have been reported - 6,857,187,374 cycles # 2.930 GHz - 14,190,515,168 instructions # 2.07 insn per cycle - 2.396988151 seconds time elapsed + 6,755,266,379 cycles # 2.914 GHz + 14,596,296,050 instructions # 2.16 insn per cycle + 2.374551989 seconds time elapsed ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttggg.mad/SubProcesses/P1_gg_ttxggg/build.cuda_f_inl0_hrd1/runTest_cuda.exe INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: FE_DIVBYZERO, FE_INVALID, FE_OVERFLOW @@ -81,15 +81,15 @@ Process = SIGMA_SM_GG_TTXGGG_CPP [gcc 11.3.1] [inlineHel=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/none+NAVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = SCALAR ('none': ~vector[1], no SIMD) -EvtsPerSec[Rmb+ME] (23) = ( 8.597266e+01 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 8.597536e+01 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 8.597536e+01 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 8.596387e+01 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 8.596675e+01 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 8.596675e+01 ) sec^-1 MeanMatrixElemValue = ( 1.187013e-05 +- 9.825040e-06 ) GeV^-6 -TOTAL : 6.144692 sec +TOTAL : 6.144911 sec INFO: No Floating Point Exceptions have been reported - 18,086,727,911 cycles # 2.942 GHz - 53,895,836,183 instructions # 2.98 insn per cycle - 6.148512893 seconds time elapsed + 18,050,189,832 cycles # 2.936 GHz + 53,895,700,840 instructions # 2.99 insn per cycle + 6.148976589 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4:20141) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttggg.mad/SubProcesses/P1_gg_ttxggg/build.none_f_inl0_hrd1/runTest_cpp.exe @@ -113,15 +113,15 @@ Process = SIGMA_SM_GG_TTXGGG_CPP [gcc 11.3.1] [inlineHel=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('sse4': SSE4.2, 128bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 3.388656e+02 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.389069e+02 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.389069e+02 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 3.418371e+02 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.418782e+02 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.418782e+02 ) sec^-1 MeanMatrixElemValue = ( 1.187013e-05 +- 9.825037e-06 ) GeV^-6 -TOTAL : 1.560721 sec +TOTAL : 1.545917 sec INFO: No Floating Point Exceptions have been reported - 4,571,260,015 cycles # 2.924 GHz - 13,800,942,063 instructions # 3.02 insn per cycle - 1.564719207 seconds time elapsed + 4,569,497,435 cycles # 2.950 GHz + 13,801,513,556 instructions # 3.02 insn per cycle + 1.550025915 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4:96651) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttggg.mad/SubProcesses/P1_gg_ttxggg/build.sse4_f_inl0_hrd1/runTest_cpp.exe @@ -145,15 +145,15 @@ Process = SIGMA_SM_GG_TTXGGG_CPP [gcc 11.3.1] [inlineHel=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('avx2': AVX2, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 6.702410e+02 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 6.704003e+02 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 6.704003e+02 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 6.821733e+02 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 6.823414e+02 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 6.823414e+02 ) sec^-1 MeanMatrixElemValue = ( 1.187188e-05 +- 9.826767e-06 ) GeV^-6 -TOTAL : 0.789887 sec +TOTAL : 0.775925 sec INFO: No Floating Point Exceptions have been reported - 2,151,012,254 cycles # 2.712 GHz - 4,840,938,021 instructions # 2.25 insn per cycle - 0.793816354 seconds time elapsed + 2,142,487,327 cycles # 2.749 GHz + 4,840,946,294 instructions # 2.26 insn per cycle + 0.780095046 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2:85884) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttggg.mad/SubProcesses/P1_gg_ttxggg/build.avx2_f_inl0_hrd1/runTest_cpp.exe @@ -177,15 +177,15 @@ Process = SIGMA_SM_GG_TTXGGG_CPP [gcc 11.3.1] [inlineHel=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('512y': AVX512, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 7.657646e+02 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 7.659745e+02 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 7.659745e+02 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 7.690038e+02 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 7.692093e+02 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 7.692093e+02 ) sec^-1 MeanMatrixElemValue = ( 1.187188e-05 +- 9.826767e-06 ) GeV^-6 -TOTAL : 0.691425 sec +TOTAL : 0.688303 sec INFO: No Floating Point Exceptions have been reported - 1,894,431,690 cycles # 2.727 GHz - 4,294,884,277 instructions # 2.27 insn per cycle - 0.695223368 seconds time elapsed + 1,889,212,473 cycles # 2.731 GHz + 4,294,832,216 instructions # 2.27 insn per cycle + 0.692479627 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2:81725) (512y: 25) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttggg.mad/SubProcesses/P1_gg_ttxggg/build.512y_f_inl0_hrd1/runTest_cpp.exe @@ -209,15 +209,15 @@ Process = SIGMA_SM_GG_TTXGGG_CPP [gcc 11.3.1] [inlineHel=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[16] ('512z': AVX512, 512bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 6.673392e+02 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 6.675470e+02 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 6.675470e+02 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 6.892742e+02 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 6.894791e+02 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 6.894791e+02 ) sec^-1 MeanMatrixElemValue = ( 1.187188e-05 +- 9.826771e-06 ) GeV^-6 -TOTAL : 0.793743 sec +TOTAL : 0.768147 sec INFO: No Floating Point Exceptions have been reported - 1,366,656,580 cycles # 1.715 GHz - 2,169,713,805 instructions # 1.59 insn per cycle - 0.797745119 seconds time elapsed + 1,358,169,653 cycles # 1.760 GHz + 2,169,697,562 instructions # 1.60 insn per cycle + 0.772513578 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 4092) (512y: 32) (512z:79551) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttggg.mad/SubProcesses/P1_gg_ttxggg/build.512z_f_inl0_hrd1/runTest_cpp.exe diff --git a/epochX/cudacpp/tput/logs_ggttggg_mad/log_ggttggg_mad_m_inl0_hrd0.txt b/epochX/cudacpp/tput/logs_ggttggg_mad/log_ggttggg_mad_m_inl0_hrd0.txt index a2a6307c02..20d12e1049 100644 --- a/epochX/cudacpp/tput/logs_ggttggg_mad/log_ggttggg_mad_m_inl0_hrd0.txt +++ b/epochX/cudacpp/tput/logs_ggttggg_mad/log_ggttggg_mad_m_inl0_hrd0.txt @@ -21,7 +21,7 @@ make: Nothing to be done for 'all'. make: Nothing to be done for 'all'. -DATE: 2024-10-06_09:10:06 +DATE: 2024-10-06_13:39:16 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: ========================================================================= @@ -30,15 +30,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_SM_GG_TTXGGG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:MIX+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 4.665934e+02 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 4.666477e+02 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 4.666666e+02 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 4.670546e+02 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 4.671064e+02 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 4.671284e+02 ) sec^-1 MeanMatrixElemValue = ( 1.187066e-05 +- 9.825548e-06 ) GeV^-6 -TOTAL : 2.202831 sec +TOTAL : 2.182847 sec INFO: No Floating Point Exceptions have been reported - 7,373,914,452 cycles # 2.913 GHz - 16,351,055,335 instructions # 2.22 insn per cycle - 2.588547453 seconds time elapsed + 7,339,013,303 cycles # 2.923 GHz + 16,206,540,336 instructions # 2.21 insn per cycle + 2.567572063 seconds time elapsed runNcu /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttggg.mad/SubProcesses/P1_gg_ttxggg/build.cuda_m_inl0_hrd0/check_cuda.exe -p 1 256 1 ==PROF== Profiling "sigmaKin": launch__registers_per_thread 255 ==PROF== Profiling "sigmaKin": sm__sass_average_branch_targets_threads_uniform.pct 100% @@ -48,15 +48,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_SM_GG_TTXGGG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:MIX+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 1.110897e+04 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.111188e+04 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.111222e+04 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.110741e+04 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.111040e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.111071e+04 ) sec^-1 MeanMatrixElemValue = ( 1.856249e-04 +- 8.329951e-05 ) GeV^-6 -TOTAL : 3.442430 sec +TOTAL : 3.418518 sec INFO: No Floating Point Exceptions have been reported - 11,070,694,428 cycles # 2.924 GHz - 25,628,142,124 instructions # 2.31 insn per cycle - 3.841933628 seconds time elapsed + 11,000,599,863 cycles # 2.932 GHz + 24,071,664,037 instructions # 2.19 insn per cycle + 3.807460255 seconds time elapsed ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttggg.mad/SubProcesses/P1_gg_ttxggg/build.cuda_m_inl0_hrd0/runTest_cuda.exe INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: FE_DIVBYZERO, FE_INVALID, FE_OVERFLOW @@ -81,15 +81,15 @@ Process = SIGMA_SM_GG_TTXGGG_CPP [gcc 11.3.1] [inlineHel=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/none+NAVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) Internal loops fptype_sv = SCALAR ('none': ~vector[1], no SIMD) -EvtsPerSec[Rmb+ME] (23) = ( 7.567548e+01 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 7.567783e+01 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 7.567783e+01 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 7.723475e+01 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 7.723699e+01 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 7.723699e+01 ) sec^-1 MeanMatrixElemValue = ( 1.187066e-05 +- 9.825549e-06 ) GeV^-6 -TOTAL : 6.978728 sec +TOTAL : 6.833597 sec INFO: No Floating Point Exceptions have been reported - 19,201,924,470 cycles # 2.751 GHz - 54,137,446,015 instructions # 2.82 insn per cycle - 6.982563293 seconds time elapsed + 19,221,657,171 cycles # 2.811 GHz + 54,133,604,262 instructions # 2.82 insn per cycle + 6.837834193 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4:32000) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttggg.mad/SubProcesses/P1_gg_ttxggg/build.none_m_inl0_hrd0/runTest_cpp.exe @@ -113,15 +113,15 @@ Process = SIGMA_SM_GG_TTXGGG_CPP [gcc 11.3.1] [inlineHel=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[2] ('sse4': SSE4.2, 128bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 1.526848e+02 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.526939e+02 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.526939e+02 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.542063e+02 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.542147e+02 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.542147e+02 ) sec^-1 MeanMatrixElemValue = ( 1.187066e-05 +- 9.825548e-06 ) GeV^-6 -TOTAL : 3.460419 sec +TOTAL : 3.426925 sec INFO: No Floating Point Exceptions have been reported - 9,442,620,757 cycles # 2.727 GHz - 26,188,001,033 instructions # 2.77 insn per cycle - 3.464377416 seconds time elapsed + 9,390,083,285 cycles # 2.738 GHz + 26,189,253,872 instructions # 2.79 insn per cycle + 3.431097664 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4:96049) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttggg.mad/SubProcesses/P1_gg_ttxggg/build.sse4_m_inl0_hrd0/runTest_cpp.exe @@ -145,15 +145,15 @@ Process = SIGMA_SM_GG_TTXGGG_CPP [gcc 11.3.1] [inlineHel=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('avx2': AVX2, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 3.548969e+02 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.549418e+02 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.549418e+02 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 3.569406e+02 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.569847e+02 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.569847e+02 ) sec^-1 MeanMatrixElemValue = ( 1.187066e-05 +- 9.825548e-06 ) GeV^-6 -TOTAL : 1.489614 sec +TOTAL : 1.481620 sec INFO: No Floating Point Exceptions have been reported - 4,075,741,004 cycles # 2.731 GHz - 9,249,825,182 instructions # 2.27 insn per cycle - 1.493453651 seconds time elapsed + 4,065,692,206 cycles # 2.738 GHz + 9,250,036,134 instructions # 2.28 insn per cycle + 1.485829823 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2:84390) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttggg.mad/SubProcesses/P1_gg_ttxggg/build.avx2_m_inl0_hrd0/runTest_cpp.exe @@ -177,15 +177,15 @@ Process = SIGMA_SM_GG_TTXGGG_CPP [gcc 11.3.1] [inlineHel=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('512y': AVX512, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 4.098256e+02 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 4.098850e+02 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 4.098850e+02 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 4.125484e+02 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 4.126111e+02 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 4.126111e+02 ) sec^-1 MeanMatrixElemValue = ( 1.187066e-05 +- 9.825548e-06 ) GeV^-6 -TOTAL : 1.290484 sec +TOTAL : 1.282256 sec INFO: No Floating Point Exceptions have been reported - 3,523,951,603 cycles # 2.724 GHz - 8,183,239,467 instructions # 2.32 insn per cycle - 1.294382992 seconds time elapsed + 3,499,188,086 cycles # 2.721 GHz + 8,183,379,655 instructions # 2.34 insn per cycle + 1.286627749 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2:80015) (512y: 80) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttggg.mad/SubProcesses/P1_gg_ttxggg/build.512y_m_inl0_hrd0/runTest_cpp.exe @@ -209,15 +209,15 @@ Process = SIGMA_SM_GG_TTXGGG_CPP [gcc 11.3.1] [inlineHel=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('512z': AVX512, 512bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 3.495372e+02 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.495944e+02 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.495944e+02 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 3.473280e+02 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.473814e+02 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.473814e+02 ) sec^-1 MeanMatrixElemValue = ( 1.187066e-05 +- 9.825548e-06 ) GeV^-6 -TOTAL : 1.513924 sec +TOTAL : 1.522209 sec INFO: No Floating Point Exceptions have been reported - 2,658,314,764 cycles # 1.752 GHz - 4,173,156,780 instructions # 1.57 insn per cycle - 1.517996809 seconds time elapsed + 2,664,185,066 cycles # 1.746 GHz + 4,173,234,337 instructions # 1.57 insn per cycle + 1.526490808 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 2615) (512y: 92) (512z:78910) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttggg.mad/SubProcesses/P1_gg_ttxggg/build.512z_m_inl0_hrd0/runTest_cpp.exe diff --git a/epochX/cudacpp/tput/logs_ggttggg_mad/log_ggttggg_mad_m_inl0_hrd1.txt b/epochX/cudacpp/tput/logs_ggttggg_mad/log_ggttggg_mad_m_inl0_hrd1.txt index 67fff86657..23825a4c65 100644 --- a/epochX/cudacpp/tput/logs_ggttggg_mad/log_ggttggg_mad_m_inl0_hrd1.txt +++ b/epochX/cudacpp/tput/logs_ggttggg_mad/log_ggttggg_mad_m_inl0_hrd1.txt @@ -21,7 +21,7 @@ make: Nothing to be done for 'all'. make: Nothing to be done for 'all'. -DATE: 2024-10-06_09:11:33 +DATE: 2024-10-06_13:40:42 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: ========================================================================= @@ -30,15 +30,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_SM_GG_TTXGGG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=1] Workflow summary = CUD:MIX+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 4.667678e+02 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 4.668217e+02 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 4.668387e+02 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 4.666875e+02 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 4.667372e+02 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 4.667596e+02 ) sec^-1 MeanMatrixElemValue = ( 1.187066e-05 +- 9.825548e-06 ) GeV^-6 -TOTAL : 2.202686 sec +TOTAL : 2.187635 sec INFO: No Floating Point Exceptions have been reported - 7,336,606,843 cycles # 2.899 GHz - 15,241,236,080 instructions # 2.08 insn per cycle - 2.586897924 seconds time elapsed + 7,341,930,161 cycles # 2.920 GHz + 16,443,581,345 instructions # 2.24 insn per cycle + 2.571225982 seconds time elapsed runNcu /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttggg.mad/SubProcesses/P1_gg_ttxggg/build.cuda_m_inl0_hrd1/check_cuda.exe -p 1 256 1 ==PROF== Profiling "sigmaKin": launch__registers_per_thread 255 ==PROF== Profiling "sigmaKin": sm__sass_average_branch_targets_threads_uniform.pct 100% @@ -48,15 +48,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_SM_GG_TTXGGG_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=1] Workflow summary = CUD:MIX+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 1.107552e+04 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.107855e+04 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.107889e+04 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.103846e+04 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.104190e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.104224e+04 ) sec^-1 MeanMatrixElemValue = ( 1.856249e-04 +- 8.329951e-05 ) GeV^-6 -TOTAL : 3.440073 sec +TOTAL : 3.428277 sec INFO: No Floating Point Exceptions have been reported - 11,052,276,434 cycles # 2.923 GHz - 25,411,180,343 instructions # 2.30 insn per cycle - 3.836365671 seconds time elapsed + 10,980,344,266 cycles # 2.924 GHz + 25,761,501,781 instructions # 2.35 insn per cycle + 3.814324829 seconds time elapsed ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttggg.mad/SubProcesses/P1_gg_ttxggg/build.cuda_m_inl0_hrd1/runTest_cuda.exe INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: FE_DIVBYZERO, FE_INVALID, FE_OVERFLOW @@ -81,15 +81,15 @@ Process = SIGMA_SM_GG_TTXGGG_CPP [gcc 11.3.1] [inlineHel=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/none+NAVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) Internal loops fptype_sv = SCALAR ('none': ~vector[1], no SIMD) -EvtsPerSec[Rmb+ME] (23) = ( 7.653903e+01 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 7.654105e+01 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 7.654105e+01 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 7.792374e+01 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 7.792594e+01 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 7.792594e+01 ) sec^-1 MeanMatrixElemValue = ( 1.187066e-05 +- 9.825549e-06 ) GeV^-6 -TOTAL : 6.889827 sec +TOTAL : 6.771406 sec INFO: No Floating Point Exceptions have been reported - 19,201,166,017 cycles # 2.786 GHz - 54,161,677,415 instructions # 2.82 insn per cycle - 6.893652512 seconds time elapsed + 19,102,420,778 cycles # 2.820 GHz + 54,161,990,407 instructions # 2.84 insn per cycle + 6.775484339 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4:32202) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttggg.mad/SubProcesses/P1_gg_ttxggg/build.none_m_inl0_hrd1/runTest_cpp.exe @@ -113,15 +113,15 @@ Process = SIGMA_SM_GG_TTXGGG_CPP [gcc 11.3.1] [inlineHel=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[2] ('sse4': SSE4.2, 128bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 1.552412e+02 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.552503e+02 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.552503e+02 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.563445e+02 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.563533e+02 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.563533e+02 ) sec^-1 MeanMatrixElemValue = ( 1.187066e-05 +- 9.825548e-06 ) GeV^-6 -TOTAL : 3.403221 sec +TOTAL : 3.378955 sec INFO: No Floating Point Exceptions have been reported - 9,295,420,050 cycles # 2.729 GHz - 26,089,296,035 instructions # 2.81 insn per cycle - 3.407123949 seconds time elapsed + 9,269,967,876 cycles # 2.741 GHz + 26,087,426,972 instructions # 2.81 insn per cycle + 3.383190937 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4:95935) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttggg.mad/SubProcesses/P1_gg_ttxggg/build.sse4_m_inl0_hrd1/runTest_cpp.exe @@ -145,15 +145,15 @@ Process = SIGMA_SM_GG_TTXGGG_CPP [gcc 11.3.1] [inlineHel=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('avx2': AVX2, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 3.556434e+02 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.556900e+02 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.556900e+02 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 3.582521e+02 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.582984e+02 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.582984e+02 ) sec^-1 MeanMatrixElemValue = ( 1.187066e-05 +- 9.825548e-06 ) GeV^-6 -TOTAL : 1.488620 sec +TOTAL : 1.476148 sec INFO: No Floating Point Exceptions have been reported - 4,059,104,235 cycles # 2.721 GHz - 9,213,839,753 instructions # 2.27 insn per cycle - 1.492560916 seconds time elapsed + 4,059,613,552 cycles # 2.744 GHz + 9,213,793,706 instructions # 2.27 insn per cycle + 1.480250671 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2:83864) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttggg.mad/SubProcesses/P1_gg_ttxggg/build.avx2_m_inl0_hrd1/runTest_cpp.exe @@ -177,15 +177,15 @@ Process = SIGMA_SM_GG_TTXGGG_CPP [gcc 11.3.1] [inlineHel=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('512y': AVX512, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 4.125241e+02 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 4.125840e+02 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 4.125840e+02 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 4.104692e+02 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 4.105310e+02 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 4.105310e+02 ) sec^-1 MeanMatrixElemValue = ( 1.187066e-05 +- 9.825548e-06 ) GeV^-6 -TOTAL : 1.282211 sec +TOTAL : 1.288792 sec INFO: No Floating Point Exceptions have been reported - 3,511,408,538 cycles # 2.732 GHz - 8,168,208,932 instructions # 2.33 insn per cycle - 1.286095846 seconds time elapsed + 3,513,320,349 cycles # 2.719 GHz + 8,169,210,464 instructions # 2.33 insn per cycle + 1.293071063 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2:79421) (512y: 230) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttggg.mad/SubProcesses/P1_gg_ttxggg/build.512y_m_inl0_hrd1/runTest_cpp.exe @@ -209,15 +209,15 @@ Process = SIGMA_SM_GG_TTXGGG_CPP [gcc 11.3.1] [inlineHel=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('512z': AVX512, 512bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 3.517573e+02 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.518129e+02 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.518129e+02 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 3.476024e+02 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.476542e+02 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.476542e+02 ) sec^-1 MeanMatrixElemValue = ( 1.187066e-05 +- 9.825548e-06 ) GeV^-6 -TOTAL : 1.503444 sec +TOTAL : 1.520796 sec INFO: No Floating Point Exceptions have been reported - 2,622,176,822 cycles # 1.740 GHz - 4,167,750,292 instructions # 1.59 insn per cycle - 1.507552292 seconds time elapsed + 2,617,904,633 cycles # 1.718 GHz + 4,167,935,855 instructions # 1.59 insn per cycle + 1.525160346 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 1879) (512y: 174) (512z:78884) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gg_ttggg.mad/SubProcesses/P1_gg_ttxggg/build.512z_m_inl0_hrd1/runTest_cpp.exe diff --git a/epochX/cudacpp/tput/logs_gqttq_mad/log_gqttq_mad_d_inl0_hrd0.txt b/epochX/cudacpp/tput/logs_gqttq_mad/log_gqttq_mad_d_inl0_hrd0.txt index 468f6865a8..520482438a 100644 --- a/epochX/cudacpp/tput/logs_gqttq_mad/log_gqttq_mad_d_inl0_hrd0.txt +++ b/epochX/cudacpp/tput/logs_gqttq_mad/log_gqttq_mad_d_inl0_hrd0.txt @@ -40,7 +40,7 @@ make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gq_ttq.mad/SubProcesses/P1_gux_ttxux' -DATE: 2024-10-06_09:05:51 +DATE: 2024-10-06_13:35:01 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: ========================================================================= @@ -49,15 +49,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_SM_GUX_TTXUX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 1.906944e+07 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 2.902591e+07 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.013821e+07 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 2.048397e+07 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 2.994676e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.094819e+07 ) sec^-1 MeanMatrixElemValue = ( 2.018083e+01 +- 1.429474e+01 ) GeV^-2 -TOTAL : 0.458221 sec +TOTAL : 0.453017 sec INFO: No Floating Point Exceptions have been reported - 1,930,997,109 cycles # 2.858 GHz - 2,724,198,211 instructions # 1.41 insn per cycle - 0.805328419 seconds time elapsed + 1,926,308,850 cycles # 2.874 GHz + 2,732,181,056 instructions # 1.42 insn per cycle + 0.727830103 seconds time elapsed runNcu /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gq_ttq.mad/SubProcesses/P1_gux_ttxux/build.cuda_d_inl0_hrd0/check_cuda.exe -p 64 256 1 ==PROF== Profiling "sigmaKin": launch__registers_per_thread 255 ==PROF== Profiling "sigmaKin": sm__sass_average_branch_targets_threads_uniform.pct 100% @@ -67,15 +67,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_SM_GUX_TTXUX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 3.002453e+07 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 5.463176e+07 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 5.675243e+07 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 3.164016e+07 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 5.544257e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 5.751306e+07 ) sec^-1 MeanMatrixElemValue = ( 2.602505e+02 +- 2.116328e+02 ) GeV^-2 -TOTAL : 0.544639 sec +TOTAL : 0.535677 sec INFO: No Floating Point Exceptions have been reported - 2,250,691,324 cycles # 2.871 GHz - 3,190,813,390 instructions # 1.42 insn per cycle - 0.843484638 seconds time elapsed + 2,227,471,943 cycles # 2.876 GHz + 3,203,792,390 instructions # 1.44 insn per cycle + 0.832408028 seconds time elapsed ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gq_ttq.mad/SubProcesses/P1_gux_ttxux/build.cuda_d_inl0_hrd0/runTest_cuda.exe INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: FE_DIVBYZERO, FE_INVALID, FE_OVERFLOW @@ -100,15 +100,15 @@ Process = SIGMA_SM_GUX_TTXUX_CPP [gcc 11.3.1] [inlineHel=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/none+NAVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = SCALAR ('none': ~vector[1], no SIMD) -EvtsPerSec[Rmb+ME] (23) = ( 1.052668e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.075406e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.075406e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.059559e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.082131e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.082131e+05 ) sec^-1 MeanMatrixElemValue = ( 2.018083e+01 +- 1.429474e+01 ) GeV^-2 -TOTAL : 1.578445 sec +TOTAL : 1.565777 sec INFO: No Floating Point Exceptions have been reported - 4,629,037,835 cycles # 2.928 GHz - 13,193,545,970 instructions # 2.85 insn per cycle - 1.584589009 seconds time elapsed + 4,619,472,901 cycles # 2.944 GHz + 13,191,195,947 instructions # 2.86 insn per cycle + 1.569869034 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 707) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gq_ttq.mad/SubProcesses/P1_gux_ttxux/build.none_d_inl0_hrd0/runTest_cpp.exe @@ -132,15 +132,15 @@ Process = SIGMA_SM_GUX_TTXUX_CPP [gcc 11.3.1] [inlineHel=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[2] ('sse4': SSE4.2, 128bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 1.869817e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.940106e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.940106e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.880201e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.950351e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.950351e+05 ) sec^-1 MeanMatrixElemValue = ( 2.018083e+01 +- 1.429474e+01 ) GeV^-2 -TOTAL : 0.895982 sec +TOTAL : 0.889927 sec INFO: No Floating Point Exceptions have been reported - 2,636,174,950 cycles # 2.931 GHz - 7,556,706,256 instructions # 2.87 insn per cycle - 0.901753059 seconds time elapsed + 2,632,104,591 cycles # 2.946 GHz + 7,554,750,135 instructions # 2.87 insn per cycle + 0.894038050 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 3099) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gq_ttq.mad/SubProcesses/P1_gux_ttxux/build.sse4_d_inl0_hrd0/runTest_cpp.exe @@ -164,15 +164,15 @@ Process = SIGMA_SM_GUX_TTXUX_CPP [gcc 11.3.1] [inlineHel=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('avx2': AVX2, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 3.170738e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.377041e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.377041e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 3.154246e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.358397e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.358397e+05 ) sec^-1 MeanMatrixElemValue = ( 2.018083e+01 +- 1.429474e+01 ) GeV^-2 -TOTAL : 0.538337 sec +TOTAL : 0.538774 sec INFO: No Floating Point Exceptions have been reported - 1,492,365,440 cycles # 2.760 GHz - 3,161,633,609 instructions # 2.12 insn per cycle - 0.543901971 seconds time elapsed + 1,490,596,175 cycles # 2.748 GHz + 3,159,351,923 instructions # 2.12 insn per cycle + 0.543025923 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 2991) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gq_ttq.mad/SubProcesses/P1_gux_ttxux/build.avx2_d_inl0_hrd0/runTest_cpp.exe @@ -196,15 +196,15 @@ Process = SIGMA_SM_GUX_TTXUX_CPP [gcc 11.3.1] [inlineHel=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('512y': AVX512, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 3.502118e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.753079e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.753079e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 3.538102e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.793543e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.793543e+05 ) sec^-1 MeanMatrixElemValue = ( 2.018083e+01 +- 1.429474e+01 ) GeV^-2 -TOTAL : 0.488254 sec +TOTAL : 0.482381 sec INFO: No Floating Point Exceptions have been reported - 1,345,193,436 cycles # 2.734 GHz - 3,015,805,712 instructions # 2.24 insn per cycle - 0.494320620 seconds time elapsed + 1,342,559,160 cycles # 2.763 GHz + 3,013,840,490 instructions # 2.24 insn per cycle + 0.486569020 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 2749) (512y: 104) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gq_ttq.mad/SubProcesses/P1_gux_ttxux/build.512y_d_inl0_hrd0/runTest_cpp.exe @@ -228,15 +228,15 @@ Process = SIGMA_SM_GUX_TTXUX_CPP [gcc 11.3.1] [inlineHel=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('512z': AVX512, 512bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 2.340176e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 2.450488e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 2.450488e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 2.350585e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 2.461907e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.461907e+05 ) sec^-1 MeanMatrixElemValue = ( 2.018083e+01 +- 1.429474e+01 ) GeV^-2 -TOTAL : 0.722488 sec +TOTAL : 0.718441 sec INFO: No Floating Point Exceptions have been reported - 1,326,137,037 cycles # 1.826 GHz - 1,964,340,659 instructions # 1.48 insn per cycle - 0.728328312 seconds time elapsed + 1,325,555,101 cycles # 1.836 GHz + 1,962,680,791 instructions # 1.48 insn per cycle + 0.722760267 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 1379) (512y: 106) (512z: 2218) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gq_ttq.mad/SubProcesses/P1_gux_ttxux/build.512z_d_inl0_hrd0/runTest_cpp.exe diff --git a/epochX/cudacpp/tput/logs_gqttq_mad/log_gqttq_mad_d_inl0_hrd0_bridge.txt b/epochX/cudacpp/tput/logs_gqttq_mad/log_gqttq_mad_d_inl0_hrd0_bridge.txt index a32e85fd77..6f472ea0e9 100644 --- a/epochX/cudacpp/tput/logs_gqttq_mad/log_gqttq_mad_d_inl0_hrd0_bridge.txt +++ b/epochX/cudacpp/tput/logs_gqttq_mad/log_gqttq_mad_d_inl0_hrd0_bridge.txt @@ -40,7 +40,7 @@ make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gq_ttq.mad/SubProcesses/P1_gux_ttxux' -DATE: 2024-10-06_09:39:10 +DATE: 2024-10-06_13:57:43 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: ========================================================================= @@ -53,16 +53,16 @@ WARNING! Set grid in Bridge (nevt=16384, gpublocks=64, gputhreads=256, gpublocks Process = SIGMA_SM_GUX_TTXUX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 3.313371e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.590831e+07 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.590831e+07 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 3.330506e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.593544e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.593544e+07 ) sec^-1 MeanMatrixElemValue = ( 2.018083e+01 +- 1.429474e+01 ) GeV^-2 -TOTAL : 0.487212 sec +TOTAL : 0.486415 sec INFO: No Floating Point Exceptions have been reported INFO: No Floating Point Exceptions have been reported - 2,046,207,140 cycles # 2.880 GHz - 3,015,907,255 instructions # 1.47 insn per cycle - 0.769534809 seconds time elapsed + 2,049,654,055 cycles # 2.881 GHz + 3,021,030,657 instructions # 1.47 insn per cycle + 0.770600249 seconds time elapsed runNcu /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gq_ttq.mad/SubProcesses/P1_gux_ttxux/build.cuda_d_inl0_hrd0/check_cuda.exe -p 64 256 1 --bridge WARNING! Bridge selected: cannot use RamboDevice, will use RamboHost WARNING! RamboHost selected: cannot use CurandDevice, will use CurandHost @@ -80,16 +80,16 @@ WARNING! Set grid in Bridge (nevt=524288, gpublocks=2048, gputhreads=256, gpublo Process = SIGMA_SM_GUX_TTXUX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 3.228660e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 2.270938e+07 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 2.270938e+07 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 3.147684e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 2.226483e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.226483e+07 ) sec^-1 MeanMatrixElemValue = ( 2.602505e+02 +- 2.116328e+02 ) GeV^-2 -TOTAL : 0.758730 sec +TOTAL : 0.766931 sec INFO: No Floating Point Exceptions have been reported INFO: No Floating Point Exceptions have been reported - 2,917,079,859 cycles # 2.883 GHz - 4,489,082,127 instructions # 1.54 insn per cycle - 1.069078440 seconds time elapsed + 2,950,026,047 cycles # 2.878 GHz + 4,494,672,422 instructions # 1.52 insn per cycle + 1.082245629 seconds time elapsed ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gq_ttq.mad/SubProcesses/P1_gux_ttxux/build.cuda_d_inl0_hrd0/runTest_cuda.exe INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: FE_DIVBYZERO, FE_INVALID, FE_OVERFLOW @@ -115,16 +115,16 @@ Process = SIGMA_SM_GUX_TTXUX_CPP [gcc 11.3.1] [inlineHel=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+BRDHST/none+NAVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = SCALAR ('none': ~vector[1], no SIMD) -EvtsPerSec[Rmb+ME] (23) = ( 1.058535e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.081557e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.081557e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.060713e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.083398e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.083398e+05 ) sec^-1 MeanMatrixElemValue = ( 2.018083e+01 +- 1.429474e+01 ) GeV^-2 -TOTAL : 1.574537 sec +TOTAL : 1.569711 sec INFO: No Floating Point Exceptions have been reported INFO: No Floating Point Exceptions have been reported - 4,656,483,821 cycles # 2.950 GHz - 13,198,201,576 instructions # 2.83 insn per cycle - 1.579077435 seconds time elapsed + 4,648,055,821 cycles # 2.954 GHz + 13,196,511,317 instructions # 2.84 insn per cycle + 1.574166724 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 707) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gq_ttq.mad/SubProcesses/P1_gux_ttxux/build.none_d_inl0_hrd0/runTest_cpp.exe @@ -149,16 +149,16 @@ Process = SIGMA_SM_GUX_TTXUX_CPP [gcc 11.3.1] [inlineHel=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+BRDHST/sse4+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[2] ('sse4': SSE4.2, 128bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 1.861172e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.931943e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.931943e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.877214e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.948259e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.948259e+05 ) sec^-1 MeanMatrixElemValue = ( 2.018083e+01 +- 1.429474e+01 ) GeV^-2 -TOTAL : 0.907508 sec +TOTAL : 0.899366 sec INFO: No Floating Point Exceptions have been reported INFO: No Floating Point Exceptions have been reported - 2,678,662,656 cycles # 2.939 GHz - 7,605,263,564 instructions # 2.84 insn per cycle - 0.912202227 seconds time elapsed + 2,671,243,847 cycles # 2.957 GHz + 7,604,713,965 instructions # 2.85 insn per cycle + 0.903982545 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 3099) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gq_ttq.mad/SubProcesses/P1_gux_ttxux/build.sse4_d_inl0_hrd0/runTest_cpp.exe @@ -183,16 +183,16 @@ Process = SIGMA_SM_GUX_TTXUX_CPP [gcc 11.3.1] [inlineHel=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+BRDHST/avx2+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('avx2': AVX2, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 3.153263e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.357026e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.357026e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 3.141068e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.348173e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.348173e+05 ) sec^-1 MeanMatrixElemValue = ( 2.018083e+01 +- 1.429474e+01 ) GeV^-2 -TOTAL : 0.547067 sec +TOTAL : 0.548234 sec INFO: No Floating Point Exceptions have been reported INFO: No Floating Point Exceptions have been reported - 1,524,781,245 cycles # 2.767 GHz - 3,210,388,287 instructions # 2.11 insn per cycle - 0.551691801 seconds time elapsed + 1,529,313,779 cycles # 2.769 GHz + 3,208,620,185 instructions # 2.10 insn per cycle + 0.552751319 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 2991) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gq_ttq.mad/SubProcesses/P1_gux_ttxux/build.avx2_d_inl0_hrd0/runTest_cpp.exe @@ -217,16 +217,16 @@ Process = SIGMA_SM_GUX_TTXUX_CPP [gcc 11.3.1] [inlineHel=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+BRDHST/512y+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('512y': AVX512, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 3.508777e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.767060e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.767060e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 3.517931e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.771338e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.771338e+05 ) sec^-1 MeanMatrixElemValue = ( 2.018083e+01 +- 1.429474e+01 ) GeV^-2 -TOTAL : 0.494747 sec +TOTAL : 0.491910 sec INFO: No Floating Point Exceptions have been reported INFO: No Floating Point Exceptions have been reported - 1,383,177,469 cycles # 2.773 GHz - 3,064,481,068 instructions # 2.22 insn per cycle - 0.499446571 seconds time elapsed + 1,375,637,436 cycles # 2.774 GHz + 3,062,876,691 instructions # 2.23 insn per cycle + 0.496350430 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 2749) (512y: 104) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gq_ttq.mad/SubProcesses/P1_gux_ttxux/build.512y_d_inl0_hrd0/runTest_cpp.exe @@ -251,16 +251,16 @@ Process = SIGMA_SM_GUX_TTXUX_CPP [gcc 11.3.1] [inlineHel=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+BRDHST/512z+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('512z': AVX512, 512bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 2.351157e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 2.462501e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 2.462501e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 2.321636e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 2.432365e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.432365e+05 ) sec^-1 MeanMatrixElemValue = ( 2.018083e+01 +- 1.429474e+01 ) GeV^-2 -TOTAL : 0.725065 sec +TOTAL : 0.735791 sec INFO: No Floating Point Exceptions have been reported INFO: No Floating Point Exceptions have been reported - 1,357,891,290 cycles # 1.863 GHz - 2,000,455,329 instructions # 1.47 insn per cycle - 0.729577819 seconds time elapsed + 1,367,392,504 cycles # 1.848 GHz + 2,002,188,212 instructions # 1.46 insn per cycle + 0.740426058 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 1379) (512y: 106) (512z: 2218) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gq_ttq.mad/SubProcesses/P1_gux_ttxux/build.512z_d_inl0_hrd0/runTest_cpp.exe diff --git a/epochX/cudacpp/tput/logs_gqttq_mad/log_gqttq_mad_d_inl0_hrd1.txt b/epochX/cudacpp/tput/logs_gqttq_mad/log_gqttq_mad_d_inl0_hrd1.txt index 67eac99bab..fba628f407 100644 --- a/epochX/cudacpp/tput/logs_gqttq_mad/log_gqttq_mad_d_inl0_hrd1.txt +++ b/epochX/cudacpp/tput/logs_gqttq_mad/log_gqttq_mad_d_inl0_hrd1.txt @@ -40,7 +40,7 @@ make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gq_ttq.mad/SubProcesses/P1_gux_ttxux' -DATE: 2024-10-06_09:06:04 +DATE: 2024-10-06_13:35:15 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: ========================================================================= @@ -49,15 +49,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_SM_GUX_TTXUX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=1] Workflow summary = CUD:DBL+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 1.866343e+07 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 2.840904e+07 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 2.947003e+07 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 2.019170e+07 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 2.942351e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.037515e+07 ) sec^-1 MeanMatrixElemValue = ( 2.018083e+01 +- 1.429474e+01 ) GeV^-2 -TOTAL : 0.463809 sec +TOTAL : 0.454983 sec INFO: No Floating Point Exceptions have been reported - 1,942,418,108 cycles # 2.861 GHz - 2,721,411,859 instructions # 1.40 insn per cycle - 0.812650633 seconds time elapsed + 1,928,787,785 cycles # 2.870 GHz + 2,713,497,110 instructions # 1.41 insn per cycle + 0.729832990 seconds time elapsed runNcu /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gq_ttq.mad/SubProcesses/P1_gux_ttxux/build.cuda_d_inl0_hrd1/check_cuda.exe -p 64 256 1 ==PROF== Profiling "sigmaKin": launch__registers_per_thread 255 ==PROF== Profiling "sigmaKin": sm__sass_average_branch_targets_threads_uniform.pct 100% @@ -67,15 +67,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_SM_GUX_TTXUX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=1] Workflow summary = CUD:DBL+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 2.997280e+07 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 5.399599e+07 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 5.603946e+07 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 3.150935e+07 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 5.503666e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 5.706891e+07 ) sec^-1 MeanMatrixElemValue = ( 2.602505e+02 +- 2.116328e+02 ) GeV^-2 -TOTAL : 0.538885 sec +TOTAL : 0.535771 sec INFO: No Floating Point Exceptions have been reported - 2,239,160,610 cycles # 2.873 GHz - 3,203,384,758 instructions # 1.43 insn per cycle - 0.836856412 seconds time elapsed + 2,230,619,844 cycles # 2.885 GHz + 3,221,530,769 instructions # 1.44 insn per cycle + 0.830892871 seconds time elapsed ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gq_ttq.mad/SubProcesses/P1_gux_ttxux/build.cuda_d_inl0_hrd1/runTest_cuda.exe INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: FE_DIVBYZERO, FE_INVALID, FE_OVERFLOW @@ -100,15 +100,15 @@ Process = SIGMA_SM_GUX_TTXUX_CPP [gcc 11.3.1] [inlineHel=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/none+NAVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = SCALAR ('none': ~vector[1], no SIMD) -EvtsPerSec[Rmb+ME] (23) = ( 1.060643e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.083213e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.083213e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.063194e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.085695e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.085695e+05 ) sec^-1 MeanMatrixElemValue = ( 2.018083e+01 +- 1.429474e+01 ) GeV^-2 -TOTAL : 1.565121 sec +TOTAL : 1.560224 sec INFO: No Floating Point Exceptions have been reported - 4,623,795,988 cycles # 2.948 GHz - 13,181,888,102 instructions # 2.85 insn per cycle - 1.571833324 seconds time elapsed + 4,618,325,769 cycles # 2.954 GHz + 13,179,878,696 instructions # 2.85 insn per cycle + 1.564335895 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 692) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gq_ttq.mad/SubProcesses/P1_gux_ttxux/build.none_d_inl0_hrd1/runTest_cpp.exe @@ -132,15 +132,15 @@ Process = SIGMA_SM_GUX_TTXUX_CPP [gcc 11.3.1] [inlineHel=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[2] ('sse4': SSE4.2, 128bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 1.878003e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.949625e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.949625e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.872263e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.942328e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.942328e+05 ) sec^-1 MeanMatrixElemValue = ( 2.018083e+01 +- 1.429474e+01 ) GeV^-2 -TOTAL : 0.892385 sec +TOTAL : 0.893695 sec INFO: No Floating Point Exceptions have been reported - 2,641,116,720 cycles # 2.947 GHz - 7,555,506,374 instructions # 2.86 insn per cycle - 0.899472366 seconds time elapsed + 2,647,903,994 cycles # 2.952 GHz + 7,552,789,844 instructions # 2.85 insn per cycle + 0.897780782 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 3093) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gq_ttq.mad/SubProcesses/P1_gux_ttxux/build.sse4_d_inl0_hrd1/runTest_cpp.exe @@ -164,15 +164,15 @@ Process = SIGMA_SM_GUX_TTXUX_CPP [gcc 11.3.1] [inlineHel=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('avx2': AVX2, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 3.178148e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.383095e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.383095e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 3.170482e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.375745e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.375745e+05 ) sec^-1 MeanMatrixElemValue = ( 2.018083e+01 +- 1.429474e+01 ) GeV^-2 -TOTAL : 0.535311 sec +TOTAL : 0.535518 sec INFO: No Floating Point Exceptions have been reported - 1,491,222,481 cycles # 2.767 GHz - 3,161,019,864 instructions # 2.12 insn per cycle - 0.541387025 seconds time elapsed + 1,489,046,799 cycles # 2.763 GHz + 3,158,894,672 instructions # 2.12 insn per cycle + 0.539597683 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 2976) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gq_ttq.mad/SubProcesses/P1_gux_ttxux/build.avx2_d_inl0_hrd1/runTest_cpp.exe @@ -196,15 +196,15 @@ Process = SIGMA_SM_GUX_TTXUX_CPP [gcc 11.3.1] [inlineHel=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('512y': AVX512, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 3.523592e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.778898e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.778898e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 3.523271e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.778257e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.778257e+05 ) sec^-1 MeanMatrixElemValue = ( 2.018083e+01 +- 1.429474e+01 ) GeV^-2 -TOTAL : 0.485060 sec +TOTAL : 0.484275 sec INFO: No Floating Point Exceptions have been reported - 1,349,314,232 cycles # 2.763 GHz - 3,012,812,614 instructions # 2.23 insn per cycle - 0.489068736 seconds time elapsed + 1,343,553,927 cycles # 2.754 GHz + 3,011,112,120 instructions # 2.24 insn per cycle + 0.488513655 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 2726) (512y: 104) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gq_ttq.mad/SubProcesses/P1_gux_ttxux/build.512y_d_inl0_hrd1/runTest_cpp.exe @@ -228,15 +228,15 @@ Process = SIGMA_SM_GUX_TTXUX_CPP [gcc 11.3.1] [inlineHel=0] Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('512z': AVX512, 512bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 2.347943e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 2.459729e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 2.459729e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 2.327043e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 2.437484e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.437484e+05 ) sec^-1 MeanMatrixElemValue = ( 2.018083e+01 +- 1.429474e+01 ) GeV^-2 -TOTAL : 0.720112 sec +TOTAL : 0.725580 sec INFO: No Floating Point Exceptions have been reported - 1,326,103,986 cycles # 1.833 GHz - 1,962,664,460 instructions # 1.48 insn per cycle - 0.726078775 seconds time elapsed + 1,322,335,339 cycles # 1.814 GHz + 1,961,024,993 instructions # 1.48 insn per cycle + 0.729737228 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 1356) (512y: 106) (512z: 2218) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gq_ttq.mad/SubProcesses/P1_gux_ttxux/build.512z_d_inl0_hrd1/runTest_cpp.exe diff --git a/epochX/cudacpp/tput/logs_gqttq_mad/log_gqttq_mad_f_inl0_hrd0.txt b/epochX/cudacpp/tput/logs_gqttq_mad/log_gqttq_mad_f_inl0_hrd0.txt index fa95ebd131..83d4db8597 100644 --- a/epochX/cudacpp/tput/logs_gqttq_mad/log_gqttq_mad_f_inl0_hrd0.txt +++ b/epochX/cudacpp/tput/logs_gqttq_mad/log_gqttq_mad_f_inl0_hrd0.txt @@ -40,7 +40,7 @@ make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gq_ttq.mad/SubProcesses/P1_gux_ttxux' -DATE: 2024-10-06_09:06:45 +DATE: 2024-10-06_13:35:55 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: ========================================================================= @@ -49,15 +49,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_SM_GUX_TTXUX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 2.818001e+07 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.982501e+07 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 4.122889e+07 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 3.015184e+07 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 4.128373e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 4.283451e+07 ) sec^-1 MeanMatrixElemValue = ( 2.018174e+01 +- 1.429492e+01 ) GeV^-2 -TOTAL : 0.452282 sec +TOTAL : 0.452034 sec INFO: No Floating Point Exceptions have been reported - 1,920,727,034 cycles # 2.860 GHz - 2,694,517,558 instructions # 1.40 insn per cycle - 0.728408510 seconds time elapsed + 1,917,841,079 cycles # 2.879 GHz + 2,715,243,569 instructions # 1.42 insn per cycle + 0.725140925 seconds time elapsed runNcu /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gq_ttq.mad/SubProcesses/P1_gux_ttxux/build.cuda_f_inl0_hrd0/check_cuda.exe -p 64 256 1 ==PROF== Profiling "sigmaKin": launch__registers_per_thread 169 ==PROF== Profiling "sigmaKin": sm__sass_average_branch_targets_threads_uniform.pct 100% @@ -67,15 +67,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_SM_GUX_TTXUX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 5.287877e+07 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 9.320334e+07 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 9.683236e+07 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 5.798431e+07 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 9.693554e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.001954e+08 ) sec^-1 MeanMatrixElemValue = ( 2.571360e+02 +- 2.114020e+02 ) GeV^-2 -TOTAL : 0.495314 sec +TOTAL : 0.488668 sec INFO: No Floating Point Exceptions have been reported - 2,079,539,950 cycles # 2.850 GHz - 2,952,237,418 instructions # 1.42 insn per cycle - 0.786339466 seconds time elapsed + 2,053,886,214 cycles # 2.871 GHz + 2,904,537,428 instructions # 1.41 insn per cycle + 0.774511097 seconds time elapsed ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gq_ttq.mad/SubProcesses/P1_gux_ttxux/build.cuda_f_inl0_hrd0/runTest_cuda.exe INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: FE_DIVBYZERO, FE_INVALID, FE_OVERFLOW @@ -100,15 +100,15 @@ Process = SIGMA_SM_GUX_TTXUX_CPP [gcc 11.3.1] [inlineHel=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/none+NAVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = SCALAR ('none': ~vector[1], no SIMD) -EvtsPerSec[Rmb+ME] (23) = ( 1.109567e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.134660e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.134660e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.112314e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.137517e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.137517e+05 ) sec^-1 MeanMatrixElemValue = ( 2.018564e+01 +- 1.429903e+01 ) GeV^-2 -TOTAL : 1.494759 sec +TOTAL : 1.490416 sec INFO: No Floating Point Exceptions have been reported - 4,403,081,916 cycles # 2.940 GHz - 12,951,948,710 instructions # 2.94 insn per cycle - 1.498420981 seconds time elapsed + 4,402,849,992 cycles # 2.947 GHz + 12,951,963,668 instructions # 2.94 insn per cycle + 1.494500070 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 645) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gq_ttq.mad/SubProcesses/P1_gux_ttxux/build.none_f_inl0_hrd0/runTest_cpp.exe @@ -132,15 +132,15 @@ Process = SIGMA_SM_GUX_TTXUX_CPP [gcc 11.3.1] [inlineHel=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('sse4': SSE4.2, 128bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 2.886806e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.066754e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.066754e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 2.891990e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.071050e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.071050e+05 ) sec^-1 MeanMatrixElemValue = ( 2.018564e+01 +- 1.429903e+01 ) GeV^-2 -TOTAL : 0.584675 sec +TOTAL : 0.583740 sec INFO: No Floating Point Exceptions have been reported - 1,726,276,919 cycles # 2.937 GHz - 4,542,407,737 instructions # 2.63 insn per cycle - 0.588476135 seconds time elapsed + 1,725,311,864 cycles # 2.938 GHz + 4,541,931,327 instructions # 2.63 insn per cycle + 0.587861356 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 3627) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gq_ttq.mad/SubProcesses/P1_gux_ttxux/build.sse4_f_inl0_hrd0/runTest_cpp.exe @@ -164,15 +164,15 @@ Process = SIGMA_SM_GUX_TTXUX_CPP [gcc 11.3.1] [inlineHel=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('avx2': AVX2, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 5.651382e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 6.346145e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 6.346145e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 5.684050e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 6.384017e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 6.384017e+05 ) sec^-1 MeanMatrixElemValue = ( 2.018828e+01 +- 1.429922e+01 ) GeV^-2 -TOTAL : 0.307816 sec +TOTAL : 0.305885 sec INFO: No Floating Point Exceptions have been reported - 856,647,676 cycles # 2.754 GHz - 1,917,830,464 instructions # 2.24 insn per cycle - 0.311794908 seconds time elapsed + 855,280,295 cycles # 2.765 GHz + 1,917,692,111 instructions # 2.24 insn per cycle + 0.309823391 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 3580) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gq_ttq.mad/SubProcesses/P1_gux_ttxux/build.avx2_f_inl0_hrd0/runTest_cpp.exe @@ -196,15 +196,15 @@ Process = SIGMA_SM_GUX_TTXUX_CPP [gcc 11.3.1] [inlineHel=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('512y': AVX512, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 6.083995e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 6.890169e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 6.890169e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 6.042139e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 6.847415e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 6.847415e+05 ) sec^-1 MeanMatrixElemValue = ( 2.018828e+01 +- 1.429922e+01 ) GeV^-2 -TOTAL : 0.287118 sec +TOTAL : 0.288949 sec INFO: No Floating Point Exceptions have been reported - 801,284,784 cycles # 2.760 GHz - 1,834,043,941 instructions # 2.29 insn per cycle - 0.290894624 seconds time elapsed + 804,082,758 cycles # 2.748 GHz + 1,833,934,465 instructions # 2.28 insn per cycle + 0.293173162 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 3400) (512y: 22) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gq_ttq.mad/SubProcesses/P1_gux_ttxux/build.512y_f_inl0_hrd0/runTest_cpp.exe @@ -228,15 +228,15 @@ Process = SIGMA_SM_GUX_TTXUX_CPP [gcc 11.3.1] [inlineHel=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[16] ('512z': AVX512, 512bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 4.500723e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 4.948038e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 4.948038e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 4.600281e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 5.058606e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 5.058606e+05 ) sec^-1 MeanMatrixElemValue = ( 2.018829e+01 +- 1.429922e+01 ) GeV^-2 -TOTAL : 0.384030 sec +TOTAL : 0.375604 sec INFO: No Floating Point Exceptions have been reported - 726,928,592 cycles # 1.877 GHz - 1,308,660,654 instructions # 1.80 insn per cycle - 0.387900268 seconds time elapsed + 727,230,617 cycles # 1.918 GHz + 1,308,665,243 instructions # 1.80 insn per cycle + 0.379697417 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 1964) (512y: 24) (512z: 2435) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gq_ttq.mad/SubProcesses/P1_gux_ttxux/build.512z_f_inl0_hrd0/runTest_cpp.exe diff --git a/epochX/cudacpp/tput/logs_gqttq_mad/log_gqttq_mad_f_inl0_hrd0_bridge.txt b/epochX/cudacpp/tput/logs_gqttq_mad/log_gqttq_mad_f_inl0_hrd0_bridge.txt index 5a6a874489..ba97831ae2 100644 --- a/epochX/cudacpp/tput/logs_gqttq_mad/log_gqttq_mad_f_inl0_hrd0_bridge.txt +++ b/epochX/cudacpp/tput/logs_gqttq_mad/log_gqttq_mad_f_inl0_hrd0_bridge.txt @@ -40,7 +40,7 @@ make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gq_ttq.mad/SubProcesses/P1_gux_ttxux' -DATE: 2024-10-06_09:39:24 +DATE: 2024-10-06_13:57:57 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: ========================================================================= @@ -53,16 +53,16 @@ WARNING! Set grid in Bridge (nevt=16384, gpublocks=64, gputhreads=256, gpublocks Process = SIGMA_SM_GUX_TTXUX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 4.958276e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 2.362856e+07 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 2.362856e+07 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 4.968170e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 2.353795e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.353795e+07 ) sec^-1 MeanMatrixElemValue = ( 2.017654e+01 +- 1.429183e+01 ) GeV^-2 -TOTAL : 0.467586 sec +TOTAL : 0.464855 sec INFO: No Floating Point Exceptions have been reported INFO: No Floating Point Exceptions have been reported - 1,958,269,299 cycles # 2.868 GHz - 2,873,921,299 instructions # 1.47 insn per cycle - 0.741370031 seconds time elapsed + 1,947,114,530 cycles # 2.869 GHz + 2,812,979,826 instructions # 1.44 insn per cycle + 0.735891933 seconds time elapsed runNcu /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gq_ttq.mad/SubProcesses/P1_gux_ttxux/build.cuda_f_inl0_hrd0/check_cuda.exe -p 64 256 1 --bridge WARNING! Bridge selected: cannot use RamboDevice, will use RamboHost WARNING! RamboHost selected: cannot use CurandDevice, will use CurandHost @@ -80,16 +80,16 @@ WARNING! Set grid in Bridge (nevt=524288, gpublocks=2048, gputhreads=256, gpublo Process = SIGMA_SM_GUX_TTXUX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:CURHST+RMBHST+BRDDEV/none+NAVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 4.867040e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.953002e+07 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.953002e+07 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 4.927625e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.941183e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.941183e+07 ) sec^-1 MeanMatrixElemValue = ( 2.609941e+02 +- 2.115589e+02 ) GeV^-2 -TOTAL : 0.638465 sec +TOTAL : 0.638481 sec INFO: No Floating Point Exceptions have been reported INFO: No Floating Point Exceptions have been reported - 2,513,600,406 cycles # 2.877 GHz - 3,810,036,638 instructions # 1.52 insn per cycle - 0.930171723 seconds time elapsed + 2,526,010,695 cycles # 2.889 GHz + 3,812,828,641 instructions # 1.51 insn per cycle + 0.931309931 seconds time elapsed ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gq_ttq.mad/SubProcesses/P1_gux_ttxux/build.cuda_f_inl0_hrd0/runTest_cuda.exe INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: FE_DIVBYZERO, FE_INVALID, FE_OVERFLOW @@ -115,16 +115,16 @@ Process = SIGMA_SM_GUX_TTXUX_CPP [gcc 11.3.1] [inlineHel=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+BRDHST/none+NAVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = SCALAR ('none': ~vector[1], no SIMD) -EvtsPerSec[Rmb+ME] (23) = ( 1.115307e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.140507e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.140507e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.109893e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.135240e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.135240e+05 ) sec^-1 MeanMatrixElemValue = ( 2.018564e+01 +- 1.429903e+01 ) GeV^-2 -TOTAL : 1.490082 sec +TOTAL : 1.497429 sec INFO: No Floating Point Exceptions have been reported INFO: No Floating Point Exceptions have been reported - 4,418,597,373 cycles # 2.958 GHz - 12,956,387,401 instructions # 2.93 insn per cycle - 1.494530314 seconds time elapsed + 4,421,092,621 cycles # 2.946 GHz + 12,956,529,994 instructions # 2.93 insn per cycle + 1.501652979 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 645) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gq_ttq.mad/SubProcesses/P1_gux_ttxux/build.none_f_inl0_hrd0/runTest_cpp.exe @@ -149,16 +149,16 @@ Process = SIGMA_SM_GUX_TTXUX_CPP [gcc 11.3.1] [inlineHel=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+BRDHST/sse4+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('sse4': SSE4.2, 128bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 2.871197e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.051268e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.051268e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 2.887812e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.066132e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.066132e+05 ) sec^-1 MeanMatrixElemValue = ( 2.018564e+01 +- 1.429903e+01 ) GeV^-2 -TOTAL : 0.592243 sec +TOTAL : 0.589116 sec INFO: No Floating Point Exceptions have been reported INFO: No Floating Point Exceptions have been reported - 1,749,393,716 cycles # 2.936 GHz - 4,590,457,409 instructions # 2.62 insn per cycle - 0.596762261 seconds time elapsed + 1,747,478,329 cycles # 2.948 GHz + 4,590,082,236 instructions # 2.63 insn per cycle + 0.593426277 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 3627) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gq_ttq.mad/SubProcesses/P1_gux_ttxux/build.sse4_f_inl0_hrd0/runTest_cpp.exe @@ -183,16 +183,16 @@ Process = SIGMA_SM_GUX_TTXUX_CPP [gcc 11.3.1] [inlineHel=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+BRDHST/avx2+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('avx2': AVX2, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 5.650062e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 6.340176e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 6.340176e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 5.642602e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 6.345494e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 6.345494e+05 ) sec^-1 MeanMatrixElemValue = ( 2.018828e+01 +- 1.429922e+01 ) GeV^-2 -TOTAL : 0.311783 sec +TOTAL : 0.312730 sec INFO: No Floating Point Exceptions have been reported INFO: No Floating Point Exceptions have been reported - 875,769,738 cycles # 2.776 GHz - 1,954,803,706 instructions # 2.23 insn per cycle - 0.316080972 seconds time elapsed + 879,374,823 cycles # 2.779 GHz + 1,954,675,082 instructions # 2.22 insn per cycle + 0.317116302 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 3580) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gq_ttq.mad/SubProcesses/P1_gux_ttxux/build.avx2_f_inl0_hrd0/runTest_cpp.exe @@ -217,16 +217,16 @@ Process = SIGMA_SM_GUX_TTXUX_CPP [gcc 11.3.1] [inlineHel=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+BRDHST/512y+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('512y': AVX512, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 6.042794e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 6.845843e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 6.845843e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 6.067249e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 6.858266e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 6.858266e+05 ) sec^-1 MeanMatrixElemValue = ( 2.018828e+01 +- 1.429922e+01 ) GeV^-2 -TOTAL : 0.293361 sec +TOTAL : 0.291525 sec INFO: No Floating Point Exceptions have been reported INFO: No Floating Point Exceptions have been reported - 825,335,769 cycles # 2.779 GHz - 1,870,845,111 instructions # 2.27 insn per cycle - 0.297556229 seconds time elapsed + 822,214,522 cycles # 2.786 GHz + 1,871,156,858 instructions # 2.28 insn per cycle + 0.295718449 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 3400) (512y: 22) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gq_ttq.mad/SubProcesses/P1_gux_ttxux/build.512y_f_inl0_hrd0/runTest_cpp.exe @@ -251,16 +251,16 @@ Process = SIGMA_SM_GUX_TTXUX_CPP [gcc 11.3.1] [inlineHel=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+BRDHST/512z+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[16] ('512z': AVX512, 512bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 4.484934e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 4.935540e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 4.935540e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 4.490240e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 4.934780e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 4.934780e+05 ) sec^-1 MeanMatrixElemValue = ( 2.018829e+01 +- 1.429922e+01 ) GeV^-2 -TOTAL : 0.390040 sec +TOTAL : 0.389351 sec INFO: No Floating Point Exceptions have been reported INFO: No Floating Point Exceptions have been reported - 749,752,693 cycles # 1.904 GHz - 1,350,296,093 instructions # 1.80 insn per cycle - 0.394449871 seconds time elapsed + 748,615,547 cycles # 1.904 GHz + 1,350,159,991 instructions # 1.80 insn per cycle + 0.393716896 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 1964) (512y: 24) (512z: 2435) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gq_ttq.mad/SubProcesses/P1_gux_ttxux/build.512z_f_inl0_hrd0/runTest_cpp.exe diff --git a/epochX/cudacpp/tput/logs_gqttq_mad/log_gqttq_mad_f_inl0_hrd1.txt b/epochX/cudacpp/tput/logs_gqttq_mad/log_gqttq_mad_f_inl0_hrd1.txt index cea07bf7e8..9758043c8a 100644 --- a/epochX/cudacpp/tput/logs_gqttq_mad/log_gqttq_mad_f_inl0_hrd1.txt +++ b/epochX/cudacpp/tput/logs_gqttq_mad/log_gqttq_mad_f_inl0_hrd1.txt @@ -40,7 +40,7 @@ make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gq_ttq.mad/SubProcesses/P1_gux_ttxux' -DATE: 2024-10-06_09:06:58 +DATE: 2024-10-06_13:36:07 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: ========================================================================= @@ -49,15 +49,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_SM_GUX_TTXUX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=1] Workflow summary = CUD:FLT+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 2.801672e+07 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.945717e+07 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 4.092440e+07 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 3.027644e+07 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 4.147650e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 4.289277e+07 ) sec^-1 MeanMatrixElemValue = ( 2.018174e+01 +- 1.429492e+01 ) GeV^-2 -TOTAL : 0.453252 sec +TOTAL : 0.452960 sec INFO: No Floating Point Exceptions have been reported - 1,914,636,683 cycles # 2.859 GHz - 2,699,162,883 instructions # 1.41 insn per cycle - 0.727606605 seconds time elapsed + 1,911,004,784 cycles # 2.862 GHz + 2,684,486,422 instructions # 1.40 insn per cycle + 0.726065606 seconds time elapsed runNcu /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gq_ttq.mad/SubProcesses/P1_gux_ttxux/build.cuda_f_inl0_hrd1/check_cuda.exe -p 64 256 1 ==PROF== Profiling "sigmaKin": launch__registers_per_thread 169 ==PROF== Profiling "sigmaKin": sm__sass_average_branch_targets_threads_uniform.pct 100% @@ -67,15 +67,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_SM_GUX_TTXUX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=1] Workflow summary = CUD:FLT+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 5.322683e+07 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 9.438723e+07 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 9.801307e+07 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 5.769930e+07 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 9.616487e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 9.939241e+07 ) sec^-1 MeanMatrixElemValue = ( 2.571360e+02 +- 2.114020e+02 ) GeV^-2 -TOTAL : 0.493317 sec +TOTAL : 0.485216 sec INFO: No Floating Point Exceptions have been reported - 2,100,361,107 cycles # 2.862 GHz - 2,955,351,040 instructions # 1.41 insn per cycle - 0.791031778 seconds time elapsed + 2,056,020,993 cycles # 2.882 GHz + 2,953,835,853 instructions # 1.44 insn per cycle + 0.770247695 seconds time elapsed ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gq_ttq.mad/SubProcesses/P1_gux_ttxux/build.cuda_f_inl0_hrd1/runTest_cuda.exe INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: FE_DIVBYZERO, FE_INVALID, FE_OVERFLOW @@ -100,15 +100,15 @@ Process = SIGMA_SM_GUX_TTXUX_CPP [gcc 11.3.1] [inlineHel=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/none+NAVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = SCALAR ('none': ~vector[1], no SIMD) -EvtsPerSec[Rmb+ME] (23) = ( 1.112466e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.138003e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.138003e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.114166e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.139511e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.139511e+05 ) sec^-1 MeanMatrixElemValue = ( 2.018564e+01 +- 1.429903e+01 ) GeV^-2 -TOTAL : 1.490381 sec +TOTAL : 1.488259 sec INFO: No Floating Point Exceptions have been reported - 4,405,341,411 cycles # 2.950 GHz - 12,928,117,316 instructions # 2.93 insn per cycle - 1.494164072 seconds time elapsed + 4,405,201,327 cycles # 2.954 GHz + 12,928,516,415 instructions # 2.93 insn per cycle + 1.492295774 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 630) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gq_ttq.mad/SubProcesses/P1_gux_ttxux/build.none_f_inl0_hrd1/runTest_cpp.exe @@ -132,15 +132,15 @@ Process = SIGMA_SM_GUX_TTXUX_CPP [gcc 11.3.1] [inlineHel=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('sse4': SSE4.2, 128bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 2.897278e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.076728e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.076728e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 2.888546e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.068875e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.068875e+05 ) sec^-1 MeanMatrixElemValue = ( 2.018564e+01 +- 1.429903e+01 ) GeV^-2 -TOTAL : 0.582482 sec +TOTAL : 0.584150 sec INFO: No Floating Point Exceptions have been reported - 1,724,294,786 cycles # 2.945 GHz - 4,536,655,836 instructions # 2.63 insn per cycle - 0.586223274 seconds time elapsed + 1,725,547,672 cycles # 2.936 GHz + 4,536,655,755 instructions # 2.63 insn per cycle + 0.588211275 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 3611) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gq_ttq.mad/SubProcesses/P1_gux_ttxux/build.sse4_f_inl0_hrd1/runTest_cpp.exe @@ -164,15 +164,15 @@ Process = SIGMA_SM_GUX_TTXUX_CPP [gcc 11.3.1] [inlineHel=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('avx2': AVX2, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 5.690817e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 6.397497e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 6.397497e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 5.675767e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 6.369890e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 6.369890e+05 ) sec^-1 MeanMatrixElemValue = ( 2.018828e+01 +- 1.429922e+01 ) GeV^-2 -TOTAL : 0.305315 sec +TOTAL : 0.306142 sec INFO: No Floating Point Exceptions have been reported - 857,155,838 cycles # 2.779 GHz - 1,914,615,212 instructions # 2.23 insn per cycle - 0.309003061 seconds time elapsed + 856,403,889 cycles # 2.767 GHz + 1,914,544,525 instructions # 2.24 insn per cycle + 0.310033119 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 3549) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gq_ttq.mad/SubProcesses/P1_gux_ttxux/build.avx2_f_inl0_hrd1/runTest_cpp.exe @@ -196,15 +196,15 @@ Process = SIGMA_SM_GUX_TTXUX_CPP [gcc 11.3.1] [inlineHel=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('512y': AVX512, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 6.056800e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 6.870570e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 6.870570e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 6.078617e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 6.893594e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 6.893594e+05 ) sec^-1 MeanMatrixElemValue = ( 2.018828e+01 +- 1.429922e+01 ) GeV^-2 -TOTAL : 0.288177 sec +TOTAL : 0.286981 sec INFO: No Floating Point Exceptions have been reported - 804,254,194 cycles # 2.761 GHz - 1,829,977,116 instructions # 2.28 insn per cycle - 0.291930002 seconds time elapsed + 804,164,677 cycles # 2.769 GHz + 1,830,058,381 instructions # 2.28 insn per cycle + 0.290981913 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 3364) (512y: 22) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gq_ttq.mad/SubProcesses/P1_gux_ttxux/build.512y_f_inl0_hrd1/runTest_cpp.exe @@ -228,15 +228,15 @@ Process = SIGMA_SM_GUX_TTXUX_CPP [gcc 11.3.1] [inlineHel=0] Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[16] ('512z': AVX512, 512bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 4.550897e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 4.994144e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 4.994144e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 4.477418e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 4.915707e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 4.915707e+05 ) sec^-1 MeanMatrixElemValue = ( 2.018829e+01 +- 1.429922e+01 ) GeV^-2 -TOTAL : 0.380837 sec +TOTAL : 0.385829 sec INFO: No Floating Point Exceptions have been reported - 727,485,601 cycles # 1.894 GHz - 1,306,171,995 instructions # 1.80 insn per cycle - 0.384559776 seconds time elapsed + 730,832,173 cycles # 1.877 GHz + 1,306,232,233 instructions # 1.79 insn per cycle + 0.389993396 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 1928) (512y: 24) (512z: 2435) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gq_ttq.mad/SubProcesses/P1_gux_ttxux/build.512z_f_inl0_hrd1/runTest_cpp.exe diff --git a/epochX/cudacpp/tput/logs_gqttq_mad/log_gqttq_mad_m_inl0_hrd0.txt b/epochX/cudacpp/tput/logs_gqttq_mad/log_gqttq_mad_m_inl0_hrd0.txt index cb0b82e9a4..4a9daae74f 100644 --- a/epochX/cudacpp/tput/logs_gqttq_mad/log_gqttq_mad_m_inl0_hrd0.txt +++ b/epochX/cudacpp/tput/logs_gqttq_mad/log_gqttq_mad_m_inl0_hrd0.txt @@ -40,7 +40,7 @@ make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gq_ttq.mad/SubProcesses/P1_gux_ttxux' -DATE: 2024-10-06_09:06:18 +DATE: 2024-10-06_13:35:28 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: ========================================================================= @@ -49,15 +49,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_SM_GUX_TTXUX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:MIX+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 1.883484e+07 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 2.876597e+07 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 2.990293e+07 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 2.023295e+07 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 2.956243e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.055376e+07 ) sec^-1 MeanMatrixElemValue = ( 2.018083e+01 +- 1.429474e+01 ) GeV^-2 -TOTAL : 0.458412 sec +TOTAL : 0.454495 sec INFO: No Floating Point Exceptions have been reported - 1,935,066,146 cycles # 2.866 GHz - 2,699,989,812 instructions # 1.40 insn per cycle - 0.733387527 seconds time elapsed + 1,932,672,375 cycles # 2.876 GHz + 2,702,475,145 instructions # 1.40 insn per cycle + 0.728822580 seconds time elapsed runNcu /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gq_ttq.mad/SubProcesses/P1_gux_ttxux/build.cuda_m_inl0_hrd0/check_cuda.exe -p 64 256 1 ==PROF== Profiling "sigmaKin": launch__registers_per_thread 255 ==PROF== Profiling "sigmaKin": sm__sass_average_branch_targets_threads_uniform.pct 100% @@ -67,15 +67,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_SM_GUX_TTXUX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:MIX+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 3.013974e+07 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 5.497451e+07 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 5.709351e+07 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 3.143232e+07 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 5.483141e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 5.685928e+07 ) sec^-1 MeanMatrixElemValue = ( 2.602505e+02 +- 2.116328e+02 ) GeV^-2 -TOTAL : 0.541801 sec +TOTAL : 0.536111 sec INFO: No Floating Point Exceptions have been reported - 2,287,504,645 cycles # 2.883 GHz - 3,220,826,671 instructions # 1.41 insn per cycle - 0.850636557 seconds time elapsed + 2,224,349,119 cycles # 2.881 GHz + 3,204,057,049 instructions # 1.44 insn per cycle + 0.830477062 seconds time elapsed ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gq_ttq.mad/SubProcesses/P1_gux_ttxux/build.cuda_m_inl0_hrd0/runTest_cuda.exe INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: FE_DIVBYZERO, FE_INVALID, FE_OVERFLOW @@ -100,15 +100,15 @@ Process = SIGMA_SM_GUX_TTXUX_CPP [gcc 11.3.1] [inlineHel=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/none+NAVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) Internal loops fptype_sv = SCALAR ('none': ~vector[1], no SIMD) -EvtsPerSec[Rmb+ME] (23) = ( 1.050634e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.073472e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.073472e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.056173e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.078772e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.078772e+05 ) sec^-1 MeanMatrixElemValue = ( 2.018083e+01 +- 1.429474e+01 ) GeV^-2 -TOTAL : 1.580404 sec +TOTAL : 1.570997 sec INFO: No Floating Point Exceptions have been reported - 4,643,189,098 cycles # 2.932 GHz - 13,180,741,468 instructions # 2.84 insn per cycle - 1.584505840 seconds time elapsed + 4,635,391,836 cycles # 2.944 GHz + 13,177,958,853 instructions # 2.84 insn per cycle + 1.575084158 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 681) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gq_ttq.mad/SubProcesses/P1_gux_ttxux/build.none_m_inl0_hrd0/runTest_cpp.exe @@ -132,15 +132,15 @@ Process = SIGMA_SM_GUX_TTXUX_CPP [gcc 11.3.1] [inlineHel=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[2] ('sse4': SSE4.2, 128bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 1.871761e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.941517e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.941517e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.870970e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.942619e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.942619e+05 ) sec^-1 MeanMatrixElemValue = ( 2.018083e+01 +- 1.429474e+01 ) GeV^-2 -TOTAL : 0.895197 sec +TOTAL : 0.894666 sec INFO: No Floating Point Exceptions have been reported - 2,647,990,030 cycles # 2.947 GHz - 7,474,565,418 instructions # 2.82 insn per cycle - 0.899253220 seconds time elapsed + 2,645,935,532 cycles # 2.946 GHz + 7,472,837,476 instructions # 2.82 insn per cycle + 0.898850651 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 3152) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gq_ttq.mad/SubProcesses/P1_gux_ttxux/build.sse4_m_inl0_hrd0/runTest_cpp.exe @@ -164,15 +164,15 @@ Process = SIGMA_SM_GUX_TTXUX_CPP [gcc 11.3.1] [inlineHel=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('avx2': AVX2, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 3.201825e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.415489e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.415489e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 3.213141e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.421543e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.421543e+05 ) sec^-1 MeanMatrixElemValue = ( 2.018083e+01 +- 1.429474e+01 ) GeV^-2 -TOTAL : 0.532048 sec +TOTAL : 0.529186 sec INFO: No Floating Point Exceptions have been reported - 1,472,019,476 cycles # 2.748 GHz - 3,129,064,583 instructions # 2.13 insn per cycle - 0.536341858 seconds time elapsed + 1,469,002,894 cycles # 2.757 GHz + 3,126,864,855 instructions # 2.13 insn per cycle + 0.533283838 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 3133) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gq_ttq.mad/SubProcesses/P1_gux_ttxux/build.avx2_m_inl0_hrd0/runTest_cpp.exe @@ -196,15 +196,15 @@ Process = SIGMA_SM_GUX_TTXUX_CPP [gcc 11.3.1] [inlineHel=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('512y': AVX512, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 3.569463e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.831852e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.831852e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 3.524182e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.783990e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.783990e+05 ) sec^-1 MeanMatrixElemValue = ( 2.018083e+01 +- 1.429474e+01 ) GeV^-2 -TOTAL : 0.479328 sec +TOTAL : 0.484555 sec INFO: No Floating Point Exceptions have been reported - 1,320,483,901 cycles # 2.736 GHz - 2,983,197,107 instructions # 2.26 insn per cycle - 0.483280271 seconds time elapsed + 1,319,427,608 cycles # 2.704 GHz + 2,981,111,929 instructions # 2.26 insn per cycle + 0.488670915 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 2895) (512y: 110) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gq_ttq.mad/SubProcesses/P1_gux_ttxux/build.512y_m_inl0_hrd0/runTest_cpp.exe @@ -228,15 +228,15 @@ Process = SIGMA_SM_GUX_TTXUX_CPP [gcc 11.3.1] [inlineHel=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('512z': AVX512, 512bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 2.268192e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 2.372574e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 2.372574e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 2.286281e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 2.389269e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.389269e+05 ) sec^-1 MeanMatrixElemValue = ( 2.018083e+01 +- 1.429474e+01 ) GeV^-2 -TOTAL : 0.745303 sec +TOTAL : 0.739207 sec INFO: No Floating Point Exceptions have been reported - 1,365,795,021 cycles # 1.824 GHz - 1,991,870,632 instructions # 1.46 insn per cycle - 0.749335143 seconds time elapsed + 1,360,659,298 cycles # 1.832 GHz + 1,989,937,554 instructions # 1.46 insn per cycle + 0.743427364 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 1679) (512y: 108) (512z: 2251) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gq_ttq.mad/SubProcesses/P1_gux_ttxux/build.512z_m_inl0_hrd0/runTest_cpp.exe diff --git a/epochX/cudacpp/tput/logs_gqttq_mad/log_gqttq_mad_m_inl0_hrd1.txt b/epochX/cudacpp/tput/logs_gqttq_mad/log_gqttq_mad_m_inl0_hrd1.txt index 222758fe32..ad1ee4eec5 100644 --- a/epochX/cudacpp/tput/logs_gqttq_mad/log_gqttq_mad_m_inl0_hrd1.txt +++ b/epochX/cudacpp/tput/logs_gqttq_mad/log_gqttq_mad_m_inl0_hrd1.txt @@ -40,7 +40,7 @@ make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gq_ttq.mad/SubProcesses/P1_gux_ttxux' -DATE: 2024-10-06_09:06:32 +DATE: 2024-10-06_13:35:42 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: ========================================================================= @@ -49,15 +49,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_SM_GUX_TTXUX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=1] Workflow summary = CUD:MIX+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 1.879429e+07 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 2.807541e+07 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 2.902111e+07 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 2.024624e+07 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 2.959886e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.065351e+07 ) sec^-1 MeanMatrixElemValue = ( 2.018083e+01 +- 1.429474e+01 ) GeV^-2 -TOTAL : 0.462530 sec +TOTAL : 0.456010 sec INFO: No Floating Point Exceptions have been reported - 1,930,179,746 cycles # 2.847 GHz - 2,724,788,037 instructions # 1.41 insn per cycle - 0.736957830 seconds time elapsed + 1,940,270,478 cycles # 2.888 GHz + 2,737,757,784 instructions # 1.41 insn per cycle + 0.730548114 seconds time elapsed runNcu /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gq_ttq.mad/SubProcesses/P1_gux_ttxux/build.cuda_m_inl0_hrd1/check_cuda.exe -p 64 256 1 ==PROF== Profiling "sigmaKin": launch__registers_per_thread 255 ==PROF== Profiling "sigmaKin": sm__sass_average_branch_targets_threads_uniform.pct 100% @@ -67,15 +67,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_SM_GUX_TTXUX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=1] Workflow summary = CUD:MIX+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 2.958663e+07 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 5.373563e+07 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 5.576283e+07 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 3.134783e+07 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 5.457936e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 5.660002e+07 ) sec^-1 MeanMatrixElemValue = ( 2.602505e+02 +- 2.116328e+02 ) GeV^-2 -TOTAL : 0.543078 sec +TOTAL : 0.538228 sec INFO: No Floating Point Exceptions have been reported - 2,226,045,922 cycles # 2.831 GHz - 3,151,460,121 instructions # 1.42 insn per cycle - 0.843097781 seconds time elapsed + 2,216,434,303 cycles # 2.863 GHz + 3,167,665,700 instructions # 1.43 insn per cycle + 0.833502645 seconds time elapsed ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gq_ttq.mad/SubProcesses/P1_gux_ttxux/build.cuda_m_inl0_hrd1/runTest_cuda.exe INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: FE_DIVBYZERO, FE_INVALID, FE_OVERFLOW @@ -100,15 +100,15 @@ Process = SIGMA_SM_GUX_TTXUX_CPP [gcc 11.3.1] [inlineHel=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/none+NAVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) Internal loops fptype_sv = SCALAR ('none': ~vector[1], no SIMD) -EvtsPerSec[Rmb+ME] (23) = ( 1.049471e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.072251e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.072251e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.052226e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.074646e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.074646e+05 ) sec^-1 MeanMatrixElemValue = ( 2.018083e+01 +- 1.429474e+01 ) GeV^-2 -TOTAL : 1.581795 sec +TOTAL : 1.576575 sec INFO: No Floating Point Exceptions have been reported - 4,647,850,638 cycles # 2.932 GHz - 13,168,659,581 instructions # 2.83 insn per cycle - 1.585735048 seconds time elapsed + 4,643,803,675 cycles # 2.939 GHz + 13,166,876,645 instructions # 2.84 insn per cycle + 1.580770513 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 666) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gq_ttq.mad/SubProcesses/P1_gux_ttxux/build.none_m_inl0_hrd1/runTest_cpp.exe @@ -132,15 +132,15 @@ Process = SIGMA_SM_GUX_TTXUX_CPP [gcc 11.3.1] [inlineHel=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[2] ('sse4': SSE4.2, 128bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 1.863863e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.934907e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.934907e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.877302e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.947583e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.947583e+05 ) sec^-1 MeanMatrixElemValue = ( 2.018083e+01 +- 1.429474e+01 ) GeV^-2 -TOTAL : 0.898950 sec +TOTAL : 0.891665 sec INFO: No Floating Point Exceptions have been reported - 2,647,565,316 cycles # 2.935 GHz - 7,477,127,209 instructions # 2.82 insn per cycle - 0.902852166 seconds time elapsed + 2,637,192,700 cycles # 2.946 GHz + 7,475,357,079 instructions # 2.83 insn per cycle + 0.895879146 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 3141) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gq_ttq.mad/SubProcesses/P1_gux_ttxux/build.sse4_m_inl0_hrd1/runTest_cpp.exe @@ -164,15 +164,15 @@ Process = SIGMA_SM_GUX_TTXUX_CPP [gcc 11.3.1] [inlineHel=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('avx2': AVX2, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 3.193877e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.403471e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.403471e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 3.224081e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.435884e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.435884e+05 ) sec^-1 MeanMatrixElemValue = ( 2.018083e+01 +- 1.429474e+01 ) GeV^-2 -TOTAL : 0.533092 sec +TOTAL : 0.527265 sec INFO: No Floating Point Exceptions have been reported - 1,474,101,191 cycles # 2.747 GHz - 3,129,731,788 instructions # 2.12 insn per cycle - 0.537323582 seconds time elapsed + 1,465,095,708 cycles # 2.760 GHz + 3,127,397,376 instructions # 2.13 insn per cycle + 0.531472493 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 3111) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gq_ttq.mad/SubProcesses/P1_gux_ttxux/build.avx2_m_inl0_hrd1/runTest_cpp.exe @@ -196,15 +196,15 @@ Process = SIGMA_SM_GUX_TTXUX_CPP [gcc 11.3.1] [inlineHel=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('512y': AVX512, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 3.595782e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.860984e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.860984e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 3.554574e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.814984e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.814984e+05 ) sec^-1 MeanMatrixElemValue = ( 2.018083e+01 +- 1.429474e+01 ) GeV^-2 -TOTAL : 0.475643 sec +TOTAL : 0.480080 sec INFO: No Floating Point Exceptions have been reported - 1,319,166,719 cycles # 2.754 GHz - 2,983,572,989 instructions # 2.26 insn per cycle - 0.479589426 seconds time elapsed + 1,315,453,571 cycles # 2.719 GHz + 2,981,734,344 instructions # 2.27 insn per cycle + 0.484288954 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 2871) (512y: 110) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gq_ttq.mad/SubProcesses/P1_gux_ttxux/build.512y_m_inl0_hrd1/runTest_cpp.exe @@ -228,15 +228,15 @@ Process = SIGMA_SM_GUX_TTXUX_CPP [gcc 11.3.1] [inlineHel=0] Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('512z': AVX512, 512bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 2.265955e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 2.372021e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 2.372021e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 2.281059e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 2.386520e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.386520e+05 ) sec^-1 MeanMatrixElemValue = ( 2.018083e+01 +- 1.429474e+01 ) GeV^-2 -TOTAL : 0.745301 sec +TOTAL : 0.739619 sec INFO: No Floating Point Exceptions have been reported - 1,365,993,831 cycles # 1.825 GHz - 1,991,757,917 instructions # 1.46 insn per cycle - 0.749395729 seconds time elapsed + 1,363,668,494 cycles # 1.835 GHz + 1,990,032,976 instructions # 1.46 insn per cycle + 0.743830616 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 1655) (512y: 108) (512z: 2251) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/gq_ttq.mad/SubProcesses/P1_gux_ttxux/build.512z_m_inl0_hrd1/runTest_cpp.exe diff --git a/epochX/cudacpp/tput/logs_heftggbb_mad/log_heftggbb_mad_d_inl0_hrd0.txt b/epochX/cudacpp/tput/logs_heftggbb_mad/log_heftggbb_mad_d_inl0_hrd0.txt index 9b3f75797b..c129573e5e 100644 --- a/epochX/cudacpp/tput/logs_heftggbb_mad/log_heftggbb_mad_d_inl0_hrd0.txt +++ b/epochX/cudacpp/tput/logs_heftggbb_mad/log_heftggbb_mad_d_inl0_hrd0.txt @@ -40,7 +40,7 @@ make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/heft_gg_bb.mad/SubProcesses/P1_gg_bbx' -DATE: 2024-10-06_10:01:13 +DATE: 2024-10-06_14:17:07 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: ========================================================================= @@ -49,15 +49,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_HEFT_GG_BBX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 4.147069e+07 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 8.778623e+07 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 9.394888e+07 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 4.352444e+07 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 8.884348e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 9.427260e+07 ) sec^-1 MeanMatrixElemValue = ( 7.148017e+00 +- 1.609110e-01 ) GeV^0 -TOTAL : 0.535117 sec +TOTAL : 0.539608 sec INFO: No Floating Point Exceptions have been reported - 2,222,375,781 cycles # 2.890 GHz - 3,181,150,200 instructions # 1.43 insn per cycle - 0.828824866 seconds time elapsed + 2,196,997,773 cycles # 2.835 GHz + 3,081,004,981 instructions # 1.40 insn per cycle + 0.833746839 seconds time elapsed runNcu /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/heft_gg_bb.mad/SubProcesses/P1_gg_bbx/build.cuda_d_inl0_hrd0/check_cuda.exe -p 2048 256 1 ==PROF== Profiling "sigmaKin": launch__registers_per_thread 228 ==PROF== Profiling "sigmaKin": sm__sass_average_branch_targets_threads_uniform.pct 100% @@ -85,15 +85,15 @@ Process = SIGMA_HEFT_GG_BBX_CPP [gcc 11.3.1] [inlineHel=0] [ Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/none+NAVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = SCALAR ('none': ~vector[1], no SIMD) -EvtsPerSec[Rmb+ME] (23) = ( 1.628496e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.666122e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.666122e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.591587e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.627477e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.627477e+05 ) sec^-1 MeanMatrixElemValue = ( 7.148017e+00 +- 1.609110e-01 ) GeV^0 -TOTAL : 6.554723 sec +TOTAL : 6.700276 sec INFO: No Floating Point Exceptions have been reported - 19,293,957,259 cycles # 2.941 GHz - 51,936,518,995 instructions # 2.69 insn per cycle - 6.561734499 seconds time elapsed + 19,287,927,863 cycles # 2.877 GHz + 51,926,715,239 instructions # 2.69 insn per cycle + 6.706323299 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 668) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/heft_gg_bb.mad/SubProcesses/P1_gg_bbx/build.none_d_inl0_hrd0/runTest_cpp.exe @@ -117,15 +117,15 @@ Process = SIGMA_HEFT_GG_BBX_CPP [gcc 11.3.1] [inlineHel=0] [ Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[2] ('sse4': SSE4.2, 128bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 2.914767e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.044981e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.044981e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 2.875073e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.005987e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.005987e+05 ) sec^-1 MeanMatrixElemValue = ( 7.148017e+00 +- 1.609110e-01 ) GeV^0 -TOTAL : 3.713846 sec +TOTAL : 3.760307 sec INFO: No Floating Point Exceptions have been reported - 10,942,394,234 cycles # 2.942 GHz - 30,809,451,561 instructions # 2.82 insn per cycle - 3.720459537 seconds time elapsed + 10,897,184,952 cycles # 2.894 GHz + 30,796,111,249 instructions # 2.83 insn per cycle + 3.766313423 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 2915) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/heft_gg_bb.mad/SubProcesses/P1_gg_bbx/build.sse4_d_inl0_hrd0/runTest_cpp.exe @@ -149,15 +149,15 @@ Process = SIGMA_HEFT_GG_BBX_CPP [gcc 11.3.1] [inlineHel=0] [ Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('avx2': AVX2, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 4.701521e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 5.038587e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 5.038587e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 4.607991e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 4.940718e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 4.940718e+05 ) sec^-1 MeanMatrixElemValue = ( 7.148017e+00 +- 1.609110e-01 ) GeV^0 -TOTAL : 2.349075 sec +TOTAL : 2.388725 sec INFO: No Floating Point Exceptions have been reported - 6,518,044,155 cycles # 2.767 GHz - 13,691,830,614 instructions # 2.10 insn per cycle - 2.356266703 seconds time elapsed + 6,487,069,047 cycles # 2.710 GHz + 13,666,506,383 instructions # 2.11 insn per cycle + 2.394735283 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 2941) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/heft_gg_bb.mad/SubProcesses/P1_gg_bbx/build.avx2_d_inl0_hrd0/runTest_cpp.exe @@ -181,15 +181,15 @@ Process = SIGMA_HEFT_GG_BBX_CPP [gcc 11.3.1] [inlineHel=0] [ Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('512y': AVX512, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 5.169544e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 5.582169e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 5.582169e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 5.161792e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 5.567543e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 5.567543e+05 ) sec^-1 MeanMatrixElemValue = ( 7.148017e+00 +- 1.609110e-01 ) GeV^0 -TOTAL : 2.151067 sec +TOTAL : 2.140740 sec INFO: No Floating Point Exceptions have been reported - 5,973,431,908 cycles # 2.768 GHz - 13,032,735,919 instructions # 2.18 insn per cycle - 2.158817844 seconds time elapsed + 5,944,170,558 cycles # 2.771 GHz + 13,006,788,495 instructions # 2.19 insn per cycle + 2.146411515 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 2667) (512y: 146) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/heft_gg_bb.mad/SubProcesses/P1_gg_bbx/build.512y_d_inl0_hrd0/runTest_cpp.exe @@ -213,15 +213,15 @@ Process = SIGMA_HEFT_GG_BBX_CPP [gcc 11.3.1] [inlineHel=0] [ Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('512z': AVX512, 512bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 3.442417e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.620453e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.620453e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 3.403950e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.576817e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.576817e+05 ) sec^-1 MeanMatrixElemValue = ( 7.148017e+00 +- 1.609110e-01 ) GeV^0 -TOTAL : 3.166223 sec +TOTAL : 3.192193 sec INFO: No Floating Point Exceptions have been reported - 5,879,580,303 cycles # 1.853 GHz - 8,614,888,302 instructions # 1.47 insn per cycle - 3.173636028 seconds time elapsed + 5,838,081,645 cycles # 1.826 GHz + 8,588,771,173 instructions # 1.47 insn per cycle + 3.197977434 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 1506) (512y: 128) (512z: 1946) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/heft_gg_bb.mad/SubProcesses/P1_gg_bbx/build.512z_d_inl0_hrd0/runTest_cpp.exe diff --git a/epochX/cudacpp/tput/logs_heftggbb_mad/log_heftggbb_mad_d_inl0_hrd1.txt b/epochX/cudacpp/tput/logs_heftggbb_mad/log_heftggbb_mad_d_inl0_hrd1.txt index fe94934cb0..98e4fa7cdc 100644 --- a/epochX/cudacpp/tput/logs_heftggbb_mad/log_heftggbb_mad_d_inl0_hrd1.txt +++ b/epochX/cudacpp/tput/logs_heftggbb_mad/log_heftggbb_mad_d_inl0_hrd1.txt @@ -40,7 +40,7 @@ make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/heft_gg_bb.mad/SubProcesses/P1_gg_bbx' -DATE: 2024-10-06_10:01:39 +DATE: 2024-10-06_14:17:33 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: ========================================================================= @@ -49,15 +49,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_HEFT_GG_BBX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=1] Workflow summary = CUD:DBL+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 4.155696e+07 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 8.811430e+07 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 9.416776e+07 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 4.364286e+07 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 8.881439e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 9.428326e+07 ) sec^-1 MeanMatrixElemValue = ( 7.148017e+00 +- 1.609110e-01 ) GeV^0 -TOTAL : 0.531667 sec +TOTAL : 0.533700 sec INFO: No Floating Point Exceptions have been reported - 2,222,115,079 cycles # 2.893 GHz - 3,196,008,298 instructions # 1.44 insn per cycle - 0.825144177 seconds time elapsed + 2,211,201,509 cycles # 2.883 GHz + 3,185,256,021 instructions # 1.44 insn per cycle + 0.826193510 seconds time elapsed runNcu /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/heft_gg_bb.mad/SubProcesses/P1_gg_bbx/build.cuda_d_inl0_hrd1/check_cuda.exe -p 2048 256 1 ==PROF== Profiling "sigmaKin": launch__registers_per_thread 216 ==PROF== Profiling "sigmaKin": sm__sass_average_branch_targets_threads_uniform.pct 100% @@ -85,15 +85,15 @@ Process = SIGMA_HEFT_GG_BBX_CPP [gcc 11.3.1] [inlineHel=0] [ Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/none+NAVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = SCALAR ('none': ~vector[1], no SIMD) -EvtsPerSec[Rmb+ME] (23) = ( 1.710634e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.751435e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.751435e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.715322e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.756563e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.756563e+05 ) sec^-1 MeanMatrixElemValue = ( 7.148017e+00 +- 1.609110e-01 ) GeV^0 -TOTAL : 6.246482 sec +TOTAL : 6.223105 sec INFO: No Floating Point Exceptions have been reported - 18,390,828,933 cycles # 2.942 GHz - 50,070,723,541 instructions # 2.72 insn per cycle - 6.253313848 seconds time elapsed + 18,377,588,405 cycles # 2.951 GHz + 50,053,125,059 instructions # 2.72 insn per cycle + 6.228818351 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 626) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/heft_gg_bb.mad/SubProcesses/P1_gg_bbx/build.none_d_inl0_hrd1/runTest_cpp.exe @@ -117,15 +117,15 @@ Process = SIGMA_HEFT_GG_BBX_CPP [gcc 11.3.1] [inlineHel=0] [ Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[2] ('sse4': SSE4.2, 128bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 3.069031e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.214398e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.214398e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 3.083881e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.230416e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.230416e+05 ) sec^-1 MeanMatrixElemValue = ( 7.148017e+00 +- 1.609110e-01 ) GeV^0 -TOTAL : 3.535906 sec +TOTAL : 3.510743 sec INFO: No Floating Point Exceptions have been reported - 10,415,008,507 cycles # 2.940 GHz - 29,198,189,749 instructions # 2.80 insn per cycle - 3.543300262 seconds time elapsed + 10,377,341,832 cycles # 2.952 GHz + 29,173,086,326 instructions # 2.81 insn per cycle + 3.516422038 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 2733) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/heft_gg_bb.mad/SubProcesses/P1_gg_bbx/build.sse4_d_inl0_hrd1/runTest_cpp.exe @@ -149,15 +149,15 @@ Process = SIGMA_HEFT_GG_BBX_CPP [gcc 11.3.1] [inlineHel=0] [ Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('avx2': AVX2, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 4.327920e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 4.613203e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 4.613203e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 4.375583e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 4.665483e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 4.665483e+05 ) sec^-1 MeanMatrixElemValue = ( 7.148017e+00 +- 1.609110e-01 ) GeV^0 -TOTAL : 2.541514 sec +TOTAL : 2.506207 sec INFO: No Floating Point Exceptions have been reported - 7,032,477,509 cycles # 2.760 GHz - 15,175,173,386 instructions # 2.16 insn per cycle - 2.548867076 seconds time elapsed + 6,970,070,462 cycles # 2.776 GHz + 15,150,854,388 instructions # 2.17 insn per cycle + 2.511934117 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 3020) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/heft_gg_bb.mad/SubProcesses/P1_gg_bbx/build.avx2_d_inl0_hrd1/runTest_cpp.exe @@ -181,15 +181,15 @@ Process = SIGMA_HEFT_GG_BBX_CPP [gcc 11.3.1] [inlineHel=0] [ Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('512y': AVX512, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 4.529226e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 4.840126e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 4.840126e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 4.507449e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 4.812396e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 4.812396e+05 ) sec^-1 MeanMatrixElemValue = ( 7.148017e+00 +- 1.609110e-01 ) GeV^0 -TOTAL : 2.433970 sec +TOTAL : 2.436022 sec INFO: No Floating Point Exceptions have been reported - 6,732,593,285 cycles # 2.759 GHz - 14,647,151,783 instructions # 2.18 insn per cycle - 2.441354685 seconds time elapsed + 6,775,044,568 cycles # 2.776 GHz + 14,623,241,099 instructions # 2.16 insn per cycle + 2.441677902 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 2621) (512y: 302) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/heft_gg_bb.mad/SubProcesses/P1_gg_bbx/build.512y_d_inl0_hrd1/runTest_cpp.exe @@ -213,15 +213,15 @@ Process = SIGMA_HEFT_GG_BBX_CPP [gcc 11.3.1] [inlineHel=0] [ Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('512z': AVX512, 512bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 3.326729e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.490201e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.490201e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 3.339115e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.503356e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.503356e+05 ) sec^-1 MeanMatrixElemValue = ( 7.148017e+00 +- 1.609110e-01 ) GeV^0 -TOTAL : 3.271504 sec +TOTAL : 3.251023 sec INFO: No Floating Point Exceptions have been reported - 6,070,928,941 cycles # 1.852 GHz - 10,360,391,243 instructions # 1.71 insn per cycle - 3.278977914 seconds time elapsed + 6,043,112,424 cycles # 1.856 GHz + 10,338,862,018 instructions # 1.71 insn per cycle + 3.256819215 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 1268) (512y: 214) (512z: 2129) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/heft_gg_bb.mad/SubProcesses/P1_gg_bbx/build.512z_d_inl0_hrd1/runTest_cpp.exe diff --git a/epochX/cudacpp/tput/logs_heftggbb_mad/log_heftggbb_mad_f_inl0_hrd0.txt b/epochX/cudacpp/tput/logs_heftggbb_mad/log_heftggbb_mad_f_inl0_hrd0.txt index 8cd2c74f38..de52b46f62 100644 --- a/epochX/cudacpp/tput/logs_heftggbb_mad/log_heftggbb_mad_f_inl0_hrd0.txt +++ b/epochX/cudacpp/tput/logs_heftggbb_mad/log_heftggbb_mad_f_inl0_hrd0.txt @@ -40,7 +40,7 @@ make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/heft_gg_bb.mad/SubProcesses/P1_gg_bbx' -DATE: 2024-10-06_10:03:02 +DATE: 2024-10-06_14:18:55 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: ========================================================================= @@ -49,15 +49,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_HEFT_GG_BBX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 7.465620e+07 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.510965e+08 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.608079e+08 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 7.907635e+07 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.544689e+08 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.633317e+08 ) sec^-1 MeanMatrixElemValue = ( 7.154219e+00 +- 1.620281e-01 ) GeV^0 -TOTAL : 0.492110 sec +TOTAL : 0.489866 sec INFO: No Floating Point Exceptions have been reported - 2,084,727,455 cycles # 2.877 GHz - 2,955,736,176 instructions # 1.42 insn per cycle - 0.784112386 seconds time elapsed + 2,060,427,926 cycles # 2.873 GHz + 2,944,112,438 instructions # 1.43 insn per cycle + 0.775002366 seconds time elapsed runNcu /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/heft_gg_bb.mad/SubProcesses/P1_gg_bbx/build.cuda_f_inl0_hrd0/check_cuda.exe -p 2048 256 1 ==PROF== Profiling "sigmaKin": launch__registers_per_thread 131 ==PROF== Profiling "sigmaKin": sm__sass_average_branch_targets_threads_uniform.pct 100% @@ -85,15 +85,15 @@ Process = SIGMA_HEFT_GG_BBX_CPP [gcc 11.3.1] [inlineHel=0] [ Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/none+NAVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = SCALAR ('none': ~vector[1], no SIMD) -EvtsPerSec[Rmb+ME] (23) = ( 1.686557e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.727704e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.727704e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.682147e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.723458e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.723458e+05 ) sec^-1 MeanMatrixElemValue = ( 7.175644e+00 +- 1.658767e-01 ) GeV^0 -TOTAL : 6.305463 sec +TOTAL : 6.322130 sec INFO: The following Floating Point Exceptions have been reported: FE_UNDERFLOW - 18,635,302,225 cycles # 2.953 GHz - 51,219,407,083 instructions # 2.75 insn per cycle - 6.310992251 seconds time elapsed + 18,614,376,939 cycles # 2.942 GHz + 51,215,780,111 instructions # 2.75 insn per cycle + 6.327463200 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 625) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/heft_gg_bb.mad/SubProcesses/P1_gg_bbx/build.none_f_inl0_hrd0/runTest_cpp.exe @@ -119,15 +119,15 @@ Process = SIGMA_HEFT_GG_BBX_CPP [gcc 11.3.1] [inlineHel=0] [ Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('sse4': SSE4.2, 128bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 4.043062e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 4.307407e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 4.307407e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 4.024196e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 4.287100e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 4.287100e+05 ) sec^-1 MeanMatrixElemValue = ( 7.175642e+00 +- 1.658767e-01 ) GeV^0 -TOTAL : 2.681205 sec +TOTAL : 2.693463 sec INFO: The following Floating Point Exceptions have been reported: FE_UNDERFLOW - 7,932,523,130 cycles # 2.953 GHz - 19,317,767,787 instructions # 2.44 insn per cycle - 2.686665617 seconds time elapsed + 7,934,402,522 cycles # 2.941 GHz + 19,316,775,708 instructions # 2.43 insn per cycle + 2.698889520 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 3542) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/heft_gg_bb.mad/SubProcesses/P1_gg_bbx/build.sse4_f_inl0_hrd0/runTest_cpp.exe @@ -153,15 +153,15 @@ Process = SIGMA_HEFT_GG_BBX_CPP [gcc 11.3.1] [inlineHel=0] [ Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('avx2': AVX2, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 7.901471e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 8.926003e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 8.926003e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 7.882187e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 8.893081e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 8.893081e+05 ) sec^-1 MeanMatrixElemValue = ( 7.198861e+00 +- 1.710281e-01 ) GeV^0 -TOTAL : 1.413719 sec +TOTAL : 1.417418 sec INFO: The following Floating Point Exceptions have been reported: FE_UNDERFLOW - 3,953,020,388 cycles # 2.786 GHz - 8,832,668,299 instructions # 2.23 insn per cycle - 1.419629254 seconds time elapsed + 3,949,456,899 cycles # 2.777 GHz + 8,833,030,082 instructions # 2.24 insn per cycle + 1.422869467 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 3715) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/heft_gg_bb.mad/SubProcesses/P1_gg_bbx/build.avx2_f_inl0_hrd0/runTest_cpp.exe @@ -185,15 +185,15 @@ Process = SIGMA_HEFT_GG_BBX_CPP [gcc 11.3.1] [inlineHel=0] [ Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('512y': AVX512, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 8.392997e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 9.544307e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 9.544307e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 8.392229e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 9.542300e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 9.542300e+05 ) sec^-1 MeanMatrixElemValue = ( 7.198861e+00 +- 1.710281e-01 ) GeV^0 -TOTAL : 1.337803 sec +TOTAL : 1.336631 sec INFO: The following Floating Point Exceptions have been reported: FE_UNDERFLOW - 3,735,491,375 cycles # 2.782 GHz - 8,430,906,889 instructions # 2.26 insn per cycle - 1.343508069 seconds time elapsed + 3,728,846,825 cycles # 2.780 GHz + 8,432,007,283 instructions # 2.26 insn per cycle + 1.342009108 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 3541) (512y: 20) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/heft_gg_bb.mad/SubProcesses/P1_gg_bbx/build.512y_f_inl0_hrd0/runTest_cpp.exe @@ -217,15 +217,15 @@ Process = SIGMA_HEFT_GG_BBX_CPP [gcc 11.3.1] [inlineHel=0] [ Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[16] ('512z': AVX512, 512bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 6.024352e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 6.578236e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 6.578236e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 5.998337e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 6.552324e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 6.552324e+05 ) sec^-1 MeanMatrixElemValue = ( 7.198861e+00 +- 1.710281e-01 ) GeV^0 -TOTAL : 1.827995 sec +TOTAL : 1.837052 sec INFO: The following Floating Point Exceptions have been reported: FE_UNDERFLOW - 3,508,723,607 cycles # 1.915 GHz - 6,244,798,669 instructions # 1.78 insn per cycle - 1.833521857 seconds time elapsed + 3,511,638,256 cycles # 1.907 GHz + 6,244,160,027 instructions # 1.78 insn per cycle + 1.842859397 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 2325) (512y: 22) (512z: 2290) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/heft_gg_bb.mad/SubProcesses/P1_gg_bbx/build.512z_f_inl0_hrd0/runTest_cpp.exe diff --git a/epochX/cudacpp/tput/logs_heftggbb_mad/log_heftggbb_mad_f_inl0_hrd1.txt b/epochX/cudacpp/tput/logs_heftggbb_mad/log_heftggbb_mad_f_inl0_hrd1.txt index 1ff1d26090..b805a8089f 100644 --- a/epochX/cudacpp/tput/logs_heftggbb_mad/log_heftggbb_mad_f_inl0_hrd1.txt +++ b/epochX/cudacpp/tput/logs_heftggbb_mad/log_heftggbb_mad_f_inl0_hrd1.txt @@ -40,7 +40,7 @@ make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/heft_gg_bb.mad/SubProcesses/P1_gg_bbx' -DATE: 2024-10-06_10:03:24 +DATE: 2024-10-06_14:19:17 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: ========================================================================= @@ -49,15 +49,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_HEFT_GG_BBX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=1] Workflow summary = CUD:FLT+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 7.690902e+07 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.615208e+08 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.727767e+08 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 8.173423e+07 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.618569e+08 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.718334e+08 ) sec^-1 MeanMatrixElemValue = ( 7.154219e+00 +- 1.620281e-01 ) GeV^0 -TOTAL : 0.493976 sec +TOTAL : 0.492260 sec INFO: No Floating Point Exceptions have been reported - 2,066,790,877 cycles # 2.843 GHz - 2,969,404,210 instructions # 1.44 insn per cycle - 0.785535997 seconds time elapsed + 2,100,738,771 cycles # 2.867 GHz + 2,988,527,648 instructions # 1.42 insn per cycle + 0.791849586 seconds time elapsed runNcu /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/heft_gg_bb.mad/SubProcesses/P1_gg_bbx/build.cuda_f_inl0_hrd1/check_cuda.exe -p 2048 256 1 ==PROF== Profiling "sigmaKin": launch__registers_per_thread 125 ==PROF== Profiling "sigmaKin": sm__sass_average_branch_targets_threads_uniform.pct 100% @@ -85,15 +85,15 @@ Process = SIGMA_HEFT_GG_BBX_CPP [gcc 11.3.1] [inlineHel=0] [ Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/none+NAVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = SCALAR ('none': ~vector[1], no SIMD) -EvtsPerSec[Rmb+ME] (23) = ( 1.736131e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.779781e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.779781e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.734827e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.778210e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.778210e+05 ) sec^-1 MeanMatrixElemValue = ( 7.175644e+00 +- 1.658767e-01 ) GeV^0 -TOTAL : 6.127979 sec +TOTAL : 6.131688 sec INFO: The following Floating Point Exceptions have been reported: FE_UNDERFLOW - 18,032,140,147 cycles # 2.940 GHz - 49,602,643,371 instructions # 2.75 insn per cycle - 6.133935412 seconds time elapsed + 18,026,994,248 cycles # 2.938 GHz + 49,602,406,837 instructions # 2.75 insn per cycle + 6.137132214 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 613) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/heft_gg_bb.mad/SubProcesses/P1_gg_bbx/build.none_f_inl0_hrd1/runTest_cpp.exe @@ -119,15 +119,15 @@ Process = SIGMA_HEFT_GG_BBX_CPP [gcc 11.3.1] [inlineHel=0] [ Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('sse4': SSE4.2, 128bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 4.506367e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 4.839198e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 4.839198e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 4.514519e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 4.846890e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 4.846890e+05 ) sec^-1 MeanMatrixElemValue = ( 7.175642e+00 +- 1.658767e-01 ) GeV^0 -TOTAL : 2.414203 sec +TOTAL : 2.409494 sec INFO: The following Floating Point Exceptions have been reported: FE_UNDERFLOW - 7,115,995,603 cycles # 2.942 GHz - 18,533,869,751 instructions # 2.60 insn per cycle - 2.419892180 seconds time elapsed + 7,114,287,703 cycles # 2.947 GHz + 18,533,588,521 instructions # 2.61 insn per cycle + 2.414848743 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 3252) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/heft_gg_bb.mad/SubProcesses/P1_gg_bbx/build.sse4_f_inl0_hrd1/runTest_cpp.exe @@ -153,15 +153,15 @@ Process = SIGMA_HEFT_GG_BBX_CPP [gcc 11.3.1] [inlineHel=0] [ Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('avx2': AVX2, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 5.374488e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 5.825683e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 5.825683e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 5.393445e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 5.845739e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 5.845739e+05 ) sec^-1 MeanMatrixElemValue = ( 7.198861e+00 +- 1.710281e-01 ) GeV^0 -TOTAL : 2.037733 sec +TOTAL : 2.031270 sec INFO: The following Floating Point Exceptions have been reported: FE_UNDERFLOW - 5,644,200,229 cycles # 2.763 GHz - 10,848,148,808 instructions # 1.92 insn per cycle - 2.043741542 seconds time elapsed + 5,638,835,972 cycles # 2.770 GHz + 10,848,746,460 instructions # 1.92 insn per cycle + 2.036781930 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 4274) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/heft_gg_bb.mad/SubProcesses/P1_gg_bbx/build.avx2_f_inl0_hrd1/runTest_cpp.exe @@ -187,15 +187,15 @@ Process = SIGMA_HEFT_GG_BBX_CPP [gcc 11.3.1] [inlineHel=0] [ Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('512y': AVX512, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 5.433283e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 5.894901e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 5.894901e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 5.419940e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 5.879978e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 5.879978e+05 ) sec^-1 MeanMatrixElemValue = ( 7.198861e+00 +- 1.710281e-01 ) GeV^0 -TOTAL : 2.017462 sec +TOTAL : 2.022838 sec INFO: The following Floating Point Exceptions have been reported: FE_UNDERFLOW - 5,594,464,289 cycles # 2.767 GHz - 10,554,918,385 instructions # 1.89 insn per cycle - 2.022782231 seconds time elapsed + 5,562,665,173 cycles # 2.743 GHz + 10,552,623,489 instructions # 1.90 insn per cycle + 2.028380700 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 4138) (512y: 12) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/heft_gg_bb.mad/SubProcesses/P1_gg_bbx/build.512y_f_inl0_hrd1/runTest_cpp.exe @@ -221,15 +221,15 @@ Process = SIGMA_HEFT_GG_BBX_CPP [gcc 11.3.1] [inlineHel=0] [ Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[16] ('512z': AVX512, 512bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 4.364066e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 4.648223e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 4.648223e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 4.406284e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 4.696184e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 4.696184e+05 ) sec^-1 MeanMatrixElemValue = ( 7.198861e+00 +- 1.710281e-01 ) GeV^0 -TOTAL : 2.491143 sec +TOTAL : 2.467524 sec INFO: The following Floating Point Exceptions have been reported: FE_UNDERFLOW - 4,639,687,839 cycles # 1.859 GHz - 8,661,216,579 instructions # 1.87 insn per cycle - 2.496647539 seconds time elapsed + 4,636,499,805 cycles # 1.876 GHz + 8,661,003,486 instructions # 1.87 insn per cycle + 2.472956612 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 2799) (512y: 0) (512z: 2885) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/heft_gg_bb.mad/SubProcesses/P1_gg_bbx/build.512z_f_inl0_hrd1/runTest_cpp.exe diff --git a/epochX/cudacpp/tput/logs_heftggbb_mad/log_heftggbb_mad_m_inl0_hrd0.txt b/epochX/cudacpp/tput/logs_heftggbb_mad/log_heftggbb_mad_m_inl0_hrd0.txt index 12c9da87af..5ab1657948 100644 --- a/epochX/cudacpp/tput/logs_heftggbb_mad/log_heftggbb_mad_m_inl0_hrd0.txt +++ b/epochX/cudacpp/tput/logs_heftggbb_mad/log_heftggbb_mad_m_inl0_hrd0.txt @@ -40,7 +40,7 @@ make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/heft_gg_bb.mad/SubProcesses/P1_gg_bbx' -DATE: 2024-10-06_10:02:06 +DATE: 2024-10-06_14:18:00 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: ========================================================================= @@ -49,15 +49,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_HEFT_GG_BBX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:MIX+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 4.131914e+07 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 8.755854e+07 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 9.359452e+07 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 4.359978e+07 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 8.913887e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 9.462941e+07 ) sec^-1 MeanMatrixElemValue = ( 7.148017e+00 +- 1.609110e-01 ) GeV^0 -TOTAL : 0.535565 sec +TOTAL : 0.531770 sec INFO: No Floating Point Exceptions have been reported - 2,204,224,001 cycles # 2.864 GHz - 3,121,247,303 instructions # 1.42 insn per cycle - 0.828499405 seconds time elapsed + 2,202,828,813 cycles # 2.870 GHz + 3,143,293,807 instructions # 1.43 insn per cycle + 0.824228233 seconds time elapsed runNcu /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/heft_gg_bb.mad/SubProcesses/P1_gg_bbx/build.cuda_m_inl0_hrd0/check_cuda.exe -p 2048 256 1 ==PROF== Profiling "sigmaKin": launch__registers_per_thread 228 ==PROF== Profiling "sigmaKin": sm__sass_average_branch_targets_threads_uniform.pct 100% @@ -85,15 +85,15 @@ Process = SIGMA_HEFT_GG_BBX_CPP [gcc 11.3.1] [inlineHel=0] [ Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/none+NAVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) Internal loops fptype_sv = SCALAR ('none': ~vector[1], no SIMD) -EvtsPerSec[Rmb+ME] (23) = ( 1.529079e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.561968e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.561968e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.536364e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.569126e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.569126e+05 ) sec^-1 MeanMatrixElemValue = ( 7.148017e+00 +- 1.609110e-01 ) GeV^0 -TOTAL : 6.973239 sec +TOTAL : 6.935785 sec INFO: The following Floating Point Exceptions have been reported: FE_UNDERFLOW - 20,550,139,482 cycles # 2.945 GHz - 51,941,635,065 instructions # 2.53 insn per cycle - 6.980082779 seconds time elapsed + 20,525,668,454 cycles # 2.958 GHz + 51,927,486,017 instructions # 2.53 insn per cycle + 6.941493282 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 655) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/heft_gg_bb.mad/SubProcesses/P1_gg_bbx/build.none_m_inl0_hrd0/runTest_cpp.exe @@ -119,15 +119,15 @@ Process = SIGMA_HEFT_GG_BBX_CPP [gcc 11.3.1] [inlineHel=0] [ Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[2] ('sse4': SSE4.2, 128bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 2.672019e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 2.782339e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 2.782339e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 2.728645e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 2.842683e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.842683e+05 ) sec^-1 MeanMatrixElemValue = ( 7.148017e+00 +- 1.609110e-01 ) GeV^0 -TOTAL : 4.043433 sec +TOTAL : 3.954425 sec INFO: The following Floating Point Exceptions have been reported: FE_UNDERFLOW - 11,521,778,322 cycles # 2.845 GHz - 30,615,090,868 instructions # 2.66 insn per cycle - 4.050715703 seconds time elapsed + 11,504,880,916 cycles # 2.906 GHz + 30,593,949,279 instructions # 2.66 insn per cycle + 3.960123980 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 2972) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/heft_gg_bb.mad/SubProcesses/P1_gg_bbx/build.sse4_m_inl0_hrd0/runTest_cpp.exe @@ -153,15 +153,15 @@ Process = SIGMA_HEFT_GG_BBX_CPP [gcc 11.3.1] [inlineHel=0] [ Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('avx2': AVX2, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 4.474164e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 4.781347e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 4.781347e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 4.543107e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 4.856603e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 4.856603e+05 ) sec^-1 MeanMatrixElemValue = ( 7.148017e+00 +- 1.609110e-01 ) GeV^0 -TOTAL : 2.469295 sec +TOTAL : 2.419181 sec INFO: The following Floating Point Exceptions have been reported: FE_UNDERFLOW - 6,758,530,167 cycles # 2.729 GHz - 13,653,357,404 instructions # 2.02 insn per cycle - 2.477625143 seconds time elapsed + 6,710,477,575 cycles # 2.768 GHz + 13,609,034,415 instructions # 2.03 insn per cycle + 2.424821035 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 3118) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/heft_gg_bb.mad/SubProcesses/P1_gg_bbx/build.avx2_m_inl0_hrd0/runTest_cpp.exe @@ -187,15 +187,15 @@ Process = SIGMA_HEFT_GG_BBX_CPP [gcc 11.3.1] [inlineHel=0] [ Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('512y': AVX512, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 4.946193e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 5.312777e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 5.312777e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 4.999405e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 5.375352e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 5.375352e+05 ) sec^-1 MeanMatrixElemValue = ( 7.148017e+00 +- 1.609110e-01 ) GeV^0 -TOTAL : 2.239110 sec +TOTAL : 2.206783 sec INFO: The following Floating Point Exceptions have been reported: FE_UNDERFLOW - 6,206,585,291 cycles # 2.765 GHz - 13,005,835,459 instructions # 2.10 insn per cycle - 2.246664710 seconds time elapsed + 6,139,944,867 cycles # 2.776 GHz + 12,973,475,838 instructions # 2.11 insn per cycle + 2.212624917 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 2851) (512y: 150) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/heft_gg_bb.mad/SubProcesses/P1_gg_bbx/build.512y_m_inl0_hrd0/runTest_cpp.exe @@ -221,15 +221,15 @@ Process = SIGMA_HEFT_GG_BBX_CPP [gcc 11.3.1] [inlineHel=0] [ Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('512z': AVX512, 512bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 3.130780e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.276017e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.276017e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 3.117166e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.258832e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.258832e+05 ) sec^-1 MeanMatrixElemValue = ( 7.148017e+00 +- 1.609110e-01 ) GeV^0 -TOTAL : 3.470623 sec +TOTAL : 3.474968 sec INFO: The following Floating Point Exceptions have been reported: FE_UNDERFLOW - 6,429,525,372 cycles # 1.849 GHz - 8,729,822,669 instructions # 1.36 insn per cycle - 3.478318009 seconds time elapsed + 6,393,098,053 cycles # 1.837 GHz + 8,700,735,124 instructions # 1.36 insn per cycle + 3.480700719 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 1792) (512y: 130) (512z: 2014) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/heft_gg_bb.mad/SubProcesses/P1_gg_bbx/build.512z_m_inl0_hrd0/runTest_cpp.exe diff --git a/epochX/cudacpp/tput/logs_heftggbb_mad/log_heftggbb_mad_m_inl0_hrd1.txt b/epochX/cudacpp/tput/logs_heftggbb_mad/log_heftggbb_mad_m_inl0_hrd1.txt index 90c964242c..ccfb92ecf7 100644 --- a/epochX/cudacpp/tput/logs_heftggbb_mad/log_heftggbb_mad_m_inl0_hrd1.txt +++ b/epochX/cudacpp/tput/logs_heftggbb_mad/log_heftggbb_mad_m_inl0_hrd1.txt @@ -40,7 +40,7 @@ make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/heft_gg_bb.mad/SubProcesses/P1_gg_bbx' -DATE: 2024-10-06_10:02:34 +DATE: 2024-10-06_14:18:27 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: ========================================================================= @@ -49,15 +49,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_HEFT_GG_BBX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=1] Workflow summary = CUD:MIX+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 4.143359e+07 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 8.817002e+07 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 9.430401e+07 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 4.328765e+07 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 8.835394e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 9.378009e+07 ) sec^-1 MeanMatrixElemValue = ( 7.148017e+00 +- 1.609110e-01 ) GeV^0 -TOTAL : 0.533031 sec +TOTAL : 0.534562 sec INFO: No Floating Point Exceptions have been reported - 2,222,154,822 cycles # 2.885 GHz - 3,215,427,054 instructions # 1.45 insn per cycle - 0.826924367 seconds time elapsed + 2,207,582,594 cycles # 2.868 GHz + 3,151,098,683 instructions # 1.43 insn per cycle + 0.828792705 seconds time elapsed runNcu /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/heft_gg_bb.mad/SubProcesses/P1_gg_bbx/build.cuda_m_inl0_hrd1/check_cuda.exe -p 2048 256 1 ==PROF== Profiling "sigmaKin": launch__registers_per_thread 216 ==PROF== Profiling "sigmaKin": sm__sass_average_branch_targets_threads_uniform.pct 100% @@ -85,15 +85,15 @@ Process = SIGMA_HEFT_GG_BBX_CPP [gcc 11.3.1] [inlineHel=0] [ Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/none+NAVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) Internal loops fptype_sv = SCALAR ('none': ~vector[1], no SIMD) -EvtsPerSec[Rmb+ME] (23) = ( 1.616471e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.652773e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.652773e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.611997e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.648398e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.648398e+05 ) sec^-1 MeanMatrixElemValue = ( 7.148017e+00 +- 1.609110e-01 ) GeV^0 -TOTAL : 6.603326 sec +TOTAL : 6.615569 sec INFO: The following Floating Point Exceptions have been reported: FE_UNDERFLOW - 19,494,406,226 cycles # 2.950 GHz - 49,966,413,800 instructions # 2.56 insn per cycle - 6.609959024 seconds time elapsed + 19,473,625,141 cycles # 2.941 GHz + 49,954,160,679 instructions # 2.57 insn per cycle + 6.621411044 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 599) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/heft_gg_bb.mad/SubProcesses/P1_gg_bbx/build.none_m_inl0_hrd1/runTest_cpp.exe @@ -119,15 +119,15 @@ Process = SIGMA_HEFT_GG_BBX_CPP [gcc 11.3.1] [inlineHel=0] [ Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[2] ('sse4': SSE4.2, 128bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 2.890177e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.018164e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.018164e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 2.810409e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 2.931525e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.931525e+05 ) sec^-1 MeanMatrixElemValue = ( 7.148017e+00 +- 1.609110e-01 ) GeV^0 -TOTAL : 3.745798 sec +TOTAL : 3.842162 sec INFO: The following Floating Point Exceptions have been reported: FE_UNDERFLOW - 11,068,643,232 cycles # 2.950 GHz - 29,164,471,893 instructions # 2.63 insn per cycle - 3.753005329 seconds time elapsed + 11,039,048,584 cycles # 2.870 GHz + 29,139,279,953 instructions # 2.64 insn per cycle + 3.847873903 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 2815) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/heft_gg_bb.mad/SubProcesses/P1_gg_bbx/build.sse4_m_inl0_hrd1/runTest_cpp.exe @@ -153,15 +153,15 @@ Process = SIGMA_HEFT_GG_BBX_CPP [gcc 11.3.1] [inlineHel=0] [ Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('avx2': AVX2, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 3.744994e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.955254e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.955254e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 3.732297e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.945355e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.945355e+05 ) sec^-1 MeanMatrixElemValue = ( 7.148017e+00 +- 1.609110e-01 ) GeV^0 -TOTAL : 2.917714 sec +TOTAL : 2.921117 sec INFO: The following Floating Point Exceptions have been reported: FE_UNDERFLOW - 8,087,123,435 cycles # 2.766 GHz - 15,210,355,188 instructions # 1.88 insn per cycle - 2.924634632 seconds time elapsed + 8,064,813,604 cycles # 2.756 GHz + 15,189,142,111 instructions # 1.88 insn per cycle + 2.926843017 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 3203) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/heft_gg_bb.mad/SubProcesses/P1_gg_bbx/build.avx2_m_inl0_hrd1/runTest_cpp.exe @@ -187,15 +187,15 @@ Process = SIGMA_HEFT_GG_BBX_CPP [gcc 11.3.1] [inlineHel=0] [ Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('512y': AVX512, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 3.909194e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 4.140218e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 4.140218e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 3.918374e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 4.151810e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 4.151810e+05 ) sec^-1 MeanMatrixElemValue = ( 7.148017e+00 +- 1.609110e-01 ) GeV^0 -TOTAL : 2.798673 sec +TOTAL : 2.786260 sec INFO: The following Floating Point Exceptions have been reported: FE_UNDERFLOW - 7,730,347,780 cycles # 2.756 GHz - 14,498,978,915 instructions # 1.88 insn per cycle - 2.805768338 seconds time elapsed + 7,689,827,935 cycles # 2.755 GHz + 14,485,079,110 instructions # 1.88 insn per cycle + 2.791938242 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 2775) (512y: 304) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/heft_gg_bb.mad/SubProcesses/P1_gg_bbx/build.512y_m_inl0_hrd1/runTest_cpp.exe @@ -221,15 +221,15 @@ Process = SIGMA_HEFT_GG_BBX_CPP [gcc 11.3.1] [inlineHel=0] [ Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('512z': AVX512, 512bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 3.049249e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.186111e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.186111e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 3.056271e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.192380e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.192380e+05 ) sec^-1 MeanMatrixElemValue = ( 7.148017e+00 +- 1.609110e-01 ) GeV^0 -TOTAL : 3.561293 sec +TOTAL : 3.541215 sec INFO: The following Floating Point Exceptions have been reported: FE_UNDERFLOW - 6,578,699,260 cycles # 1.844 GHz - 9,927,155,424 instructions # 1.51 insn per cycle - 3.569129809 seconds time elapsed + 6,524,678,245 cycles # 1.840 GHz + 9,894,099,208 instructions # 1.52 insn per cycle + 3.546878744 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 1565) (512y: 216) (512z: 2216) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/heft_gg_bb.mad/SubProcesses/P1_gg_bbx/build.512z_m_inl0_hrd1/runTest_cpp.exe diff --git a/epochX/cudacpp/tput/logs_smeftggtttt_mad/log_smeftggtttt_mad_d_inl0_hrd0.txt b/epochX/cudacpp/tput/logs_smeftggtttt_mad/log_smeftggtttt_mad_d_inl0_hrd0.txt index 2b34ea67ad..d807ab0a6e 100644 --- a/epochX/cudacpp/tput/logs_smeftggtttt_mad/log_smeftggtttt_mad_d_inl0_hrd0.txt +++ b/epochX/cudacpp/tput/logs_smeftggtttt_mad/log_smeftggtttt_mad_d_inl0_hrd0.txt @@ -40,7 +40,7 @@ make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/smeft_gg_tttt.mad/SubProcesses/P1_gg_ttxttx' -DATE: 2024-10-06_10:00:07 +DATE: 2024-10-06_14:16:00 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: ========================================================================= @@ -49,15 +49,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_SMEFTSIM_TOPU3L_MWSCHEME_UFO_GG_TTXTTX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 2.760509e+04 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 2.779507e+04 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 2.782702e+04 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 2.748231e+04 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 2.767058e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.770539e+04 ) sec^-1 MeanMatrixElemValue = ( 7.185537e-04 +- 6.562553e-04 ) GeV^-4 -TOTAL : 0.473450 sec +TOTAL : 0.474293 sec INFO: No Floating Point Exceptions have been reported - 1,994,326,240 cycles # 2.874 GHz - 2,845,102,706 instructions # 1.43 insn per cycle - 0.753810347 seconds time elapsed + 1,996,308,044 cycles # 2.869 GHz + 2,926,968,206 instructions # 1.47 insn per cycle + 0.754801737 seconds time elapsed runNcu /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/smeft_gg_tttt.mad/SubProcesses/P1_gg_ttxttx/build.cuda_d_inl0_hrd0/check_cuda.exe -p 1 256 1 ==PROF== Profiling "sigmaKin": launch__registers_per_thread 255 ==PROF== Profiling "sigmaKin": sm__sass_average_branch_targets_threads_uniform.pct 100% @@ -67,15 +67,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_SMEFTSIM_TOPU3L_MWSCHEME_UFO_GG_TTXTTX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 7.019067e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 7.126130e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 7.133988e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 6.989035e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 7.098592e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 7.107611e+05 ) sec^-1 MeanMatrixElemValue = ( 8.048215e-03 +- 4.042405e-03 ) GeV^-4 -TOTAL : 0.490494 sec +TOTAL : 0.488675 sec INFO: No Floating Point Exceptions have been reported - 2,031,600,016 cycles # 2.857 GHz - 2,995,319,726 instructions # 1.47 insn per cycle - 0.772627668 seconds time elapsed + 2,042,802,013 cycles # 2.873 GHz + 3,020,112,467 instructions # 1.48 insn per cycle + 0.770645683 seconds time elapsed ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/smeft_gg_tttt.mad/SubProcesses/P1_gg_ttxttx/build.cuda_d_inl0_hrd0/runTest_cuda.exe INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: FE_DIVBYZERO, FE_INVALID, FE_OVERFLOW @@ -100,15 +100,15 @@ Process = SIGMA_SMEFTSIM_TOPU3L_MWSCHEME_UFO_GG_TTXTTX_CPP [ Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/none+NAVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = SCALAR ('none': ~vector[1], no SIMD) -EvtsPerSec[Rmb+ME] (23) = ( 3.383469e+03 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.386752e+03 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.386752e+03 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 3.421940e+03 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.425187e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.425187e+03 ) sec^-1 MeanMatrixElemValue = ( 7.185537e-04 +- 6.562553e-04 ) GeV^-4 -TOTAL : 0.158264 sec +TOTAL : 0.156494 sec INFO: No Floating Point Exceptions have been reported - 469,342,334 cycles # 2.906 GHz - 1,390,298,076 instructions # 2.96 insn per cycle - 0.162106230 seconds time elapsed + 468,759,287 cycles # 2.937 GHz + 1,389,793,011 instructions # 2.96 insn per cycle + 0.160190784 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 3908) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/smeft_gg_tttt.mad/SubProcesses/P1_gg_ttxttx/build.none_d_inl0_hrd0/runTest_cpp.exe @@ -132,15 +132,15 @@ Process = SIGMA_SMEFTSIM_TOPU3L_MWSCHEME_UFO_GG_TTXTTX_CPP [ Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[2] ('sse4': SSE4.2, 128bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 6.476358e+03 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 6.488167e+03 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 6.488167e+03 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 6.307424e+03 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 6.318627e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 6.318627e+03 ) sec^-1 MeanMatrixElemValue = ( 7.185537e-04 +- 6.562553e-04 ) GeV^-4 -TOTAL : 0.083612 sec +TOTAL : 0.085600 sec INFO: No Floating Point Exceptions have been reported - 240,584,825 cycles # 2.769 GHz - 693,113,903 instructions # 2.88 insn per cycle - 0.087424946 seconds time elapsed + 241,344,896 cycles # 2.708 GHz + 693,048,135 instructions # 2.87 insn per cycle + 0.089754910 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 9482) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/smeft_gg_tttt.mad/SubProcesses/P1_gg_ttxttx/build.sse4_d_inl0_hrd0/runTest_cpp.exe @@ -164,15 +164,15 @@ Process = SIGMA_SMEFTSIM_TOPU3L_MWSCHEME_UFO_GG_TTXTTX_CPP [ Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('avx2': AVX2, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 1.432068e+04 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.438681e+04 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.438681e+04 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.406160e+04 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.412006e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.412006e+04 ) sec^-1 MeanMatrixElemValue = ( 7.185537e-04 +- 6.562553e-04 ) GeV^-4 -TOTAL : 0.038906 sec +TOTAL : 0.039600 sec INFO: No Floating Point Exceptions have been reported - 114,140,366 cycles # 2.711 GHz - 257,891,266 instructions # 2.26 insn per cycle - 0.042661267 seconds time elapsed + 114,730,167 cycles # 2.669 GHz + 258,023,133 instructions # 2.25 insn per cycle + 0.043596864 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 8501) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/smeft_gg_tttt.mad/SubProcesses/P1_gg_ttxttx/build.avx2_d_inl0_hrd0/runTest_cpp.exe @@ -196,15 +196,15 @@ Process = SIGMA_SMEFTSIM_TOPU3L_MWSCHEME_UFO_GG_TTXTTX_CPP [ Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('512y': AVX512, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 1.618386e+04 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.625883e+04 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.625883e+04 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.607323e+04 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.615807e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.615807e+04 ) sec^-1 MeanMatrixElemValue = ( 7.185537e-04 +- 6.562553e-04 ) GeV^-4 -TOTAL : 0.034671 sec +TOTAL : 0.034861 sec INFO: No Floating Point Exceptions have been reported - 102,555,024 cycles # 2.705 GHz - 240,017,026 instructions # 2.34 insn per cycle - 0.038425016 seconds time elapsed + 102,526,881 cycles # 2.692 GHz + 240,021,399 instructions # 2.34 insn per cycle + 0.038637040 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 8143) (512y: 150) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/smeft_gg_tttt.mad/SubProcesses/P1_gg_ttxttx/build.512y_d_inl0_hrd0/runTest_cpp.exe @@ -228,15 +228,15 @@ Process = SIGMA_SMEFTSIM_TOPU3L_MWSCHEME_UFO_GG_TTXTTX_CPP [ Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('512z': AVX512, 512bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 1.192893e+04 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.198052e+04 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.198052e+04 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.193500e+04 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.198547e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.198547e+04 ) sec^-1 MeanMatrixElemValue = ( 7.185537e-04 +- 6.562553e-04 ) GeV^-4 -TOTAL : 0.046494 sec +TOTAL : 0.046444 sec INFO: No Floating Point Exceptions have been reported - 90,048,800 cycles # 1.806 GHz - 134,302,710 instructions # 1.49 insn per cycle - 0.050438224 seconds time elapsed + 90,172,252 cycles # 1.810 GHz + 134,326,079 instructions # 1.49 insn per cycle + 0.050348620 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 1943) (512y: 126) (512z: 7086) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/smeft_gg_tttt.mad/SubProcesses/P1_gg_ttxttx/build.512z_d_inl0_hrd0/runTest_cpp.exe diff --git a/epochX/cudacpp/tput/logs_smeftggtttt_mad/log_smeftggtttt_mad_d_inl0_hrd1.txt b/epochX/cudacpp/tput/logs_smeftggtttt_mad/log_smeftggtttt_mad_d_inl0_hrd1.txt index dc41fe503f..26a8ac7c0b 100644 --- a/epochX/cudacpp/tput/logs_smeftggtttt_mad/log_smeftggtttt_mad_d_inl0_hrd1.txt +++ b/epochX/cudacpp/tput/logs_smeftggtttt_mad/log_smeftggtttt_mad_d_inl0_hrd1.txt @@ -40,7 +40,7 @@ make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/smeft_gg_tttt.mad/SubProcesses/P1_gg_ttxttx' -DATE: 2024-10-06_10:00:18 +DATE: 2024-10-06_14:16:12 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: ========================================================================= @@ -49,15 +49,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_SMEFTSIM_TOPU3L_MWSCHEME_UFO_GG_TTXTTX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=1] Workflow summary = CUD:DBL+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 2.797107e+04 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 2.816023e+04 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 2.819423e+04 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 2.797717e+04 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 2.815890e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.819075e+04 ) sec^-1 MeanMatrixElemValue = ( 7.185537e-04 +- 6.562553e-04 ) GeV^-4 -TOTAL : 0.469966 sec +TOTAL : 0.473644 sec INFO: No Floating Point Exceptions have been reported - 2,001,057,465 cycles # 2.881 GHz - 2,930,552,926 instructions # 1.46 insn per cycle - 0.752195966 seconds time elapsed + 1,990,807,875 cycles # 2.855 GHz + 2,905,693,421 instructions # 1.46 insn per cycle + 0.756003360 seconds time elapsed runNcu /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/smeft_gg_tttt.mad/SubProcesses/P1_gg_ttxttx/build.cuda_d_inl0_hrd1/check_cuda.exe -p 1 256 1 ==PROF== Profiling "sigmaKin": launch__registers_per_thread 255 ==PROF== Profiling "sigmaKin": sm__sass_average_branch_targets_threads_uniform.pct 100% @@ -67,15 +67,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_SMEFTSIM_TOPU3L_MWSCHEME_UFO_GG_TTXTTX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=1] Workflow summary = CUD:DBL+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 7.121137e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 7.233030e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 7.241027e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 7.088721e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 7.200972e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 7.209783e+05 ) sec^-1 MeanMatrixElemValue = ( 8.048215e-03 +- 4.042405e-03 ) GeV^-4 -TOTAL : 0.489610 sec +TOTAL : 0.491650 sec INFO: No Floating Point Exceptions have been reported - 2,050,200,483 cycles # 2.873 GHz - 3,056,241,818 instructions # 1.49 insn per cycle - 0.771808178 seconds time elapsed + 2,050,276,459 cycles # 2.860 GHz + 3,029,162,988 instructions # 1.48 insn per cycle + 0.775119967 seconds time elapsed ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/smeft_gg_tttt.mad/SubProcesses/P1_gg_ttxttx/build.cuda_d_inl0_hrd1/runTest_cuda.exe INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: FE_DIVBYZERO, FE_INVALID, FE_OVERFLOW @@ -100,15 +100,15 @@ Process = SIGMA_SMEFTSIM_TOPU3L_MWSCHEME_UFO_GG_TTXTTX_CPP [ Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/none+NAVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = SCALAR ('none': ~vector[1], no SIMD) -EvtsPerSec[Rmb+ME] (23) = ( 3.406266e+03 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.409565e+03 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.409565e+03 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 3.413489e+03 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.416788e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.416788e+03 ) sec^-1 MeanMatrixElemValue = ( 7.185537e-04 +- 6.562553e-04 ) GeV^-4 -TOTAL : 0.156466 sec +TOTAL : 0.156282 sec INFO: No Floating Point Exceptions have been reported - 465,689,745 cycles # 2.917 GHz - 1,385,079,930 instructions # 2.97 insn per cycle - 0.160315659 seconds time elapsed + 466,053,605 cycles # 2.922 GHz + 1,385,165,667 instructions # 2.97 insn per cycle + 0.159974886 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 3796) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/smeft_gg_tttt.mad/SubProcesses/P1_gg_ttxttx/build.none_d_inl0_hrd1/runTest_cpp.exe @@ -132,15 +132,15 @@ Process = SIGMA_SMEFTSIM_TOPU3L_MWSCHEME_UFO_GG_TTXTTX_CPP [ Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[2] ('sse4': SSE4.2, 128bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 6.388983e+03 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 6.401822e+03 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 6.401822e+03 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 6.486310e+03 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 6.498291e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 6.498291e+03 ) sec^-1 MeanMatrixElemValue = ( 7.185537e-04 +- 6.562553e-04 ) GeV^-4 -TOTAL : 0.083799 sec +TOTAL : 0.082577 sec INFO: No Floating Point Exceptions have been reported - 238,961,924 cycles # 2.745 GHz - 689,073,758 instructions # 2.88 insn per cycle - 0.087593094 seconds time elapsed + 239,403,854 cycles # 2.783 GHz + 689,070,859 instructions # 2.88 insn per cycle + 0.086650213 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 9525) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/smeft_gg_tttt.mad/SubProcesses/P1_gg_ttxttx/build.sse4_d_inl0_hrd1/runTest_cpp.exe @@ -164,15 +164,15 @@ Process = SIGMA_SMEFTSIM_TOPU3L_MWSCHEME_UFO_GG_TTXTTX_CPP [ Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('avx2': AVX2, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 1.419818e+04 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.425419e+04 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.425419e+04 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.442708e+04 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.448468e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.448468e+04 ) sec^-1 MeanMatrixElemValue = ( 7.185537e-04 +- 6.562553e-04 ) GeV^-4 -TOTAL : 0.038479 sec +TOTAL : 0.037887 sec INFO: No Floating Point Exceptions have been reported - 111,800,811 cycles # 2.682 GHz - 253,484,287 instructions # 2.27 insn per cycle - 0.042138594 seconds time elapsed + 111,507,695 cycles # 2.714 GHz + 253,516,531 instructions # 2.27 insn per cycle + 0.041654426 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 8457) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/smeft_gg_tttt.mad/SubProcesses/P1_gg_ttxttx/build.avx2_d_inl0_hrd1/runTest_cpp.exe @@ -196,15 +196,15 @@ Process = SIGMA_SMEFTSIM_TOPU3L_MWSCHEME_UFO_GG_TTXTTX_CPP [ Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('512y': AVX512, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 1.620452e+04 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.628839e+04 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.628839e+04 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.610850e+04 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.618243e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.618243e+04 ) sec^-1 MeanMatrixElemValue = ( 7.185537e-04 +- 6.562553e-04 ) GeV^-4 -TOTAL : 0.033872 sec +TOTAL : 0.034086 sec INFO: No Floating Point Exceptions have been reported - 100,998,379 cycles # 2.706 GHz - 235,641,730 instructions # 2.33 insn per cycle - 0.037957581 seconds time elapsed + 100,481,810 cycles # 2.694 GHz + 235,711,337 instructions # 2.35 insn per cycle + 0.037878094 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 8101) (512y: 150) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/smeft_gg_tttt.mad/SubProcesses/P1_gg_ttxttx/build.512y_d_inl0_hrd1/runTest_cpp.exe @@ -228,15 +228,15 @@ Process = SIGMA_SMEFTSIM_TOPU3L_MWSCHEME_UFO_GG_TTXTTX_CPP [ Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('512z': AVX512, 512bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 1.156678e+04 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.161477e+04 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.161477e+04 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.195034e+04 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.200065e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.200065e+04 ) sec^-1 MeanMatrixElemValue = ( 7.185537e-04 +- 6.562553e-04 ) GeV^-4 -TOTAL : 0.047111 sec +TOTAL : 0.045722 sec INFO: No Floating Point Exceptions have been reported - 88,066,978 cycles # 1.743 GHz - 129,735,533 instructions # 1.47 insn per cycle - 0.051105123 seconds time elapsed + 88,668,580 cycles # 1.801 GHz + 129,733,084 instructions # 1.46 insn per cycle + 0.049851936 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 1899) (512y: 126) (512z: 7084) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/smeft_gg_tttt.mad/SubProcesses/P1_gg_ttxttx/build.512z_d_inl0_hrd1/runTest_cpp.exe diff --git a/epochX/cudacpp/tput/logs_smeftggtttt_mad/log_smeftggtttt_mad_f_inl0_hrd0.txt b/epochX/cudacpp/tput/logs_smeftggtttt_mad/log_smeftggtttt_mad_f_inl0_hrd0.txt index 4b10dcf1d1..20186cca44 100644 --- a/epochX/cudacpp/tput/logs_smeftggtttt_mad/log_smeftggtttt_mad_f_inl0_hrd0.txt +++ b/epochX/cudacpp/tput/logs_smeftggtttt_mad/log_smeftggtttt_mad_f_inl0_hrd0.txt @@ -40,7 +40,7 @@ make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/smeft_gg_tttt.mad/SubProcesses/P1_gg_ttxttx' -DATE: 2024-10-06_10:00:52 +DATE: 2024-10-06_14:16:45 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: ========================================================================= @@ -49,15 +49,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_SMEFTSIM_TOPU3L_MWSCHEME_UFO_GG_TTXTTX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 2.214342e+04 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 2.224285e+04 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 2.226222e+04 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 2.219281e+04 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 2.228333e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.230354e+04 ) sec^-1 MeanMatrixElemValue = ( 7.188141e-04 +- 6.565202e-04 ) GeV^-4 -TOTAL : 0.476842 sec +TOTAL : 0.476574 sec INFO: No Floating Point Exceptions have been reported - 1,989,613,876 cycles # 2.873 GHz - 2,928,089,356 instructions # 1.47 insn per cycle - 0.750924959 seconds time elapsed + 1,987,229,611 cycles # 2.865 GHz + 2,865,120,608 instructions # 1.44 insn per cycle + 0.752273753 seconds time elapsed runNcu /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/smeft_gg_tttt.mad/SubProcesses/P1_gg_ttxttx/build.cuda_f_inl0_hrd0/check_cuda.exe -p 1 256 1 ==PROF== Profiling "sigmaKin": launch__registers_per_thread 255 ==PROF== Profiling "sigmaKin": sm__sass_average_branch_targets_threads_uniform.pct 100% @@ -67,15 +67,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_SMEFTSIM_TOPU3L_MWSCHEME_UFO_GG_TTXTTX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 7.950242e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 8.029144e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 8.036217e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 7.943809e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 8.019801e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 8.027422e+05 ) sec^-1 MeanMatrixElemValue = ( 8.020494e-03 +- 4.025605e-03 ) GeV^-4 -TOTAL : 0.473909 sec +TOTAL : 0.480123 sec INFO: No Floating Point Exceptions have been reported - 1,995,145,721 cycles # 2.886 GHz - 2,912,342,089 instructions # 1.46 insn per cycle - 0.748274226 seconds time elapsed + 1,975,257,418 cycles # 2.835 GHz + 2,890,245,285 instructions # 1.46 insn per cycle + 0.755777609 seconds time elapsed ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/smeft_gg_tttt.mad/SubProcesses/P1_gg_ttxttx/build.cuda_f_inl0_hrd0/runTest_cuda.exe INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: FE_DIVBYZERO, FE_INVALID, FE_OVERFLOW @@ -100,15 +100,15 @@ Process = SIGMA_SMEFTSIM_TOPU3L_MWSCHEME_UFO_GG_TTXTTX_CPP [ Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/none+NAVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = SCALAR ('none': ~vector[1], no SIMD) -EvtsPerSec[Rmb+ME] (23) = ( 3.462777e+03 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.466245e+03 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.466245e+03 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 3.370251e+03 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.373527e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.373527e+03 ) sec^-1 MeanMatrixElemValue = ( 7.177153e-04 +- 6.554185e-04 ) GeV^-4 -TOTAL : 0.154509 sec +TOTAL : 0.158665 sec INFO: No Floating Point Exceptions have been reported - 463,950,135 cycles # 2.942 GHz - 1,382,102,782 instructions # 2.98 insn per cycle - 0.158280886 seconds time elapsed + 464,695,039 cycles # 2.867 GHz + 1,382,030,576 instructions # 2.97 insn per cycle + 0.162645518 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 3058) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/smeft_gg_tttt.mad/SubProcesses/P1_gg_ttxttx/build.none_f_inl0_hrd0/runTest_cpp.exe @@ -132,15 +132,15 @@ Process = SIGMA_SMEFTSIM_TOPU3L_MWSCHEME_UFO_GG_TTXTTX_CPP [ Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('sse4': SSE4.2, 128bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 1.221716e+04 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.226773e+04 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.226773e+04 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.190562e+04 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.194783e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.194783e+04 ) sec^-1 MeanMatrixElemValue = ( 7.177152e-04 +- 6.554185e-04 ) GeV^-4 -TOTAL : 0.045148 sec +TOTAL : 0.046246 sec INFO: No Floating Point Exceptions have been reported - 132,927,826 cycles # 2.743 GHz - 372,156,154 instructions # 2.80 insn per cycle - 0.049041087 seconds time elapsed + 133,098,115 cycles # 2.681 GHz + 372,094,339 instructions # 2.80 insn per cycle + 0.050241047 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4:10141) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/smeft_gg_tttt.mad/SubProcesses/P1_gg_ttxttx/build.sse4_f_inl0_hrd0/runTest_cpp.exe @@ -164,15 +164,15 @@ Process = SIGMA_SMEFTSIM_TOPU3L_MWSCHEME_UFO_GG_TTXTTX_CPP [ Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('avx2': AVX2, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 2.776220e+04 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 2.801025e+04 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 2.801025e+04 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 2.741192e+04 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 2.766585e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.766585e+04 ) sec^-1 MeanMatrixElemValue = ( 7.165746e-04 +- 6.542823e-04 ) GeV^-4 -TOTAL : 0.021005 sec +TOTAL : 0.021282 sec INFO: No Floating Point Exceptions have been reported - 65,153,242 cycles # 2.690 GHz - 142,838,093 instructions # 2.19 insn per cycle - 0.024771930 seconds time elapsed + 65,559,059 cycles # 2.662 GHz + 142,834,726 instructions # 2.18 insn per cycle + 0.025223609 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 9241) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/smeft_gg_tttt.mad/SubProcesses/P1_gg_ttxttx/build.avx2_f_inl0_hrd0/runTest_cpp.exe @@ -196,15 +196,15 @@ Process = SIGMA_SMEFTSIM_TOPU3L_MWSCHEME_UFO_GG_TTXTTX_CPP [ Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('512y': AVX512, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 3.070417e+04 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.098717e+04 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.098717e+04 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 3.015767e+04 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.043123e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.043123e+04 ) sec^-1 MeanMatrixElemValue = ( 7.165746e-04 +- 6.542823e-04 ) GeV^-4 -TOTAL : 0.019184 sec +TOTAL : 0.019536 sec INFO: No Floating Point Exceptions have been reported - 60,296,621 cycles # 2.678 GHz - 132,772,434 instructions # 2.20 insn per cycle - 0.023065155 seconds time elapsed + 60,482,290 cycles # 2.643 GHz + 132,828,591 instructions # 2.20 insn per cycle + 0.023450810 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 8959) (512y: 28) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/smeft_gg_tttt.mad/SubProcesses/P1_gg_ttxttx/build.512y_f_inl0_hrd0/runTest_cpp.exe @@ -228,15 +228,15 @@ Process = SIGMA_SMEFTSIM_TOPU3L_MWSCHEME_UFO_GG_TTXTTX_CPP [ Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[16] ('512z': AVX512, 512bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 2.324469e+04 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 2.345673e+04 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 2.345673e+04 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 2.283679e+04 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 2.303294e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.303294e+04 ) sec^-1 MeanMatrixElemValue = ( 7.165747e-04 +- 6.542824e-04 ) GeV^-4 -TOTAL : 0.024875 sec +TOTAL : 0.025293 sec INFO: No Floating Point Exceptions have been reported - 52,411,208 cycles # 1.857 GHz - 79,637,147 instructions # 1.52 insn per cycle - 0.028776798 seconds time elapsed + 52,767,726 cycles # 1.836 GHz + 79,568,985 instructions # 1.51 insn per cycle + 0.029353704 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 2836) (512y: 30) (512z: 7437) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/smeft_gg_tttt.mad/SubProcesses/P1_gg_ttxttx/build.512z_f_inl0_hrd0/runTest_cpp.exe diff --git a/epochX/cudacpp/tput/logs_smeftggtttt_mad/log_smeftggtttt_mad_f_inl0_hrd1.txt b/epochX/cudacpp/tput/logs_smeftggtttt_mad/log_smeftggtttt_mad_f_inl0_hrd1.txt index 67a7328c67..dd11c14c66 100644 --- a/epochX/cudacpp/tput/logs_smeftggtttt_mad/log_smeftggtttt_mad_f_inl0_hrd1.txt +++ b/epochX/cudacpp/tput/logs_smeftggtttt_mad/log_smeftggtttt_mad_f_inl0_hrd1.txt @@ -40,7 +40,7 @@ make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/smeft_gg_tttt.mad/SubProcesses/P1_gg_ttxttx' -DATE: 2024-10-06_10:01:02 +DATE: 2024-10-06_14:16:56 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: ========================================================================= @@ -49,15 +49,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_SMEFTSIM_TOPU3L_MWSCHEME_UFO_GG_TTXTTX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=1] Workflow summary = CUD:FLT+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 2.235104e+04 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 2.244507e+04 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 2.246621e+04 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 2.229962e+04 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 2.239566e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.241669e+04 ) sec^-1 MeanMatrixElemValue = ( 7.188141e-04 +- 6.565202e-04 ) GeV^-4 -TOTAL : 0.477845 sec +TOTAL : 0.479037 sec INFO: No Floating Point Exceptions have been reported - 1,997,911,903 cycles # 2.876 GHz - 2,886,764,809 instructions # 1.44 insn per cycle - 0.753229194 seconds time elapsed + 1,972,941,176 cycles # 2.828 GHz + 2,877,624,846 instructions # 1.46 insn per cycle + 0.755169595 seconds time elapsed runNcu /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/smeft_gg_tttt.mad/SubProcesses/P1_gg_ttxttx/build.cuda_f_inl0_hrd1/check_cuda.exe -p 1 256 1 ==PROF== Profiling "sigmaKin": launch__registers_per_thread 255 ==PROF== Profiling "sigmaKin": sm__sass_average_branch_targets_threads_uniform.pct 100% @@ -67,15 +67,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_SMEFTSIM_TOPU3L_MWSCHEME_UFO_GG_TTXTTX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=1] Workflow summary = CUD:FLT+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 8.096496e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 8.193422e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 8.206590e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 8.023200e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 8.106275e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 8.114385e+05 ) sec^-1 MeanMatrixElemValue = ( 8.020496e-03 +- 4.025606e-03 ) GeV^-4 -TOTAL : 0.477844 sec +TOTAL : 0.480018 sec INFO: No Floating Point Exceptions have been reported - 2,000,227,335 cycles # 2.879 GHz - 2,887,661,973 instructions # 1.44 insn per cycle - 0.753759254 seconds time elapsed + 1,981,487,889 cycles # 2.840 GHz + 2,901,456,143 instructions # 1.46 insn per cycle + 0.756475922 seconds time elapsed ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/smeft_gg_tttt.mad/SubProcesses/P1_gg_ttxttx/build.cuda_f_inl0_hrd1/runTest_cuda.exe INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: FE_DIVBYZERO, FE_INVALID, FE_OVERFLOW @@ -100,15 +100,15 @@ Process = SIGMA_SMEFTSIM_TOPU3L_MWSCHEME_UFO_GG_TTXTTX_CPP [ Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/none+NAVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = SCALAR ('none': ~vector[1], no SIMD) -EvtsPerSec[Rmb+ME] (23) = ( 3.435869e+03 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.439325e+03 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.439325e+03 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 3.382050e+03 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.385282e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.385282e+03 ) sec^-1 MeanMatrixElemValue = ( 7.177153e-04 +- 6.554185e-04 ) GeV^-4 -TOTAL : 0.154994 sec +TOTAL : 0.157401 sec INFO: No Floating Point Exceptions have been reported - 461,652,768 cycles # 2.918 GHz - 1,376,807,565 instructions # 2.98 insn per cycle - 0.158786297 seconds time elapsed + 462,383,689 cycles # 2.875 GHz + 1,376,799,523 instructions # 2.98 insn per cycle + 0.161485208 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 2930) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/smeft_gg_tttt.mad/SubProcesses/P1_gg_ttxttx/build.none_f_inl0_hrd1/runTest_cpp.exe @@ -132,15 +132,15 @@ Process = SIGMA_SMEFTSIM_TOPU3L_MWSCHEME_UFO_GG_TTXTTX_CPP [ Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('sse4': SSE4.2, 128bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 1.215601e+04 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.220158e+04 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.220158e+04 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.196381e+04 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.200694e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.200694e+04 ) sec^-1 MeanMatrixElemValue = ( 7.177152e-04 +- 6.554185e-04 ) GeV^-4 -TOTAL : 0.044587 sec +TOTAL : 0.046098 sec INFO: No Floating Point Exceptions have been reported - 130,364,411 cycles # 2.725 GHz - 367,274,419 instructions # 2.82 insn per cycle - 0.048380365 seconds time elapsed + 131,749,091 cycles # 2.656 GHz + 367,321,639 instructions # 2.79 insn per cycle + 0.050262320 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4:10124) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/smeft_gg_tttt.mad/SubProcesses/P1_gg_ttxttx/build.sse4_f_inl0_hrd1/runTest_cpp.exe @@ -164,15 +164,15 @@ Process = SIGMA_SMEFTSIM_TOPU3L_MWSCHEME_UFO_GG_TTXTTX_CPP [ Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('avx2': AVX2, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 2.799777e+04 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 2.825160e+04 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 2.825160e+04 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 2.726552e+04 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 2.749509e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.749509e+04 ) sec^-1 MeanMatrixElemValue = ( 7.165746e-04 +- 6.542823e-04 ) GeV^-4 -TOTAL : 0.020201 sec +TOTAL : 0.020656 sec INFO: No Floating Point Exceptions have been reported - 63,211,215 cycles # 2.704 GHz - 138,063,768 instructions # 2.18 insn per cycle - 0.023985955 seconds time elapsed + 64,169,817 cycles # 2.653 GHz + 138,019,393 instructions # 2.15 insn per cycle + 0.024831516 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 9196) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/smeft_gg_tttt.mad/SubProcesses/P1_gg_ttxttx/build.avx2_f_inl0_hrd1/runTest_cpp.exe @@ -196,15 +196,15 @@ Process = SIGMA_SMEFTSIM_TOPU3L_MWSCHEME_UFO_GG_TTXTTX_CPP [ Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('512y': AVX512, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 3.035669e+04 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.062918e+04 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.062918e+04 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 3.016981e+04 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.044462e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.044462e+04 ) sec^-1 MeanMatrixElemValue = ( 7.165746e-04 +- 6.542823e-04 ) GeV^-4 -TOTAL : 0.018625 sec +TOTAL : 0.018754 sec INFO: No Floating Point Exceptions have been reported - 57,993,332 cycles # 2.658 GHz - 127,990,808 instructions # 2.21 insn per cycle - 0.022353301 seconds time elapsed + 58,730,786 cycles # 2.651 GHz + 127,995,867 instructions # 2.18 insn per cycle + 0.022749447 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 8910) (512y: 28) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/smeft_gg_tttt.mad/SubProcesses/P1_gg_ttxttx/build.512y_f_inl0_hrd1/runTest_cpp.exe @@ -228,15 +228,15 @@ Process = SIGMA_SMEFTSIM_TOPU3L_MWSCHEME_UFO_GG_TTXTTX_CPP [ Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[16] ('512z': AVX512, 512bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 2.344103e+04 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 2.363443e+04 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 2.363443e+04 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 2.294226e+04 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 2.314507e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.314507e+04 ) sec^-1 MeanMatrixElemValue = ( 7.165747e-04 +- 6.542824e-04 ) GeV^-4 -TOTAL : 0.024010 sec +TOTAL : 0.024468 sec INFO: No Floating Point Exceptions have been reported - 50,268,269 cycles # 1.840 GHz - 74,785,740 instructions # 1.49 insn per cycle - 0.027917015 seconds time elapsed + 50,463,446 cycles # 1.809 GHz + 74,790,769 instructions # 1.48 insn per cycle + 0.028490560 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 2791) (512y: 30) (512z: 7439) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/smeft_gg_tttt.mad/SubProcesses/P1_gg_ttxttx/build.512z_f_inl0_hrd1/runTest_cpp.exe diff --git a/epochX/cudacpp/tput/logs_smeftggtttt_mad/log_smeftggtttt_mad_m_inl0_hrd0.txt b/epochX/cudacpp/tput/logs_smeftggtttt_mad/log_smeftggtttt_mad_m_inl0_hrd0.txt index 50cf2d796e..91e4dc2f69 100644 --- a/epochX/cudacpp/tput/logs_smeftggtttt_mad/log_smeftggtttt_mad_m_inl0_hrd0.txt +++ b/epochX/cudacpp/tput/logs_smeftggtttt_mad/log_smeftggtttt_mad_m_inl0_hrd0.txt @@ -40,7 +40,7 @@ make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/smeft_gg_tttt.mad/SubProcesses/P1_gg_ttxttx' -DATE: 2024-10-06_10:00:29 +DATE: 2024-10-06_14:16:23 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: ========================================================================= @@ -49,15 +49,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_SMEFTSIM_TOPU3L_MWSCHEME_UFO_GG_TTXTTX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:MIX+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 2.754018e+04 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 2.771557e+04 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 2.774637e+04 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 2.753026e+04 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 2.771501e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.774577e+04 ) sec^-1 MeanMatrixElemValue = ( 7.185537e-04 +- 6.562553e-04 ) GeV^-4 -TOTAL : 0.469482 sec +TOTAL : 0.471544 sec INFO: No Floating Point Exceptions have been reported - 1,992,256,665 cycles # 2.872 GHz - 2,888,484,617 instructions # 1.45 insn per cycle - 0.750839241 seconds time elapsed + 1,996,960,982 cycles # 2.874 GHz + 2,937,743,800 instructions # 1.47 insn per cycle + 0.751265732 seconds time elapsed runNcu /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/smeft_gg_tttt.mad/SubProcesses/P1_gg_ttxttx/build.cuda_m_inl0_hrd0/check_cuda.exe -p 1 256 1 ==PROF== Profiling "sigmaKin": launch__registers_per_thread 255 ==PROF== Profiling "sigmaKin": sm__sass_average_branch_targets_threads_uniform.pct 100% @@ -67,15 +67,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_SMEFTSIM_TOPU3L_MWSCHEME_UFO_GG_TTXTTX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:MIX+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 6.962737e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 7.089994e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 7.098896e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 6.958904e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 7.070330e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 7.078356e+05 ) sec^-1 MeanMatrixElemValue = ( 8.048215e-03 +- 4.042405e-03 ) GeV^-4 -TOTAL : 0.485052 sec +TOTAL : 0.488223 sec INFO: No Floating Point Exceptions have been reported - 2,027,704,407 cycles # 2.871 GHz - 3,029,735,278 instructions # 1.49 insn per cycle - 0.765353713 seconds time elapsed + 2,032,139,788 cycles # 2.868 GHz + 3,011,355,323 instructions # 1.48 insn per cycle + 0.769138182 seconds time elapsed ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/smeft_gg_tttt.mad/SubProcesses/P1_gg_ttxttx/build.cuda_m_inl0_hrd0/runTest_cuda.exe INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: FE_DIVBYZERO, FE_INVALID, FE_OVERFLOW @@ -100,15 +100,15 @@ Process = SIGMA_SMEFTSIM_TOPU3L_MWSCHEME_UFO_GG_TTXTTX_CPP [ Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/none+NAVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) Internal loops fptype_sv = SCALAR ('none': ~vector[1], no SIMD) -EvtsPerSec[Rmb+ME] (23) = ( 3.401289e+03 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.404577e+03 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.404577e+03 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 3.394834e+03 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.398121e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.398121e+03 ) sec^-1 MeanMatrixElemValue = ( 7.185537e-04 +- 6.562553e-04 ) GeV^-4 -TOTAL : 0.157429 sec +TOTAL : 0.157679 sec INFO: No Floating Point Exceptions have been reported - 471,621,611 cycles # 2.936 GHz - 1,398,387,891 instructions # 2.97 insn per cycle - 0.161191989 seconds time elapsed + 471,655,191 cycles # 2.930 GHz + 1,398,435,971 instructions # 2.96 insn per cycle + 0.161525548 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 3899) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/smeft_gg_tttt.mad/SubProcesses/P1_gg_ttxttx/build.none_m_inl0_hrd0/runTest_cpp.exe @@ -132,15 +132,15 @@ Process = SIGMA_SMEFTSIM_TOPU3L_MWSCHEME_UFO_GG_TTXTTX_CPP [ Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[2] ('sse4': SSE4.2, 128bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 6.729709e+03 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 6.743939e+03 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 6.743939e+03 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 6.698724e+03 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 6.711279e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 6.711279e+03 ) sec^-1 MeanMatrixElemValue = ( 7.185537e-04 +- 6.562553e-04 ) GeV^-4 -TOTAL : 0.080446 sec +TOTAL : 0.080981 sec INFO: No Floating Point Exceptions have been reported - 237,178,815 cycles # 2.833 GHz - 688,220,781 instructions # 2.90 insn per cycle - 0.084309693 seconds time elapsed + 237,545,954 cycles # 2.817 GHz + 688,175,235 instructions # 2.90 insn per cycle + 0.084837532 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 9334) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/smeft_gg_tttt.mad/SubProcesses/P1_gg_ttxttx/build.sse4_m_inl0_hrd0/runTest_cpp.exe @@ -164,15 +164,15 @@ Process = SIGMA_SMEFTSIM_TOPU3L_MWSCHEME_UFO_GG_TTXTTX_CPP [ Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('avx2': AVX2, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 1.409119e+04 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.415451e+04 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.415451e+04 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.429367e+04 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.435354e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.435354e+04 ) sec^-1 MeanMatrixElemValue = ( 7.185537e-04 +- 6.562553e-04 ) GeV^-4 -TOTAL : 0.039507 sec +TOTAL : 0.039048 sec INFO: No Floating Point Exceptions have been reported - 114,068,471 cycles # 2.665 GHz - 253,096,543 instructions # 2.22 insn per cycle - 0.043335126 seconds time elapsed + 113,835,727 cycles # 2.687 GHz + 253,188,651 instructions # 2.22 insn per cycle + 0.042923124 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 8363) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/smeft_gg_tttt.mad/SubProcesses/P1_gg_ttxttx/build.avx2_m_inl0_hrd0/runTest_cpp.exe @@ -196,15 +196,15 @@ Process = SIGMA_SMEFTSIM_TOPU3L_MWSCHEME_UFO_GG_TTXTTX_CPP [ Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('512y': AVX512, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 1.680681e+04 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.688641e+04 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.688641e+04 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.619526e+04 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.627466e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.627466e+04 ) sec^-1 MeanMatrixElemValue = ( 7.185537e-04 +- 6.562553e-04 ) GeV^-4 -TOTAL : 0.033493 sec +TOTAL : 0.034648 sec INFO: No Floating Point Exceptions have been reported - 101,334,967 cycles # 2.753 GHz - 233,610,113 instructions # 2.31 insn per cycle - 0.037380618 seconds time elapsed + 101,297,175 cycles # 2.672 GHz + 233,656,753 instructions # 2.31 insn per cycle + 0.038451828 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 7501) (512y: 146) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/smeft_gg_tttt.mad/SubProcesses/P1_gg_ttxttx/build.512y_m_inl0_hrd0/runTest_cpp.exe @@ -228,15 +228,15 @@ Process = SIGMA_SMEFTSIM_TOPU3L_MWSCHEME_UFO_GG_TTXTTX_CPP [ Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('512z': AVX512, 512bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 1.194656e+04 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.199944e+04 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.199944e+04 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.149362e+04 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.154045e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.154045e+04 ) sec^-1 MeanMatrixElemValue = ( 7.185537e-04 +- 6.562553e-04 ) GeV^-4 -TOTAL : 0.046435 sec +TOTAL : 0.048182 sec INFO: No Floating Point Exceptions have been reported - 91,210,419 cycles # 1.827 GHz - 133,172,431 instructions # 1.46 insn per cycle - 0.050429905 seconds time elapsed + 91,585,533 cycles # 1.771 GHz + 133,173,816 instructions # 1.45 insn per cycle + 0.052356525 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 2084) (512y: 122) (512z: 6354) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/smeft_gg_tttt.mad/SubProcesses/P1_gg_ttxttx/build.512z_m_inl0_hrd0/runTest_cpp.exe diff --git a/epochX/cudacpp/tput/logs_smeftggtttt_mad/log_smeftggtttt_mad_m_inl0_hrd1.txt b/epochX/cudacpp/tput/logs_smeftggtttt_mad/log_smeftggtttt_mad_m_inl0_hrd1.txt index e1fc789bed..ebb5cbbdf3 100644 --- a/epochX/cudacpp/tput/logs_smeftggtttt_mad/log_smeftggtttt_mad_m_inl0_hrd1.txt +++ b/epochX/cudacpp/tput/logs_smeftggtttt_mad/log_smeftggtttt_mad_m_inl0_hrd1.txt @@ -40,7 +40,7 @@ make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/smeft_gg_tttt.mad/SubProcesses/P1_gg_ttxttx' -DATE: 2024-10-06_10:00:41 +DATE: 2024-10-06_14:16:34 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: ========================================================================= @@ -49,15 +49,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_SMEFTSIM_TOPU3L_MWSCHEME_UFO_GG_TTXTTX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=1] Workflow summary = CUD:MIX+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 2.793622e+04 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 2.811451e+04 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 2.814397e+04 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 2.773277e+04 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 2.791162e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.794087e+04 ) sec^-1 MeanMatrixElemValue = ( 7.185537e-04 +- 6.562553e-04 ) GeV^-4 -TOTAL : 0.470592 sec +TOTAL : 0.472050 sec INFO: No Floating Point Exceptions have been reported - 1,997,502,547 cycles # 2.880 GHz - 2,923,476,215 instructions # 1.46 insn per cycle - 0.750818094 seconds time elapsed + 1,995,616,436 cycles # 2.869 GHz + 2,893,983,428 instructions # 1.45 insn per cycle + 0.752820451 seconds time elapsed runNcu /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/smeft_gg_tttt.mad/SubProcesses/P1_gg_ttxttx/build.cuda_m_inl0_hrd1/check_cuda.exe -p 1 256 1 ==PROF== Profiling "sigmaKin": launch__registers_per_thread 255 ==PROF== Profiling "sigmaKin": sm__sass_average_branch_targets_threads_uniform.pct 100% @@ -67,15 +67,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_SMEFTSIM_TOPU3L_MWSCHEME_UFO_GG_TTXTTX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=1] Workflow summary = CUD:MIX+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 7.055830e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 7.165646e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 7.173712e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 7.072170e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 7.194698e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 7.203289e+05 ) sec^-1 MeanMatrixElemValue = ( 8.048215e-03 +- 4.042405e-03 ) GeV^-4 -TOTAL : 0.491474 sec +TOTAL : 0.486909 sec INFO: No Floating Point Exceptions have been reported - 2,044,918,526 cycles # 2.859 GHz - 3,006,189,896 instructions # 1.47 insn per cycle - 0.774360899 seconds time elapsed + 2,041,565,483 cycles # 2.873 GHz + 2,935,495,035 instructions # 1.44 insn per cycle + 0.769807281 seconds time elapsed ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/smeft_gg_tttt.mad/SubProcesses/P1_gg_ttxttx/build.cuda_m_inl0_hrd1/runTest_cuda.exe INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: FE_DIVBYZERO, FE_INVALID, FE_OVERFLOW @@ -100,15 +100,15 @@ Process = SIGMA_SMEFTSIM_TOPU3L_MWSCHEME_UFO_GG_TTXTTX_CPP [ Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/none+NAVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) Internal loops fptype_sv = SCALAR ('none': ~vector[1], no SIMD) -EvtsPerSec[Rmb+ME] (23) = ( 3.402707e+03 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.406541e+03 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.406541e+03 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 3.372631e+03 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.375838e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.375838e+03 ) sec^-1 MeanMatrixElemValue = ( 7.185537e-04 +- 6.562553e-04 ) GeV^-4 -TOTAL : 0.156609 sec +TOTAL : 0.157921 sec INFO: No Floating Point Exceptions have been reported - 468,766,259 cycles # 2.933 GHz - 1,393,706,102 instructions # 2.97 insn per cycle - 0.160398151 seconds time elapsed + 469,993,924 cycles # 2.916 GHz + 1,393,770,617 instructions # 2.97 insn per cycle + 0.161782917 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 3800) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/smeft_gg_tttt.mad/SubProcesses/P1_gg_ttxttx/build.none_m_inl0_hrd1/runTest_cpp.exe @@ -132,15 +132,15 @@ Process = SIGMA_SMEFTSIM_TOPU3L_MWSCHEME_UFO_GG_TTXTTX_CPP [ Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[2] ('sse4': SSE4.2, 128bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 6.728046e+03 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 6.740604e+03 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 6.740604e+03 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 6.721580e+03 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 6.735210e+03 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 6.735210e+03 ) sec^-1 MeanMatrixElemValue = ( 7.185537e-04 +- 6.562553e-04 ) GeV^-4 -TOTAL : 0.079699 sec +TOTAL : 0.079886 sec INFO: No Floating Point Exceptions have been reported - 235,148,851 cycles # 2.837 GHz - 684,201,633 instructions # 2.91 insn per cycle - 0.083458032 seconds time elapsed + 235,176,255 cycles # 2.830 GHz + 684,227,192 instructions # 2.91 insn per cycle + 0.083716444 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 9368) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/smeft_gg_tttt.mad/SubProcesses/P1_gg_ttxttx/build.sse4_m_inl0_hrd1/runTest_cpp.exe @@ -164,15 +164,15 @@ Process = SIGMA_SMEFTSIM_TOPU3L_MWSCHEME_UFO_GG_TTXTTX_CPP [ Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('avx2': AVX2, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 1.447554e+04 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.453499e+04 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.453499e+04 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.423416e+04 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.429146e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.429146e+04 ) sec^-1 MeanMatrixElemValue = ( 7.185537e-04 +- 6.562553e-04 ) GeV^-4 -TOTAL : 0.037781 sec +TOTAL : 0.038397 sec INFO: No Floating Point Exceptions have been reported - 111,660,471 cycles # 2.716 GHz - 248,651,696 instructions # 2.23 insn per cycle - 0.041691428 seconds time elapsed + 111,438,850 cycles # 2.677 GHz + 248,629,035 instructions # 2.23 insn per cycle + 0.042176688 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 8316) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/smeft_gg_tttt.mad/SubProcesses/P1_gg_ttxttx/build.avx2_m_inl0_hrd1/runTest_cpp.exe @@ -196,15 +196,15 @@ Process = SIGMA_SMEFTSIM_TOPU3L_MWSCHEME_UFO_GG_TTXTTX_CPP [ Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('512y': AVX512, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 1.634149e+04 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.641617e+04 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.641617e+04 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.666243e+04 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.674346e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.674346e+04 ) sec^-1 MeanMatrixElemValue = ( 7.185537e-04 +- 6.562553e-04 ) GeV^-4 -TOTAL : 0.033571 sec +TOTAL : 0.032964 sec INFO: No Floating Point Exceptions have been reported - 99,219,938 cycles # 2.697 GHz - 229,292,514 instructions # 2.31 insn per cycle - 0.037291206 seconds time elapsed + 99,571,120 cycles # 2.749 GHz + 229,181,934 instructions # 2.30 insn per cycle + 0.036770040 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 7452) (512y: 146) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/smeft_gg_tttt.mad/SubProcesses/P1_gg_ttxttx/build.512y_m_inl0_hrd1/runTest_cpp.exe @@ -228,15 +228,15 @@ Process = SIGMA_SMEFTSIM_TOPU3L_MWSCHEME_UFO_GG_TTXTTX_CPP [ Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('512z': AVX512, 512bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 1.191988e+04 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.196872e+04 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.196872e+04 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.197023e+04 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.201930e+04 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.201930e+04 ) sec^-1 MeanMatrixElemValue = ( 7.185537e-04 +- 6.562553e-04 ) GeV^-4 -TOTAL : 0.045809 sec +TOTAL : 0.045558 sec INFO: No Floating Point Exceptions have been reported - 88,834,257 cycles # 1.806 GHz - 128,615,199 instructions # 1.45 insn per cycle - 0.049747357 seconds time elapsed + 89,003,514 cycles # 1.819 GHz + 128,586,186 instructions # 1.44 insn per cycle + 0.049541890 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 2035) (512y: 122) (512z: 6355) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/smeft_gg_tttt.mad/SubProcesses/P1_gg_ttxttx/build.512z_m_inl0_hrd1/runTest_cpp.exe diff --git a/epochX/cudacpp/tput/logs_susyggt1t1_mad/log_susyggt1t1_mad_d_inl0_hrd0.txt b/epochX/cudacpp/tput/logs_susyggt1t1_mad/log_susyggt1t1_mad_d_inl0_hrd0.txt index 107a77153b..1d16bf4e31 100644 --- a/epochX/cudacpp/tput/logs_susyggt1t1_mad/log_susyggt1t1_mad_d_inl0_hrd0.txt +++ b/epochX/cudacpp/tput/logs_susyggt1t1_mad/log_susyggt1t1_mad_d_inl0_hrd0.txt @@ -40,7 +40,7 @@ make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_t1t1.mad/SubProcesses/P1_gg_t1t1x' -DATE: 2024-10-06_09:58:55 +DATE: 2024-10-06_14:14:49 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: ========================================================================= @@ -49,15 +49,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_MSSM_SLHA2_GG_T1T1X_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 5.910300e+07 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 2.325267e+08 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 2.783205e+08 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 6.665762e+07 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 2.427501e+08 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.869408e+08 ) sec^-1 MeanMatrixElemValue = ( 1.486736e-01 +- 3.293564e-05 ) GeV^0 -TOTAL : 0.523085 sec +TOTAL : 0.513979 sec INFO: No Floating Point Exceptions have been reported - 2,188,593,202 cycles # 2.883 GHz - 3,112,954,096 instructions # 1.42 insn per cycle - 0.817031478 seconds time elapsed + 2,152,531,070 cycles # 2.882 GHz + 3,097,168,287 instructions # 1.44 insn per cycle + 0.804495761 seconds time elapsed runNcu /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_t1t1.mad/SubProcesses/P1_gg_t1t1x/build.cuda_d_inl0_hrd0/check_cuda.exe -p 2048 256 1 ==PROF== Profiling "sigmaKin": launch__registers_per_thread 130 ==PROF== Profiling "sigmaKin": sm__sass_average_branch_targets_threads_uniform.pct 100% @@ -85,15 +85,15 @@ Process = SIGMA_MSSM_SLHA2_GG_T1T1X_CPP [gcc 11.3.1] [inline Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/none+NAVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = SCALAR ('none': ~vector[1], no SIMD) -EvtsPerSec[Rmb+ME] (23) = ( 9.066686e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.035589e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.035589e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 9.130375e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.042201e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.042201e+06 ) sec^-1 MeanMatrixElemValue = ( 1.486736e-01 +- 3.293564e-05 ) GeV^0 -TOTAL : 1.278019 sec +TOTAL : 1.260750 sec INFO: No Floating Point Exceptions have been reported - 3,764,987,469 cycles # 2.931 GHz - 9,752,169,319 instructions # 2.59 insn per cycle - 1.285199771 seconds time elapsed + 3,728,759,094 cycles # 2.947 GHz + 9,729,238,165 instructions # 2.61 insn per cycle + 1.266117079 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 341) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_t1t1.mad/SubProcesses/P1_gg_t1t1x/build.none_d_inl0_hrd0/runTest_cpp.exe @@ -117,15 +117,15 @@ Process = SIGMA_MSSM_SLHA2_GG_T1T1X_CPP [gcc 11.3.1] [inline Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[2] ('sse4': SSE4.2, 128bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 1.478889e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.890818e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.890818e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.479998e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.887572e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.887572e+06 ) sec^-1 MeanMatrixElemValue = ( 1.486736e-01 +- 3.293564e-05 ) GeV^0 -TOTAL : 0.830635 sec +TOTAL : 0.821705 sec INFO: No Floating Point Exceptions have been reported - 2,356,582,684 cycles # 2.814 GHz - 5,959,230,788 instructions # 2.53 insn per cycle - 0.838030934 seconds time elapsed + 2,323,677,760 cycles # 2.812 GHz + 5,934,281,966 instructions # 2.55 insn per cycle + 0.827074513 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 1369) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_t1t1.mad/SubProcesses/P1_gg_t1t1x/build.sse4_d_inl0_hrd0/runTest_cpp.exe @@ -149,15 +149,15 @@ Process = SIGMA_MSSM_SLHA2_GG_T1T1X_CPP [gcc 11.3.1] [inline Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('avx2': AVX2, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 2.229956e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.271002e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.271002e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 2.186738e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.183853e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.183853e+06 ) sec^-1 MeanMatrixElemValue = ( 1.486736e-01 +- 3.293564e-05 ) GeV^0 -TOTAL : 0.594206 sec +TOTAL : 0.593252 sec INFO: No Floating Point Exceptions have been reported - 1,695,017,656 cycles # 2.820 GHz - 3,345,002,918 instructions # 1.97 insn per cycle - 0.601755215 seconds time elapsed + 1,671,043,752 cycles # 2.794 GHz + 3,316,680,081 instructions # 1.98 insn per cycle + 0.598664722 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 1499) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_t1t1.mad/SubProcesses/P1_gg_t1t1x/build.avx2_d_inl0_hrd0/runTest_cpp.exe @@ -181,15 +181,15 @@ Process = SIGMA_MSSM_SLHA2_GG_T1T1X_CPP [gcc 11.3.1] [inline Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('512y': AVX512, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 2.272289e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.349942e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.349942e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 2.302155e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.409887e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.409887e+06 ) sec^-1 MeanMatrixElemValue = ( 1.486736e-01 +- 3.293564e-05 ) GeV^0 -TOTAL : 0.586413 sec +TOTAL : 0.570148 sec INFO: No Floating Point Exceptions have been reported - 1,670,913,790 cycles # 2.815 GHz - 3,318,759,581 instructions # 1.99 insn per cycle - 0.594196558 seconds time elapsed + 1,624,753,303 cycles # 2.825 GHz + 3,285,579,460 instructions # 2.02 insn per cycle + 0.575835819 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 1375) (512y: 96) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_t1t1.mad/SubProcesses/P1_gg_t1t1x/build.512y_d_inl0_hrd0/runTest_cpp.exe @@ -213,15 +213,15 @@ Process = SIGMA_MSSM_SLHA2_GG_T1T1X_CPP [gcc 11.3.1] [inline Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('512z': AVX512, 512bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 2.146635e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.068698e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.068698e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 2.164859e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.106649e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.106649e+06 ) sec^-1 MeanMatrixElemValue = ( 1.486736e-01 +- 3.293564e-05 ) GeV^0 -TOTAL : 0.618076 sec +TOTAL : 0.599006 sec INFO: No Floating Point Exceptions have been reported - 1,426,424,228 cycles # 2.279 GHz - 2,470,718,173 instructions # 1.73 insn per cycle - 0.626622796 seconds time elapsed + 1,367,383,975 cycles # 2.263 GHz + 2,425,368,514 instructions # 1.77 insn per cycle + 0.604825010 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 580) (512y: 60) (512z: 1021) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_t1t1.mad/SubProcesses/P1_gg_t1t1x/build.512z_d_inl0_hrd0/runTest_cpp.exe diff --git a/epochX/cudacpp/tput/logs_susyggt1t1_mad/log_susyggt1t1_mad_d_inl0_hrd1.txt b/epochX/cudacpp/tput/logs_susyggt1t1_mad/log_susyggt1t1_mad_d_inl0_hrd1.txt index 00276091a3..6bd6aed927 100644 --- a/epochX/cudacpp/tput/logs_susyggt1t1_mad/log_susyggt1t1_mad_d_inl0_hrd1.txt +++ b/epochX/cudacpp/tput/logs_susyggt1t1_mad/log_susyggt1t1_mad_d_inl0_hrd1.txt @@ -40,7 +40,7 @@ make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_t1t1.mad/SubProcesses/P1_gg_t1t1x' -DATE: 2024-10-06_09:59:08 +DATE: 2024-10-06_14:15:01 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: ========================================================================= @@ -49,15 +49,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_MSSM_SLHA2_GG_T1T1X_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=1] Workflow summary = CUD:DBL+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 5.969963e+07 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 2.427733e+08 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 2.936447e+08 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 6.670616e+07 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 2.520242e+08 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.001211e+08 ) sec^-1 MeanMatrixElemValue = ( 1.486736e-01 +- 3.293564e-05 ) GeV^0 -TOTAL : 0.519852 sec +TOTAL : 0.520206 sec INFO: No Floating Point Exceptions have been reported - 2,172,307,830 cycles # 2.872 GHz - 3,081,950,905 instructions # 1.42 insn per cycle - 0.813507263 seconds time elapsed + 2,184,654,038 cycles # 2.889 GHz + 3,064,792,816 instructions # 1.40 insn per cycle + 0.813226837 seconds time elapsed runNcu /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_t1t1.mad/SubProcesses/P1_gg_t1t1x/build.cuda_d_inl0_hrd1/check_cuda.exe -p 2048 256 1 ==PROF== Profiling "sigmaKin": launch__registers_per_thread 124 ==PROF== Profiling "sigmaKin": sm__sass_average_branch_targets_threads_uniform.pct 100% @@ -85,15 +85,15 @@ Process = SIGMA_MSSM_SLHA2_GG_T1T1X_CPP [gcc 11.3.1] [inline Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/none+NAVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = SCALAR ('none': ~vector[1], no SIMD) -EvtsPerSec[Rmb+ME] (23) = ( 9.156288e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.045734e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.045734e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 9.158284e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.045552e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.045552e+06 ) sec^-1 MeanMatrixElemValue = ( 1.486736e-01 +- 3.293564e-05 ) GeV^0 -TOTAL : 1.265578 sec +TOTAL : 1.257018 sec INFO: No Floating Point Exceptions have been reported - 3,747,828,201 cycles # 2.946 GHz - 9,632,221,913 instructions # 2.57 insn per cycle - 1.272810702 seconds time elapsed + 3,714,778,405 cycles # 2.944 GHz + 9,612,055,230 instructions # 2.59 insn per cycle + 1.262787111 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 359) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_t1t1.mad/SubProcesses/P1_gg_t1t1x/build.none_d_inl0_hrd1/runTest_cpp.exe @@ -117,15 +117,15 @@ Process = SIGMA_MSSM_SLHA2_GG_T1T1X_CPP [gcc 11.3.1] [inline Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[2] ('sse4': SSE4.2, 128bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 1.494739e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.931280e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.931280e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.542462e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.999174e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.999174e+06 ) sec^-1 MeanMatrixElemValue = ( 1.486736e-01 +- 3.293564e-05 ) GeV^0 -TOTAL : 0.827695 sec +TOTAL : 0.793291 sec INFO: No Floating Point Exceptions have been reported - 2,378,817,913 cycles # 2.850 GHz - 5,912,991,474 instructions # 2.49 insn per cycle - 0.835517705 seconds time elapsed + 2,337,878,530 cycles # 2.929 GHz + 5,879,791,997 instructions # 2.52 insn per cycle + 0.798863449 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 1340) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_t1t1.mad/SubProcesses/P1_gg_t1t1x/build.sse4_d_inl0_hrd1/runTest_cpp.exe @@ -149,15 +149,15 @@ Process = SIGMA_MSSM_SLHA2_GG_T1T1X_CPP [gcc 11.3.1] [inline Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('avx2': AVX2, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 2.079942e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 2.957305e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 2.957305e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 2.213213e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.244341e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.244341e+06 ) sec^-1 MeanMatrixElemValue = ( 1.486736e-01 +- 3.293564e-05 ) GeV^0 -TOTAL : 0.628333 sec +TOTAL : 0.587861 sec INFO: No Floating Point Exceptions have been reported - 1,788,933,654 cycles # 2.817 GHz - 3,328,376,953 instructions # 1.86 insn per cycle - 0.635862534 seconds time elapsed + 1,651,052,093 cycles # 2.785 GHz + 3,288,347,468 instructions # 1.99 insn per cycle + 0.593562571 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 1436) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_t1t1.mad/SubProcesses/P1_gg_t1t1x/build.avx2_d_inl0_hrd1/runTest_cpp.exe @@ -181,15 +181,15 @@ Process = SIGMA_MSSM_SLHA2_GG_T1T1X_CPP [gcc 11.3.1] [inline Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('512y': AVX512, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 2.320640e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.437091e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.437091e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 2.301299e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.409948e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.409948e+06 ) sec^-1 MeanMatrixElemValue = ( 1.486736e-01 +- 3.293564e-05 ) GeV^0 -TOTAL : 0.574324 sec +TOTAL : 0.568665 sec INFO: No Floating Point Exceptions have been reported - 1,653,934,067 cycles # 2.845 GHz - 3,291,054,827 instructions # 1.99 insn per cycle - 0.581926884 seconds time elapsed + 1,614,706,343 cycles # 2.814 GHz + 3,262,647,746 instructions # 2.02 insn per cycle + 0.574381531 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 1328) (512y: 96) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_t1t1.mad/SubProcesses/P1_gg_t1t1x/build.512y_d_inl0_hrd1/runTest_cpp.exe @@ -213,15 +213,15 @@ Process = SIGMA_MSSM_SLHA2_GG_T1T1X_CPP [gcc 11.3.1] [inline Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('512z': AVX512, 512bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 2.152026e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.087565e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.087565e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 2.159455e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.102047e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.102047e+06 ) sec^-1 MeanMatrixElemValue = ( 1.486736e-01 +- 3.293564e-05 ) GeV^0 -TOTAL : 0.611501 sec +TOTAL : 0.599363 sec INFO: No Floating Point Exceptions have been reported - 1,420,414,146 cycles # 2.296 GHz - 2,439,626,449 instructions # 1.72 insn per cycle - 0.619276325 seconds time elapsed + 1,375,467,385 cycles # 2.275 GHz + 2,410,556,486 instructions # 1.75 insn per cycle + 0.605336952 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 547) (512y: 60) (512z: 1007) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_t1t1.mad/SubProcesses/P1_gg_t1t1x/build.512z_d_inl0_hrd1/runTest_cpp.exe diff --git a/epochX/cudacpp/tput/logs_susyggt1t1_mad/log_susyggt1t1_mad_f_inl0_hrd0.txt b/epochX/cudacpp/tput/logs_susyggt1t1_mad/log_susyggt1t1_mad_f_inl0_hrd0.txt index bd2093b69b..bc85ea8b80 100644 --- a/epochX/cudacpp/tput/logs_susyggt1t1_mad/log_susyggt1t1_mad_f_inl0_hrd0.txt +++ b/epochX/cudacpp/tput/logs_susyggt1t1_mad/log_susyggt1t1_mad_f_inl0_hrd0.txt @@ -40,7 +40,7 @@ make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_t1t1.mad/SubProcesses/P1_gg_t1t1x' -DATE: 2024-10-06_09:59:44 +DATE: 2024-10-06_14:15:38 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: ========================================================================= @@ -49,15 +49,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_MSSM_SLHA2_GG_T1T1X_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 1.032605e+08 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.087100e+08 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.501992e+08 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.100454e+08 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.150353e+08 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.521607e+08 ) sec^-1 MeanMatrixElemValue = ( 1.486732e-01 +- 3.293572e-05 ) GeV^0 -TOTAL : 0.487961 sec +TOTAL : 0.483765 sec INFO: No Floating Point Exceptions have been reported - 2,048,884,733 cycles # 2.866 GHz - 2,915,076,407 instructions # 1.42 insn per cycle - 0.773529382 seconds time elapsed + 2,035,555,079 cycles # 2.861 GHz + 2,898,811,913 instructions # 1.42 insn per cycle + 0.768167291 seconds time elapsed runNcu /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_t1t1.mad/SubProcesses/P1_gg_t1t1x/build.cuda_f_inl0_hrd0/check_cuda.exe -p 2048 256 1 ==PROF== Profiling "sigmaKin": launch__registers_per_thread 97 ==PROF== Profiling "sigmaKin": sm__sass_average_branch_targets_threads_uniform.pct 100% @@ -85,15 +85,15 @@ Process = SIGMA_MSSM_SLHA2_GG_T1T1X_CPP [gcc 11.3.1] [inline Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/none+NAVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = SCALAR ('none': ~vector[1], no SIMD) -EvtsPerSec[Rmb+ME] (23) = ( 9.070270e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.039772e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.039772e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 9.177391e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.052998e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.052998e+06 ) sec^-1 MeanMatrixElemValue = ( 1.486735e-01 +- 3.293563e-05 ) GeV^0 -TOTAL : 1.246364 sec +TOTAL : 1.233473 sec INFO: No Floating Point Exceptions have been reported - 3,688,263,957 cycles # 2.948 GHz - 9,604,598,454 instructions # 2.60 insn per cycle - 1.251819600 seconds time elapsed + 3,657,643,761 cycles # 2.953 GHz + 9,602,348,063 instructions # 2.63 insn per cycle + 1.239280985 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 463) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_t1t1.mad/SubProcesses/P1_gg_t1t1x/build.none_f_inl0_hrd0/runTest_cpp.exe @@ -117,15 +117,15 @@ Process = SIGMA_MSSM_SLHA2_GG_T1T1X_CPP [gcc 11.3.1] [inline Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('sse4': SSE4.2, 128bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 2.214709e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.338045e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.338045e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 2.264027e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.464295e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.464295e+06 ) sec^-1 MeanMatrixElemValue = ( 1.486735e-01 +- 3.293563e-05 ) GeV^0 -TOTAL : 0.563106 sec +TOTAL : 0.553896 sec INFO: No Floating Point Exceptions have been reported - 1,636,975,072 cycles # 2.881 GHz - 3,967,404,939 instructions # 2.42 insn per cycle - 0.568812477 seconds time elapsed + 1,638,557,809 cycles # 2.933 GHz + 3,967,726,100 instructions # 2.42 insn per cycle + 0.559443183 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 1579) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_t1t1.mad/SubProcesses/P1_gg_t1t1x/build.sse4_f_inl0_hrd0/runTest_cpp.exe @@ -149,15 +149,15 @@ Process = SIGMA_MSSM_SLHA2_GG_T1T1X_CPP [gcc 11.3.1] [inline Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('avx2': AVX2, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 2.994371e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 5.295152e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 5.295152e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 3.023713e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 5.335107e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 5.335107e+06 ) sec^-1 MeanMatrixElemValue = ( 1.486735e-01 +- 3.293562e-05 ) GeV^0 -TOTAL : 0.439594 sec +TOTAL : 0.436853 sec INFO: No Floating Point Exceptions have been reported - 1,256,321,725 cycles # 2.826 GHz - 2,497,438,777 instructions # 1.99 insn per cycle - 0.445252542 seconds time elapsed + 1,255,835,282 cycles # 2.843 GHz + 2,497,607,113 instructions # 1.99 insn per cycle + 0.442365788 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 1924) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_t1t1.mad/SubProcesses/P1_gg_t1t1x/build.avx2_f_inl0_hrd0/runTest_cpp.exe @@ -181,15 +181,15 @@ Process = SIGMA_MSSM_SLHA2_GG_T1T1X_CPP [gcc 11.3.1] [inline Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('512y': AVX512, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 3.098864e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 5.632832e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 5.632832e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 3.115926e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 5.626426e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 5.626426e+06 ) sec^-1 MeanMatrixElemValue = ( 1.486735e-01 +- 3.293562e-05 ) GeV^0 -TOTAL : 0.427898 sec +TOTAL : 0.425934 sec INFO: No Floating Point Exceptions have been reported - 1,236,536,318 cycles # 2.855 GHz - 2,473,365,360 instructions # 2.00 insn per cycle - 0.433705293 seconds time elapsed + 1,227,981,284 cycles # 2.850 GHz + 2,473,454,778 instructions # 2.01 insn per cycle + 0.431514140 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 1870) (512y: 1) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_t1t1.mad/SubProcesses/P1_gg_t1t1x/build.512y_f_inl0_hrd0/runTest_cpp.exe @@ -213,15 +213,15 @@ Process = SIGMA_MSSM_SLHA2_GG_T1T1X_CPP [gcc 11.3.1] [inline Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[16] ('512z': AVX512, 512bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 2.931142e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 4.994223e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 4.994223e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 2.921042e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 4.965272e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 4.965272e+06 ) sec^-1 MeanMatrixElemValue = ( 1.486735e-01 +- 3.293561e-05 ) GeV^0 -TOTAL : 0.448530 sec +TOTAL : 0.449651 sec INFO: No Floating Point Exceptions have been reported - 1,079,279,667 cycles # 2.379 GHz - 2,073,684,661 instructions # 1.92 insn per cycle - 0.454351959 seconds time elapsed + 1,080,311,210 cycles # 2.376 GHz + 2,073,766,122 instructions # 1.92 insn per cycle + 0.455236555 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 1011) (512y: 5) (512z: 1292) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_t1t1.mad/SubProcesses/P1_gg_t1t1x/build.512z_f_inl0_hrd0/runTest_cpp.exe diff --git a/epochX/cudacpp/tput/logs_susyggt1t1_mad/log_susyggt1t1_mad_f_inl0_hrd1.txt b/epochX/cudacpp/tput/logs_susyggt1t1_mad/log_susyggt1t1_mad_f_inl0_hrd1.txt index 2473496911..28ffe75db0 100644 --- a/epochX/cudacpp/tput/logs_susyggt1t1_mad/log_susyggt1t1_mad_f_inl0_hrd1.txt +++ b/epochX/cudacpp/tput/logs_susyggt1t1_mad/log_susyggt1t1_mad_f_inl0_hrd1.txt @@ -40,7 +40,7 @@ make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_t1t1.mad/SubProcesses/P1_gg_t1t1x' -DATE: 2024-10-06_09:59:56 +DATE: 2024-10-06_14:15:49 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: ========================================================================= @@ -49,15 +49,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_MSSM_SLHA2_GG_T1T1X_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=1] Workflow summary = CUD:FLT+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 1.032625e+08 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.129649e+08 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.575777e+08 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.092879e+08 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.103265e+08 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.461433e+08 ) sec^-1 MeanMatrixElemValue = ( 1.486732e-01 +- 3.293572e-05 ) GeV^0 -TOTAL : 0.481858 sec +TOTAL : 0.482634 sec INFO: No Floating Point Exceptions have been reported - 2,051,512,664 cycles # 2.885 GHz - 2,948,723,179 instructions # 1.44 insn per cycle - 0.768027645 seconds time elapsed + 2,045,108,402 cycles # 2.876 GHz + 2,894,483,136 instructions # 1.42 insn per cycle + 0.767852674 seconds time elapsed runNcu /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_t1t1.mad/SubProcesses/P1_gg_t1t1x/build.cuda_f_inl0_hrd1/check_cuda.exe -p 2048 256 1 ==PROF== Profiling "sigmaKin": launch__registers_per_thread 86 ==PROF== Profiling "sigmaKin": sm__sass_average_branch_targets_threads_uniform.pct 100% @@ -85,15 +85,15 @@ Process = SIGMA_MSSM_SLHA2_GG_T1T1X_CPP [gcc 11.3.1] [inline Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/none+NAVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = SCALAR ('none': ~vector[1], no SIMD) -EvtsPerSec[Rmb+ME] (23) = ( 9.212337e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.061006e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.061006e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 9.157074e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.053130e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.053130e+06 ) sec^-1 MeanMatrixElemValue = ( 1.486735e-01 +- 3.293563e-05 ) GeV^0 -TOTAL : 1.227579 sec +TOTAL : 1.233672 sec INFO: No Floating Point Exceptions have been reported - 3,620,291,769 cycles # 2.937 GHz - 9,471,544,557 instructions # 2.62 insn per cycle - 1.233302650 seconds time elapsed + 3,622,912,681 cycles # 2.926 GHz + 9,471,798,171 instructions # 2.61 insn per cycle + 1.238988196 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 367) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_t1t1.mad/SubProcesses/P1_gg_t1t1x/build.none_f_inl0_hrd1/runTest_cpp.exe @@ -117,15 +117,15 @@ Process = SIGMA_MSSM_SLHA2_GG_T1T1X_CPP [gcc 11.3.1] [inline Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('sse4': SSE4.2, 128bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 2.220343e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.350531e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.350531e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 2.158903e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.218276e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.218276e+06 ) sec^-1 MeanMatrixElemValue = ( 1.486735e-01 +- 3.293563e-05 ) GeV^0 -TOTAL : 0.560958 sec +TOTAL : 0.574739 sec INFO: No Floating Point Exceptions have been reported - 1,637,220,191 cycles # 2.892 GHz - 3,933,324,289 instructions # 2.40 insn per cycle - 0.566799529 seconds time elapsed + 1,639,202,062 cycles # 2.828 GHz + 3,933,578,666 instructions # 2.40 insn per cycle + 0.580246352 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 1517) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_t1t1.mad/SubProcesses/P1_gg_t1t1x/build.sse4_f_inl0_hrd1/runTest_cpp.exe @@ -149,15 +149,15 @@ Process = SIGMA_MSSM_SLHA2_GG_T1T1X_CPP [gcc 11.3.1] [inline Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('avx2': AVX2, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 2.995950e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 5.312007e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 5.312007e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 3.010428e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 5.298313e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 5.298313e+06 ) sec^-1 MeanMatrixElemValue = ( 1.486735e-01 +- 3.293562e-05 ) GeV^0 -TOTAL : 0.438140 sec +TOTAL : 0.436519 sec INFO: No Floating Point Exceptions have been reported - 1,255,613,659 cycles # 2.833 GHz - 2,482,092,959 instructions # 1.98 insn per cycle - 0.443764126 seconds time elapsed + 1,254,374,622 cycles # 2.842 GHz + 2,482,005,788 instructions # 1.98 insn per cycle + 0.442037121 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 1817) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_t1t1.mad/SubProcesses/P1_gg_t1t1x/build.avx2_f_inl0_hrd1/runTest_cpp.exe @@ -181,15 +181,15 @@ Process = SIGMA_MSSM_SLHA2_GG_T1T1X_CPP [gcc 11.3.1] [inline Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('512y': AVX512, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 3.087645e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 5.599722e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 5.599722e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 3.115288e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 5.617306e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 5.617306e+06 ) sec^-1 MeanMatrixElemValue = ( 1.486735e-01 +- 3.293562e-05 ) GeV^0 -TOTAL : 0.428178 sec +TOTAL : 0.425579 sec INFO: No Floating Point Exceptions have been reported - 1,231,320,501 cycles # 2.843 GHz - 2,457,271,461 instructions # 2.00 insn per cycle - 0.433769891 seconds time elapsed + 1,225,775,270 cycles # 2.847 GHz + 2,457,345,493 instructions # 2.00 insn per cycle + 0.431111165 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 1773) (512y: 1) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_t1t1.mad/SubProcesses/P1_gg_t1t1x/build.512y_f_inl0_hrd1/runTest_cpp.exe @@ -213,15 +213,15 @@ Process = SIGMA_MSSM_SLHA2_GG_T1T1X_CPP [gcc 11.3.1] [inline Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[16] ('512z': AVX512, 512bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 2.945345e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 5.024652e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 5.024652e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 2.928995e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 5.022491e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 5.022491e+06 ) sec^-1 MeanMatrixElemValue = ( 1.486735e-01 +- 3.293561e-05 ) GeV^0 -TOTAL : 0.444653 sec +TOTAL : 0.447926 sec INFO: No Floating Point Exceptions have been reported - 1,073,447,692 cycles # 2.387 GHz - 2,057,517,401 instructions # 1.92 insn per cycle - 0.450271011 seconds time elapsed + 1,077,388,330 cycles # 2.378 GHz + 2,057,768,427 instructions # 1.91 insn per cycle + 0.453760812 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 906) (512y: 5) (512z: 1273) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_t1t1.mad/SubProcesses/P1_gg_t1t1x/build.512z_f_inl0_hrd1/runTest_cpp.exe diff --git a/epochX/cudacpp/tput/logs_susyggt1t1_mad/log_susyggt1t1_mad_m_inl0_hrd0.txt b/epochX/cudacpp/tput/logs_susyggt1t1_mad/log_susyggt1t1_mad_m_inl0_hrd0.txt index 5ae4907c26..45924a2ad4 100644 --- a/epochX/cudacpp/tput/logs_susyggt1t1_mad/log_susyggt1t1_mad_m_inl0_hrd0.txt +++ b/epochX/cudacpp/tput/logs_susyggt1t1_mad/log_susyggt1t1_mad_m_inl0_hrd0.txt @@ -40,7 +40,7 @@ make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_t1t1.mad/SubProcesses/P1_gg_t1t1x' -DATE: 2024-10-06_09:59:20 +DATE: 2024-10-06_14:15:13 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: ========================================================================= @@ -49,15 +49,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_MSSM_SLHA2_GG_T1T1X_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:MIX+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 6.059495e+07 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 2.307970e+08 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 2.770458e+08 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 6.498097e+07 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 2.421044e+08 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.862158e+08 ) sec^-1 MeanMatrixElemValue = ( 1.486736e-01 +- 3.293564e-05 ) GeV^0 -TOTAL : 0.521991 sec +TOTAL : 0.518852 sec INFO: No Floating Point Exceptions have been reported - 2,182,804,723 cycles # 2.882 GHz - 3,091,712,352 instructions # 1.42 insn per cycle - 0.814546737 seconds time elapsed + 2,181,751,185 cycles # 2.892 GHz + 3,067,627,339 instructions # 1.41 insn per cycle + 0.811262491 seconds time elapsed runNcu /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_t1t1.mad/SubProcesses/P1_gg_t1t1x/build.cuda_m_inl0_hrd0/check_cuda.exe -p 2048 256 1 ==PROF== Profiling "sigmaKin": launch__registers_per_thread 130 ==PROF== Profiling "sigmaKin": sm__sass_average_branch_targets_threads_uniform.pct 100% @@ -85,15 +85,15 @@ Process = SIGMA_MSSM_SLHA2_GG_T1T1X_CPP [gcc 11.3.1] [inline Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/none+NAVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) Internal loops fptype_sv = SCALAR ('none': ~vector[1], no SIMD) -EvtsPerSec[Rmb+ME] (23) = ( 9.006175e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.025890e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.025890e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 9.016471e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.028681e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.028681e+06 ) sec^-1 MeanMatrixElemValue = ( 1.486736e-01 +- 3.293564e-05 ) GeV^0 -TOTAL : 1.286472 sec +TOTAL : 1.276697 sec INFO: No Floating Point Exceptions have been reported - 3,808,533,169 cycles # 2.945 GHz - 9,779,238,528 instructions # 2.57 insn per cycle - 1.294044616 seconds time elapsed + 3,778,626,359 cycles # 2.948 GHz + 9,754,693,663 instructions # 2.58 insn per cycle + 1.282493224 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 341) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_t1t1.mad/SubProcesses/P1_gg_t1t1x/build.none_m_inl0_hrd0/runTest_cpp.exe @@ -117,15 +117,15 @@ Process = SIGMA_MSSM_SLHA2_GG_T1T1X_CPP [gcc 11.3.1] [inline Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[2] ('sse4': SSE4.2, 128bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 1.477969e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.892042e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.892042e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.532120e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.977008e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.977008e+06 ) sec^-1 MeanMatrixElemValue = ( 1.486736e-01 +- 3.293564e-05 ) GeV^0 -TOTAL : 0.834785 sec +TOTAL : 0.799182 sec INFO: No Floating Point Exceptions have been reported - 2,360,159,801 cycles # 2.803 GHz - 5,954,715,990 instructions # 2.52 insn per cycle - 0.842708021 seconds time elapsed + 2,317,007,493 cycles # 2.881 GHz + 5,921,331,810 instructions # 2.56 insn per cycle + 0.805063913 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 1412) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_t1t1.mad/SubProcesses/P1_gg_t1t1x/build.sse4_m_inl0_hrd0/runTest_cpp.exe @@ -149,15 +149,15 @@ Process = SIGMA_MSSM_SLHA2_GG_T1T1X_CPP [gcc 11.3.1] [inline Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('avx2': AVX2, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 2.260391e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.350498e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.350498e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 2.222547e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.258596e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.258596e+06 ) sec^-1 MeanMatrixElemValue = ( 1.486736e-01 +- 3.293564e-05 ) GeV^0 -TOTAL : 0.587860 sec +TOTAL : 0.586954 sec INFO: No Floating Point Exceptions have been reported - 1,670,861,769 cycles # 2.810 GHz - 3,283,918,691 instructions # 1.97 insn per cycle - 0.595426943 seconds time elapsed + 1,668,054,727 cycles # 2.818 GHz + 3,257,028,002 instructions # 1.95 insn per cycle + 0.592592912 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 1567) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_t1t1.mad/SubProcesses/P1_gg_t1t1x/build.avx2_m_inl0_hrd0/runTest_cpp.exe @@ -181,15 +181,15 @@ Process = SIGMA_MSSM_SLHA2_GG_T1T1X_CPP [gcc 11.3.1] [inline Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('512y': AVX512, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 2.348300e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.498815e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.498815e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 2.349931e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.510050e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.510050e+06 ) sec^-1 MeanMatrixElemValue = ( 1.486736e-01 +- 3.293564e-05 ) GeV^0 -TOTAL : 0.573278 sec +TOTAL : 0.559254 sec INFO: No Floating Point Exceptions have been reported - 1,645,784,221 cycles # 2.835 GHz - 3,247,832,958 instructions # 1.97 insn per cycle - 0.581347619 seconds time elapsed + 1,601,238,220 cycles # 2.836 GHz + 3,210,413,202 instructions # 2.00 insn per cycle + 0.565147247 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 1446) (512y: 101) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_t1t1.mad/SubProcesses/P1_gg_t1t1x/build.512y_m_inl0_hrd0/runTest_cpp.exe @@ -213,15 +213,15 @@ Process = SIGMA_MSSM_SLHA2_GG_T1T1X_CPP [gcc 11.3.1] [inline Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('512z': AVX512, 512bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 2.143317e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.068862e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.068862e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 2.202125e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.186949e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.186949e+06 ) sec^-1 MeanMatrixElemValue = ( 1.486736e-01 +- 3.293564e-05 ) GeV^0 -TOTAL : 0.613179 sec +TOTAL : 0.591100 sec INFO: No Floating Point Exceptions have been reported - 1,394,199,360 cycles # 2.248 GHz - 2,406,597,613 instructions # 1.73 insn per cycle - 0.620673412 seconds time elapsed + 1,353,281,726 cycles # 2.269 GHz + 2,377,285,937 instructions # 1.76 insn per cycle + 0.596988135 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 768) (512y: 64) (512z: 1063) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_t1t1.mad/SubProcesses/P1_gg_t1t1x/build.512z_m_inl0_hrd0/runTest_cpp.exe diff --git a/epochX/cudacpp/tput/logs_susyggt1t1_mad/log_susyggt1t1_mad_m_inl0_hrd1.txt b/epochX/cudacpp/tput/logs_susyggt1t1_mad/log_susyggt1t1_mad_m_inl0_hrd1.txt index 3e507cd882..3045d7bd9f 100644 --- a/epochX/cudacpp/tput/logs_susyggt1t1_mad/log_susyggt1t1_mad_m_inl0_hrd1.txt +++ b/epochX/cudacpp/tput/logs_susyggt1t1_mad/log_susyggt1t1_mad_m_inl0_hrd1.txt @@ -40,7 +40,7 @@ make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_t1t1.mad/SubProcesses/P1_gg_t1t1x' -DATE: 2024-10-06_09:59:32 +DATE: 2024-10-06_14:15:26 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: ========================================================================= @@ -49,15 +49,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_MSSM_SLHA2_GG_T1T1X_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=1] Workflow summary = CUD:MIX+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 6.080757e+07 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 2.449829e+08 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 2.987143e+08 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 6.515360e+07 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 2.504380e+08 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.003121e+08 ) sec^-1 MeanMatrixElemValue = ( 1.486736e-01 +- 3.293564e-05 ) GeV^0 -TOTAL : 0.520335 sec +TOTAL : 0.520822 sec INFO: No Floating Point Exceptions have been reported - 2,182,231,478 cycles # 2.885 GHz - 3,097,447,003 instructions # 1.42 insn per cycle - 0.813407395 seconds time elapsed + 2,178,225,681 cycles # 2.883 GHz + 3,041,368,306 instructions # 1.40 insn per cycle + 0.813152763 seconds time elapsed runNcu /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_t1t1.mad/SubProcesses/P1_gg_t1t1x/build.cuda_m_inl0_hrd1/check_cuda.exe -p 2048 256 1 ==PROF== Profiling "sigmaKin": launch__registers_per_thread 124 ==PROF== Profiling "sigmaKin": sm__sass_average_branch_targets_threads_uniform.pct 100% @@ -85,15 +85,15 @@ Process = SIGMA_MSSM_SLHA2_GG_T1T1X_CPP [gcc 11.3.1] [inline Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/none+NAVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) Internal loops fptype_sv = SCALAR ('none': ~vector[1], no SIMD) -EvtsPerSec[Rmb+ME] (23) = ( 8.967180e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.023779e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.023779e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 9.026476e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.029346e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.029346e+06 ) sec^-1 MeanMatrixElemValue = ( 1.486736e-01 +- 3.293564e-05 ) GeV^0 -TOTAL : 1.289771 sec +TOTAL : 1.275067 sec INFO: No Floating Point Exceptions have been reported - 3,794,201,935 cycles # 2.927 GHz - 9,666,542,351 instructions # 2.55 insn per cycle - 1.297077628 seconds time elapsed + 3,764,430,893 cycles # 2.940 GHz + 9,645,104,810 instructions # 2.56 insn per cycle + 1.281132919 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 359) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_t1t1.mad/SubProcesses/P1_gg_t1t1x/build.none_m_inl0_hrd1/runTest_cpp.exe @@ -117,15 +117,15 @@ Process = SIGMA_MSSM_SLHA2_GG_T1T1X_CPP [gcc 11.3.1] [inline Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[2] ('sse4': SSE4.2, 128bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 1.583493e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 2.064503e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 2.064503e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.506599e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.933281e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.933281e+06 ) sec^-1 MeanMatrixElemValue = ( 1.486736e-01 +- 3.293564e-05 ) GeV^0 -TOTAL : 0.784715 sec +TOTAL : 0.809579 sec INFO: No Floating Point Exceptions have been reported - 2,328,374,642 cycles # 2.942 GHz - 5,878,440,022 instructions # 2.52 insn per cycle - 0.792155161 seconds time elapsed + 2,297,531,150 cycles # 2.820 GHz + 5,849,370,273 instructions # 2.55 insn per cycle + 0.815395414 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 1371) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_t1t1.mad/SubProcesses/P1_gg_t1t1x/build.sse4_m_inl0_hrd1/runTest_cpp.exe @@ -149,15 +149,15 @@ Process = SIGMA_MSSM_SLHA2_GG_T1T1X_CPP [gcc 11.3.1] [inline Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('avx2': AVX2, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 2.254464e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.329047e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.329047e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 2.264779e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.338675e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.338675e+06 ) sec^-1 MeanMatrixElemValue = ( 1.486736e-01 +- 3.293564e-05 ) GeV^0 -TOTAL : 0.590226 sec +TOTAL : 0.576247 sec INFO: No Floating Point Exceptions have been reported - 1,689,754,472 cycles # 2.827 GHz - 3,255,343,739 instructions # 1.93 insn per cycle - 0.598325338 seconds time elapsed + 1,643,312,247 cycles # 2.827 GHz + 3,217,931,348 instructions # 1.96 insn per cycle + 0.581994486 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 1483) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_t1t1.mad/SubProcesses/P1_gg_t1t1x/build.avx2_m_inl0_hrd1/runTest_cpp.exe @@ -181,15 +181,15 @@ Process = SIGMA_MSSM_SLHA2_GG_T1T1X_CPP [gcc 11.3.1] [inline Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('512y': AVX512, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 2.345727e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.502859e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.502859e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 2.349416e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.509529e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.509529e+06 ) sec^-1 MeanMatrixElemValue = ( 1.486736e-01 +- 3.293564e-05 ) GeV^0 -TOTAL : 0.572126 sec +TOTAL : 0.558847 sec INFO: No Floating Point Exceptions have been reported - 1,634,040,486 cycles # 2.820 GHz - 3,219,951,921 instructions # 1.97 insn per cycle - 0.580193189 seconds time elapsed + 1,594,083,427 cycles # 2.826 GHz + 3,182,361,820 instructions # 2.00 insn per cycle + 0.564618029 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 1382) (512y: 101) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_t1t1.mad/SubProcesses/P1_gg_t1t1x/build.512y_m_inl0_hrd1/runTest_cpp.exe @@ -213,15 +213,15 @@ Process = SIGMA_MSSM_SLHA2_GG_T1T1X_CPP [gcc 11.3.1] [inline Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('512z': AVX512, 512bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 2.168828e+06 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.118471e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.118471e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 2.174712e+06 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.118675e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.118675e+06 ) sec^-1 MeanMatrixElemValue = ( 1.486736e-01 +- 3.293564e-05 ) GeV^0 -TOTAL : 0.609357 sec +TOTAL : 0.594905 sec INFO: No Floating Point Exceptions have been reported - 1,417,478,840 cycles # 2.299 GHz - 2,399,490,515 instructions # 1.69 insn per cycle - 0.617376810 seconds time elapsed + 1,378,433,407 cycles # 2.297 GHz + 2,362,422,299 instructions # 1.71 insn per cycle + 0.600675181 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 716) (512y: 64) (512z: 1056) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_t1t1.mad/SubProcesses/P1_gg_t1t1x/build.512z_m_inl0_hrd1/runTest_cpp.exe diff --git a/epochX/cudacpp/tput/logs_susyggtt_mad/log_susyggtt_mad_d_inl0_hrd0.txt b/epochX/cudacpp/tput/logs_susyggtt_mad/log_susyggtt_mad_d_inl0_hrd0.txt index 607647c622..57fd476c08 100644 --- a/epochX/cudacpp/tput/logs_susyggtt_mad/log_susyggtt_mad_d_inl0_hrd0.txt +++ b/epochX/cudacpp/tput/logs_susyggtt_mad/log_susyggtt_mad_d_inl0_hrd0.txt @@ -40,7 +40,7 @@ make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_tt.mad/SubProcesses/P1_gg_ttx' -DATE: 2024-10-06_09:56:33 +DATE: 2024-10-06_14:12:26 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: ========================================================================= @@ -49,15 +49,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_MSSM_SLHA2_GG_TTX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:DBL+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 4.270000e+07 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 9.214418e+07 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 9.893995e+07 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 4.547164e+07 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 8.998299e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 9.543568e+07 ) sec^-1 MeanMatrixElemValue = ( 2.072848e+00 +- 3.360985e-03 ) GeV^0 -TOTAL : 0.542175 sec +TOTAL : 0.530156 sec INFO: No Floating Point Exceptions have been reported - 2,178,993,269 cycles # 2.803 GHz - 3,108,059,533 instructions # 1.43 insn per cycle - 0.838052893 seconds time elapsed + 2,190,821,065 cycles # 2.834 GHz + 3,161,610,915 instructions # 1.44 insn per cycle + 0.829392557 seconds time elapsed runNcu /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_tt.mad/SubProcesses/P1_gg_ttx/build.cuda_d_inl0_hrd0/check_cuda.exe -p 2048 256 1 ==PROF== Profiling "sigmaKin": launch__registers_per_thread 214 ==PROF== Profiling "sigmaKin": sm__sass_average_branch_targets_threads_uniform.pct 100% @@ -85,15 +85,15 @@ Process = SIGMA_MSSM_SLHA2_GG_TTX_CPP [gcc 11.3.1] [inlineHe Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/none+NAVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = SCALAR ('none': ~vector[1], no SIMD) -EvtsPerSec[Rmb+ME] (23) = ( 1.830273e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.876984e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.876984e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.822725e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.868835e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.868835e+05 ) sec^-1 MeanMatrixElemValue = ( 2.072848e+00 +- 3.360985e-03 ) GeV^0 -TOTAL : 5.836443 sec +TOTAL : 5.860059 sec INFO: No Floating Point Exceptions have been reported - 17,247,101,824 cycles # 2.952 GHz - 45,921,478,129 instructions # 2.66 insn per cycle - 5.842453521 seconds time elapsed + 17,254,224,990 cycles # 2.942 GHz + 45,925,329,897 instructions # 2.66 insn per cycle + 5.865465576 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 622) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_tt.mad/SubProcesses/P1_gg_ttx/build.none_d_inl0_hrd0/runTest_cpp.exe @@ -117,15 +117,15 @@ Process = SIGMA_MSSM_SLHA2_GG_TTX_CPP [gcc 11.3.1] [inlineHe Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[2] ('sse4': SSE4.2, 128bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 3.179372e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.338539e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.338539e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 3.171496e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.325752e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.325752e+05 ) sec^-1 MeanMatrixElemValue = ( 2.072848e+00 +- 3.360985e-03 ) GeV^0 -TOTAL : 3.409251 sec +TOTAL : 3.414159 sec INFO: No Floating Point Exceptions have been reported - 10,038,815,546 cycles # 2.940 GHz - 27,809,165,185 instructions # 2.77 insn per cycle - 3.415697404 seconds time elapsed + 10,034,587,484 cycles # 2.935 GHz + 27,804,206,177 instructions # 2.77 insn per cycle + 3.419727831 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 2537) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_tt.mad/SubProcesses/P1_gg_ttx/build.sse4_d_inl0_hrd0/runTest_cpp.exe @@ -149,15 +149,15 @@ Process = SIGMA_MSSM_SLHA2_GG_TTX_CPP [gcc 11.3.1] [inlineHe Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('avx2': AVX2, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 5.016017e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 5.397611e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 5.397611e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 5.046776e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 5.438466e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 5.438466e+05 ) sec^-1 MeanMatrixElemValue = ( 2.072848e+00 +- 3.360985e-03 ) GeV^0 -TOTAL : 2.202025 sec +TOTAL : 2.185519 sec INFO: No Floating Point Exceptions have been reported - 6,083,216,423 cycles # 2.757 GHz - 12,595,496,799 instructions # 2.07 insn per cycle - 2.208459235 seconds time elapsed + 6,077,131,953 cycles # 2.775 GHz + 12,586,644,788 instructions # 2.07 insn per cycle + 2.191009024 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 2620) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_tt.mad/SubProcesses/P1_gg_ttx/build.avx2_d_inl0_hrd0/runTest_cpp.exe @@ -181,15 +181,15 @@ Process = SIGMA_MSSM_SLHA2_GG_TTX_CPP [gcc 11.3.1] [inlineHe Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('512y': AVX512, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 5.491994e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 5.947919e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 5.947919e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 5.417064e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 5.868788e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 5.868788e+05 ) sec^-1 MeanMatrixElemValue = ( 2.072848e+00 +- 3.360985e-03 ) GeV^0 -TOTAL : 2.018419 sec +TOTAL : 2.043714 sec INFO: No Floating Point Exceptions have been reported - 5,588,215,007 cycles # 2.761 GHz - 12,004,808,489 instructions # 2.15 insn per cycle - 2.024606102 seconds time elapsed + 5,561,903,884 cycles # 2.715 GHz + 12,001,535,017 instructions # 2.16 insn per cycle + 2.049034524 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 2365) (512y: 144) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_tt.mad/SubProcesses/P1_gg_ttx/build.512y_d_inl0_hrd0/runTest_cpp.exe @@ -213,15 +213,15 @@ Process = SIGMA_MSSM_SLHA2_GG_TTX_CPP [gcc 11.3.1] [inlineHe Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('512z': AVX512, 512bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 3.529303e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.713663e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.713663e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 3.352961e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.523680e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.523680e+05 ) sec^-1 MeanMatrixElemValue = ( 2.072848e+00 +- 3.360985e-03 ) GeV^0 -TOTAL : 3.082414 sec +TOTAL : 3.236525 sec INFO: No Floating Point Exceptions have been reported - 5,763,724,377 cycles # 1.867 GHz - 8,350,228,242 instructions # 1.45 insn per cycle - 3.088879573 seconds time elapsed + 5,746,916,255 cycles # 1.773 GHz + 8,345,602,626 instructions # 1.45 insn per cycle + 3.241918296 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 1468) (512y: 122) (512z: 1806) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_tt.mad/SubProcesses/P1_gg_ttx/build.512z_d_inl0_hrd0/runTest_cpp.exe diff --git a/epochX/cudacpp/tput/logs_susyggtt_mad/log_susyggtt_mad_d_inl0_hrd1.txt b/epochX/cudacpp/tput/logs_susyggtt_mad/log_susyggtt_mad_d_inl0_hrd1.txt index 3ed4c3c5ff..e51da4de0a 100644 --- a/epochX/cudacpp/tput/logs_susyggtt_mad/log_susyggtt_mad_d_inl0_hrd1.txt +++ b/epochX/cudacpp/tput/logs_susyggtt_mad/log_susyggtt_mad_d_inl0_hrd1.txt @@ -40,7 +40,7 @@ make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_tt.mad/SubProcesses/P1_gg_ttx' -DATE: 2024-10-06_09:56:58 +DATE: 2024-10-06_14:12:51 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: ========================================================================= @@ -49,15 +49,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_MSSM_SLHA2_GG_TTX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=1] Workflow summary = CUD:DBL+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 4.306886e+07 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 9.297289e+07 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 9.977845e+07 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 4.670548e+07 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 9.480120e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.009137e+08 ) sec^-1 MeanMatrixElemValue = ( 2.072848e+00 +- 3.360985e-03 ) GeV^0 -TOTAL : 0.530204 sec +TOTAL : 0.524003 sec INFO: No Floating Point Exceptions have been reported - 2,211,323,980 cycles # 2.884 GHz - 3,201,430,578 instructions # 1.45 insn per cycle - 0.823926524 seconds time elapsed + 2,184,007,023 cycles # 2.878 GHz + 3,162,002,812 instructions # 1.45 insn per cycle + 0.815834933 seconds time elapsed runNcu /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_tt.mad/SubProcesses/P1_gg_ttx/build.cuda_d_inl0_hrd1/check_cuda.exe -p 2048 256 1 ==PROF== Profiling "sigmaKin": launch__registers_per_thread 212 ==PROF== Profiling "sigmaKin": sm__sass_average_branch_targets_threads_uniform.pct 100% @@ -85,15 +85,15 @@ Process = SIGMA_MSSM_SLHA2_GG_TTX_CPP [gcc 11.3.1] [inlineHe Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/none+NAVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = SCALAR ('none': ~vector[1], no SIMD) -EvtsPerSec[Rmb+ME] (23) = ( 1.872475e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.921622e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.921622e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.878364e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.927708e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.927708e+05 ) sec^-1 MeanMatrixElemValue = ( 2.072848e+00 +- 3.360985e-03 ) GeV^0 -TOTAL : 5.706755 sec +TOTAL : 5.688828 sec INFO: No Floating Point Exceptions have been reported - 16,797,600,798 cycles # 2.941 GHz - 44,912,592,336 instructions # 2.67 insn per cycle - 5.712473159 seconds time elapsed + 16,754,360,316 cycles # 2.943 GHz + 44,907,567,532 instructions # 2.68 insn per cycle + 5.694077130 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 566) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_tt.mad/SubProcesses/P1_gg_ttx/build.none_d_inl0_hrd1/runTest_cpp.exe @@ -117,15 +117,15 @@ Process = SIGMA_MSSM_SLHA2_GG_TTX_CPP [gcc 11.3.1] [inlineHe Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[2] ('sse4': SSE4.2, 128bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 3.376254e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.552215e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.552215e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 3.366768e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.541874e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.541874e+05 ) sec^-1 MeanMatrixElemValue = ( 2.072848e+00 +- 3.360985e-03 ) GeV^0 -TOTAL : 3.215396 sec +TOTAL : 3.222459 sec INFO: No Floating Point Exceptions have been reported - 9,523,990,060 cycles # 2.957 GHz - 26,686,144,259 instructions # 2.80 insn per cycle - 3.221864250 seconds time elapsed + 9,504,243,300 cycles # 2.945 GHz + 26,679,789,068 instructions # 2.81 insn per cycle + 3.227749508 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 2326) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_tt.mad/SubProcesses/P1_gg_ttx/build.sse4_d_inl0_hrd1/runTest_cpp.exe @@ -149,15 +149,15 @@ Process = SIGMA_MSSM_SLHA2_GG_TTX_CPP [gcc 11.3.1] [inlineHe Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('avx2': AVX2, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 4.628485e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 4.953785e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 4.953785e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 4.613939e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 4.942469e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 4.942469e+05 ) sec^-1 MeanMatrixElemValue = ( 2.072848e+00 +- 3.360985e-03 ) GeV^0 -TOTAL : 2.376456 sec +TOTAL : 2.379785 sec INFO: No Floating Point Exceptions have been reported - 6,603,885,103 cycles # 2.772 GHz - 14,117,515,687 instructions # 2.14 insn per cycle - 2.382952116 seconds time elapsed + 6,579,416,075 cycles # 2.760 GHz + 14,110,711,216 instructions # 2.14 insn per cycle + 2.385135341 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 2705) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_tt.mad/SubProcesses/P1_gg_ttx/build.avx2_d_inl0_hrd1/runTest_cpp.exe @@ -181,15 +181,15 @@ Process = SIGMA_MSSM_SLHA2_GG_TTX_CPP [gcc 11.3.1] [inlineHe Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('512y': AVX512, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 4.799064e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 5.148539e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 5.148539e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 4.801047e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 5.147341e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 5.147341e+05 ) sec^-1 MeanMatrixElemValue = ( 2.072848e+00 +- 3.360985e-03 ) GeV^0 -TOTAL : 2.297050 sec +TOTAL : 2.290406 sec INFO: No Floating Point Exceptions have been reported - 6,386,723,525 cycles # 2.773 GHz - 13,726,619,432 instructions # 2.15 insn per cycle - 2.304339219 seconds time elapsed + 6,362,907,076 cycles # 2.772 GHz + 13,711,959,883 instructions # 2.15 insn per cycle + 2.295820466 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 2356) (512y: 298) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_tt.mad/SubProcesses/P1_gg_ttx/build.512y_d_inl0_hrd1/runTest_cpp.exe @@ -213,15 +213,15 @@ Process = SIGMA_MSSM_SLHA2_GG_TTX_CPP [gcc 11.3.1] [inlineHe Workflow summary = CPP:DBL+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK FP precision = DOUBLE (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('512z': AVX512, 512bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 3.339110e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.504311e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.504311e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 3.370270e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.538227e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.538227e+05 ) sec^-1 MeanMatrixElemValue = ( 2.072848e+00 +- 3.360985e-03 ) GeV^0 -TOTAL : 3.254444 sec +TOTAL : 3.218828 sec INFO: No Floating Point Exceptions have been reported - 5,974,020,045 cycles # 1.833 GHz - 10,122,964,274 instructions # 1.69 insn per cycle - 3.261538649 seconds time elapsed + 5,963,892,646 cycles # 1.850 GHz + 10,102,555,737 instructions # 1.69 insn per cycle + 3.224163022 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 1318) (512y: 208) (512z: 1986) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_tt.mad/SubProcesses/P1_gg_ttx/build.512z_d_inl0_hrd1/runTest_cpp.exe diff --git a/epochX/cudacpp/tput/logs_susyggtt_mad/log_susyggtt_mad_f_inl0_hrd0.txt b/epochX/cudacpp/tput/logs_susyggtt_mad/log_susyggtt_mad_f_inl0_hrd0.txt index 7bd4c9bca6..e54ea4343e 100644 --- a/epochX/cudacpp/tput/logs_susyggtt_mad/log_susyggtt_mad_f_inl0_hrd0.txt +++ b/epochX/cudacpp/tput/logs_susyggtt_mad/log_susyggtt_mad_f_inl0_hrd0.txt @@ -40,7 +40,7 @@ make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_tt.mad/SubProcesses/P1_gg_ttx' -DATE: 2024-10-06_09:58:13 +DATE: 2024-10-06_14:14:07 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: ========================================================================= @@ -49,15 +49,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_MSSM_SLHA2_GG_TTX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:FLT+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 8.178914e+07 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.740854e+08 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.866078e+08 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 8.999490e+07 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.793835e+08 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.906021e+08 ) sec^-1 MeanMatrixElemValue = ( 2.072877e+00 +- 3.361153e-03 ) GeV^0 -TOTAL : 0.492488 sec +TOTAL : 0.483764 sec INFO: No Floating Point Exceptions have been reported - 2,067,407,730 cycles # 2.879 GHz - 2,921,575,837 instructions # 1.41 insn per cycle - 0.777094459 seconds time elapsed + 2,047,667,106 cycles # 2.884 GHz + 2,968,289,378 instructions # 1.45 insn per cycle + 0.766679954 seconds time elapsed runNcu /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_tt.mad/SubProcesses/P1_gg_ttx/build.cuda_f_inl0_hrd0/check_cuda.exe -p 2048 256 1 ==PROF== Profiling "sigmaKin": launch__registers_per_thread 125 ==PROF== Profiling "sigmaKin": sm__sass_average_branch_targets_threads_uniform.pct 100% @@ -85,15 +85,15 @@ Process = SIGMA_MSSM_SLHA2_GG_TTX_CPP [gcc 11.3.1] [inlineHe Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/none+NAVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = SCALAR ('none': ~vector[1], no SIMD) -EvtsPerSec[Rmb+ME] (23) = ( 1.933137e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.988210e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.988210e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.935430e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.990076e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.990076e+05 ) sec^-1 MeanMatrixElemValue = ( 2.072937e+00 +- 3.361545e-03 ) GeV^0 -TOTAL : 5.510560 sec +TOTAL : 5.503490 sec INFO: No Floating Point Exceptions have been reported - 16,216,363,781 cycles # 2.940 GHz - 45,321,064,348 instructions # 2.79 insn per cycle - 5.516237540 seconds time elapsed + 16,207,286,106 cycles # 2.943 GHz + 45,321,018,575 instructions # 2.80 insn per cycle + 5.508602791 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 600) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_tt.mad/SubProcesses/P1_gg_ttx/build.none_f_inl0_hrd0/runTest_cpp.exe @@ -117,15 +117,15 @@ Process = SIGMA_MSSM_SLHA2_GG_TTX_CPP [gcc 11.3.1] [inlineHe Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('sse4': SSE4.2, 128bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 4.554782e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 4.893509e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 4.893509e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 4.547104e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 4.885282e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 4.885282e+05 ) sec^-1 MeanMatrixElemValue = ( 2.072937e+00 +- 3.361544e-03 ) GeV^0 -TOTAL : 2.389253 sec +TOTAL : 2.392169 sec INFO: No Floating Point Exceptions have been reported - 7,056,712,623 cycles # 2.947 GHz - 17,792,064,584 instructions # 2.52 insn per cycle - 2.395009745 seconds time elapsed + 7,049,859,218 cycles # 2.942 GHz + 17,793,027,419 instructions # 2.52 insn per cycle + 2.397313519 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 3147) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_tt.mad/SubProcesses/P1_gg_ttx/build.sse4_f_inl0_hrd0/runTest_cpp.exe @@ -149,15 +149,15 @@ Process = SIGMA_MSSM_SLHA2_GG_TTX_CPP [gcc 11.3.1] [inlineHe Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('avx2': AVX2, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 8.351394e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 9.496890e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 9.496890e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 8.377816e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 9.516033e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 9.516033e+05 ) sec^-1 MeanMatrixElemValue = ( 2.072967e+00 +- 3.361967e-03 ) GeV^0 -TOTAL : 1.343765 sec +TOTAL : 1.336936 sec INFO: No Floating Point Exceptions have been reported - 3,745,450,403 cycles # 2.777 GHz - 8,262,540,860 instructions # 2.21 insn per cycle - 1.349671424 seconds time elapsed + 3,737,612,310 cycles # 2.788 GHz + 8,264,128,335 instructions # 2.21 insn per cycle + 1.342431742 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 3371) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_tt.mad/SubProcesses/P1_gg_ttx/build.avx2_f_inl0_hrd0/runTest_cpp.exe @@ -181,15 +181,15 @@ Process = SIGMA_MSSM_SLHA2_GG_TTX_CPP [gcc 11.3.1] [inlineHe Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('512y': AVX512, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 8.821818e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.011140e+06 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.011140e+06 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 8.815674e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.009235e+06 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.009235e+06 ) sec^-1 MeanMatrixElemValue = ( 2.072967e+00 +- 3.361967e-03 ) GeV^0 -TOTAL : 1.275053 sec +TOTAL : 1.274864 sec INFO: No Floating Point Exceptions have been reported - 3,558,622,083 cycles # 2.780 GHz - 7,915,407,710 instructions # 2.22 insn per cycle - 1.280856743 seconds time elapsed + 3,546,590,301 cycles # 2.773 GHz + 7,915,079,917 instructions # 2.23 insn per cycle + 1.279865807 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 3214) (512y: 20) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_tt.mad/SubProcesses/P1_gg_ttx/build.512y_f_inl0_hrd0/runTest_cpp.exe @@ -213,15 +213,15 @@ Process = SIGMA_MSSM_SLHA2_GG_TTX_CPP [gcc 11.3.1] [inlineHe Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[16] ('512z': AVX512, 512bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 6.584138e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 7.256759e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 7.256759e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 6.423567e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 7.063619e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 7.063619e+05 ) sec^-1 MeanMatrixElemValue = ( 2.072967e+00 +- 3.361967e-03 ) GeV^0 -TOTAL : 1.679646 sec +TOTAL : 1.719047 sec INFO: No Floating Point Exceptions have been reported - 3,255,689,642 cycles # 1.933 GHz - 6,101,216,288 instructions # 1.87 insn per cycle - 1.685383243 seconds time elapsed + 3,268,558,552 cycles # 1.897 GHz + 6,101,864,476 instructions # 1.87 insn per cycle + 1.724253151 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 2258) (512y: 22) (512z: 2156) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_tt.mad/SubProcesses/P1_gg_ttx/build.512z_f_inl0_hrd0/runTest_cpp.exe diff --git a/epochX/cudacpp/tput/logs_susyggtt_mad/log_susyggtt_mad_f_inl0_hrd1.txt b/epochX/cudacpp/tput/logs_susyggtt_mad/log_susyggtt_mad_f_inl0_hrd1.txt index bd2def4f48..91992027a9 100644 --- a/epochX/cudacpp/tput/logs_susyggtt_mad/log_susyggtt_mad_f_inl0_hrd1.txt +++ b/epochX/cudacpp/tput/logs_susyggtt_mad/log_susyggtt_mad_f_inl0_hrd1.txt @@ -40,7 +40,7 @@ make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_tt.mad/SubProcesses/P1_gg_ttx' -DATE: 2024-10-06_09:58:33 +DATE: 2024-10-06_14:14:27 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: ========================================================================= @@ -49,15 +49,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_MSSM_SLHA2_GG_TTX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=1] Workflow summary = CUD:FLT+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 8.136229e+07 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.747823e+08 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.880709e+08 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 8.901471e+07 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.756934e+08 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.864870e+08 ) sec^-1 MeanMatrixElemValue = ( 2.072877e+00 +- 3.361153e-03 ) GeV^0 -TOTAL : 0.488528 sec +TOTAL : 0.487146 sec INFO: No Floating Point Exceptions have been reported - 2,057,813,122 cycles # 2.874 GHz - 2,903,563,490 instructions # 1.41 insn per cycle - 0.774040886 seconds time elapsed + 2,032,124,904 cycles # 2.840 GHz + 2,919,882,412 instructions # 1.44 insn per cycle + 0.792835745 seconds time elapsed runNcu /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_tt.mad/SubProcesses/P1_gg_ttx/build.cuda_f_inl0_hrd1/check_cuda.exe -p 2048 256 1 ==PROF== Profiling "sigmaKin": launch__registers_per_thread 124 ==PROF== Profiling "sigmaKin": sm__sass_average_branch_targets_threads_uniform.pct 100% @@ -85,15 +85,15 @@ Process = SIGMA_MSSM_SLHA2_GG_TTX_CPP [gcc 11.3.1] [inlineHe Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/none+NAVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = SCALAR ('none': ~vector[1], no SIMD) -EvtsPerSec[Rmb+ME] (23) = ( 1.970300e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 2.026987e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 2.026987e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.975253e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 2.032105e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 2.032105e+05 ) sec^-1 MeanMatrixElemValue = ( 2.072937e+00 +- 3.361545e-03 ) GeV^0 -TOTAL : 5.407589 sec +TOTAL : 5.393536 sec INFO: No Floating Point Exceptions have been reported - 15,991,185,925 cycles # 2.955 GHz - 44,429,993,623 instructions # 2.78 insn per cycle - 5.412895968 seconds time elapsed + 15,934,362,915 cycles # 2.952 GHz + 44,425,195,468 instructions # 2.79 insn per cycle + 5.398578435 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 533) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_tt.mad/SubProcesses/P1_gg_ttx/build.none_f_inl0_hrd1/runTest_cpp.exe @@ -117,15 +117,15 @@ Process = SIGMA_MSSM_SLHA2_GG_TTX_CPP [gcc 11.3.1] [inlineHe Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('sse4': SSE4.2, 128bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 5.328908e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 5.798682e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 5.798682e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 5.343645e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 5.827664e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 5.827664e+05 ) sec^-1 MeanMatrixElemValue = ( 2.072937e+00 +- 3.361544e-03 ) GeV^0 -TOTAL : 2.053409 sec +TOTAL : 2.047786 sec INFO: No Floating Point Exceptions have been reported - 6,061,427,520 cycles # 2.945 GHz - 17,076,312,832 instructions # 2.82 insn per cycle - 2.059026016 seconds time elapsed + 6,058,825,095 cycles # 2.952 GHz + 17,075,922,771 instructions # 2.82 insn per cycle + 2.052915572 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 2862) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_tt.mad/SubProcesses/P1_gg_ttx/build.sse4_f_inl0_hrd1/runTest_cpp.exe @@ -149,15 +149,15 @@ Process = SIGMA_MSSM_SLHA2_GG_TTX_CPP [gcc 11.3.1] [inlineHe Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('avx2': AVX2, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 6.019252e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 6.594125e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 6.594125e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 6.083653e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 6.669156e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 6.669156e+05 ) sec^-1 MeanMatrixElemValue = ( 2.072967e+00 +- 3.361967e-03 ) GeV^0 -TOTAL : 1.827330 sec +TOTAL : 1.807992 sec INFO: No Floating Point Exceptions have been reported - 5,036,041,688 cycles # 2.749 GHz - 10,223,391,747 instructions # 2.03 insn per cycle - 1.833165934 seconds time elapsed + 5,026,811,725 cycles # 2.774 GHz + 10,224,011,314 instructions # 2.03 insn per cycle + 1.813053215 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 3906) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_tt.mad/SubProcesses/P1_gg_ttx/build.avx2_f_inl0_hrd1/runTest_cpp.exe @@ -181,15 +181,15 @@ Process = SIGMA_MSSM_SLHA2_GG_TTX_CPP [gcc 11.3.1] [inlineHe Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('512y': AVX512, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 6.156943e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 6.756865e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 6.756865e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 6.174767e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 6.772193e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 6.772193e+05 ) sec^-1 MeanMatrixElemValue = ( 2.072967e+00 +- 3.361967e-03 ) GeV^0 -TOTAL : 1.789449 sec +TOTAL : 1.781471 sec INFO: No Floating Point Exceptions have been reported - 4,972,642,094 cycles # 2.772 GHz - 9,995,367,434 instructions # 2.01 insn per cycle - 1.795052964 seconds time elapsed + 4,966,627,592 cycles # 2.781 GHz + 9,994,873,985 instructions # 2.01 insn per cycle + 1.786602286 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 3805) (512y: 2) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_tt.mad/SubProcesses/P1_gg_ttx/build.512y_f_inl0_hrd1/runTest_cpp.exe @@ -213,15 +213,15 @@ Process = SIGMA_MSSM_SLHA2_GG_TTX_CPP [gcc 11.3.1] [inlineHe Workflow summary = CPP:FLT+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK FP precision = FLOAT (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[16] ('512z': AVX512, 512bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 4.670992e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 5.000057e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 5.000057e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 4.669971e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 4.998509e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 4.998509e+05 ) sec^-1 MeanMatrixElemValue = ( 2.072967e+00 +- 3.361967e-03 ) GeV^0 -TOTAL : 2.331763 sec +TOTAL : 2.329966 sec INFO: No Floating Point Exceptions have been reported - 4,369,500,962 cycles # 1.870 GHz - 8,444,287,674 instructions # 1.93 insn per cycle - 2.337616992 seconds time elapsed + 4,364,868,939 cycles # 1.870 GHz + 8,446,067,590 instructions # 1.94 insn per cycle + 2.334969199 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 2744) (512y: 4) (512z: 2754) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_tt.mad/SubProcesses/P1_gg_ttx/build.512z_f_inl0_hrd1/runTest_cpp.exe diff --git a/epochX/cudacpp/tput/logs_susyggtt_mad/log_susyggtt_mad_m_inl0_hrd0.txt b/epochX/cudacpp/tput/logs_susyggtt_mad/log_susyggtt_mad_m_inl0_hrd0.txt index 9029ad668b..93ece29df2 100644 --- a/epochX/cudacpp/tput/logs_susyggtt_mad/log_susyggtt_mad_m_inl0_hrd0.txt +++ b/epochX/cudacpp/tput/logs_susyggtt_mad/log_susyggtt_mad_m_inl0_hrd0.txt @@ -40,7 +40,7 @@ make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_tt.mad/SubProcesses/P1_gg_ttx' -DATE: 2024-10-06_09:57:23 +DATE: 2024-10-06_14:13:17 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: ========================================================================= @@ -49,15 +49,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_MSSM_SLHA2_GG_TTX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=0] Workflow summary = CUD:MIX+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 4.278122e+07 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 9.299718e+07 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 9.972605e+07 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 4.676389e+07 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 9.500057e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.010573e+08 ) sec^-1 MeanMatrixElemValue = ( 2.072848e+00 +- 3.360985e-03 ) GeV^0 -TOTAL : 0.535533 sec +TOTAL : 0.525492 sec INFO: No Floating Point Exceptions have been reported - 2,218,013,615 cycles # 2.871 GHz - 3,167,587,965 instructions # 1.43 insn per cycle - 0.830721869 seconds time elapsed + 2,179,870,995 cycles # 2.873 GHz + 3,103,997,010 instructions # 1.42 insn per cycle + 0.815696217 seconds time elapsed runNcu /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_tt.mad/SubProcesses/P1_gg_ttx/build.cuda_m_inl0_hrd0/check_cuda.exe -p 2048 256 1 ==PROF== Profiling "sigmaKin": launch__registers_per_thread 214 ==PROF== Profiling "sigmaKin": sm__sass_average_branch_targets_threads_uniform.pct 100% @@ -85,15 +85,15 @@ Process = SIGMA_MSSM_SLHA2_GG_TTX_CPP [gcc 11.3.1] [inlineHe Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/none+NAVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) Internal loops fptype_sv = SCALAR ('none': ~vector[1], no SIMD) -EvtsPerSec[Rmb+ME] (23) = ( 1.807535e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.852925e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.852925e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.806988e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.852521e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.852521e+05 ) sec^-1 MeanMatrixElemValue = ( 2.072848e+00 +- 3.360985e-03 ) GeV^0 -TOTAL : 5.910224 sec +TOTAL : 5.910240 sec INFO: No Floating Point Exceptions have been reported - 17,388,420,068 cycles # 2.940 GHz - 46,077,588,135 instructions # 2.65 insn per cycle - 5.916245730 seconds time elapsed + 17,371,441,694 cycles # 2.937 GHz + 46,072,970,411 instructions # 2.65 insn per cycle + 5.915521321 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 622) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_tt.mad/SubProcesses/P1_gg_ttx/build.none_m_inl0_hrd0/runTest_cpp.exe @@ -117,15 +117,15 @@ Process = SIGMA_MSSM_SLHA2_GG_TTX_CPP [gcc 11.3.1] [inlineHe Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[2] ('sse4': SSE4.2, 128bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 3.226882e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.387878e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.387878e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 3.224599e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.385700e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.385700e+05 ) sec^-1 MeanMatrixElemValue = ( 2.072848e+00 +- 3.360985e-03 ) GeV^0 -TOTAL : 3.360909 sec +TOTAL : 3.359088 sec INFO: No Floating Point Exceptions have been reported - 9,940,043,952 cycles # 2.953 GHz - 27,598,360,403 instructions # 2.78 insn per cycle - 3.367569953 seconds time elapsed + 9,899,358,837 cycles # 2.943 GHz + 27,589,824,139 instructions # 2.79 insn per cycle + 3.364658660 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 2581) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_tt.mad/SubProcesses/P1_gg_ttx/build.sse4_m_inl0_hrd0/runTest_cpp.exe @@ -149,15 +149,15 @@ Process = SIGMA_MSSM_SLHA2_GG_TTX_CPP [gcc 11.3.1] [inlineHe Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('avx2': AVX2, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 5.038546e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 5.426797e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 5.426797e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 5.068926e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 5.463695e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 5.463695e+05 ) sec^-1 MeanMatrixElemValue = ( 2.072848e+00 +- 3.360985e-03 ) GeV^0 -TOTAL : 2.194996 sec +TOTAL : 2.176587 sec INFO: No Floating Point Exceptions have been reported - 6,084,814,623 cycles # 2.765 GHz - 12,511,133,896 instructions # 2.06 insn per cycle - 2.201688699 seconds time elapsed + 6,029,594,261 cycles # 2.765 GHz + 12,489,346,153 instructions # 2.07 insn per cycle + 2.182015832 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 2776) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_tt.mad/SubProcesses/P1_gg_ttx/build.avx2_m_inl0_hrd0/runTest_cpp.exe @@ -181,15 +181,15 @@ Process = SIGMA_MSSM_SLHA2_GG_TTX_CPP [gcc 11.3.1] [inlineHe Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('512y': AVX512, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 5.589922e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 6.068248e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 6.068248e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 5.578303e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 6.052444e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 6.052444e+05 ) sec^-1 MeanMatrixElemValue = ( 2.072848e+00 +- 3.360985e-03 ) GeV^0 -TOTAL : 1.988387 sec +TOTAL : 1.986535 sec INFO: No Floating Point Exceptions have been reported - 5,540,380,764 cycles # 2.778 GHz - 11,938,541,192 instructions # 2.15 insn per cycle - 1.995322896 seconds time elapsed + 5,524,507,194 cycles # 2.775 GHz + 11,925,746,101 instructions # 2.16 insn per cycle + 1.991739584 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 2521) (512y: 146) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_tt.mad/SubProcesses/P1_gg_ttx/build.512y_m_inl0_hrd0/runTest_cpp.exe @@ -213,15 +213,15 @@ Process = SIGMA_MSSM_SLHA2_GG_TTX_CPP [gcc 11.3.1] [inlineHe Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('512z': AVX512, 512bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 3.615006e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.807457e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.807457e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 3.622403e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.818212e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.818212e+05 ) sec^-1 MeanMatrixElemValue = ( 2.072848e+00 +- 3.360985e-03 ) GeV^0 -TOTAL : 3.015683 sec +TOTAL : 3.002982 sec INFO: No Floating Point Exceptions have been reported - 5,630,115,254 cycles # 1.863 GHz - 8,130,918,173 instructions # 1.44 insn per cycle - 3.022730001 seconds time elapsed + 5,596,074,707 cycles # 1.861 GHz + 8,112,022,621 instructions # 1.45 insn per cycle + 3.008304266 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 1671) (512y: 126) (512z: 1865) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_tt.mad/SubProcesses/P1_gg_ttx/build.512z_m_inl0_hrd0/runTest_cpp.exe diff --git a/epochX/cudacpp/tput/logs_susyggtt_mad/log_susyggtt_mad_m_inl0_hrd1.txt b/epochX/cudacpp/tput/logs_susyggtt_mad/log_susyggtt_mad_m_inl0_hrd1.txt index 44aa1a6a94..1f58d7323c 100644 --- a/epochX/cudacpp/tput/logs_susyggtt_mad/log_susyggtt_mad_m_inl0_hrd1.txt +++ b/epochX/cudacpp/tput/logs_susyggtt_mad/log_susyggtt_mad_m_inl0_hrd1.txt @@ -40,7 +40,7 @@ make[1]: Entering directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp make[1]: Nothing to be done for 'all'. make[1]: Leaving directory '/data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_tt.mad/SubProcesses/P1_gg_ttx' -DATE: 2024-10-06_09:57:48 +DATE: 2024-10-06_14:13:42 On itscrd90.cern.ch [CPU: Intel(R) Xeon(R) Silver 4216 CPU] [GPU: 1x Tesla V100S-PCIE-32GB]: ========================================================================= @@ -49,15 +49,15 @@ INFO: The following Floating Point Exceptions will cause SIGFPE program aborts: Process = SIGMA_MSSM_SLHA2_GG_TTX_CUDA [nvcc 12.0.140 (gcc 11.3.1)] [inlineHel=0] [hardcodePARAM=1] Workflow summary = CUD:MIX+THX:CURDEV+RMBDEV+MESDEV/none+NAVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) -EvtsPerSec[Rmb+ME] (23) = ( 4.308177e+07 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 9.314026e+07 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 9.965515e+07 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 4.673424e+07 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 9.488015e+07 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.009479e+08 ) sec^-1 MeanMatrixElemValue = ( 2.072848e+00 +- 3.360985e-03 ) GeV^0 -TOTAL : 0.530653 sec +TOTAL : 0.527649 sec INFO: No Floating Point Exceptions have been reported - 2,220,013,015 cycles # 2.891 GHz - 3,185,773,009 instructions # 1.44 insn per cycle - 0.824701846 seconds time elapsed + 2,222,247,787 cycles # 2.872 GHz + 3,175,400,593 instructions # 1.43 insn per cycle + 0.832433547 seconds time elapsed runNcu /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_tt.mad/SubProcesses/P1_gg_ttx/build.cuda_m_inl0_hrd1/check_cuda.exe -p 2048 256 1 ==PROF== Profiling "sigmaKin": launch__registers_per_thread 212 ==PROF== Profiling "sigmaKin": sm__sass_average_branch_targets_threads_uniform.pct 100% @@ -85,15 +85,15 @@ Process = SIGMA_MSSM_SLHA2_GG_TTX_CPP [gcc 11.3.1] [inlineHe Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/none+NAVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) Internal loops fptype_sv = SCALAR ('none': ~vector[1], no SIMD) -EvtsPerSec[Rmb+ME] (23) = ( 1.857128e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 1.905464e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 1.905464e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 1.861686e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 1.909888e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 1.909888e+05 ) sec^-1 MeanMatrixElemValue = ( 2.072848e+00 +- 3.360985e-03 ) GeV^0 -TOTAL : 5.753526 sec +TOTAL : 5.738703 sec INFO: No Floating Point Exceptions have been reported - 16,958,834,547 cycles # 2.945 GHz - 45,095,701,979 instructions # 2.66 insn per cycle - 5.759360611 seconds time elapsed + 16,935,741,187 cycles # 2.949 GHz + 45,092,665,978 instructions # 2.66 insn per cycle + 5.743960838 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 567) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_tt.mad/SubProcesses/P1_gg_ttx/build.none_m_inl0_hrd1/runTest_cpp.exe @@ -117,15 +117,15 @@ Process = SIGMA_MSSM_SLHA2_GG_TTX_CPP [gcc 11.3.1] [inlineHe Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/sse4+CXVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[2] ('sse4': SSE4.2, 128bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 3.365466e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.544754e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.544754e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 3.288167e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.454893e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.454893e+05 ) sec^-1 MeanMatrixElemValue = ( 2.072848e+00 +- 3.360985e-03 ) GeV^0 -TOTAL : 3.232551 sec +TOTAL : 3.295721 sec INFO: No Floating Point Exceptions have been reported - 9,533,065,833 cycles # 2.943 GHz - 26,273,852,197 instructions # 2.76 insn per cycle - 3.239846074 seconds time elapsed + 9,499,277,299 cycles # 2.878 GHz + 26,250,814,415 instructions # 2.76 insn per cycle + 3.301150355 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 2386) (avx2: 0) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_tt.mad/SubProcesses/P1_gg_ttx/build.sse4_m_inl0_hrd1/runTest_cpp.exe @@ -149,15 +149,15 @@ Process = SIGMA_MSSM_SLHA2_GG_TTX_CPP [gcc 11.3.1] [inlineHe Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/avx2+CXVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('avx2': AVX2, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 4.514012e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 4.821697e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 4.821697e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 4.500682e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 4.809631e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 4.809631e+05 ) sec^-1 MeanMatrixElemValue = ( 2.072848e+00 +- 3.360985e-03 ) GeV^0 -TOTAL : 2.435584 sec +TOTAL : 2.436969 sec INFO: No Floating Point Exceptions have been reported - 6,758,526,375 cycles # 2.768 GHz - 14,047,168,742 instructions # 2.08 insn per cycle - 2.442338814 seconds time elapsed + 6,719,638,924 cycles # 2.752 GHz + 14,030,407,169 instructions # 2.09 insn per cycle + 2.442293094 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 2895) (512y: 0) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_tt.mad/SubProcesses/P1_gg_ttx/build.avx2_m_inl0_hrd1/runTest_cpp.exe @@ -181,15 +181,15 @@ Process = SIGMA_MSSM_SLHA2_GG_TTX_CPP [gcc 11.3.1] [inlineHe Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/512y+CXVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[4] ('512y': AVX512, 256bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 4.791737e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 5.138604e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 5.138604e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 4.761028e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 5.103276e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 5.103276e+05 ) sec^-1 MeanMatrixElemValue = ( 2.072848e+00 +- 3.360985e-03 ) GeV^0 -TOTAL : 2.301242 sec +TOTAL : 2.308424 sec INFO: No Floating Point Exceptions have been reported - 6,403,253,635 cycles # 2.776 GHz - 13,529,712,107 instructions # 2.11 insn per cycle - 2.307614270 seconds time elapsed + 6,376,412,980 cycles # 2.757 GHz + 13,515,459,389 instructions # 2.12 insn per cycle + 2.313683857 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 2531) (512y: 302) (512z: 0) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_tt.mad/SubProcesses/P1_gg_ttx/build.512y_m_inl0_hrd1/runTest_cpp.exe @@ -213,15 +213,15 @@ Process = SIGMA_MSSM_SLHA2_GG_TTX_CPP [gcc 11.3.1] [inlineHe Workflow summary = CPP:MIX+CXS:CURHST+RMBHST+MESHST/512z+CXVBRK FP precision = MIXED (NaN/abnormal=0, zero=0) Internal loops fptype_sv = VECTOR[8] ('512z': AVX512, 512bit) [cxtype_ref=YES] -EvtsPerSec[Rmb+ME] (23) = ( 3.627313e+05 ) sec^-1 -EvtsPerSec[MatrixElems] (3) = ( 3.823087e+05 ) sec^-1 -EvtsPerSec[MECalcOnly] (3a) = ( 3.823087e+05 ) sec^-1 +EvtsPerSec[Rmb+ME] (23) = ( 3.607877e+05 ) sec^-1 +EvtsPerSec[MatrixElems] (3) = ( 3.798879e+05 ) sec^-1 +EvtsPerSec[MECalcOnly] (3a) = ( 3.798879e+05 ) sec^-1 MeanMatrixElemValue = ( 2.072848e+00 +- 3.360985e-03 ) GeV^0 -TOTAL : 3.002431 sec +TOTAL : 3.013448 sec INFO: No Floating Point Exceptions have been reported - 5,614,669,392 cycles # 1.866 GHz - 9,218,497,811 instructions # 1.64 insn per cycle - 3.009264991 seconds time elapsed + 5,603,771,116 cycles # 1.857 GHz + 9,209,433,254 instructions # 1.64 insn per cycle + 3.018845208 seconds time elapsed =Symbols in CPPProcess_cpp.o= (~sse4: 0) (avx2: 1456) (512y: 212) (512z: 2059) ------------------------------------------------------------------------- runTest /data/avalassi/GPU2023/madgraph4gpuX/epochX/cudacpp/susy_gg_tt.mad/SubProcesses/P1_gg_ttx/build.512z_m_inl0_hrd1/runTest_cpp.exe
Graph Cross-Section
G1 47.1 0.222 98.304 3292.0 69.9 46.82 0.388 16.384 5181.0 111
G2