References

The LDC software project aim at addressing the needs of LISA in phase A regarding simulation and data analysis pipelines.

Those needs have been indentified in several contexts:

  • in the LISA Data Challenge working group
  • in the LISA Science Group
  • in the LISA Data Processing group

The LDC software project (https://gitlab.in2p3.fr/LISA/LDC) is a python package, it provides common tools like orbits, waveforms, noise model subpackages, etc.

The data generation folder contains a set of python scripts used to run data generation pipelines and test them (content of their output and their performance).

From LISA-LCST-SGS-WPD-001

WP2 includes a number of work packages associated with data analysis tools in general, and in particular, those which form a common toolkit that can be used to assemble pipelines, produce higher level data products, analysis instrument performance, and extract scientific conclusions from LISA data. Wherever possible, effort should be made to produce common tools with broad enough interfaces to allow them to be used in multiple contexts, thus minimising duplication and easing maintenance and testing

From MLDC wiki

https://gitlab.in2p3.fr/stas/MLDC/wikis/home

Role and tasks of DPC in LDC:

  • DPC will host the astrophysical catalogues of GW sources (several realisations per astrophysical model).
  • DPC will be used to generate the simulated data set and store them with a dedicated database system.
  • DPC will provide the web-interface with description of data sets, how to get it, and people/groups working on this particular data set.
  • DPC will provide facilities to store and evaluate the results of the search methods.
  • DPC will provide the git-facilities for developing the search software packages.
  • DPC will point to the available computing facilities to run the tests.
  • DPC will provide collaboration facilities for the teleconferences, storing the documents, teleconference minutes.
  • DPC will provide pre-processed data to enable participants that don't want to develop full pipeline but just concentrate on a particular point; examples: data with the galaxy reduced, data with the high SNR sources removed. To do this the DPC will run the "best" available pipeline, to start with the one developed during the previous LDC and later the new methods.
  • DPC is cool

From LDPG WP spreadsheet

Objective 5: Supporting consortium activities

The goal is to support the consortium activities and coherently connect them. First frameworks will have to be set for simulation and LDC data generation, for calibration pipelines, for pre-processing pipeline (TDI & before) and for data analysis pipeline. To interface these activities, an implementation of a first data model for simulation and for the various data level will be needed. This should build a coherent end-to-end data flow that could be used for objective 3. We also should provide computing and storage ressources for the data analysis pipeline and for simulations via proto-DCC. The performances management tool (LIG) should also be supported.

Objective 3: Prototyping for benchmarking

The goal is to do benchmarking using available prototypes best representing what the DDPC will have to compute. This would include individual LDC pipeline, a first pipeline orchestration and simulations. To run this prototypes a preliminary frameworks with a preliminary common execution on first DCCs will be used. Regarding the data, various data models with versioning will be tested as well as a full tracking of configuration and data products.