KEYWORDS: Data acquisition, Telescopes, Atmospheric Cherenkov telescopes, Observatories, Databases, Gamma radiation, Data processing, Data analysis, Control systems, Reconstruction algorithms
With dozens of telescopes in both hemispheres, the Cherenkov Telescope Array Observatory (CTAO) will be the largest ground-based gamma-ray observatory and will offer extensive energy coverage from 20 GeV to 300 TeV. Its large effective area, wide field-of-view, rapid slewing capability, and exceptional sensitivity will make CTAO an essential instrument for the future of ground-based gamma-ray astronomy. Furthermore, its two arrays will send alerts on transient and variable phenomena (e.g., gamma-ray bursts, active galactic nuclei, gamma-ray binaries, and serendipitous sources) to maximise the scientific return. Effective and rapid communication with the community requires a reliable automated system to detect and issue candidate science alerts. This automation will be achieved by the Science Alert Generation (SAG) pipeline, a core system of the CTA Observatory. The SAG is part of the Array Control and Data Acquisition (ACADA) system. The SAG working group develops pipelines for data reconstruction, data quality monitoring, science monitoring, and real-time alert issuance to the Transients Handler system of ACADA. The SAG performs the first real-time scientific analysis during data acquisition. The system analyzes data on multiple time scales (from seconds to hours) and must issue candidate science alerts with 20 seconds of latency and at least half the CTAO nominal sensitivity. Dedicated, highly optimized software and hardware architectures must be designed and tested to satisfy these stringent requirements and manage trigger rates of tens of kHz from both arrays. In this work, we present the general architecture and current development status of the ACADA/SAG system.
The Cherenkov Telescope Array Observatory (CTAO) is the next-generation atmospheric Cherenkov gamma-ray project. CTAO will be deployed at two sites, one in the Northern and the other in the Southern Hemisphere, containing telescopes of three different sizes for covering different energy domains. The commissioning of the first CTAO Large-sized Telescope (LST-1) is being finalized at the CTAO Northern site. Additional calibration and environmental monitoring instruments such as laser imaging detection and ranging (LIDAR) instruments and weather stations will support the telescope operations. The Array Control and Data Acquisition (ACADA) system is the central element for onsite CTAO operations. ACADA controls, supervises, and handles the data generated by the telescopes and the auxiliary instruments. It will drive the efficient planning and execution of observations while handling the several Gb/s camera data generated by each CTAO telescope. The ACADA system contains the CTAO Science Alert Generation Pipeline – a real-time data processing and analysis pipeline, dedicated to the automatic generation of science alert candidates as data are being acquired. These science alerts, together with external alerts arriving from other scientific instruments, will be managed by the Transients Handler (TH) component. The TH informs the Short-term Scheduler of ACADA about interesting science alerts, enabling the modification of ongoing observations at sub-minute timescales. The capacity for such fast reactions – together with the fast movement of CTAO telescopes – makes CTAO an excellent instrument for studying high-impact astronomical transient phenomena. The ACADA software is based on the Alma Common Software (ACS) framework, and written in C++, Java, Python, and Javascript. The first release of the ACADA software, ACADA REL1, was finalized in July 2023, and integrated after a testing campaign with the LST-1 finalized in October 2023. This contribution describes the design and status of the ACADA software system.
The Cherenkov Telescope Array Observatory (CTAO) is the next-generation atmospheric Cherenkov gammaray Observatory. CTAO will be constructed on two sites, one array in the Northern and the other in the Southern hemisphere, containing telescopes of three different sizes, for covering different energy domains. To combine and orchestrate the different telescopes and auxiliary instruments (array elements), the Array Control and Data Acquisition (ACADA) system is the central element for the Observatory on-site operations: it controls, supervises, and handles the data generated by the array elements. Considering the criticality of the ACADA system for future Observatory operations, corresponding quality assurance provisions have been made at the different steps of the software development lifecycle, with focus on continuous integration and testing at all levels. To enable higher-level tests of the software deployed on a distributed system, an ACADA test cluster has been set up to facilitate testing and debugging of issues in a more realistic environment. Furthermore, a separate software integration and test cluster has also been established that allows for the off-site testing of the integrated software packages of ACADA and of the corresponding array elements. Here the software integration can be prepared, interfaces and interactions can be tested, and on-site procedures that are required later in the process can be checked beforehand, only limited by the simulation capabilities that are delivered as part of the software packages. Once preparations and testing with the off-site test cluster are completed, the integrated software can be deployed at the target site. The software packages and setup parameters are kept under configuration control at all stages, and deployment steps are documented to ensure that installations are reproducible. This methodology has been applied for the first time in the context of the integration of ACADA with the first CTAO Large-sized Telescope (LST-1) in October 2023.
KEYWORDS: Atmospheric Cherenkov telescopes, Data acquisition, Cameras, Control systems, Telescopes, Interferometry, Data centers, Software development, Computer architecture, Quality systems
The ASTRI Mini-Array is an international collaboration led by the Italian National Institute for Astrophysics. This project aims to construct and operate an array of nine Imaging Atmospheric Cherenkov Telescopes to study gamma-ray sources at very high energy (TeV) and perform stellar intensity interferometry observations. We describe the software architecture and the technologies used to implement the Online Observation Quality System (OOQS) for the ASTRI Mini-Array project. The OOQS aims to execute data quality checks on the data acquired in real-time by the Cherenkov cameras and intensity interferometry instruments, and provides feedback to both the Central Control System and the Operator about abnormal conditions detected. The OOQS can notify other sub-systems, triggering their reaction to promptly correct anomalies. The results from the data quality analyses (e.g. camera plots, histograms, tables, and more) are stored in the Quality Archive for further investigation and they are summarised in reports available to the Operator. Once the OOQS results are stored, the operator can visualize them using the Human Machine Interface. The OOQS is designed to manage the high data rate generated by the instruments (up to 4.5 GB/s) and received from the Array Data Acquisition System through the Kafka service. The data are serialized and deserialized during the transmission using the Avro framework. The Slurm workload scheduler executes the analyses exploiting key features such as parallel analyses and scalability.
Gamma-Flash is an Italian project funded by the Italian Space Agency (ASI) and led by the National Institute for Astrophysics (INAF), devoted to the observation and study of high-energy phenomena, such as terrestrial gamma-ray flashes and gamma-ray glows produced in Earth’s atmosphere during thunderstorms. The project represents the ground-based supplement to the work of the ASI AGILE satellite in this particular field. This contribution presents the architecture of the Gamma-Flash data pipeline placed at the Osservatorio Climatico “O. Vittori” on the top of Mt. Cimone (2165 m a.s.l., Northern-Central Italy). It consists of RedPitaya ARM-FPGA boards designed for acquiring events at different energies from scintillator crystals coupled to photomultiplier tubes, and a main computer that executes a real-time software pipeline. The software performs several data processing steps, data acquisition, data reduction level, algorithms for waveform selection, and finally it produces the cumulative energy spectrum of the gamma radiation collected by the photomultipliers. Data is stored in different layers, each with a different purpose, and it is available to the scientific community as HDF5 files. The pipeline has a modular architecture to provide good maintenance and flexibility, allowing for easy extensions in the future. A specific subset of data is stored in a database connected to a real-time graphical dashboard for quick-look analysis, showing the acquisition products and the environmental telemetry data.
Access to the requested content is limited to institutions that have purchased or subscribe to SPIE eBooks.
You are receiving this notice because your organization may not have SPIE eBooks access.*
*Shibboleth/Open Athens users─please
sign in
to access your institution's subscriptions.
To obtain this item, you may purchase the complete book in print or electronic format on
SPIE.org.
INSTITUTIONAL Select your institution to access the SPIE Digital Library.
PERSONAL Sign in with your SPIE account to access your personal subscriptions or to use specific features such as save to my library, sign up for alerts, save searches, etc.