Keyword: Linux
Paper Title Other Keywords Page
MOAR02 Modernizing Digital Video Systems at the National Ignition Facility (NIF): Success Stories, Open Challenges and Future Directions Windows, controls, hardware, software 26
  • V.K. Gopalan, A.I. Barnes, G.K. Brunton, J. Dixon, C.M. Estes, M. Fedorov, M.S. Flegel, B. Hackel, D.J. Koning, S.L. Townsend, D. Tucker, J.L. Vaher
    LLNL, Livermore, USA
  Funding: This work was performed under the auspices of the U.S. Department of Energy by Lawrence Livermore National Laboratory under Contract DE-AC52-07NA27344.
The National Ignition Facility (NIF), the world’s most energetic laser, completed a multi-year project for migrating control software platforms from Ada to Java in 2019. Following that work, a technology refresh of NIF’s Digital Video (DVID) systems was identified as the next important step. The DVIDs were facing long-term maintenance risk due to its obsolete Window XP platform, with over 500 computers to be individually upgraded and patched, 24 camera types with a variety of I/O interfaces and proprietary drivers/software with their licensing needs. In this presentation, we discuss how we leveraged the strengths of NIF’s distributed, cross platform architecture and our system migration expertise to migrate the DVID platforms to diskless clients booting off a single purpose-built immutable Linux image, and replacing proprietary camera drivers with open-source drivers. The in-place upgrades with well-defined fallback strategies ensured minimal impact to the continuous 24/7 shot operations. We will also present our strategy for continuous build, test, and release of the Linux OS image to keep up with future security patches and package upgrades.
LLNL IM Document Release Number: LLNL-ABS-822092
slides icon Slides MOAR02 [0.872 MB]  
DOI • reference for this paper ※  
About • Received ※ 08 October 2021       Revised ※ 14 October 2021       Accepted ※ 11 November 2021       Issue date ※ 28 February 2022
Cite • reference for this paper using ※ BibTeX, ※ LaTeX, ※ Text/Word, ※ RIS, ※ EndNote (xml)  
MOPV020 Digitisation of the Analogue Waveform System at ISIS controls, timing, real-time, diagnostics 169
  • W.A. Frank, B.R. Aljamal, R.A. Washington
    STFC/RAL/ISIS, Chilton, Didcot, Oxon, United Kingdom
  Funding: UKRI/STFC
The Analogue Waveform System (AWS) at the ISIS Neutron and Muon Source is a distributed system that allows operators to select and monitor analogue waveforms from equipment throughout the facility on oscilloscopes in the Main Control Room (MCR). These signals originate from key accelerator systems in the linear accelerator and synchrotron such as the ion source, magnets, beam diagnostics, and radio frequency (RF) systems. Historical data for ISIS is available on the control system for many relevant channels. However, at present, to avoid disrupting the oscilloscope displays in the MCR, only an hourly image capture of the AWS waveforms is stored. This is largely inadequate for potential data-intensive applications such as anomaly detection, predictive maintenance, post-mortem analysis, or (semi-)automated machine setup, optimization, and control. To address this, a new digital data acquisition (DAQ) system is under development based on the principle of large channel count, simultaneous DAQ. This paper details the proposed architecture of the system and the results of initial prototyping, testing, and commissioning.
DOI • reference for this paper ※  
About • Received ※ 08 October 2021       Revised ※ 21 October 2021       Accepted ※ 16 December 2021       Issue date ※ 04 February 2022
Cite • reference for this paper using ※ BibTeX, ※ LaTeX, ※ Text/Word, ※ RIS, ※ EndNote (xml)  
MOPV044 Lessons Learned Moving from Pharlap to Linux RT timing, network, hardware, Windows 257
  • C. Charrondière, O.O. Andreassen, D. Sierra-Maíllo Martínez, J. Tagg, T. Zilliox
    CERN, Meyrin, Switzerland
  The start of the Advanced Proton Driven Plasma Wakefield Acceleration Experiment (AWAKE) facility at CERN in 2016 came with the need for a continuous image acquisition system. The international scientific collaboration responsible for this project requested low and high resolution acquisition at a capture rate of 10Hz and 1 Hz respectively. To match these requirements, GigE digital cameras were connected to a PXI system running PharLap, a real-time operating system, using dual port 1GB/s network cards. With new requirements for a faster acquisition with higher resolution, it was decided to add 10GB/s network cards and a Network Attached Storage (NAS) directly connected to the PXI system to avoid saturating the network. There was also a request to acquire high-resolution images on several cameras during a limited duration, typically 30 seconds, in a burst acquisition mode. To comply with these new requirements PharLap had to be abandoned and replaced with Linux RT. This paper describes the limitation of the PharLap system and the lessons learned during the transition to Linux RT. We will show the improvement of CPU stability and data throughput reached.  
poster icon Poster MOPV044 [0.525 MB]  
DOI • reference for this paper ※  
About • Received ※ 08 October 2021       Revised ※ 18 October 2021       Accepted ※ 20 November 2021       Issue date ※ 28 February 2022
Cite • reference for this paper using ※ BibTeX, ※ LaTeX, ※ Text/Word, ※ RIS, ※ EndNote (xml)  
MOPV050 DevOps and CI/CD for WinCC Open Architecture Applications and Frameworks framework, controls, software, Windows 281
  • R.P.I. Silvola
    CERN, Meyrin, Switzerland
  • L. Sargsyan
    ANSL, Yerevan, Armenia
  This paper presents the Continuous Integration and Continuous Deployment (CI/CD) tool chain for WinCC Open Architecture applications and frameworks developed at CERN, enabling a DevOps oriented approach of working. By identifying common patterns and time consuming procedures, and by agreeing on standard repository structures, naming conventions and tooling, we have gained a turnkey solution which automates the compilation of binaries and generation of documentation, thus guaranteeing they are up to date and match the source code in the repository. The pipelines generate deployment-ready software releases, which pass through both static code analysis and unit tests before automatically being deployed to short and long-term repositories. The tool chain leverages industry standard technologies, such as GitLab, Docker and Nexus. The technologies chosen for the tool chain are well understood and have a long, solid track record, reducing the effort in maintenance and potential long term risk. The setup has reduced the expert time needed for testing and releases, while generally improving the release quality.  
poster icon Poster MOPV050 [0.923 MB]  
DOI • reference for this paper ※  
About • Received ※ 08 October 2021       Revised ※ 13 October 2021       Accepted ※ 23 February 2022       Issue date ※ 11 March 2022
Cite • reference for this paper using ※ BibTeX, ※ LaTeX, ※ Text/Word, ※ RIS, ※ EndNote (xml)  
THBL01 Control System Management and Deployment at MAX IV software, controls, TANGO, GUI 819
  • B. Bertrand, A. Freitas, V. Hardion
    MAX IV Laboratory, Lund University, Lund, Sweden
  The control systems of big research facilities like synchrotron are composed of many different hardware and software parts. Deploying and maintaining such systems require proper workflows and tools. MAX IV has been using Ansible to manage and deploy its full control system, both software and infrastructure, for quite some time with great success. All required software (i.e. tango devices, GUIs…) used to be packaged as RPMs (Red Hat Package Manager) making deployment and dependencies management easy. Using RPMs brings many advantages (big community, well tested packages, stability) but also comes with a few drawbacks, mainly the dependency to the release cycle of the Operating System. The Python ecosystem is changing quickly and using recent modules can become challenging with RPMs. We have been investigating conda as an alternative package manager. Conda is a popular open-source package, dependency and environment management system. This paper will describe our workflow and experience working with both package managers.  
slides icon Slides THBL01 [5.899 MB]  
DOI • reference for this paper ※  
About • Received ※ 10 October 2021       Accepted ※ 21 November 2021       Issue date ※ 12 February 2022  
Cite • reference for this paper using ※ BibTeX, ※ LaTeX, ※ Text/Word, ※ RIS, ※ EndNote (xml)  
THBL03 The State of Containerization in CERN Accelerator Controls controls, software, operation, hardware 829
  • R. Voirin, T. Oulevey, M. Vanden Eynden
    CERN, Geneva, Switzerland
  In industry, containers have dramatically changed the way system administrators deploy and manage applications. Developers are gradually switching from delivering monolithic applications to microservices. Using containerization solutions provides many advantages, such as: applications running in an isolated manner, decoupled from the operating system and its libraries; run-time dependencies, including access to persistent storage, are clearly declared. However, introducing these new techniques requires significant modifications of existing computing infrastructure as well as a cultural change. This contribution will explore practical use cases for containers and container orchestration within the CERN Accelerator Controls domain. We will explore challenges that have been arising in this field for the past two years and technical choices that we have made to tackle them. We will also outline the foreseen future developments.  
slides icon Slides THBL03 [0.863 MB]  
DOI • reference for this paper ※  
About • Received ※ 08 October 2021       Revised ※ 24 October 2021       Accepted ※ 06 January 2022       Issue date ※ 28 February 2022
Cite • reference for this paper using ※ BibTeX, ※ LaTeX, ※ Text/Word, ※ RIS, ※ EndNote (xml)  
THPV033 Reusable Real-Time Software Components for the SPS Low Level RF Control System hardware, software, controls, interface 939
  • M. Sumiński, K. Adrianek, B. Bielawski, A.C. Butterworth, J. Egli, G. Hagmann, P. Kuzmanović, S. Novel González, A. Rey, A. Spierer
    CERN, Geneva, Switzerland
  In 2021 the Super Proton Synchrotron has been recommissioned after a complete renovation of its low level RF system (LLRF). The new system has largely moved to digital signal processing implemented as a set of functional blocks (IP cores) in Field Programmable Gate Arrays (FPGAs) with associated software to control them. Some of these IP cores provide generic functionalities such as timing, function generation, data resampling and signal acquisition, and are reused in several components, with a potential application in other accelerators. To take full advantage of the modular approach, IP core flexibility must be complemented by the software stack. In this paper we present steps we have taken to reach this goal from the software point of view, and describe the custom tools and procedures used to implement the various software layers.  
poster icon Poster THPV033 [1.234 MB]  
DOI • reference for this paper ※  
About • Received ※ 09 October 2021       Accepted ※ 25 February 2022       Issue date ※ 28 February 2022  
Cite • reference for this paper using ※ BibTeX, ※ LaTeX, ※ Text/Word, ※ RIS, ※ EndNote (xml)