Design

The following sections provide a big picture of Lima2 Design.

Library Layers

Lima2 is composed of 3 main layers that are further documented in their own sections:

  • the Camera layer is an abstraction layer that implement the detector specific code

  • the Processing layer control is a tasks system and image processing algorithms library used to build the processing pipeline

  • the Control layer, built on top of the previous two, is in charge of preparing and running the acquisition and building and executing the processing pipeline.

Big picture of Lima2

Distributed computing

High performance detectors generate more data than a single acquisition backend can handle. The solution consists in running multiple coordinated acquisition backends, each backend supporting only part of dataflow.

Topologies

Multiple topologies and frame dispatching strategies are to be considered for different uses cases.

Single computer

Use case: Legacy cameras.

_images/topologies_single.png

Distributed Processing

Use case: PCO Edge / Dectris Eiger2.

_images/topologies_manyproc.png

Distributed Acquisition with partial frame dispatching

Use case: PSI Eiger / Jungfrau.

_images/topologies_partial.png

Distributed Acquisition with full frame dispatching

Use case: ESRF RASHPA / Smartpix.

_images/topologies_full.png

MPI Introduction

Message Passing Interface (MPI) is a standardized message-passing library interface specification.

There are a lot of tutorials on MPI such as this excellent Introduction to distributed computing with MPI,only the basic concepts are recap in the following paragraphs.

The message passing programming model

An MPI application is a group of autonomous processes, each executing its own code written in a classic language (C++, Python). All the program variables are private and reside in the local memory of each process. Each process has the possibility of executing different parts of a program. A variable is exchanged between two or several processes via a programmed call to specific MPI subroutines.

Multiple Data Single Program

In this configuration, every processes in the cluster run the same program. The program can have multiple facets or roles which are assigned according to process rank.

Multiple Data Multiple Program

MPI also have the ability run tasks that can be a different program that executes different part of the workflow. For example, one task can be a C++ program and another a python.

Communications

When a program is ran with MPI all the processes are grouped in what we call a communicator. Every communication is linked to a communicator, allowing the communication to reach different processes.

Point to Point

TODO.

_images/mpi_send_rcv.png

MPI defines four communication modes:

  • Synchronous

  • Buffered

  • Ready

  • Standard

that exists in both blocking/nonblocking variants.

Collective

TODO.

MPI Streams

MPIStream is an extension of MPI that supports data streams, a sequence of data flowing between source and destination processes. Streaming is widely used for signal, image and video processing for its efficiency in pipelining and effectiveness in reducing demand for memory.

Resources

Papers

MPI Streams for HPC Applications

Talks

Talks that influenced the design of Lima2.

Image processing

Boost Generic Image Library (GIL)

Data structures

CppCon 2014: Herb Sutter “Lock-Free Programming (or, Juggling Razor Blades), Part I” CppCon 2014: Herb Sutter “Lock-Free Programming (or, Juggling Razor Blades), Part II”

General techniques

NDC 2017: Sean Parent “Better Code: Concurrency” C++Now 2018: Louis Dionne “Runtime Polymorphism: Back to the Basics”

Meta programming

C++Now 2017: Louis Dionne “Fun with Boost.Hana”

Types

CppCon 2016: Ben Deane “Using Types Effectively” Zero overhead utilities for more type safety