Quriosity's ambition is to extend the application horizon of quantum information science by addressing novel questions positioned at the intersection between theoretical research in quantum information and the engineering of quantum devices, with a focus on approaches combining digital and quantum photonics technologies.

The overarching goal of the project-team will be to push forward our ability to harness and exploit high-dimensional complex quantum systems for quantum information processing and quantum communications purposes.

Leveraging a dual approach combining fundamental research in quantum information with quantum photonics expertise, Quriosity will strive to take advantage of and develop strong synergies with the unique quantum ecosystem of Saclay and to pursue objectives that have the potential to bring radical advances to several application domains of quantum technologies, ranging from cryptography to computing:

The research program that we aim to lead in the Quriosity project-team intends to embrace a relatively wide area of theoretical questions, ranging from quantum cryptography, that we ambition to combine with complexity-based schemes and establish as a framework to enhance hardware security, to the mathematical foundations of quantum information and quantum computing. Conversely, we also intend to develop research capable of leveraging photonics and digital information processing technologies to design systems capable of producing high-dimensional and controllable quantum states of light in order to push forward the frontiers of quantum information processing advantage.

This axis aims to identify and solve frontier research topics in quantum cryptography, from two main perspectives. First by exploring the interplay between security models - including computational ones - and theoretical quantum cryptography, allowing to build protocols with stronger security properties and lesser resource requirements. Second by laying a special emphasis on interplay between quantum cryptography and hardware security, with the need to develop extended techniques for quantum cryptographic hardware security certification, but also the idea to strengthen hardware security and its resilience to information leakage by resorting to quantum cryptographic constructions.

We proposed in 2015 a security model that we later coined a Quantum Computational Timelock (QCT) security model. It consists in assuming that computationally secure encryption may only be broken after time much longer than the coherence time of quantum memories available at the time of protocol execution. The QCT security model opens the possibility to propose new quantum cryptographic constructions and in particular to make use of encoding and security proof techniques that strongly depart from “traditional” quantum conjugate coding that is a central ingredient in most quantum cryptographic protocols.

The QCT security model opens towards a rich variety of fascinating questions, that we have certainly not all identified. In the coming years we intend to push forward the theoretical analysis of several of these questions, that relate to the computational frontiers of quantum cryptography. One ongoing direction consists in studying key agreement constructions whose security can be reduced to distributed computational problems that exhibit an exponential separation in terms of quantum or classical communication complexity.

As an alternative way to build secure protocols in the QCT model, we also intend to investigate pseudo-random quantum states, which can be seen as a computational variant of a

Device-independent cryptography allows one to perform quantum cryptography with reduced or even no trust assumptions on the quantum hardware. It remains a challenge experimentally and pushing the performance (in terms of key rate, or trust reduction) of device-independent cryptography defines an active research frontier for quatum cryptography. Recent implementations of DI-QKD147, 57, 57 have shown that whilst it is now feasible, it has a relatively low rate and can only be executed over a short distance. By improving the theoretical methods for analyzing various protocols and security proofs and by improving the protocol design we can look to boost the rates of these protocols and push them towards a more viable technology. Examples of such improvements include protocol design modifications 51, 40 and improved methods to calculate rates 28, 29, 50. Our goals are to develop better designed protocols and security proofs (assessing their performance in experiments) and to investigate the fundamental limitations of DI protocol rates Overall pushing the practicality of DI forwards and improving our understanding of its limitations.

As a complementary line of research we will also investigate prospects of semi-device-independent protocols as a viable near-term alternative to device-independent security. Proposed protocols rely on assumptions of system energy 49, dimension bounds and bounded distrust 52 amongst others. We will investigate alternative assumptions and derive resulting protocols to be analyzed and subsequently implemented. We will also apply the semi-device-independent framework to the problem of hardware verification, designing tests to establish that the hardware is functioning correctly whilst placing limited trust on the components.

We will also investigate some questions placed at the intersection between classical hardware security and quantum cryptography, namely how to prove the security of a cryptographic protocols when implemented using hardware, such as processors or storage, that may leak some of the security-sensitive information.

We intend to tackle leakage-resilience cryptography from a new viewpoint, that will consist in integrating quantum cryptographic constructions as a base layer within cryptographic systems, in order to obtain security guarantees even in presence of information leakage with strictly weaker assumptions than existing classical leakage-resilience protocols. We will first consider simple cryptographic protocols such as One-Time-Pad encryption or authentication protocols relying on Physically Uncloneable Functions PUFs. We intend for example to investigate how the use of hybrid classical-quantum cryptographic hardware, comprising quantum channels to interconnect processors or secure storage sites, can lead to cryptographic protocols with provable security under some realistic information leakage models.

40 year of quantum cryptography (QC) have lead to major theoretical and technological advances, with fundamental impact on the field of information security. Market adoption however remains limited, with major challenges that practical QC still needs to be overcome in order to become widely used in real-world applications. We identify in particular two main challenges: 1) cryptographic advantage, namely the design of protocols for which the use of QC in combination with classical cryptography gives a competitive edge over classical cryptography only; 2) security certification of quantum cryptographic implementations.
Quriosity intends to actively contribute to lift these barriers and to foster the development of real-world quantum cryptography and in particular to the uptake of a French and European industry. The development of a QC industry is indeed becoming an important topic, with strategic investments from leading scientific countries (China, Korea, Japan, UK, etc. ) including also notably the EU27 supporthing the EuroQCI initiative. On the other hand, the adoption of quantum cryptography for real-world application remains often considered with skepticism by representatives of the cybersecurity community, stressing the dire need of cross-disciplinary vision combining best-in-class classical and quantum cryptography expertise.

Regarding cryptographic advantage, our conviction is that one should not aim at constructions where quantum cryptography would just functionally replace classical cryptography, but on the other hand to identify applications where the use of QC combined with post-quantum cryptography (PQC) can present strict security gain over PQC alone.

Regarding security certification, it has become a central challenge in particular in the context of the EuroQCI initiative aiming at developing a pan-European quantum communication infrastructure, together with an industry, in the next 10 years It constitutes a complex task, requiring the collaboration of experts from different fields. In future years, we intend to tackle this question from different angles: on the theory side, we intend to propose a shift in the security objective towards everlasting security, and demonstrate how this can make the security certification of key establishment based on QKD combined with ephemeral post-quantum cryptography primitives much more tractable. On the system engineering side and in resonance with Section 3.2, we intend to identify and close implementation security gaps in modern CV-QKD systems relying on digital signal processing, notably the complex interplay between calibration procedure and finite-size security, but also between Nyquist pulse shaping and leakage.

Building a quantum processor that we could use to solve real-world problems with practical benefits might constitute one of the most burning scientific and technological challenges of the beginning of the 21st century. Very interestingly, recent results indicate that quantum optical circuits constitute a very promising approach for quantum information processing, in particularly high-dimensional linear optics systems, which can form a (weaker) non-universal quantum computing platform, and yet efficiently perform tasks intractable for a classical computer, such as Boson Sampling 17.

We will actively investigate new theoretical questions related to quantum information processing with high-dimensional photonic system, and their interplay with technology and experiments.

Quantum Key Distribution (QKD) systems are among the most advanced quantum communications technologies available today. QKD therefore provides an ideal platform to test novel system designs and validate quantum communication technology over real networks Leveraging essential features of modern optical communication systems, and in particular high sampling rates and digital signal processing 43, quantum coherent communications systems constitute a recent and promising route towards high-rates, highly integrated and cost-effective quantum communication systems. They rely on two central ingredients: -Spectrally efficient modulation formats and coherent detection, exploiting phase and intensity information and able to operate a very high rates (> GHz) even with shot-noise limited receivers. - Digital signal processing that takes advantage of the high sampling rates to digitally evaluate and compensate many impairments of the communications such as optical carrier phase noise or polarization mode dispersion, using dedicated algorithms.

In collaboration with Prof. Yves Jaouen from the GTO team of Telecom Paris,and working on a state-of-the-art experimental platform, Quriosity has designed and demonstrated for the first time DSP-enhanced quantum communications, with noise control performances that allow to successfully run QKD over metropolitan distances while being jointly deployed over classical coherent optical link 21. We have also filed a patent about this general concept and our inventive system design.

In the future, we then aim to leverage digital signal processing and machine learning (ML) techniques to characterize and mitigate noise in order to push further our ability to operate quantum communications over existing optical fibers, in coexistence with classical signals.

As a complementary line of research, we intend to theoretically study multimode quantum coherent communications using multimode shaping of the local oscillator, taking inspiration from 30. We also intend to explore the possibility to rely on CV multimode encoding as a way to experimentally implement new quantum cryptographic constructions in the hybrid quantum computational security models introduced in Section 3.1.

In collaboration with the teams of Nadia Belabas and Pascale Senellart at C2N and in the context of the ParisQCI project, we study how to combine high-dimensional photonic gates in the frequency domain, to efficiently synthesize high-dimensional unitary transformations. Leveraging on the possibility to parallelize single-qubit unitaries, that we have recently analyzed 5 we intend to study how such systems could be leveraged for optical quantum information processing, and in particular for quantum metrology. In the future, we will also investigate how to scale the platform to perform information processing with high-dimensional quantum states, opening the possibility to achieve quantum computational advantage, but also implementation routes for the hybrid quantum-computational cryptographic protocols in the QCT model, studied in Section 3.1

In collaboration with the team of Sylvain Gigan at ENS Ulm, and in the context of Francesco Mazzoncini's PhD that we co-supervise, we aim to use a multimode programmable linear circuit, built around a multimode fiber (cf. Figure 2) to perform some fundamental tests and demonstrations of quantum communication advantage, related to fundamental problems such as the Vector in a Subspace 48.

The prospects of this work are very promising: first they could lead to the first experimental demonstration of a exponential communication complexity gap between one-way quantum communication and two-way classical communications and may also open towards the possibility for experimentally robust Bell inequality violations 45, with applications for quantum cryptography and also in quantum computing.

Quantum information and computation are built upon the mathematical frameworks of
functional analysis and information theory. Developing our understanding of the mathematical underpinnings of these theories can in turn lead to new insights and applications. At Quriosity, one of our aims is to explore quantum information theory through the lens of the underlying mathematics.
In a nutshell, we will parallelly develop new analytic and numerical tools for the study of quantum entropic quantities and complex quantum systems made of spin or bosonic degrees of freedom. We will in turn consider these systems to design new, physically motivated models of noise-robust quantum computing.

Convex optimization concerns the optimization of convex functions over convex sets. This family of optimization problems has several particularly nice properties, including the guarantee of global optima, which makes them particularly appealing from both the perspective of the mathematics and the applications. They are widely applicable to many domains of science but in particular they arise rather naturally in the context of quantum theory as many of the relevant objects (states, channels and measurements) form convex sets.

We will aim to develop and apply techniques in convex optimization theory to problems within quantum information and quantum computing. Recent examples of our work in this area include 28, 29 where we developed semidefinite programming relaxations for entropic optimization problems relevant to device independent cryptography. Continuing this line of research we aim to extend these techniques to other entropic quantities beyond the relative entropy, for instance to the Petz and sandwiched families of Rényi divergences. We also have the ambitious goal of understanding and characterizing what classes of functions, relevant in the context of quantum theory, are amenable to such semidefinite programming approximations. In other words, what optimization problems in quantum information theory and quantum computing can we approximate?

A well-known example concerns strengthenings of the monotonicity of the relative entropy under the action of a quantum channel or a Markovian evolution known as strong data processing and modified logarithmic Sobolev inequalities. These fundamental inequalities are known to be hard to prove analytically, even for simple random walks on

Entropies are fundamental quantities in quantum information theory, obtaining operational meanings in terms of rates of various tasks 36. By improving our understanding of these quantities, we can in turn gain new insights into the various applications in which they appear.

For example, new chain rules for Rényi entropies 35 led to a versatile framework for cryptographic security proofs 20. The result, known as the entropy accumulation theorem, effectively gives sufficient conditions under which the entropy of a large system can be accurately described by the entropy of its individual systems. At QURIOSITY we aim to understand under which conditions does entropy accumulate in this manner? By understanding the minimal requirements for entropy to accumulate we can understand the minimal requirements under which a randomness based cryptographic protocol functions securely. Moreover, we aim to investigate the connection between the entropy accumulation theorem and the related works of the quantum probability estimation framework 56. This is an alternative method to break large entropies down into smaller quantities and reports several advantages over the entropy accumulation theorem. Understanding how advantages from one technique can be transferred to the other will lead to much stronger theoretical results and would have immediate applications to improve security proofs and rates of cryptographic protocols, leading to more practical technologies.

Other types of decompositions of entropic quantities of interacting complex systems into smaller components involving marginals over subsystems include generalizations of the famous strong subadditivity of the relative entropy known as approximate tensor-stability of the relative entropy. These are at the core of most successful methods for finding the speed of convergence of Gibbs sampling algorithms based on the modified logarithmic Sobolev inequality. In previous work, we successfully extended these notions to the quantum realm 38 and applied them to problems in network quantum information theory 25, 33 and open complex quantum systems 31, 22. Extensions and refinements of these concepts will lead to new breakthroughs in both fields (see Sections 3.3.3 and 3.3.4).

A complexity theoretical definition of the quantum phase of a state

In the setting of classical Gibbs measures, analogous questions have been intensively studied from the perspective of Markov chain Monte Carlo algorithms (MCMC). On regular lattices, the analysis of the speed of convergence of MCMC for lattice spin systems is by now well-understood through the study of correlations at equilibrium. The generalization to general interaction graphs is still a very active field of research in theoretical computer science, probability theory and mathematical physics 32. The problem becomes even harder in the quantum regime, where purely quantum mechanical effects, e.g. long-range entanglement, may cause the quantum Markov chain to slow down in an unpredicted manner. For the important case of commuting interactions, which include most hitherto studied Hamiltonians for the purpose of quantum error-correction, and for physical dynamics generated by the weak coupling of the system with a large environment (Davies dynamics), general results were obtained through spectral methods. However the latter are not powerful enough to distinguish evolutions generating topologically ordered states from rapidly mixing ones. Instead, more involved techniques, e.g. entropic inequalities, are needed. In 24, 31, 23, 22, we were able to prove rapid mixing by extending one of the most successful classical approaches to prove rapid mixing based on the modified logarithmic Sobolev inequality and the approximate tensor-stability of the relative entropy (cf Section 3.3.2). Extending this novel powerful approach, we plan to conduct a systematic joint study of mixing times and thermal stability of topological quantum order in low lattice dimensions. We will conduct this research in collaboration with Daniel Stilck França from QINFO with whom we co-authored 31. We also see a clear connection with the research focus of Daniel Malz who was recently recruited as a junior professor at Inria Saclay, the mathematical and theoretical condensed matter physicists at CPhT, as well as the team PEIPS at CMAP (X).

In parallel to the previous research plan, we will conduct a mathematical analysis on the storage of quantum information and the concept of self-correction in complex quantum systems. Early work on the storage time of candidates of self-correcting quantum memories relied on the connection to the energy barrier of the system, that is the energy the system must reach for a logical error to occur, via an empirical principle called the Arrhenius law. More recently, the energy barrier was rigorously related to spectral properties of the evolution, whereas some no-go theorems showed the impossibility of an exact mathematical formulation of the Arrhenius law. Here instead, we plan to relate the memory lifetime of a device directly to properties of its thermal equilibrium state. We currently work on this research direction in the setting of lattice spin systems with Anthony Leverrier and Ivan Bardet from the team COSMIQ through the development of spectral methods, and plan to extend our framework to lattices with bosonic degrees of freedom in the near future. We also plan to initiate a dialogue with Jean-René Chazottes from CPhT (X) on refinements of our techniques using concentration and entropic inequalities which already proved their usefulness in the study of hitting times of classical Markov chains and their metastability. One of our long–term goals is to find systems with thermally stable entanglement, both stable against thermal fluctuations and robust against local perturbations. Such a theoretical result would be of very high practical interest since experimental implementations are inevitably subject to noise and errors.

As the size of quantum devices continues to increase beyond what can be easily simulated classically, new challenges have appeared concerning the robust and efficient characterization of their states. This often necessitates the preparation and destructive measurement of exponentially many copies of the quantum system, as well as the storage of measurement outcomes in a classical memory. Recently, new methods of tomography have been proposed which precisely leverage this important simplification to develop efficient state learning algorithms. One highly relevant development in this direction is that of classical shadows 41, 42. In we propose a better solution by combining classical shadows with new insights from the emerging field of quantum optimal transport. Our current first step only applies to topologically trivial quantum states such as high-temperature Gibbs states or outputs of shallow quantum circuits, and more effort is needed to adapt and generalize our algorithm to non-trivial phases. We envision three new major contributions: First, we will develop constrained versions of concentration inequalities in order to develop efficient tomography algorithms of complex quantum states, assuming the prior knowledge of their phase. This line of research is original even in the classical setting where works on constrained entropic inequalities only very recently appeared in the literature. The expertise of Jean-René Chazottes from CPhT (X) will prove crucial to the success of this project. Second, we will extend the framework of shadow tomography to CV quantum systems. The main difficulties here are two-fold: first, CV systems are infinite-dimensional in nature, and hence some physical constraints need to be imposed on the states that one can hope to learn, such as their energy. Moreover, the set of measurements (homodyne/heterodyne) available in photonic experiments further limits the type of observables that one can hope to predict. In order to ensure the wide applicability of the method and test the resulting algorithm, we will rely on the already established interactions of IQA with the groups of experimentalists at IP Paris and Saclay, and initiate a fruitful dialogue with start-up like Quandela and Pasqal. In the future, we will use these methods to devise hardware-oriented noise-learning algorithms for many-body systems. For this, we plan to get in touch with the experts on statistical learning among IP Paris, and in particular at LIX.

The theoretical study of quantum computation and its advantages has, in the past decade, opened to a new perspective: higher-order quantum computation, i.e. the way in which one can transform black-box quantum gates by inserting them into computation architectures. This is useful to study the ways in which one can query subroutines in quantum computation, a pratice that is bound to become ubiquitous, for example in delegated quantum computing. The study of higher-order quantum computation has already led to promising as well as disconcerting results, such as about the difficulty of formally defining a quantum version of the computational `if' clause 19, or the fact that one might be able to query two unknown gates in a `superposed order of application', using a computation architecture called the quantum switch 34. Using the latter leads to computational advantages for certain tasks 18. However, the mathematical study of higher-order quantum processes quickly encounters thorny formal issues related to their non-trivial compositional structure.

Overcoming these issues would require the development of a specific and robust type system, stipulating which inputs a given higher-order quantum process admits and which output it produces. Despite recent advances 44, currently available type systems are not detailed enough to provide a fully compositional view of higher-order quantum computation. Our work thus focuses on refining them, through the encoding of sectorial structure, i.e. information about how quantum channels behave with respect to certain direct-sum decompositions of their input and output spaces, using the recently developed framework of routed quantum circuits 53, 54. Progress in this direction will pave the way to computer manipulation of complex higher-order processes, for instance to numerically optimise the advantage they yield.

Many of the peculiarities of quantum theory can be tracked down to it not matching our classical notion of causal structure 55; this leads to the question of how one could develop a quantum notion of causal structure, on which some progress has been achieved recently 26. Exploring quantum theory from a causal perspective yields potential progress in understanding its structure and potential applications, in particular for the aforementioned higher-order quantum processes, whose performances are directly connected to their causal structure. In that regard, a particularly important conjecture to prove is that of causal decompositions46, which puts forward a tentative equivalence between a unitary channel's causal structure (operational data about which of its inputs can affect which of its outputs) and its compositional structure (mathematical data about how it can be written as the composition of sub-channels). If such a conjecture (which has not been proven yet in the general case) were to be true, it would yield a remarkable mathematical lever on the relationship between the operational and formal sides of quantum theory. We investigate this conjecture mathematically with the aim to prove it in more and more general cases; this involves abstract mathematical methods employing C* algebras. More generally, we explore how the latter might provide a useful formal basis for considerations of causality in a quantum context.

Quriosity positions its activity at the - fruitfull - frontier between theoretical research in quantum computer science and mathematics, and quantum technology engineering and applications.

We in particular believe that useful quantum inventions and technologies are going to emerge from the current investments in quantum information sciences and technologies, much before large scale (and error corrected) quantum computers can be built.

Our research programs opens in particular towards such perspective, on different aspects:

Quriosity members are individually, and collectively making efforts to reduce their carbon footprint, in particular by taking the plane much less than before the Covid period. Quriosity, and report to a working group at LTCI level, whose objective will be to increase the global awareness on carbon footprint, and steer the discussions to help decide on collective regulatory measures.

Quriosity aims at publishing high-impact papers in high profile journals such as Nature, Science, Physical Review, Quantum, IEEE Transactions on Information Theory, as well as top conferences in our field such as QIP, QCrypt, TQC as well as Crypto, EuroCrypt, CHES.

Telecom Paris currently holds 5 granted patents: 3 on hybrid quantum computational cryptography (axis 3.1) and 2 on quantum coherent communications (axis 3.2). We plan to patent technological innovations, including foundamental proposals for which we see a clear implementation route and possible exploitation paths.

Quriosity intends to play a vigorous role in the training of the future generation of quantum engineers and researchers. IQA and Romain Alléaume have been at the forefront of such development by opening the Quantum Engineering M2 Program in 2017. At Saclay level, and in collaboration notably with the QuACs Inria team but also with active Saclay quantum industry, we have the mid-term ambition to launch a master program on quantum computer science and engineering.

After less than a year of existence Quriosity has obtained oustanding results at QIP 2024 in terms of selected contributions, with 4 members of Quriosity that are co-authors of 5 accepted papers:

Cambyse Rouzé, Mirjam Weilenmann and Augustin Vanrietvelde have each been awarded a starting package in the context of the ANR ExcellenceS Program, on the STEP2 project sucessfully presented by IP Paris.

Tristan Nemoz has been granted the Best Poster Award (category Information Communication Electronics) at IP PAris PhD Day, on December 10 2023.

The 2-day workshop held at University of Edinburgh on Novepber 23-24 and organized between Quriosity (Romain Alléaume, Peter Brown, Francesco Mazzoncini, Tristan Nemoz) and the team of Elham Kashefi at Quantum Software Lab (Elham Kashefi, Mina Doosti, Myrto Arapinis, Yao Ma, Abbas Poshtvan, Chrirag Wadhwa) has been extremly fruitful both scientifically and in the quality and friendlyness of the exchanges. We can see strong convergence, and possible synergies and collaborations, between our visions of the role that Quantum Cryptography can play in Hardware Security, as also expressed in 3.1
)

In 11, we introduce an explicit construction for a key distribution protocol in the Quantum Computational Timelock (QCT) security model, where one assumes that computationally secure encryption may only be broken after a time much longer than the coherence time of available quantum memories.
Taking advantage of the QCT assumptions, we build a key distribution protocol called HM-QCT from the Hidden Matching problem for which there exists an exponential gap in one-way communication complexity between classical and quantum strategies.
We establish that the security of HM-QCT against arbitrary i.i.d. attacks can be reduced to the difficulty of solving the underlying Hidden Matching problem with classical information. Legitimate users, on the other hand, can use quantum communication, which gives them the possibility of sending multiple copies of the same quantum state while retaining an information advantage. This leads to an everlasting secure key distribution scheme over n bosonic modes. Such a level of security is unattainable with purely classical techniques. Remarkably, the scheme remains secure with up to

In 16, 15 we developed new constructions to certify properties of quantum systems using Bell-tests. In 16 we provide a method to stitch together bipartite Bell-inequalties to construct useful multipartite Bell-inequalities. We show that these new inequalities inherit certain self-testing prop erties of the original inequalities and we use this to produce protocols that achieve maximal device-independent randomness generation. In 15 we explore the complicated relationship between nonlocality and secret-key, showing that perfect secret key-rates can be achieved in device-independent quantum key distribution protocols with correlations that are arbitrarily close to classical. This involves developing a compact expression of a parameterized Bell-inequality that represents all possible self-tests of the maximally entangled state

We have developed a new security proof framework for standard quantum key-distribution (QKD) protocols. The framework is: (i) generic, applying to all possible round based QKD protocols; (ii) tight, providing key-rates that are optimal for any given finite number of rounds up to leading order correction and (iii) computable, we develop convex optimization methods to compute the key-rates. This is achieved in part through the development of new entropic quantities that enable tight accounting of finite-size corrections. We expect the work to have significant impact on the future development of QKD protocols and security proofs.

Building up on 21, we are performing a systematic analysis and model of excess noise in a quantum coherent communication channel, jointly operated with a classical coherent channel, taking into account the effect of digital signal processing. We are moreover validating our results with digital simulation along with experimental system demonstration. We are moreover studying how Kramers-Krönig coherent detection can be used in quantum communications.

In 5, we propose novel methods for the exact synthesis of single-qubit unitaries with high success probability and gate fidelity, considering both time-bin and frequency-bin encodings. The proposed schemes are experimentally implementable with a spectral linear-optical quantum computation (S- LOQC) platform, composed of electro-optic phase modulators and phase-only programmable filters (pulse shapers). We further investigate the parallelization of arbitrary single-qubit gates over multiple qubits with a compact experimental setup, both for spectral and temporal encodings. Our analysis positions spectral S-LOQC as a promising platform to conduct massively parallel single qubit operations, with potential applications to quantum metrology and quantum tomography.

In 10, we established an analytic upper bound on the fault-tolerance threshold for concatenated GKP-stabilizer codes with local update recovery. Our bound applies to noise channels that are tensor products of one-mode beamsplitters with arbitrary environment states, capturing, in particular, photon loss occurring independently in each mode. It shows that for loss rates above a threshold given explicitly as a function of the locality of the recovery maps, encoded information is lost at an exponential rate.

We are developing fine-grained type systems for the certification of quantum computation architectures, in particular those involving quantum control (quantum versions of the "if" clause) and indefinite causal order (application of operations in a superposed order).

We are investigating the relationship (and in particular the potential equivalence) between the causal structure of quantum dynamics and their compositional structure. We are finishing a proof (to be published soon) that there is an equivalence between the two in the case of local dynamics over a 1D array of quantum systems, at any range.

In 12, 12, 13, 27, we have developed robust, sample and computationally efficient quantum algorithms for tomography and learning of states and noise on many-body discrete and continuous variables quantum systems, including thermal and ground states of spin and Bosonic Hamiltonians, finitely correlated states, and Pauli noise channels with unknown underlying local structure.

We have kept on working on the complexity of Gibbs sampling algorithms and its applications to the stability of quantum simulation and the characterization of self-correcting quantum memories 22, 23. In particular, we have extended our previous result on the rapid mixing of Gibbs samplers from 2-local to k-local interacting systems (the article is being finalized). In particular, we resolved an old open problem, namely whether the existence of a dissipative gap implies rapid mixing of the thermalization process. As regards to applications to quantum memories, our result implies that the entire class of quantum double models (including the 2D Toric code) reaches thermal equilibrium in logarithmic time, while the previous best thermalization time scaled linearly with the system size.

A recently developed quantum algorithm known as the Quantum Singular Value Transform (QSVT) was shown to encompass the standard quantum algorithms, Shor/Grover etc., as special cases. Using the fact that the QSVT provides a correspondence between polynomials and certain families of quantum circuits, we develop methods to optimize these circuits (and hence the resulting algorithms) in terms of polynomial optimization and its resulting SDP relaxations. We also show to scale up these optimizations to large-scale circuits, demonstrating its usefulness at scales necessary for quantum advantage.

Doctoral project of Guillaume Ricard, on Quatum Coherent Communications and Digital Signal Processing, Funded by Paris funded by Paris Region (region Ile-de France) in the context of the Paris Region PhD call, with a planned collaboration with Quandela on noise mitigation in optical coherent quantum communications.

The Quantum Secure Networks Partnership (QSNP) aims at creating a sustainable European ecosystem in quantum cryptography and communication. Its 42 partners are world-leading academic groups, research and technology organizations (RTOs), quantum component and system spin-offs, cybersecurity providers, integrators, and telecommunication operators. The Partnership thus has the expertise in all technology development phases, from new designs to field deployment, making it ideal to carry out the future Specific Grant Agreement (SGA) projects.

All teaching durations are given in hetd = "heures équivalent TD".

(Apart from our own students PhD juries)