We aim to advance the second quantum revolution in the AI era by developing novel information-processing protocols and demonstrating them on physical platforms, with a dual track: harvesting classical gains now while advancing the protocols and hardware needed for scalable, fault-tolerant quantum information processing.
The second quantum revolution promises advances in communication, computation, and sensing, driven by quantum information processing that harnesses superposition and entanglement. These fundamental quantum properties were first demonstrated in pioneering experiments with photons in the 1970s and 1980s, recognized by the 2022 Nobel Prize in Physics. Photons have emerged as ideal carriers for quantum information, and integrated photonic circuits are central to scaling both quantum and classical information processing.

Quantum cryptography uses the laws of quantum mechanics to guarantee information security, not through the computational difficulty of breaking a code, but through the fundamental physics of measurement and entanglement. Any eavesdropper necessarily disturbs the quantum states being transmitted, making interception detectable in principle. This is the basis of quantum key distribution (QKD), which allows two parties to establish a shared secret key with information-theoretic security: security that holds even against an adversary with unlimited computational power. Our work in QKD integrates theory and experimentation, co-optimizing every stage from security proofs to hardware implementation. We advance high-dimensional QKD, encoding information in multiple degrees of freedom of single photons (time bins, orbital angular momentum, spatial modes) to achieve higher secret key rates, and have demonstrated protocols that double key rates using only commercial off-the-shelf components over standard telecom fiber.
Beyond secure communication, quantum mechanics also enables fundamentally new forms of secure computation. In secure multiparty computation, several parties jointly evaluate a function while keeping their inputs private, for example, a hospital classifying a patient’s data using a server-owned model without either party revealing their private information. We pioneered an information-theoretically secure computation scheme using conventional telecom components, with analytical upper bounds on information leakage for both neural-network weights and client inputs. At the quantum limit, where mode occupation approaches the vacuum level, we exploit vacuum fluctuations for secure computation via nondemolition computation based on optical nonlinearity. Our quantum-secure multiparty deep learning protocol was published in Physical Review X (2025) and covered by MIT News, Tech Briefs, ScienceDaily, Phys.org, and SciTechDaily.
A central question in quantum information science is: which problems can a quantum computer solve faster than any classical one? For structured problems like integer factoring, quantum algorithms offer an exponential advantage, while for unstructured search the speedup is only polynomial. Between these extremes lies a large, largely unmapped space of partially structured problems. A core aim of our lab is to identify which physical resources convert partial structure into provable quantum advantage.
Integrated photonics is emerging as one of the most promising platforms for building scalable quantum processors. Our experimental work is built on two complementary foundry platforms: silicon photonics (CMOS-compatible, low-loss, suitable for squeezing and photon-pair generation) and thin-film lithium niobate (high nonlinearity enabling fast modulation, spontaneous parametric down-conversion, quantum transducers, and nondemolition gates). These platforms support the generation, manipulation, and detection of quantum states of light on a chip, the essential building blocks for photonic quantum computing. We develop coherent photonic control tools for quantum information processing, including all-fiber sources and sorters for multimode correlated photons and programmable processing of high-dimensional entangled states, and are currently validating nondemolition computation in thin-film lithium niobate photonic-crystal cavities.
Quantum machine learning (QML) sits at the intersection of quantum computing and artificial intelligence. Building on theoretical speedups and a growing photonic QML community, we translate these concepts into practical experiments. Where much QML work assumes large-scale, fault-tolerant quantum computers or unrealistic models, our testbed consists of fully functional optical computers that already solve hard problems efficiently. We explore what becomes achievable with realistic squeezing and non-Gaussian operations. In spirit, this mirrors LIGO, where a robust classical detector is quantum-enhanced by modest squeezing.
Analog computing has been studied for decades, but the demand for high precision kept practical systems in the digital domain. That constraint has now relaxed: today’s machine-learning models rely on 8-bit or even 4-bit arithmetic, well within analog reach. This opens the door to computing architectures that perform machine learning directly in the physical domain, using optical and radio-frequency hardware to execute neural-network operations at the speed of light, with immediate energy-cost advantages over digital electronics.
Optics is uniquely promising for machine learning because the core operation of neural networks, matrix-vector multiplication, maps naturally onto the physics of light propagation. A beam of light passing through a mesh of interferometers or scattering through a diffractive layer performs a linear transformation at the speed of light, consuming energy only for the light source and detectors, not for the computation itself. This is why optical neural networks have attracted intense interest from both academia and industry as a path to AI inference that is orders of magnitude more energy-efficient than electronic processors.
Rather than adapting analog hardware to run digital-era models, we propose the inverse: design hardware-tailored AI models natively suited to optics. The cornerstone of our approach is digital twins, differentiable AI-based virtual replicas of physical systems synchronized in real time with experimental data. Training these twins end-to-end lets us co-optimize hardware and algorithms simultaneously, discover photonics-tailored architectures, and accelerate experimental discovery. We develop optical neural networks based on quadrature amplitude modulation for high-throughput inference and nanophotonic integrators that implement neural-network primitives directly in silicon.
Optics already plays a growing role in AI infrastructure: optical interconnects carry data between GPUs in modern data centers, and photonic tensor cores are being developed by startups and major labs to accelerate matrix operations. We push this further by exploring unconventional physical substrates for neural-network inference. In recent work, we demonstrated that a cellphone’s RF front-end (its antenna, filters, mixers, and amplifiers) can perform real-time deep-learning inference when the model is streamed over the air, eliminating on-device memory and dedicated AI hardware entirely. This disaggregated in-physics computing paradigm opens the door to AI inference on any device with a radio, from IoT sensors to satellites, and was published in Science Advances (2026).
The same principle applies broadly: any physical system whose input-output relationship is differentiable can, in principle, be trained as a computing element. We co-design analog electronics with photonics for in-physics AI accelerators that exploit the native physics of signal propagation rather than fighting it.
A fundamental question drives this entire research direction: What is the irreducible energy required to complete a computational task within finite temperature, time, and space, and how close can physical systems push us to that limit?
Classical complexity theory classifies problems in abstract, unitless time and space, largely ignoring energy, the resource that dominates advanced computing today. Physics offers a richer lens. In the spirit of Szilard, Landauer, and Bennett, we go beyond the abstractions of computer science and bring thermodynamics back into the foundation. Landauer’s principle states that erasing a single bit of information at temperature T dissipates at least kT ln 2 joules of energy, a bound set by the second law of thermodynamics. This seemingly tiny quantity, roughly 3 × 10⁻²¹ joules at room temperature, represents the absolute physical floor of computation. Today’s transistors dissipate roughly a million times more energy per operation, so the gap is enormous, but closing it requires fundamentally rethinking how we compute.
We study these thermodynamic and information-theoretic bounds on computation, connecting Landauer’s principle and finite-time thermodynamics to the physics of noise, to understand the ultimate efficiency of analog and optical processors. This theoretical thread informs the design of all our in-physics computing platforms and tells us how far we are from the physical limits of what is possible.
A core direction of our research treats AI as a partner in rigorous scientific discovery. We couple large language models with proof assistants like Lean 4: the LLM conjectures broadly while the verifier certifies strictly, producing results that are provably correct, not merely plausible. We co-founded the Center for Verifiable AI (CVAI) at MIT-Technion to pursue this research program.
We build multi-agent frameworks that equip LLMs with the tools of the Lean proof assistant. The model inspects proof goals, searches for relevant lemmas, and generates verified proof steps, all within a feedback loop where the proof assistant checks every step for correctness. This leverages the creativity of LLMs while demanding the rigor of formal verification. We have demonstrated this approach on benchmarks in abstract algebra and quantum theory.
The impact extends far beyond mathematics: formal verification can certify that a quantum protocol is truly secure, that an optimization algorithm converges, or that a physical model satisfies conservation laws. By automating this process, we aim to accelerate the cycle of conjecture, proof, and discovery across the sciences.
We recently formalized, to our knowledge, the first quantum cryptography security proof in Lean, machine-checking the mathematical argument that guarantees the security of a quantum communication protocol.
Building on this, we are assembling a Lean library for quantum information theory, currently exceeding 35,500 lines of formalized mathematics, spanning density operators, quantum channels, von Neumann entropy, and cryptographic protocols. Verified milestones include the Weyl and Ky Fan eigenvalue inequalities, the Fannes continuity bound, the Fuchs-van de Graaf inequality, and an end-to-end BB84 security proof with CSS error correction. The long-term vision is automated discovery of advanced primitives, including new quantum cryptography schemes, secure computation protocols, and error correction codes, where AI-guided formal reasoning explores proof spaces far larger than any human could navigate, with every step machine-verified.
We design benchmark suites that test whether AI-generated conjectures can survive full formalization, categorize the failure modes that block successful proofs, and track indicators such as proof completion rate and verification throughput. These metrics, maintained across successive model and library releases, provide an empirical record of what works and where the largest opportunities for improvement lie.