Julian Büchel

Title

Pre-Doc
Julian Büchel

Bio

LinkedIn | Google Scholar

Bio

Julian Büchel is a Pre-Doctoral Researcher in the Analog In-Memory Computing group at IBM Research – Zurich and a PhD student at the Department of Computer Science at ETH Zürich under the supervision of Prof. Martin Vechev. He studied Computer Science and Neural Systems and Computation at ETH Zürich from 2016 to 2022. During his Master's, he worked as a Research Engineer at SynSense, a company focused on digital neuromorphic vision. During his Master's thesis conducted at IBM Research in 2021, he demonstrated that adversarial attacks in the weight domain can be used to enhance the robustness of DNNs (see ICLR paper). After his thesis, he joined IBM Research as a Pre-Doctoral Researcher in 2022. Work co-authored by him has won the 2023 OSS award and an honorable mention in the 2024 Pat Goldberg Memorial Award. His current research interests evolve around the topic of bringing LLMs to Analog In-Memory Computing architectures based on high-density non-volatile memory technology. He primarily focuses on re-training LLMs to improve the robustness against various types of noise such as quantization noise and noise introduced by the analog hardware.

News

  • [23 May 2025] v1.0.1 of AIHWKIT-Lightning is released. Check it out here.
  • [15 May 2025] We released a new paper called "Analog Foundation Models" where we show that LLMs can be trained to tolerate weight noise introduced by AIMC hardware. We loose accuracy comparable to the 4W8I scenario in quantized LLMs.
  • [01 May 2025] Iason Chalas successfully defended his MSc thesis. Glad to announce he will join my team focusing on algorithms for analog AI.
  • [08 January 2025] The paper "Efficient scaling of large language models with mixture of experts and 3D analog in-memory computing" got the cover at Nature Computational Science.

Software

  • AIHWKIT-Lightning is a scalable and light-weight library for HW-aware training.
  • 3D-CiM is a simulator that can be used to assess the system level performance of a high-level 3D AIMC-based accelerator.
  • Analog-MoE is a triton kernel for a MoE layer equipped with HW-aware training capabilities such as input/output quantization, tiling, and noise injection.
  • Sigma-MoE is a library extending Robert Csordas original Sigma-MoE implementation by making it compatible with torch.compile and huggingface.
  • analog-foundation-models is code supporting our "Analog Foundation Models" paper and can be used as a recipe to train your own analog foundation model.

Selected publications

Analog In-Memory Computing:

Spiking Neural Networks:

Selected Patents

Reviewing

Get in touch

jub[at]zurich[dot]ibm[dot]com

Top collaborators