General information
Organisation
The French Alternative Energies and Atomic Energy Commission (CEA) is a key player in research, development and innovation in four main areas :
• defence and security,
• nuclear energy (fission and fusion),
• technological research for industry,
• fundamental research in the physical sciences and life sciences.
Drawing on its widely acknowledged expertise, and thanks to its 16000 technicians, engineers, researchers and staff, the CEA actively participates in collaborative projects with a large number of academic and industrial partners.
The CEA is established in ten centers spread throughout France
Reference
2025-37899
Description de l'unité
The internship will take place at CEA-LIST, in the Integrated Multi-Sensor Intelligence Laboratory, which brings together experts in artificial intelligence, embedded systems and sensors.
Position description
Category
Mathematics, information, scientific, software
Contract
Internship
Job title
Artificial Intelligence - Large Language Models optimizations for In-Memory Computing hardware H/F
Subject
This internship aims at putting together a software infrastructure for mapping, simulating and exploring the performance of Large Language Models (LLMs) on In-Memory Computing (IMC) hardware, starting with existing open-source tools chains and integrating functionalities dedicated to IMC hardware, such as quantization and error models.
Contract duration (months)
6
Job description
Large Language Models (LLMs), such as ChatGPT, have led to a new AI revolution with applications in every domain. However, LLMs are very resource-consuming (energy, compute…) and, hence, an important line of research focuses on optimizing these models. Existing open-source tool chains, such as LLM Compressor [1] and OpenVINO [2], enable almost-automatic optimizations to compress LLMs into smaller ones by, e.g., quantization and pruning. However, they only target conventional hardware, such as GPUs. New hardware paradigms, such as In-Memory Computing (IMC) are promising to accelerate and reduce the energy consumption of LLMs [3]. However, running LLMs on such hardware requires specific optimizations due to the characteristics of these hardware. For instance, they require extreme quantization of the model (reducing the number of bits on which data, weigths and activations are encoded), because analogue IMC fabric has a limited number of bits, and optimizing the robustness of the model, because IMC computations are prone to errors. Nevertheless, software tools and methods for mapping state-of-the-art LLMs on these hardware platforms lag behind.
This internship aims at putting together a software infrastructure for mapping, simulating and exploring the performance of LLMs on IMC hardware, starting with existing open-source tools chains and integrating functionalities dedicated to IMC hardware, such as quantization and error models. The student will be integrated within a multidisciplinary team of research engineers, PhDs, PostDocs and interns, at the heart of an ecosystem of industrial and academic partners in the world of embedded AI. He/she will have access to supercomputers infrastructure. He/she will benefit from increased expertise in LLMs, compression methods, and efficient hardware for AI. Leveraging the tools and knowledge developed during the internship, the student could be offered the opportunity to pursue a PhD on compression methods for LLMs.
[1] https://github.com/vllm-project/llm-compressor [2] https://github.com/openvinotoolkit/openvino [3] Analog Foundation Models, Büchel et al, NeurIPS 2025.
Applicant Profile
Final year engineering or Master 2 student in computer science / Artificial Intelligence / embedded systems
Strong motivation to learn and contribute to artificial intelligence research. Strong knowledge of computer science, programming environments (Unix), languages (Python) and software development tools (Git…). Strong knowledge and experience in deep learning theory and development frameworks (Pytorch or Tensorflow). Knowledge of embedded systems is a plus.
Position location
Site
Grenoble
Location
Grenoble
Candidate criteria
Prepared diploma
Bac+5 - Master 2
PhD opportunity
Oui
Requester
Position start date
02/02/2026