EuroCC@Greece in collaboration with Smart Attica EDIH, announced  the 9th Course of HPC Training Series with the subject "Running LLMs on HPC: Transformers, Inference & Deployment", that took place online on January 17th, 2025. 

Date: January 17th, 2025, at 10:00 EET 

Location: Online via Zoom

Presentation Languages: Greek & English

Audience:

  • Data scientists and machine learning engineers.  
  • NLP researchers and practitioners.  
  • HPC system administrators and engineers.  
  • Developers exploring Hugging Face Transformers and RAG.  
  • Academic researchers working on language modeling projects.  
  • Professionals interested in training or deploying LLMs on HPC.  
  • Organizations planning to adopt HPC for AI workloads.

Description: This course focused on Large Language Models running on High-Performance Computing systems. Participants gained a foundational understanding of the Hugging Face Transformers library, embeddingsmodels, and of Retrieval-Augmented Generation. They discovered how to effectively set up an inference server on HPC systems as well as a deployment process and limitations. Training of the Greek LLM Meltemi was also presented. This seminar included hands-on sessions where users were be able to run the provided code.

Learning Objectives:

  • Hands-on experience on Hugging Face Transformers 
  • Set up and troubleshoot LLM inference servers on HPC systems.  
  • Explore LLM deployment on HPC, including limitations and applications.  
  • Learn the training process of the Greek LLM Meltemi.  
  • Understand capacity and scaling challenges in LLM deployment.  
  • Experiment with real-world applications of LLMs.

Prerequisites:

  • Basic understanding of machine learning and neural networks,
  • Knowledge of Python programming,
  • Basic command-line and Linux skills.

Note: Please enter your institutional/corporate email when registering.

 

Starts
Ends
Europe/Athens
Registration
Registration for this event is currently open.