EuroCC@Greece in collaboration with Smart Attica EDIH, announced the 9th Course of HPC Training Series with the subject "Running LLMs on HPC: Transformers, Inference & Deployment", that took place online on January 17th, 2025.
Date: January 17th, 2025, at 10:00 EET
Location: Online via Zoom
Presentation Languages: Greek & English
Audience:
- Data scientists and machine learning engineers.
- NLP researchers and practitioners.
- HPC system administrators and engineers.
- Developers exploring Hugging Face Transformers and RAG.
- Academic researchers working on language modeling projects.
- Professionals interested in training or deploying LLMs on HPC.
- Organizations planning to adopt HPC for AI workloads.
Description: This course focused on Large Language Models running on High-Performance Computing systems. Participants gained a foundational understanding of the Hugging Face Transformers library, embeddings’ models, and of Retrieval-Augmented Generation. They discovered how to effectively set up an inference server on HPC systems as well as a deployment process and limitations. Training of the Greek LLM Meltemi was also presented. This seminar included hands-on sessions where users were be able to run the provided code.
Learning Objectives:
- Hands-on experience on Hugging Face Transformers
- Set up and troubleshoot LLM inference servers on HPC systems.
- Explore LLM deployment on HPC, including limitations and applications.
- Learn the training process of the Greek LLM Meltemi.
- Understand capacity and scaling challenges in LLM deployment.
- Experiment with real-world applications of LLMs.
Prerequisites:
- Basic understanding of machine learning and neural networks,
- Knowledge of Python programming,
- Basic command-line and Linux skills.
Note: Please enter your institutional/corporate email when registering.