
Description: This training event will introduce techniques for AI/ML inference on HPCMP machines, including multi-GPU/multi-node model loading, performance optimizations and inference servers. Attendees will have first access to two distributed inference servers assembled by PET. Examples will mostly involve LLMs but many of the same techniques and optimizations are also relevant for non-LLM applications.
| Presenter(s): Dr. Calvin Anderson, GDIT/PET Location: Webcast Date & Time: May 8, 2025, 1:00p - 4:00p ET Additional Notes: Enrollment closes end of the day May 7, 2025. |
Controlled by: DoD HPCMP Controlled by: PET Program CUI Category: OPSEC Limited Dissemination Control: FEDCON POC: Mr. Ronald Hedgepeth, pet@hpc.mil |
CUI
- Presenter: Calvin Anderson
Search Terms: Artificial Intelligence, AI, Natural Language
Processing, NLP, Retrieval Augmented Generation, RAG, Large Language Model, LLM,
Fine-tuning, Transformers
Course ID number for Global Search: TE1616_Archive

