IS&T RCS Tutorial - GPT & Transformers (LLMs Part 3) (Hands-on)

Training Large Language Models (LLMs) requires a large neural network, large data, and large compute. We will discuss these difficulties. We’ll look at the Transformer architecture in detail to develop a quantitative understanding of how it works and how specifically tools like ChatGPT, DeepSeek, Llama, etc. work. We will then use a pre-trained SentenceTransformer model to do a range of classification on real-world data.

When 10:00 am - 12:00 pm on 13 June 2025
Building Biological Science Center, 2 Cummington Mall, Room 107