About This Offering

We will start with the fundamentals—tokenization, embeddings, and transformer architecture—and move toward cutting-edge advancements in Large Language Models. Along the way, you will gain hands-on experience with platforms and tools like Hugging Face, LangChain, and real-world APIs.

Registration form


Suman Maity

An Assistant Professor in the Department of Computer Science at Missouri S&T. Previously, he was a postdoctoral research associate at MIT and a postdoctoral fellow at CSSI and NICO, Northwestern University. He earned his PhD from IIT Kharagpur and received IBM and Microsoft Research India PhD Fellowships.

His research focuses on Social NLP and Responsible Machine Learning, with work published in top-tier venues like WWW, EMNLP, and ACL. He has served as a reviewer for conferences such as AAAI, WWW, and ICWSM.

  • Registration: Open until March 27, 2026
  • Course Dates: April 6 & 8, 2026
  • PDH: 16
  • Price: $1,499
  • Prerequisites:
    • Basic python programming knowledge
    • Familiarity with machine learning concepts
       
  • Location: 2127 Innerbelt Business Center Drive, St. Louis, MO 63114

The second bootcamp course of our A.I. bootcamp series provides a comprehensive introduction to Large Language Models (LLMs) and their practical applications. Participants will explore the evolution of language models, learn how tokenization and embeddings work, and dive into the inner workings of transformer-based architectures. It covers essential topics like prompt engineering, advanced text generation techniques, Retrieval-Augmented Generation (RAG) and LLM Agents. Through guided exercises and real-world tools and platforms like Hugging Face, LangChain, etc. attendees will gain practical experience building LLM-powered systems.

By the end of the course, participants will be able to: 

  1. Describe the evolution of language models and the significance of LLMs in modern NLP
  2. Understand the architecture and functionality of transformer-based models
  3. Apply prompt engineering techniques for various tasks such as summarization, Q&A, and creative generation
  4. Use libraries and tools like Hugging Face Transformers, LangChain etc. to develop LLM-based solutions.
  5. Build a simple Retrieval-Augmented Generation (RAG) pipeline for domain-specific applications.
  6. Design and implement a small-scale LLM-powered project demonstrating practical skills learned in the course.
  • Introduction to Language Models
    • Evolution of language modeling: N-grams → RNNs → Transformers → LLMs
    • Why LLMs matter: zero-shot, few-shot, and chain-of-thought reasoning
    • Overview of popular models: Open-source (LLaMA, Mistral) vs proprietary (GPT, Claude)
  • Tokens and Embeddings
    • What are tokens? Tokenization strategies: Byte Pair Encoding (BPE), WordPiece, SentencePiece
    • From static to contextual embeddings: Word2Vec → BERT → GPT
    • Hands-on: Visualize tokenization and embedding spaces using dimensionality reduction (e.g., t-SNE or UMAP)
  • Transformer Model
    • Key components: self-attention, multi-head attention, positional encodings
    • Transformer variants: encoder-only (BERT), decoder-only (GPT), encoder-decoder (T5)
    • Anatomy of a layer: attention heads, feedforward layers, residual connections, layer normalization
    • Hands-on: Explore a simplified transformer architecture
  • Prompt Engineering
    • Prompting strategies: zero-shot, few-shot, chain-of-thought, and system-level prompts
    • Instruction tuning vs prompt engineering
    • Hands-on: Design prompt templates for summarization, Q&A, and reasoning
    • Build a prompt-based mini app (e.g., travel planner, story generator) using OpenAI API or open-source LLMs via Hugging Face
  • Retrieval-Augmented Generation (RAG)
    • What is RAG? Concept, motivation, and key applications
    • Core components: document chunking, text embeddings, vector stores, retrievers, re-rankers
    • Popular frameworks: LangChain, LlamaIndex, FAISS, Chroma etc.
    • Hands-on: Build a mini RAG system to answer questions from PDF files or Wikipedia articles
  • Introduction to LLM Agents
    • What are LLM Agents and how they extend LLM capabilities
    • Core components: reasoning, planning, and tool use
    • Examples of agentic workflows
    • Hands-on: Building a simple LLM-powered agent using open-source framework
  • Final Project & Wrap-Up