Staging environment

AI Systems Design & Inference Engineering

Abi Aryan

ML Research Engineer and Book Author

From AI Engineer to AI Systems Engineer

[๐ฎ๐ฉ๐๐š๐ญ๐ž๐ฌ] ๐‚๐จ๐ก๐จ๐ซ๐ญ ๐Ÿ: ๐…๐ž๐› ๐Ÿ๐Ÿ - ๐€๐ฉ๐ซ๐ข๐ฅ ๐Ÿ ๐ข๐ฌ ๐ง๐จ๐ฐ ๐Ÿ๐ฎ๐ฅ๐ฅ๐ฒ ๐ฌ๐จ๐ฅ๐ ๐จ๐ฎ๐ญ ๐š๐ง๐ ๐œ๐ฅ๐จ๐ฌ๐ž๐!โฃ

If youโ€™re an AI engineer who can already prompt or fine-tune models but youโ€™ve never been able to answer questions like:

- Why is my 70B model using 120 GB of VRAM and still slow?

- How do I serve 500 concurrent users on 4xH100s without going broke?

- What actually happens inside FlashAttention / PagedAttention / tensor parallelism?

- How do I save my company millions running open models in production?

โ€ฆ then this is the course youโ€™ve been waiting for.

What youโ€™ll learn

Most engineers load models with Hugging Face then serve something basic and get stuck on cost, latency, or scale. In this course, we will

  • Learn to serve a tiny model (e.g. Phi-3-mini or TinyLlama) on a single consumer GPU in <15 minutes.

  • Learn how to develop and deploy inference gateways, do endpoint management and learn steaming vs non-streaming architectures

  • You'll learn concrete requestโ€‘toโ€‘token mental model before hardware and optimization detail.

  • Use real tools to see exactly where time/memory is wasted.

  • You'll learn to calculate exactly how many tokens, bytes and VRAM your model will take.

  • You'll learn how to build an inference engine from scratch and the internals of vLLM, SGLang in detail and an overview of the rest..

  • You'll learn how to optimize for compute versus memory.

  • We will implement compute management tricks and then memory management optimizations incl KV Cache, Quantization etc.

  • How to decide which hardware to use, for which model and what will scale and where will different model/hardware combinations hit bottleneck

  • Scaling models doesn't just mean throwing more compute at the problem

  • You'll learn distributed systems architecture design using Ray and Kubernetes - Fleets, Multi-Node and Multi-GPU & how to scale inferencing

  • We will implement distributed inferencing for 70Bโ€“405B models.

  • The internals- what happens when your system gets a request, how tokenization, batching, memory allocation is done on your GPUs.

  • Inference Optimization techniques (speculative decoding, chunked prefill etc) both at prefill & decode stages for dense, sparse, and MoE

  • How to do concurrency management at request/model/hardware level. How to use Load balancing and parallelism (Data, Tensor, Model, Expert).

  • 1. Zach Mueller, Head of Dev Rel @ Lambda, ex-Huggging Face

  • 2. Suman Debnath, Technical Lead ML @ Anyscale

  • 3. Paul Iusztin, AI Engineer and Author (LLM Engineer's Handbook)

Learn directly from Abi

Abi Aryan

Abi Aryan

Founder and Research Engineering Lead @ Abide AI

Book Author (LLMOps, GPU Engineering for AI Systems - upcoming)
O'Reilly Media
@Packtpub
See all products from goabiaryan

Who this course is for

  • AI engineers, ML infrastructure engineers, and backend developers who own inference cost and need to 10โ€“50ร— optimize it.

  • Founders & engineers building LLM apps who are tired of burning money on OpenAI

  • AI engineers who don't understand system design for building reliable applications

    Anyone on job market for Inferencing/Solns Architect role

Prerequisites

  • You have shipped at least one non-trivial Python project

    We cannot teach you the basics of Python code.

  • You can already load and run an LLM using HF Transformers

    If you have never touched AI models, this course is not for you.

  • Comfortable with basic terminal/SSH, git, yaml & json files

    We can help with Docker and Kubernetes, don't worry.

What's included

Abi Aryan

Live sessions

Learn directly from Abi Aryan in a real-time, interactive format.

1400 USD in compute credits

A total of 1400 USD in compute credits from Modal, Anyscale and Lambda to work on the course exercises

Lifetime access

Anyone who signs up for the course will have lifetime access to the course and to the future cohort recording/materials too as I keep updating it.

Resume Review

We will have 1-on-1 resume review if you are currently on the job market or looking for a career transition

171-page system design interview guide

It covers 150 practice questions covering interview questions for 1. Systems Architecture Design 2. Inference Optimization & Serving 3. RAG & Context Engineering 4. Reliability and Guardrails 5. Observability, Evaluation & Monitoring, and finally 6. Tradeoffs, Scenarios & Integrations

Community of peers

A discord channel to stay connected with peers.

Guest Speakers

Learn from industry professionals and their experiences.

Certificate of completion

Share your new skills with your employer or on LinkedIn.

Maven Guarantee

Your purchase is backed by the Maven Guarantee.

Course syllabus

Week 1

Feb 21โ€”Feb 22

    Request Lifecycle and Inference Basics

    6 items

Week 2

Feb 23โ€”Mar 1

    Observe Before Optimize: Profiling & Cost Modelling for LLMs

    1 item

Free resources

Schedule

Live sessions

3-5 hrs / week

We will try our best to accomodate different timezones.

Weekly Project

1-3 hrs / week

Abi will be available to help you in case you get stuck.

Quizzes and Async content

2-4 hrs / week

This is optional but incredibly helpful for further reading and reference.

Frequently asked questions

Sponsors