Suggested Searches

Next AI/ML Science and Technology Interest Group (AI/ML STIG) Lecture February 9, 2026

Transformers

Speaker

Helen Qu | Flatiron Institute

Build a decoder-only transformer (GPT-style) from scratch in PyTorch. Train it on Tiny Shakespeare for character-level language modeling and generate text.

Topics covered:

  • Self-attention as a learned, data-dependent mixing operator
  • Causal (masked) self-attention for autoregressive modeling
  • Building a GPT-style Transformer block from scratch
  • Token and positional embeddings
  • Training a small autoregressive language model
  • Text generation with temperature and top-k sampling

Lecture tutorial materials and jupyter notebooks can be found here: https://tingyuansen.github.io/NASA_AI_ML_STIG/#schedule

We'll be using this slido link for Q&A during the talk: https://app.sli.do/event/kx2WYpZtkHYtbWMntJoFFj 

The link to join the meeting is here: https://science.nasa.gov/astrophysics/programs/cosmic-origins/community/ai-ml-stig-lecture-series-9-feb-2026/

News Straight to Your Inbox

Subscribe to your community email news list

We will never share your email address.

Sign Up
Angled from the upper left corner to the lower right corner is a cone-shaped orange-red cloud known as Herbig-Haro 49/50. This feature takes up about three-fourths of the length of this angle. The upper left end of this feature has a translucent, rounded end. The conical feature widens slightly from the rounded end at the upper right down to the lower right. Along the cone there are additional rounded edges, like edges of a wave, and intricate foamy-like details, as well as a clearer view of the black background of space. In the upper left, overlapping with the rounded end of Herbig-Haro 49/50, is a background spiral galaxy with a concentrated blue center that fades outward to blend with red spiral arms. The background of space is speckled with some white stars and smaller, more numerous, fainter white galaxies throughout.