Michael Leung
  • Writing
  • About

On this page

  • Michael Min Wah Leung
  • Writing
  • Edit this page
  • Report an issue
Categories
All (3)
GRPO (1)
LLMs (1)
MLX (1)
RLHF (1)
hybrid-inference (1)
interpretability (1)
neuroscience (1)
post-training (1)
seq2seq (1)
sign-language (1)
signal-processing (1)

Michael Min Wah Leung

Notes on post-training, sequence modelling, and the occasional brain. Code-first, results-honest, written by an ML engineer with a research background in neuroscience.

Writing

Why SFT learned the words but GRPO learned the rules

post-training
GRPO
RLHF
LLMs
Teaching a 14B model a proprietary equipment-naming taxonomy with a hand-tuned reward function, and why ~250 lines of reward code and a quarter-epoch of GRPO closed the gap that more SFT couldn’t.
May 2, 2026
12 min

From consuming a pretrained model to training my own

seq2seq
sign-language
MLX
hybrid-inference
Building a continuous-sign-language Copilot: a Transformer Seq2Seq trained from scratch on How2Sign, two training backends, and a hybrid runtime that reaches 93.6% sentence-level recognition.
May 1, 2026
13 min

Patient-specific filters as biomarkers

neuroscience
signal-processing
interpretability
ICA, FOOOF, and CSP for noisy EEG, and what spatial filters taught me about feature extraction in transformers.
Apr 30, 2026
10 min
No matching items

© 2026 Michael Leung

Writing · About · RSS

  • Edit this page
  • Report an issue

Built with Quarto.