01. Prompting and structured output
Instruction design, role prompting, schema-constrained responses, and prompt debugging.
Two-week intensive curriculum
This guide is built for 12 to 14 hours of study per day over two weeks. Each subject page is intentionally dense: it gives you a beginner track, an advanced engineering track, practical examples, Python code, diagrams, and a concrete to-do list covering what to learn, practice, and build before moving on.
Instruction design, role prompting, schema-constrained responses, and prompt debugging.
Latency, throughput, observability, guardrails, caching, and cost control in real systems.
When to fine-tune, LoRA-style adapters, data formatting, and evaluation.
Compression, inference efficiency, hardware trade-offs, and quality loss management.
Grounding generation with external knowledge using retrieval pipelines.
Approximate nearest neighbor search, indexing, filtering, and storage design.
Dense representations, similarity, domain adaptation, and practical evaluation.
Splitting, metadata design, and index construction for retrieval quality.
Ranking, lexical retrieval, evaluation, and hybrid search strategies.
Detection, prevention, grounding, and response shaping for higher reliability.
Cross-lingual transfer, multilingual embeddings, and language-specific pitfalls.
Cleaning, standardization, Unicode handling, and pipeline design.
Risk management, governance, privacy, fairness, and safe deployment.
Tokens, corpora, ambiguity, syntax, semantics, and task framing.
Bag-of-words, n-grams, TF-IDF, RNNs, CNNs, Seq2Seq, and Transformers.
BoW, TF-IDF, word embeddings, sentence embeddings, and document embeddings.
Wordpiece, BPE, unigram tokenization, and vocabulary construction trade-offs.
Word2Vec, GloVe, fastText, analogies, and limitations of static embeddings.
Autoregressive modeling, sequence labeling, and temporal dependence.
Recurrent modeling, encoder-decoder systems, and practical weaknesses.
Modern language modeling architecture, scaling, and training mechanics.
Masked language modeling, pretraining objectives, and transfer learning.
Soft attention, self-attention, cross-attention, and computational trade-offs.
Translation pipelines, alignment, decoding, and multilingual evaluation.
BLEU, ROUGE, perplexity, WER, accuracy, precision, recall, and F1.
Attention inspection, probing, attribution, and limits of mechanistic claims.
How performance changes with data, parameters, and compute budgets.
Capabilities, limitations, training pipeline, deployment patterns, and future directions.
What each metric measures, when accuracy misleads, the precision-recall trade-off, F-beta, and multi-class averaging.
What activation functions are, ReLU, sigmoid, and softmax explained, and how backpropagation trains a network using the chain rule.
Personal study notes: LLM fine-tuning, CNNs, RNNs, Transformers, attention, hyperparameters, KV cache, and more.
50 multiple-choice questions covering neural networks, attention, embeddings, training, and interview-focused topics.
MSE, MAE, Huber, binary cross-entropy, categorical cross-entropy, hinge, KL divergence, contrastive, and triplet loss.
Neurons, activation functions, backpropagation, CNNs, ResNet, transfer learning, Transformers, BERT, GPT, and generative models.