Andrew Fairless, Ph.D.
/About/Bio
/Projects
/Reposts
/Tags
/Categories
Entries tagged :: large language model
.
2025-08-14
What I Read: Scale
2025-08-13
What I Read: Prompt Optimization
2025-08-04
What I Read: Learning Agentic Patterns
2025-07-31
What I Read: AGI not milestone
2025-07-29
What I Read: Scaling RL
2025-07-28
What I Read: Save Product
2025-07-24
What I Read: AI Products
2025-07-22
What I Read: BM25F
2025-07-21
What I Read: Language Models
2025-07-15
What I Read: reasoning research
2025-07-02
What I Read: Generative AI
2025-07-01
What I Read: LLMOps, human
2025-06-30
What I Read: Cybernetic Teammate
2025-06-26
What I Read: Recommendation, LLMs
2025-06-24
What I Read: Small Language Models
2025-06-23
What I Read: LLM Agents
2025-06-19
What I Read: LLM Reasoning
2025-06-18
What I Read: Domain specific architectures
2025-06-12
What I Read: reinforcement learning
2025-06-11
What I Read: LLMs in medicine
2025-06-10
What I Read: Science-Literate AI
2025-06-05
What I Read: Distributed Systems Programming
2025-06-03
What I Read: Model, Product
2025-05-28
What I Read: BAML
2025-05-27
What I Read: BAML, agentic
2025-05-26
What I Read: RL, PPO, GRPO
2025-05-22
What I Read: Group relative policy optimization
2025-05-21
What I Read: reasoning LLMs
2025-05-19
What I Read: chatbot limitations
2025-05-15
What I Read: next weak learners
2025-05-12
What I Read: AI-ready data
2025-05-08
What I Read: short case, Nvidia
2025-04-30
What I Read: adaptive LLM
2025-04-29
What I Read: tensor dimensions, transformers
2025-04-28
What I Read: building AI
2025-04-27
What I Read: cosine similarity
2025-04-24
What I Read: AI HCI
2025-04-23
What I Read: AI languages
2025-04-17
What I Read: model merging
2025-04-15
What I Read: Agents
2025-04-14
What I Read: Age of Data
2025-04-09
What I Read: Building agents
2025-04-01
What I Read: reward hacking
2025-03-25
What I Read: autoencoders, interpretability
2025-03-06
What I Read: multimodal LLMs
2025-03-05
What I Read: LLMs, school math
2025-03-03
What I Read: debate, ai
2025-02-25
What I Read: llm judge
2025-02-12
What I Read: evaluation quicksand
2025-02-11
What I Read: Mamba, State Space Models
2025-01-27
What I Read: Transformers Inference Optimization
2025-01-21
What I Read: GenAI, Classify Text
2025-01-09
What I Read: LLMs, 2024
2025-01-06
What I Read: embedding models
2024-12-19
What I Read: Toy Models, Superposition?
2024-12-17
What I Watch: LLM agents, production
2024-12-16
What I Watch: How LLMs store facts
2024-12-11
What I Read: Fine Tuning LLM
2024-12-04
What I Read: passively learned, causality
2024-12-03
What I Read: Evaluating LLM-Evaluators
2024-11-21
What I Read: Classifying pdfs
2024-11-20
What I Read: Tool Retrieval, RAG
2024-11-18
What I Read: LLM Pre-training Post-training
2024-11-13
What I Read: Open-endedness, Agentic AI
2024-10-30
What I Read: Visual Guide, Quantization
2024-10-29
What I Read: Generative AI Platform
2024-10-28
What I Read: History, Transformer
2024-10-21
What I Read: LLM evaluation
2024-10-17
What I Read: Data Flywheels, LLM
2024-10-15
What I Read: Improving Language Models, Practical Size
2024-10-07
What I Read: Extrinsic Hallucinations, LLMs
2024-10-01
What I Read: What can LLMs never do?
2024-09-25
What I Read: bare metal to 70B
2024-09-23
What I Read: Detecting hallucinations, LLMs, semantic entropy
2024-09-19
What I Read: Structured Generation, LLMs
2024-09-16
What I Read: Musings on AI Engineering
2024-09-04
What I Read: LLMs train LLMs
2024-09-03
What I Read: Summarization, LLMs
2024-08-26
What I Read: neural systems understanding
2024-08-19
What I Read: What We Learned Building LLMs
2024-08-15
What I Read: Merge Large Language Models
2024-08-13
What I Read: LLM Pipelines, DSPy
2024-08-07
What I Read: LLM evaluation
2024-08-06
What I Read: LLM, DSPy Assertions and Suggestions
2024-08-05
What I Read: implicit biases, LLM
2024-07-29
What I Read: Platonic Hypothesis
2024-07-25
What I Read: Game Theory, AI
2024-07-15
What I Read: LLMs, Open Source
2024-06-20
What I Read: Structured Generation, Constrained Decoding
2024-06-18
What I Read: Attention, transformers
2024-06-12
What I Read: Data Selection, LLMs
2024-06-10
What I Read: Mamba Explained
2024-06-03
What I Read: Chain-of-Thought Reasoning
2024-05-20
What I Read: text embeddings
2024-05-14
What I Read: 1-bit LLMs, 1.58 Bits
2024-05-09
What I Read: Mamba, Easy Way
2024-04-30
What I Read: Structured State Space Sequence Models
2024-04-29
What I Read: Forgetting Can Help AI Learn
2024-04-25
What I Read: Predictive Human Preference, Model Ranking to Model Routing
2024-04-22
What I Read: Compound AI Systems
2024-04-18
What I Read: Scaling ChatGPT, Engineering Challenges
2024-04-10
What I Read: How Quickly LLMs Learn Skills?
2024-04-09
What I Read: Deploy Model
2024-04-04
What I Read: LoRA from Scratch
2024-04-03
What I Read: LLM Evaluation Metrics
2024-04-01
What I Read: Artificial, Biological Intelligence
2024-03-19
What I Read: Sampling Text Generation
2024-03-18
What I Read: Chatbots Understand Text
2024-03-04
What I Read: Self-Attention in GPT
2024-02-21
What I Read: Research Directions
2024-02-19
What I Read: Instruction Tuning
2024-02-15
What I Read: Will Scaling Solve Robotics?
2024-02-14
What I Read: 3D human pose estimation
2024-02-13
What I Read: Limits of Transformers on Compositionality
2024-02-08
What I Read: survey LLM tooling
2024-02-07
What I Read: Multi-Modal Retrieval-Augmented Generation
2024-02-06
What I Read: Adversarial Attacks on LLMs
2024-02-05
What I Read: Finetuning LLMs Using LoRA
2024-01-30
What I Read: Nvidia, GPU gold rush
2024-01-23
What I Read: Helping AI See
2024-01-11
What I Read: Enterprise AI, RAG + Fine Tuning
2024-01-04
What I Read: Multimodality
2024-01-03
What I Read: Finetuning LLMs with LoRA and QLoRA
2023-12-20
What I Read: Distributed Training, Finetuning
2023-12-19
What I Read: Artificial General Intelligence
2023-12-11
What I Read: Tiny Language Models
2023-12-07
What I Read: LLM Apps, Data Pipelines
2023-12-05
What I Read: evaluating AI systems
2023-12-04
What I Read: Problems of AI Consciousness
2023-11-27
What I Read: Privacy side channels in ML
2023-11-16
What I Read: Estimate Token Importance in LLM Prompts
2023-11-08
What I Read: How make history with LLMs
2023-11-06
What I Read: Optimizing LLM in production
2023-10-30
What I Read: LLM Training, RLHF
2023-10-23
What I Read: LLMs, single example
2023-10-12
What I Read: GPT-4, 8 Models in One
2023-10-10
What I Read: LLM research
2023-10-02
What I Read: Models Memorize or Generalize?
2023-09-28
What I Read: scaling laws, cross-entropy loss
2023-09-25
What I Read: LLMs in Planning
2023-09-21
What I Read: Economic Case for Generative AI
2023-09-19
What I Read: LLM-based Products
2023-09-13
What I Read: Attention Off By One
2023-09-11
What I Read: What Do LLMs Know About Linguistics?
2023-09-07
What I Read: LLMs
2023-08-29
What I Read: shape of AGI
2023-08-28
What I Read: What we dont talk about
2023-08-21
What I Read: Disagreement Modelling
2023-08-17
What I Read: LLM Agents
2023-08-16
What I Read: artificial intelligence really hard
2023-08-08
What I Read: Ways Digital Minds Know
2023-08-04
What I Read: Attack Impacts AI Chatbots
2023-07-27
What I Read: LLM Chatbots, Browser
2023-07-26
What I Read: Neural Networks Learn Language
2023-07-19
What I Read: Hard Stuff, Building Products, LLMs
2023-07-17
What I Read: Prompt injection
2023-07-12
What I Read: What, Why ChatGPT