Blog

Your blog category

Empowering LLMs to Think Deeper by Erasing Thoughts

Introduction Recent large language models (LLMs) — such as OpenAI’s o1/o3, DeepSeek’s R1 and Anthropic’s Claude 3.7 — demonstrate that allowing the model to think deeper and longer at test time can significantly enhance model’s reasoning capability. The core approach underlying their deep thinking capability is called chain-of-thought (CoT), where the model iteratively generates intermediate

Empowering LLMs to Think Deeper by Erasing Thoughts Read More »

How I Finally Understood MCP — and Got It Working in Real Life

Table of Content Introduction: Why I Wrote This The Evolution of Tool Integration with LLMs What Is Model Context Protocol (MCP), Really? Wait, MCP sounds like RAG… but is it? In an MCP-based setup In a traditional RAG system Traditional RAG Implementation MCP Implementation Quick recap! Core Capabilities of an MCP Server Real-World Example: Claude

How I Finally Understood MCP — and Got It Working in Real Life Read More »

The Westworld Blunder

We’re entering an interesting moment in AI development. AI systems are getting memory, reasoning chains, self-critiques, and long-context recall. These capabilities are exactly some of the things that I’ve previously written would be prerequisites for an AI system to be conscious. Just to be clear, I don’t believe today’s AI systems are self-aware, but I no longer

The Westworld Blunder Read More »

Pause Your ML Pipelines for Human Review Using AWS Step Functions + Slack

Have you ever wanted to pause an automated workflow to wait for a human decision? Maybe you need approval before provisioning cloud resources, promoting a machine learning model to production, or charging a customer’s credit card. In many data science and machine learning workflows, automation gets you 90% of the way — but that critical last

Pause Your ML Pipelines for Human Review Using AWS Step Functions + Slack Read More »

What My GPT Stylist Taught Me About Prompting Better

When I built a GPT-powered fashion assistant, I expected runway looks—not memory loss, hallucinations, or semantic déjà vu. But what unfolded became a lesson in how prompting really works—and why LLMs are more like wild animals than tools. This article builds on my previous article on TDS, where I introduced Glitter as a proof-of-concept GPT

What My GPT Stylist Taught Me About Prompting Better Read More »

Log Link vs Log Transformation in R — The Difference that Misleads Your Entire Data Analysis

Although normal distributions are the most commonly used, a lot of real-world data unfortunately is not normal. When faced with extremely skewed data, it’s tempting for us to utilize log transformations to normalize the distribution and stabilize the variance. I recently worked on a project analyzing the energy consumption of training AI models, using data

Log Link vs Log Transformation in R — The Difference that Misleads Your Entire Data Analysis Read More »

A Review of AccentFold: One of the Most Important Papers on African ASR

I really enjoyed reading this paper, not because I’ve met some of the authors before, but because it felt necessary. Most of the papers I’ve written about so far have made waves in the broader ML community, which is great. This one, though, is unapologetically African (i.e. it solves a very African problem), and I

A Review of AccentFold: One of the Most Important Papers on African ASR Read More »