Cline: The Next Generation AI Coding Assistant

An exploration of Cline, the autonomous AI coding agent that lives in your IDE and handles complex, multi-step engineering tasks through tool-use and agency.

April 9, 2026 · 3 min · James Myddelton

The Rise of Small Language Models: Why Size Isn't Everything

The Rise of Small Language Models: Why Size Isn’t Everything For years, the narrative was simple: bigger is better. GPT-4 was massive, Claude was massive, and the race seemed to be about who could train the largest model on the most data. But that story is changing. Small language models - typically under 15 billion parameters - are proving that you don’t need 175 billion parameters to solve real problems. The shift isn’t just about efficiency. It’s a fundamental change in how we think about AI deployment, cost, and what actually matters for most use cases. ...

April 9, 2026 · 7 min · James M

Open WebUI: A Polished Interface for Local and Remote LLMs

If you’ve spent time running language models locally through Ollama or another inference engine, you’ve probably discovered the same friction point: the command-line experience works, but it’s clunky. You’re juggling terminal windows, managing conversation context manually, managing files through the filesystem. Open WebUI solves this by offering what Ollama itself didn’t: a genuinely usable interface. What Open WebUI Does Open WebUI is a web-based chat interface designed to work with language models. It’s styled after ChatGPT, with a familiar conversation layout, sidebar for conversation management, and all the modern UX conveniences you’d expect. The critical difference: you control the backend entirely. ...

April 8, 2026 · 6 min · James M

Running AI Models Locally with Ollama: From Setup to OpenClaw

Running AI Models Locally with Ollama: From Setup to OpenClaw Ollama has quietly become the go-to tool for developers who want to run large language models on their own machines without relying on APIs. No cloud costs, no rate limits, no sending your prompts to third-party servers. Just you, your hardware, and a surprisingly capable AI model running locally. What is Ollama? Ollama is a lightweight platform designed to make running open-source language models accessible. It handles the complexity of model management - downloading, optimization, memory management - so you just run a command and start prompting. ...

April 8, 2026 · 4 min · James M

Following the Money: Databricks vs Snowflake vs the Open-Source Alternative

A deep dive into the shifting economics of the data landscape in 2026. Why the choice between Snowflake and Databricks is increasingly an accounting decision, and where the open-source DIY stack actually saves you money.

April 8, 2026 · 4 min · James Myddelton

List of AI GitHub Projects

A collection of significant open-source AI projects that are shaping the ecosystem. AI Agent Frameworks AutoGen - Microsoft’s multi-agent conversation framework for building complex AI systems with role-based agents CrewAI - Framework for orchestrating autonomous AI agents that work together as a crew Langchain - Foundational library for building applications with LLMs, offering chains, agents, and memory abstractions Open Interpreter - Let LLMs run code locally and interact with your computer Code & Development Auto-GPT - Early autonomous AI agent that can break down goals and execute them iteratively Aider - AI pair programmer that can edit code in your local repository Prompt Engineering Guide - Comprehensive guide with papers, techniques, and best practices Specialized Tools OpenClaw - AI agent framework for operating graphical user interfaces directly Ollama - Simple way to run large language models locally LiteLLM - Unified interface for calling all major LLM APIs with cost tracking Research & Resources Transformers - Hugging Face’s comprehensive library for state-of-the-art NLP models Papers with Code - Curated dataset linking papers with their implementations

May 27, 2023 · 1 min · James M