Shakespeare and the Bible: An AI Investigation

Summary Could the greatest playwright of all time have secretly shaped one of the most influential religious texts in history? Some believe William Shakespeare left his mark on the King James Bible hidden in plain sight. With the power of AI, we’ll investigate whether there’s any truth to this conspiracy. You can read about the conspiracy here: Did Shakespeare write the King James Bible? Shakespeare and the King James Bible – Some Tentative Conclusions

PostgreSQL for AI: Storing and Searching Embeddings with pgvector

Summary Vector databases are essential for modern AI applications like semantic search, recommendation systems, and natural language processing. They allow us to store and query high-dimensional vectors efficiently. With the pgvector extension PostgreSQL becomes a powerful vector database, enabling you to combine traditional relational data with vector-based operations. In this post, we will walk through the full process: Installing PostgreSQL and pgvector Setting up a vector-enabled database Generating embeddings using Ollama Running similarity queries with Python By the end, you’ll be able to store, query, and compare high-dimensional vectors in PostgreSQL, opening up new possibilities for AI-powered applications.

Build Smarter AI: Leveraging the Model Context Protocol for Dynamic Context

Summary The evolution of technology is driven by protocols structured ways for systems to communicate and interact. The internet, APIs, and even modern databases rely on protocols to function efficiently. Similarly, as AI becomes more powerful, it needs a structured and standardized way to manage context across interactions. Enter the Model Context Protocol (MCP) a framework designed to enhance the way AI models understand, retain, and utilize context over multiple exchanges.

Getting Started with Neo4j: Build Your First Knowledge Graph

Summary In AI and data science, knowledge graphs are powerful tools for modeling complex relationships between entities. They enable intelligent querying, recommendation systems, and semantic search. Neo4j, an open-source graph database, is one of the most popular tools for building and managing knowledge graphs. In this post, we’ll walk you through setting up Neo4j, configuring it for use as a knowledge graph, and manipulating the database using Python. A knowledge graph consists of buckets of information (entities) with intelligent links between those buckets.

Beyond Text Generation: Coding Ollama Function Calls and Tools

Summary Function calling allows Large Language Models (LLMs) to interact with APIs, databases, and other tools, making them more than just text generators. Integrating LLMs with functions enables you to harness their powerful text processing capabilities, seamlessly enhancing the technological solutions you develop. This post will explain how you can call local python functions and tools in Ollama. Introduction to Ollama Function Calling Ollama allows you to run state-of-the-art LLMs like Qwen, Llama, and others locally without relying on cloud APIs.

From +AI to AI+: Embracing AI as the Foundation of Everything You Do

Summary What if AI isn’t just a tool but the foundation of how you work, create, and think? Today there are two distinct ways to approach AI: +AI and AI+. +AI means integrating AI into an existing workflow, business, or process. It’s the approach discussed in countless books, YouTube videos, and business strategies. If you’re working, running a business, or earning money, this is the natural and logical way to use AI.

Building AI-Powered Applications with Haystack and Ollama

Summary In this post, I will demonstrate how to set up and use haystack with Ollama. haystack is a framework that helps when building applications powered by LLMs. It offers extensive LLM-related functionality. It is open source under the Apache license. It is actively developed, with numerous contributors. It is widely used in production by various clients. These are some of the key items to watch for when using a library in a project.

LiteLLM: A Lightweight Wrapper for Multi-Provider LLMs

Summary In this post I will cover LiteLLM. I used it for my implementation of Textgrad also it was using in blog posts I did about Agents. Working with multiple LLM providers is painful. Every provider has its own API, requiring custom integration, different pricing models, and maintenance overhead. LiteLLM solves this by offering a single, unified API that allows developers to switch between OpenAI, Hugging Face, Cohere, Anthropic, and others without modifying their code.

TextGrad: dynamic optimization of your LLM

Summary This post aims to be a comprehensive tutorial on Textgrad. Textgrad enables the optimization of LLM’s using their text responses. This will be part of SmartAnswer the ultimate LLM query tool which I will be blogging about shortly. Why TextGrad? Brings Gradient Descent to LLMs – Instead of numerical gradients, TextGrad leverages textual feedback to iteratively improve outputs. Automates Prompt Optimization – Eliminates the guesswork in refining LLM prompts. Works with Any LLM – From OpenAI’s GPT to local models like Ollama.

The Power of Logits: Unlocking Smarter, Safer LLM Responses

Summary In this blog post I want to fully explore logits and how they can be used to enhance AI applications I want to understand the ideas from this paper: “Is That Your Final Answer? Test-Time Scaling Improves Selective Question Answering” This paper introduces a new approach, Selective Question Answering (SQA). This introduces confidence scores to decide when an answer should be given. In this post, we’ll cover the core insights of the paper and implement a basic confidence-based selection function in Python.

Efficient Similarity Search with FAISS and SQLite in Python

Summary This is another component in SmartAnswer and enhanced LLM interface. In this blog post, we introduce a wrapper class, FaissDB, which integrates FAISS with SQLite or any database to manage document embeddings and enable efficient similarity search. This approach combines FAISS’s vector search capabilities with the storage and querying power of a database, making it ideal for applications such as Retrieval-Augmented Generation (RAG) and recommendation systems. It builds up this tool PaperSearch.

Automating Paper Retrieval and Processing with PaperSearch

Summary This is part on in a series of blog post working towards SmartAnswer a comprehensive improvement to how Large Language Models LLMs answer questions. This tool will be the source of data for SmartAnswer and allow it to find and research better data when generating answers. I want this tool to be included in that solution but I dot want all the code from this tool distracting from the SmartAnswer solution.

SQLite: the small database that packs a big punch

Summary SQLite is one of the most widely used database engines in the world, powering everything from mobile applications (Android, iOS) to browsers (Google Chrome, Mozilla Firefox), IoT devices, and even gaming consoles. Unlike traditional client-server databases (e.g., MySQL, PostgreSQL), SQLite is an embedded, serverless database that stores data in a single file, making it easy to manage and deploy. Python developers frequently choose SQLite for its inherent simplicity and portability, leveraging the built-in sqlite3 module for effortless database integration.

RAFT: Reward rAnked FineTuning - A New Approach to Generative Model Alignment

Summary This post is an explanation of this paper:RAFT: Reward rAnked FineTuning for Generative Foundation Model Alignment. Generative foundation models, such as Large Language Models (LLMs) and diffusion models, have revolutionized AI by achieving human-like content generation. However, they often suffer from Biases – Models can learn and reinforce societal biases present in the training data (e.g., gender, racial, or cultural stereotypes). Ethical Concerns – AI-generated content can be misused for misinformation, deepfakes, or spreading harmful narratives.

Faiss: A Fast, Efficient Similarity Search Library

Summary Searching through massive datasets efficiently is a challenge, whether in image retrieval, recommendation systems, or semantic search. Faiss (Facebook AI Similarity Search) is a powerful open-source library developed by Meta to handle high-dimensional similarity search at scale. It’s particularly well-suited for tasks like: Image search: Finding visually similar images in a large database. Recommendation systems: Recommending items (products, movies, etc.) to users based on their preferences. Semantic search: Finding documents or text passages that are semantically similar to a given query.

K-Means Clustering

Summary Imagine you have a dataset of customer profiles. How can you group similar customers together to tailor marketing campaigns? This is where K-Means clustering comes into play. K-Means is a popular unsupervised learning algorithm used for clustering data points into distinct groups based on their similarities. It is widely used in various domains such as customer segmentation, image compression, and anomaly detection. In this blog post, we’ll cover how K-Means works and demonstrate its implementation in Python using scikit-learn.

AI: The Future Interface to Technology

Summary Imagine a world where you simply think of a task, and invisible devices seamlessly execute it. In fact most of what used to be your daily tasks you won’t even think about they will be automatically executed. Sounds like science fiction? This I believe is the future of human technology interaction. The technology disappears behind an AI driven interface. Do we currently have Artificial Intelligence Artificial intelligence refers to computer programs designed to mimic human cognitive abilities, such as understanding natural language, recognizing patterns, learning from data, and solving complex problems.

Self-Learning LLMs for Stock Forecasting: A Python Implementation with Direct Preference Optimization

Summary Forecasting future events is a critical task in fields like finance, politics, and technology. However, improving the forecasting abilities of large language models (LLMs) often requires extensive human supervision. In this post, we explore a novel approach from the paper LLMs Can Teach Themselves to Better Predict the Future that enables LLMs to teach themselves better forecasting skills using self-play and Direct Preference Optimization (DPO). We’ll walk through a Python implementation of this method, step by step.

Using Quantization to speed up and slim down your LLM

Summary Large Language Models (LLMs) are powerful, but their size can lead to slow inference speeds and high memory consumption, hindering real-world deployment. Quantization, a technique that reduces the precision of model weights, offers a powerful solution. This post will explore how to use quantization techniques like bitsandbytes, AutoGPTQ, and AutoRound to dramatically improve LLM inference performance. What is Quantization? Quantization reduces the computational and storage demands of a model by representing its weights with lower-precision data types.

Mastering LLM Fine-Tuning: A Practical Guide with LLaMA-Factory and LoRA

Summary Large Language Models (LLMs) offer immense potential, but realizing that potential often requires fine-tuning them on task-specific data. This guide provides a comprehensive overview of LLM fine-tuning, focusing on practical implementation with LLaMA-Factory and the powerful LoRA technique. What is Fine-Tuning? Fine-tuning adapts a pre-trained model to a new, specific task or dataset. It leverages the general knowledge already learned by the model from a massive dataset (source domain) and refines it with a smaller, more specialized dataset (target domain).

Debugging Jupyter Notebooks in VS Code

Summary Visual Studio Code is the most popular editor for development. Jupyter Notebooks is the most widely used way to share, demonstrate and develop code in modern AI development. Debugging code is not just used when you have a bug. After you have written any substantial piece of code I suggest stepping through it in the debugger if possible. This can help improve you understanding and the quality of the code you have written

DeepResearch Part 3: Getting the best web data for your research

Summary This post details building a robust web data pipeline using SmolAgents. We’ll create tools to retrieve content from various web endpoints, convert it to a consistent format (Markdown), store it efficiently, and then evaluate its relevance and quality using Large Language Models (LLMs). This pipeline is crucial for building a knowledge base for LLM applications. Web Data Convertor (MarkdownConverter) We leverage the MarkdownConverter class, inspired by the one in autogen, to handle the diverse formats encountered on the web.

DeepResearch Part 2: Building a RAG Tool for arXiv PDFs

Summary In this post, we’ll build a Retrieval Augmented Generation (RAG) tool to process the PDF files downloaded from arXiv in the previous post DeepResearch Part 1. This RAG tool will be capable of loading, processing, and semantically searching the document content. It’s a versatile tool applicable to various text sources, including web pages. Building the RAG Tool Following up on our arXiv downloader, we now need a tool to process the downloaded PDFs.

DeepResearch Part 1: Building an arXiv Search Tool with SmolAgents

Summary This post kicks off a series of three where we’ll build, extend, and use the open-source DeepResearch application inspired by the Hugging Face blog post. In this first part, we’ll focus on creating an arXiv search tool that can be used with SmolAgents. DeepResearch aims to empower research by providing tools that automate and streamline the process of discovering and managing academic papers. This series will demonstrate how to build such tools, starting with a powerful arXiv search tool.

DeepResearch Part 1: Building an arXiv Search Tool with SmolAgents

Summary This post kicks off a series of three where we’ll build, extend, and use the open-source DeepResearch application inspired by the Hugging Face blog post. In this first part, we’ll focus on creating an arXiv search tool that can be used with SmolAgents. DeepResearch aims to empower research by providing tools that automate and streamline the process of discovering and managing academic papers. This series will demonstrate how to build such tools, starting with a powerful arXiv search tool.

FFmpeg: A Practical Guide to Essential Command-Line Options

Introduction FFmpeg is an incredibly versatile command-line tool for manipulating audio and video files. This post provides a practical collection of useful FFmpeg commands for common tasks. FFmpeg Command Structure The general structure of an FFmpeg command is: ffmpeg [global_options] {[input_file_options] -i input_url} ... {[output_file_options] output_url} ... Merging Video and Audio Merging video and audio, with audio re-encoding ffmpeg -i video.mp4 -i audio.wav -c:v copy -c:a aac output.mp4 Copying the audio without re-encoding ffmpeg -i video.

Writing Neural Networks with PyTorch

Summary This post provides a practical guide to building common neural network architectures using PyTorch. We’ll explore feedforward networks, convolutional neural networks (CNNs), recurrent neural networks (RNNs), LSTMs, transformers, autoencoders, and GANs, along with code examples and explanations. 1. Understanding PyTorch’s Neural Network Module PyTorch provides the torch.nn module to build neural networks. It provides classes for defining layers, activation functions, and loss functions, making it easy to create and manage complex network architectures in a structured way.

Mastering Prompt Engineering: A Practical Guide

Summary This post provides a comprehensive guide to prompt engineering, the art of crafting effective inputs for Large Language Models (LLMs). Mastering prompt engineering is crucial for maximizing the potential of LLMs and achieving desired results. Effective prompting is the easiest way to enhance your experience with Large Language Models (LLMs). The prompts we make are our interface to LLMs. This is how we communicate with them. This is why it is important to understand how to do it well.

Harnessing the Power of Stable Diffusion WebUI

Summary In this blog I aim to try building using open source tools where possible. The benefits are price, control, knowledge and eventually quality. In the shorter term though the quality will trail the paid versions. My belief is we can construct AI applications to be self correcting sort of like how your camera auto focuses for you. This process will involve a lot of computation so using a paid service could be costly.

Creating AI-Powered Paper Videos: From Research to YouTube

Summary This post demonstrates how to automatically transform a scientific paper (or any text/audio content) into a YouTube video using AI. We’ll leverage several powerful tools, including large language models (LLMs), Whisper for transcription, Stable Diffusion for image generation, and FFmpeg for video assembly. This process can streamline content creation and make research more accessible. Overview Our pipeline involves these steps: Audio Generation (Optional): If starting from a text document, we’ll use a text-to-speech service (like NotebookLM, or others) to create an audio narration.

Fast Poisson Disk Sampling in Arbitrary Dimensions

Summary In this post I explore Robert Bridson’s paper: Fast Poisson Disk Sampling in Arbitrary Dimensions and provide an example python implementation. Additionally, I introduce an alternative method using Cellular Automata to generate Poisson disk distributions. Poisson disk sampling is a widely used technique in computer graphics, particularly for applications like rendering, texture generation, and particle simulation. Its appeal lies in producing sample distributions with “blue noise” characteristics—random yet evenly spaced, avoiding clustering.

Activation Functions

Introduction Activation functions are a component of neural networks they introduce non-linearity into the model, enabling it to learn complex patterns. Without activation functions, a neural network would essentially act as a linear model, regardless of its depth. Key Properties of Activation Functions Non-linearity: Enables the model to learn complex relationships. Differentiability: Allows backpropagation to optimize weights. Range: Defines the output range, impacting gradient flow. In this post I will outline each of the most common activation functions how they are calculated and when they should be used.

SVM Support Vector Machine an introduction

Summary In this post I will implement a Support Vector Machine (SVM) in python. Then describe what it does how it does it and some applications of the instrument. What Are Support Vector Machines (SVM)? Support Vector Machines (SVM) are supervised learning algorithms used for classification and regression tasks. Their strength lies in handling both linear and non-linear problems effectively. By finding the optimal hyperplane that separates classes, SVMs maximize the margin between data points of different classes, making them highly effective in high-dimensional spaces.

Color wars: Cellular Automata fight until one domiates

Summary This post is about color wars: a grid containing dynamic automata at war until one dominates. Implementation The implementation consists of two core components: the Grid and the CellularAutomaton. 1. CellularAutomaton Class The CellularAutomaton class represents individual entities in the grid. Each automaton has: Attributes: ID, strength, age, position. Behavior: Updates itself by aging, reproducing, or dying based on simple rules. 2. Grid Class The Grid manages a collection of automata.

More Machine Learning Questions and Answers with Python examples

44. What does it mean to Fit a Model? Answer Fitting a model refers to the process of adjusting the model’s internal parameters to best match the given training data. It’s like tailoring a suit – you adjust the fabric and stitching to make it fit the wearer perfectly. Key Terms: Model: A mathematical representation that captures patterns in data. Examples include linear regression, decision trees, neural networks, etc. Parameters: These are the internal variables within the model that determine its behavior.

Cellular Automata: Traffic Flow Simulation using the Nagel-Schreckenberg Model

Summary The Nagel-Schreckenberg (NaSch) model is a traffic flow model which uses used cellular automata to simulate and predict traffic on roads. Design of the Nagel-Schreckenberg Model Discrete Space and Time: The road is divided into cells, each representing a fixed length (e.g., a few meters). Time advances in discrete steps. Vehicle Representation: Each cell is either empty or occupied by a single vehicle. Each vehicle has a velocity (an integer) which determines how many cells it moves in a single time step.

Simulate Gastropod Shell Growth Using Cellular Automata

Summary I started with this paper A developmentally descriptive method forquantifying shape in gastropod shells and bridged the results to a cellular automata approach. An example of the shell we are modelling: Steps 1. Identify the Key Biological Features The paper outlines the logarithmic helicospiral model for shell growth, where: The shell grows outward and upward in a spiral shape. Parameters like width growth ((g_w)), height growth ((g_h)), and aperture shape dictate the final form.

Cellular Automata: Introduction

Summary This page is the first in a series of posts about Cellular Automata. I believe that we could get the first evidence of AI through cellular automata. A recent paper Intelligence at the Edge of Chaos found that LLM’s trained on more complex data generate better results. Which makes sense in a human context like the harder the material is I study the smarter I get. We need to find out why this is also the case with machines.

Rag: Retrieval-Augmented Generation

Summary Retrieval-Augmented Generation (RAG) is a powerful technique that enhances large language models (LLMs) by allowing them to use external knowledge sources. An Artificial Intelligence (AI) system consists of components working together to apply knowledge learned from data. Some common components of those systems are: Large Language Model (LLM): Typically the core component of the system, often there is more than one. These are large models that have been trained on massive amounts of data and can make intelligent predictions based on their training.

CAG: Cache-Augmented Generation

Summary CAG performs better but does not solve the key reason RAG was created small context windows. Retrieval-Augmented Generation (RAG) is currently(early 2025) the most popular way to use external knowledge in current LLM opperations. RAG allows you to enhance your LLM with data beyond the data it was trained on. Ther are many great RAG solutions and products. RAG has some drawbacks - There can be significant retreival latency as it searches for and organizes the correct data.

Agents: A tutorial on building agents in python

LLM Agents Agents are used enhance and extend the functionality of LLM’s. In this tutorial, we’ll explore what LLM agents are, how they work, and how to implement them in Python. What Are LLM Agents? An agent is an autonomous process that may use the LLM and other tools multiple times to achieve a goal. The LLM output often controls the workflow of the agent(s). What is the difference between Agents and LLMs or AI?

Courses: Free course on Agentic AI

Some FREE AI courses on Agents I recommend doing Agents were important in Machine Learning development last year. These are some courses I have done and recommend they are all free. A good reason to do courses and look at youtube videos is you will learn current applications of AI and may get ideas for new applications. AI Agentic Design Patterns with AutoGen AI Agentic Design Patterns with AutoGen Topics: Agents, Microsoft, Autogen.

Ollama: The local LLM solution

Using Ollama Introduction Ollama is the best platform for running, managing, and interacting with Large Language Models (LLM) models locally. For Python programmers, Ollama offers seamless integration and robust features for querying, manipulating, and deploying LLMs. In this post I will explore how Python developers can leverage Ollama for powerful and efficient AI-based workflows. 1. What is Ollama? Ollama is a tool designed to enable local hosting and interaction with LLMs.

Hugo: A Static Site Generator

Hugo: A Static Site Generator In this post I give an introduction to what I think is the best static site generator: Hugo. What is Hugo? Hugo is an open-source static site generator written in Go. It takes structured content, often written in Markdown, and compiles it into static HTML, CSS, and JavaScript files. Setting Up Hugo: A Quickstart Guide Follow these steps to set up your first Hugo site

ChromaDB: The Lightweight Open-Source Vector Database for AI Applications

1️⃣ Introduction In the era of AI-powered search, retrieval-augmented generation (RAG), and recommendation systems, efficient vector search is a necessity. While many vector databases exist, most require heavy infrastructure. Enter ChromaDB: a lightweight, open-source vector database optimized for rapid prototyping and local AI applications. 2️⃣ What is ChromaDB? Definition: ChromaDB is a vector database for storing and querying embeddings. It provides an easy-to-use interface for AI developers to integrate similarity search into their applications.

Project 7: FX-Trader

A derivatives trading system. This is my attempt at building a derivatives trading system. In this first post I am going to outline the goals of the project and some of the early design decisions.

Project 6: Validator

A validation tool for excel files. Sometimes you need to export data from one system for loading into another. For instance you may export a report from a derivatives trading system for information for a collateral management system.

This Excel macro file validates input files to make sure that they are in a specific format. The input file can be in any format that excel can load. The workbook will load the file and check for errors in place.

Project 5: Dictator

Google have just released their speech API. One really cool feature is the ability to transcribe voice in real time. Two years ago I built an app with this idea in mind. At that at the time I could no make it work. Now it is time to resurrect that app. This post will cover the recording section of that application.

Project 4: Meth

This is another simple android application. It does one thing It keeps your phone awake.

Project 3: Site Shot

This is a really simple android application. It does one thing It takes a photo of a web site and allows you to share the photo.

A script to generate android images

This is a script I use in my android projects to generate different sized images. You can find the script here process.vbs So I build all my images in inkscape. This is brilliant application. You can find some brilliant tutorials here heathenx The script uses inkscape to convert the svg files to png images of different sizes. The script also use imagemajic to format the pngs nicely. Finally it compresses the result using two png crushing programs

Project 2: File Explorer for android

In this project I am building a file explorer library for android. As I was working on catcher it became obvious I would need a file picker and explorer solution. So I did a bit of looking on the web. I found three interesting projects that nearly did what I wanted. I put a few of them together to come up with a hybrid solution.

Project 1: Catcher

This is an android application to transfer files from your phone to somewhere else. I will be built as a PC solution but can be used for a server solution also.

AI Is the Interface: The Future of Human-Technology Interaction Technology is the bridge that transforms data into knowledge. In the coming years, artificial intelligence will evolve from being a tool that assists humans to becoming the primary interface through which we interact with technology and process information. The future of human-computer interaction will not be through keyboards, touchscreens, or even direct programming—it will be mediated by AI systems that understand, interpret, and execute our intentions seamlessly.