Being certain of epistemic uncertainty

I’ve been dancing around Probability Theory, it’s history, application, and weeding out what is Frequentist from what is Bayesian. The. Relating it to Rational Psychology. I’m not there, getting there, but not there. This isn’t what I do full time but it is what I think about when I’m not working, parenting, or socialising. Thankfully it’s not interesting to friends and family so I get a break from it myself! πŸ˜†

Continue reading β†’

Decisions decisions - Final Year Project πŸ€”πŸ€”πŸ€”

I am very torn between two possibilities : Building on my Q-Learning Maze Solving Agent I did for AI Applications (Q-Learning Maze Solving Agent) by adding a Neural Network (Sutton and Barton) Building on the Intelligent Agents work I did in AI by applying an Agent Decisions Process (Self-Consistency LLM-Agent) (the process is my interpretation of Russell and Norvig’s work) (The Douglas Adam’s extra option πŸ€“) Adding a cached “self-awareness” layer based on a Bayesian Learning Agent that stores it’s certainty on answers it gives.

Continue reading β†’

Can LLMs do Critical Thinking? Of course not. Can an AI system think critically? Why not?

A very interesting paper on Critical Thinking in an LLM (or lack thereof) Our study investigates how language models handle multiple-choice questions that have no correct answer among the options. Unlike traditional approaches that include escape options like None of the above (Wang et al., 2024a; Kadavath et al., 2022), we deliberately omit these choices to test the models’ critical thinking abilities. A model demonstrating good judgment should either point out that no correct answer is available or provide the actual correct answer, even when it’s not listed.

Continue reading β†’

The old ‘un still does the job on the MINST Handwritten dataset !

Ska-en-Provence πŸŽΆπŸŽ‰

[Being Human 3/n]: moving on from previous unmet goals...

I wish I had time to finish: my research on the Evolution of Probalisitic Reasoning in AI Particularly Dempster-Shafer and Bayesian Networks How LLMs and Bayesian networks can be used for Risk Management create an youtube/insta/tiktok vid for my latest post on LLM Agent But I don’t!! So this is me putting it to one side…

Continue reading β†’

[Being Human 2/n] Being scrappy shows we are Human in this Brave New World

Polish is cheap in this Brave New World of AI. Being scrappy is a way of being authentic and, most importantly, Being Human!

Continue reading β†’

[IA Series 7/n] Building a Self-Consistency LLM-Agent: From PEAS Analysis to Production Code

Building a Self-Consistency LLM-Agent: From PEAS Analysis to Production Code - a guide to designing an LLM-based agent.

Continue reading β†’

A refreshing AI-en-Provence 🍦

AI-en-Provence πŸ€“πŸ˜‚βœŒπŸΌβœŒπŸΌβœŒπŸΌ

[IA Series 6/n] A Bayesian Learning Agent: Bayes Theorem and Intelligent Agents

The article discusses how to implement Bayes Theorem in a learning agent that updates its beliefs about an environment based on new evidence, illustrated through a game involving guessing a number derived from a dice throw.

Continue reading β†’

It is not reasoning...

Reasoning vs Stream of Consciousness - the output of a transformer is not reasoned in the way we think it is.

Continue reading β†’

[Being Human Series 1/n] Introspection and the cusp of not knowing

What is knowledge? Wtf am I trying to learn! Claude “thinks” this post is mental masturbation πŸ˜† well even the physical version serves a good purpose! πŸ€·πŸΌβ€β™‚οΈ

Continue reading β†’

[IA Series 5/n] The Evolution from Logic to Probability to Deep Learning: A course correction to Transformers

Introduction In the previous post, I shared my view on “Why Study Logic?”, we looked at the Knowledge Representation and highlighted the importance of Logic and Reasoning in storing and accessing Knowledge. In this post I’m going to highlight a section from the book “Introduction to Artificial Intelligence” by Wolfgang Ertel. His approach with this book was to make AI more accessible than Russel and Norvig’s 1000+ page bible. It worked for me.

Continue reading β†’

[IA Series 4/n] A Big Question: Why Study Logic in a World of Probabilistic AI?

Introduction The purpose of this article is to help me answer the question “Why am I studying Logic?”. If it helps you, that’d be great, let me know! The question comes from a nagging feeling of, why don’t I see logic used more in the ‘real world’. It could be a personal bias as I more easily see the utility of Rosenblatt’s work, where he looked at both Symbolic Logic and Probability Theory to help solve a problem and choose Probability Theory ([NN Series 1/n] From Neurons to Neural Networks: The Perceptron), with that we had the birth of the Artificial Neuron and the rest is history!

Continue reading β†’

"There must be an invisible sun, giving heat to everyone"

“There must be an invisible sun, giving heat to everyone”

Continue reading β†’

[IA Series 3/n] Intelligent Agents Term Sheet

“[IA Series 3/n] Intelligent Agents Term Sheet” breaks down essential AI terminology from Russell & Norvig’s seminal textbook. Learn what makes agents rational (or irrational), understand different agent types, and follow a structured 5-step design process from environment analysis to implementation. Perfect reference for AI practitioners and students. Coming next: how agents mirror human traits. #ArtificialIntelligence #IntelligentAgents #AIDesign

Continue reading β†’

Building an Intelligent Agent

First draft in public 😱 πŸ˜†πŸ€“ What’s the best way for an agent to build a semantically sound and syntactically correct knowledge base? Dog fooding my course material means the first step is to define the task environment. /Checks notes Task Environment: The description of Performance, Environment, Actuators, and Sensors (PEAS). This provides a complete specification of the problem domain. So how can I implement this πŸ€” First I need to think on the domain, something different to the examples (e.

Continue reading β†’

[zero-RL] Summarising what LUFFY offers

Here’s a “standard” progression of training methodologies: PRE-Training - This is where the model gains broad knowledge, forming the foundation necessary for reasoning. CPT (Continued Pre-training) - Makes the model knowledgeable about specific domains. SFT (Supervised Fine-Tuning) - Makes the model skilled at specific tasks by leveraging knowledge it already has. RL (Reinforcement Learning) - Using methods like GRPO, DPO to align model behavior. Reasoning traces play different roles at each stage:

Continue reading β†’

[zero-RL] where is the exploration?

Source: Off Policy “zero RL” in simple terms Results demonstrate that LUFFY encourages the model to imitate high-quality reasoning traces while maintaining exploration of its own sampling space. Authors introduce policy shaping via regularized importance sampling, which amplifies learning signals for low-probability yet crucial actions under “off-policy” guidance. The aspect that is still not clear to me is how there is any exploration of the solution space.

Continue reading β†’