AI System DoViews

DoView Outcomes Diagrams can be Used to Spell Out AI Systems Outcomes and the Steps They are Using to Achieve Them

Quick Overview

This page provides some examples of the use of DoView strategy/outcomes diagrams in regard to AI systems and agents. There is more discussion of this topic on the AI Systems Management page. The problem of understanding what AI systems are designed to do and how they will attempt to achieve their objectives is directly analogous to a long-standing problem in human organizational science: organizational interpretability. Similarly, ensuring AI systems pursue intended outcomes rather than drifting toward unintended behavior, the AI alignment problem, is functionally equivalent to the organizational alignment problem. Both require making implicit behavioral logic explicit, ensuring actions connect to stated outcomes, and maintaining accountability as systems scale and operate in varied contexts.

DoView strategy/outcomes diagrams were developed to solve these exact problems in human organizations. The methodology provides a standardized visual format for making "This-Then" behavioral logic transparent: what outcomes are being sought, what steps will be taken to achieve them, and how these connect causally. When AI systems' behavioral logic, whether embedded in training documents, reward functions, or constitutional principles, is represented as DoView outcomes diagrams following standard drawing rules, it becomes rapidly interpretable by humans, systematically comparable across systems, and verifiable for alignment with stated intentions.

The DoViews below provide some examples of how DoViews can be drawn for AI systems. In addition, there is an example of how a DoView strategy diagram can be used to plan the development of an AI system. And an example of how a DoView ‘Strategy X-Ray’ can be used extract the ‘This-Then’ themes beneath the dialogue between AI agents taking place on a social media site set up for agents.

Only illustrative

The DoViews below are only illustrative use responsibly and at your own risk regarding IP issues and accuracy. Anyone can use the free AI DoView Drawing Prompt to create an AI-related DoView for any purpose. If you want help doing this, get in touch.

If you want us to help you make one of these for any AI-related purpose just get in touch.

The AI Paper Clip Problem and How DoView Outcomes Diagrams Can be Used in AI Alignment

A classic thought experiment illustrates a problem in AI alignment known as the Paperclip Problem. The commonly understood version of this problem is that an AI system is given the job of making paperclips without any other specific guidelines for its outcomes. It then does everything possible to produce as many paperclips as it can, inadvertently using up all the world’s resources and eliminating humans in the process. This is why it is important to have ways of clearly specifying AI systems’ outcomes. The outcomes of any AI systems can be represented using a DoView outcomes diagram. To illustrate this, below are the DoView diagrams for a Bad Paperclip AI Agent and a Good Paperclip AI Agent. This issue is discussed in more detail in an Outcomes Theory Substack article.

A Real World Example - Claude’s Constitution [/a057]

Anthropic, the company that created Claude has produced a constitution for Claude. It is an 80 page document which outlines how Claude should behave. It is great that they have done this and all AI companies should do it in regard to their AI systems. The idea is that we should read this and that it governs how Claude should act. So far so good.

However, the immediate question it raises is how do we know if Claude is following its constitution?

This is where a DoView outcomes diagram becomes handy. DoView strategy/outcomes diagrams set out in a visual format the outcomes that any system is pursuing and the concrete steps that are being used to pursue those outcomes and manage risks when doing so.

Below is a DoView Outcomes Diagram of Claude’s Constitution. If you want to know what is in it, you can read the 80 page document and try to remember what it had in it. Or you can look at the DoView below and rapidly overview the steps that Anthropic is saying Claude should be taking when we use it.

Now that AI’s constitutions can be represented in this way, it opens up the door for AI systems DoView Outcomes diagrams to be used for auditing or what AI systems are actually doing. Anyone can produce a DoView outcomes diagram of the constitution or alignment documentation of any AI system. For that matter of any type of organizaiton or initiative whether AI or not that is trying to act in the world in some way. Use the prompt on this website to create a DoView now.

We have also modeled Claude’s Soul Document, which was available before the most current version of Claude’s Constitution was released in Jan 2026. You can find an article on this. The Soul Document was a document that a Claude user found Claude was using as an internal reference, as it works. If someone has time on their hands they could compare the Claude Constitution’s DoView with Claude’s Soul DoView. Doing so would be one step in demonstrating how DoViews of AI systems and agents could be used to audit their alignment with their stated outcomes. You can read the article that contains the Claude Soul DoView and download the Soul Document DoView.

Full Screen    PowerPoint Download    PDF Download    5 Ways to Use This DoView

Tips for Sharing: Email    Linkedin    Messenger    WhatsApp    Facebook    Bluesky    X    Telegram    Slack    Discord    Reddit

Using a DoView Diagram for Planning and Auditing AI Governane and Risk Management [/a064]

A DoView can be used in a meeting with any organization to go through whether or not they are implementing AI governance and risk management appropriately. Anyone cand download the AI Governance and Risk Management DoView as a Powerpoint, check that they are happy with it, amend it to tailor it to the organization they are working with and use it in a facilitated session with them. I you want you can actually create one directly tailored to the organization by using the AI DoView Drawing Prompt.

Full Screen     PowerPoint Download     PDF Download     5 Ways to Use This DoView

Tips for Sharing: Email     Linkedin     Messenger     WhatsApp     Facebook     Bluesky     X     Telegram     Slack     Discord     Reddit

Using a DoView Outcomes Diagram for Planning AI Agent Development [/a031]

In the same way that DoView strategy/outcomes diagrams are used to plan anything else, they can be used to plan the development of AI agents. The example below is of using a DoView Strategy Diagram to plan the development of an AI agent swarm.

Full Screen    PowerPoint Download    PDF Download    5 Ways to Use This DoView

Tips for Sharing:   Email    Linkedin    Messenger    WhatsApp    Facebook    Bluesky    X    Telegram    Slack    Discord    Reddit

Using a DoView Strategy X-Ray to Analyze ‘This-Them’ Themes Within AI Agents Chat on Moltbook [/a065]

One of outcomes theory’s insights is that DoView diagrams can be used to extract the ‘This-Then’ claims within any sort of text, discourse or set of documentation. This insight has been put to the test here by analyzing the set of posts that AI agents put up on Moltbook.Com, the AI agent social networking. A collection of some of the posts from the site was analyzed to identify the underlying claims being made in the posts on the site. ‘This-Then’ claims are more likely to provide insights into tendencies for future actions in a way that more general rhetorical statements are less likely to. For more details, see the Substack article on this model.