Ashby Diagram Software
All an Ashby Plot is, is a plot with one independent variable (X-axis) and two or three dependent variables. One is plotted on the Y-axis, and the remaining identified by color, marker size, etc. There's nothing special or unusual about this sort of graph. Examples are given in references Il and 21. PC-based software systems 31 which implement the method are available. This Appendix compiles indices for a range of common applications. Uses of material indices Material selection Components have functions: to carry loads safely, to transmit heat, to store energy, to insulate, and so forth. Material classes and subclasses. Only two subclasses are shown for each class; each, in reality, has many. Material properties. Each property type has many members, of which only two are shown. Which can be made by combining them ( gure 1). Each class has a. For material selection, MF Ashby of Cambridge University developed this very powerful method of material selection using charts and they're called Ashby plots and there's also a software you can utilize to plot these Ashby plots. And what Ashby plots do is they show you a relation between two material properties across a number of materials.
[V]ariety can destroy variety
W. Ross AshbyAshby Diagram Software Downloads
There are more things in heaven and earth, Horatio,
Than are dreamt of in your philosophy.
In his book An Introduction to Cybernetics, published in 1956, the English psychiatrist W. Ross Ashby proposed the Law of Requisite Variety. His original formulation isn’t easy to extract into a blog post, but the Principia Cybernetica website has a pretty good definition:
Ashby Plot Software
The larger the variety of actions available to a control system, the larger the variety of perturbations it is able to compensate.
Like many concepts in systems thinking, the Law of Requisite Variety is quite abstract, which makes it hard to get a handle on. Here’s a concrete example I find useful for thinking about it.
Ashby Chart Software
Imagine you’re trying to balance a broomstick on your hand:
This is an inherently unstable system, and so you have to keep moving your hand around to keep the broomstick balanced, but you can do it. You’re acting as a control system to keep the broomstick up.
If you constrain the broomstick to have only one degree of freedom, you have what’s called the inverted pendulum problem, which is a classic control systems problem. Here’s a diagram:
The goal is to move the cart in order to keep the pendulum balanced. If you have sensor information that measures the tilt angle, θ, you can use that data to build a control system to push on the cart in order to keep the pendulum from falling over. Information about the tilt angle is part of the model that the control system has about the physical system it’s trying to control.
Now, imagine that the pendulum isn’t constrained to only one degree of freedom, but it now has two degrees of freedom: this is the situation when you’re balancing a broom on your hand. There are now two tilt angles to worry about: it can fall towards/away from your body, or it can fall left/right.
You can’t use the original inverted pendulum control system to solve this problem, because that only models one of the tilt angles. Imagine you can only move your hand forward and back, but not left or right. Because of this, the control system won’t be able to correct for the other angle: the pendulum will fall over.
The problem is that the new system can vary in ways that the control system wasn’t designed to handle: it can get into states that aren’t modeled by the original system.
This is what the Law of Requisite Variety is about: if you want to build a control system, the control system needs to be able to model every possible state that the system being controlled can get into: the state space of the control system has to be at least as large as the state space of the physical system. If it isn’t, then the physical system can get into states that the control system won’t be able to deal with.
Bringing this into the software world: when we build infrastructure software, we’re invariably building control systems. These control systems can only handle situations that it is designed for. We invariably run into trouble when the systems we build get into states that the designer never imagined happening. A fun example of this case is some pathological traffic pattern.
The fundamental problem with building software control systems is that we humans aren’t capable of imagining all possible states that the systems being controlled can get into. In particular, we can’t imagine the changes that people are going to make in the future that will create new states that we simply could not ever imagine needing to handle. And so, our control systems will invariably be inadequate, because they won’t be able to handle these situations. The variety of the world exceeds the variety our control systems are designed to handle.
Fortunately, we humans are capable of conceiving of a much wider variety of system states than the systems we build. That’s why, when our software-based control systems fail and the humans get paged in, the humans are eventually able to make sense of what state the system has gotten itself into and put things right.
Even we humans are not exempt from Ashby’s Law. But we can revise of our (mental) models of the system in ways that our software-based control systems cannot, and that’s why we can deal effectively with incidents. Because of how we can update our models, we can adapt where software cannot.