I often think about ancient civilizations. I’m sure you do too. I cogitate and ruminate on all the people who lived, loved, and died, and I’m saddened by the fact that we no longer remember their names. In many cases, we aren’t even aware that their civilizations existed (I base this on the fact we seem to keep on discovering previously unsuspected cultures and empires).
I learn something new every day. For example, I just discovered that, while all Sumerians were Mesopotamians, not all Mesopotamians were Sumerians. It’s obvious when you think about it because some Mesopotamians were members of other empires that existed in the Mesopotamian area at the same time as the Sumerian kingdom. I know, “Duh,” right?
The reason I mention this here is that I love devices used to perform calculations in all their incarnations, starting with the abacus, the Sumerian / Mesopotamian version of which appeared sometime between 2700 and 2300 BC. I’m also intrigued by the people who came up with cunning concepts, like the Scottish mathematician, John Napier, who invented Napier’s Bones in the early 1600s. These were multiplication symbols inscribed on strips of wood or bone. Napier also invented logarithms. In 1621, the English mathematician and clergyman, William Oughtred, used Napier’s logarithms as the basis for the slide rule.
Over time, people started to create mechanical calculators, like Blaise Pascal with his Arithmetic Machine and Gottfried von Leibniz with his Step Reckoner. These early devices evolved into full-up mechanical computers from folks like Charles Babbage and his Analytical Steam Engine (which never actually worked) and Konrad Zuse with his Z1 (which did).

Charles Babbage (Source: Max Maxfield and Alvin Brown)
Eventually, of course, people developed the electronic calculators and computers we’ve grown to know and love today. I, for one, cannot imagine a world without these little beauties. I find it increasingly hard to remember those dark, dismal times when I was forced to walk across a room to control the lights or television, as opposed to simply lounging back in my comfy command chair and saying something like, “Alexa, please turn the television on.”
I find it interesting to note that, while most of my engineering friends are familiar with the history of calculators and computers—at least when painted in broad strokes—many are unfamiliar with the fact that there has historically been a great deal of fascination with logic in general. This fascination was initially expressed in the form of logic diagrams, and later in the construction of special-purpose machines for manipulating logical expressions and representations.
I was just ambling past one of the bookshelves in my office when I spotted the book Logic Machines and Diagrams by Martin Gardner, who—among many other things—was famous for his Mathematical Games columns, which appeared in Scientific American for 25 years. I wrote about this in Bebop Bytes Back: An Unconventional Guide to Computers, which I co-authored with my friend, Alvin Brown. I just re-read what we wrote, and I decided to share it here on the off chance you would find it as interesting as does your humble narrator (as you know, I pride myself on my humility—you’ll have to go a long way to find someone humbler than me). So, sit back and make yourself comfortable because here we go…
The Tree of Porphyry
Diagrams used to represent logical concepts have been around in one form or another for a very long time. For example, Aristotle (384-322 BC) was certainly familiar with the idea of using a stylized tree figure to represent the relationships between (and successive sub-divisions of) such things as different species. Diagrams of this type, which are known as the Tree of Porphyry, are often to be found in medieval pictures.
John Venn and his Venn Diagrams
Following the Tree of Porphyry, there seems to have been a dearth of activity on the logic diagram front until 1761, when the brilliant Swiss mathematician Leonhard Euler introduced a geometric system that could generate solutions for problems in class logic. However, Euler’s work in this area didn’t really catch on because it was somewhat awkward to use, and it was eventually supplanted in the 1890s by a more polished scheme proposed by the English logician John Venn (1834-1923). Venn was heavily influenced by the work of the English mathematician George Boole (1815-1864), and his Venn Diagrams very much complemented Boolean algebra.
Venn Diagrams were strongly based on the interrelationships between overlapping circles or ellipses. The first logic diagrams based on squares or rectangles were introduced in 1881 by Allan Marquand (1853-1924). A lecturer in logic and ethics at John Hopkins University, Marquand’s diagrams spurred interest by a number of other contenders, including one offering by an English logician and author, the Reverend Charles Lutwidge Dodgson (1832-1898).
Dodgson’s diagrammatic technique first appeared in his book The Game of Logic, which was published in 1886, but he is better known to us by his pen-name, Lewis Carroll, and as being the author of Alice’s Adventures in Wonderland. Apart from anything else, these rectangular diagrams are of interest to us because they were the forerunners of a more modern form known as Karnaugh Maps. These graphical techniques, which were invented by Maurice Karnaugh in the 1950s, quickly became one of the mainstays of the digital logic designer’s tool chest.
Ramon Lull and his Ars Magna
Possibly the first person in the history of formal logic to use a mechanical device to generate (so-called) logical proofs was the Spanish theologian Ramon Lull (1232-1316). In 1274, feeling remorse for having done something naughty, Lull climbed Mount Randa in Majorca in search of spiritual sustenance. After fasting and contemplating his navel for several days, Lull experienced what he believed to be a divine revelation, and he promptly rushed back down the mountain to pen his famous Ars Magna. This magnum opus described several eccentric logical techniques, but the one of which Lull was most proud (and which received the most attention) was based on concentric disks of card, wood, or metal mounted on a central axis.

Ramon Lull-type disks (Source: Max Maxfield and Alvin Brown)
Lull’s idea was that each disk should contain a number of different words or symbols that could be combined in different ways by rotating the disks. In the case of our somewhat jocular example above, we can achieve 4 x 4 x 4 = 64 different sentences along the lines of “I love mice,” “You hate cats,” and “They [The French] eat frogs.”
Of course, Lull had a more serious purpose in mind, which was to prove the truth of everything contained within the Bible (he didn’t believe in thinking small). For example, he used his disks to show that “God’s mercy is infinite,” “God’s mercy is mysterious,” “God’s mercy is just,” and so forth.
Lull’s devices were far more complex than our simple example might suggest, with several containing as many as sixteen different words or symbols on each disk. His masterpiece was the Figura Universalis, which consisted of fourteen concentric circles. The mind boggles at the range of combinations that could be generated by this device. Strange as it may seem to us, Lull’s followers (called Lullists) flourished during the late Middle Ages and the Renaissance, and Lullism spread far and wide across Europe.
Why is all of this of interest to us? Well, by some strange quirk of fate, Lull’s work fired the imagination of several characters with whom we are already familiar, such as Gottfried von Leibniz who invented the Step Reckoner. Although Leibniz had little regard for Lull’s work in general, he believed there was a chance it could be extended to apply to formal logic. In a rare flight of fancy, Leibniz conjectured that it might be possible to create a universal algebra that could represent just about everything under the sun, including (but not limited to) moral and metaphysical truths.
In 1666, at the age of 19, Leibniz wrote his Dissertio de Arte Combinatoria, from which comes a famous quote describing the way in which he believed the world could be in the future: “If controversies were to arise,” said Leibniz, “there would be no more need of disputation between two philosophers than between two accountants. For it would suffice to take their pencils in their hands and say to each other: ‘Let us calculate.’”
Jonathan Swift and Gulliver’s Travels
Of course, it’s safe to say that Ramon Lull also has his detractors (which is a rather kind way of saying that many people considered him to be a raving lunatic). In 1726, the Anglo-Irish satirist Jonathan Swift wrote Gulliver’s Travels, which was originally intended as an attack on the hypocrisy of the establishment (including the government, the courts, and the clergy—Swift didn’t like to restrict himself unduly), but which was written so pleasingly that it immediately became a children’s favorite. (On the off chance you were wondering, Swift penned his great work nine years before the billiard cue was invented. Prior to this, players used to strike the balls with a small mace.)

Jonathan Swift (Source: Max Maxfield and Alvin Brown)
Returning to Gulliver’s Travels… In Part III, Chapter 5 of the tale, a professor of Laputa shows Gulliver a machine that generates random sequences of words. This device was based on a 20-foot square frame supporting wires threaded through wooden cubes, where each face of every cube had a piece of paper bearing a word pasted onto it. Students randomly changed the words using forty handles mounted around the frame. The students then examined the cubes, and if three or four adjacent words formed part of a sentence that made any sense, they were immediately written down by scribes. The professor told Gulliver that by means of this technique: “The most ignorant person at a reasonable charge, and with little bodily labor, may write books in philosophy, poetry, law, mathematics, and theology, without the least assistance from genius or study.”
The point is that Swift is believed to have been mocking Lull’s art when he penned this part of his story. (Having said this, computer programs have been used to create random poetry and music …… which makes you wonder what Swift would have written about us.)
In fact, Swift continues to affect us in strange and wondrous ways to this day. When a computer uses multiple bytes to represent a number, there are two main techniques for storing those bytes in memory: either the most-significant byte is stored in the location with the lowest address (in which case we might say it’s stored “big-end-first), or the least-significant byte is stored in the lowest address (in which case we might say it’s stored “little-end-first). Not surprisingly, some computer architects (designers) favor one style while others take the opposite tack.
This didn’t really matter until people became interested in creating heterogeneous computing environments in which multiple diverse machines were connected, at which point many acrimonious arguments ensued. In 1980, a famous paper written by Danny Cohen titled On Holy Wars and a Plea for Peace used the terms “big-endian” and “little-endian” to refer to the two techniques for storing data. These terms, which are still in use today, were derived from that part of Gulliver’s tale whereby two countries go to war over which end of a hard-boiled egg should be eaten first—the little end or the big end!
Lewis Carroll’s Logical Conundrums
Leaping from one subject to another with the agility of a mountain goat, we might also note that Lewis Carroll (who was mentioned in an earlier topic) enjoyed posing logical conundrums in many of his books, such as Alice’s Adventures in Wonderland (1865), Through the Looking-Glass (1872), and The Hunting of the Snark (1876).

Lewis Carroll (Source: Max Maxfield and Alvin Brown)
For example, consider this scene from the Mad Hatter’s tea party in Chapter 7 of Alice’s Adventures in Wonderland:
“Take some more tea,” the March Hare said to Alice, very earnestly.
“I’ve had nothing yet,” Alice replied in an offended tone: “so I can’t take more.”
“You mean you can’t take less,” said the hatter: “it’s very easy to take more than nothing.”
Apropos of nothing at all, the phrase “As mad as a Hatter” comes from the fact that, in ye olden tymes, the manufacturers of men’s top hats used mercury compounds as part of the process. Over time, the mercury accumulated in their bodies causing severe impairment to their mental functions.
And we would have to chastise ourselves soundly if we neglected the scene involving Tweedledum and Tweedledee in Chapter 4 of Through the Looking-Glass:
“I know what you’re thinking about,” said Tweedledum; “but it isn’t so, nohow.”
“Contrariwise,” continued Tweedledee, “if it was so, it might be; and if it were so, it would be; but as it isn’t, it ain’t. That’s logic.”
You must admit, these gems of information aren’t to be found in most of the other things you’ve read today, but I fear we are in danger of wandering off the beaten path (“No,” you cry, “tell me it isn’t so!”).
Charles Stanhope and His Stanhope Demonstrator
The world’s first real logic machine in the sense that it could actually be used to solve formal logic problems—as opposed to Ramon Lull’s, which tended to create more problems than it solved—was invented in the early 1800s by the British scientist and statesman Charles Stanhope (1753-1816).
The third Earl of Stanhope and a man of many talents, he designed a device called the Stanhope Demonstrator, which was a small box with a window in the top along with two different colored slides that the user pushed into slots in the sides. Although this doesn’t sound like much, it was a start (there was more to this device than we can cover here), but Stanhope wouldn’t publish any details and instructed his friends not to say anything about what he was doing. In fact, it wasn’t until around sixty years after his death that the Earl’s notes and one of his devices fell into the hands of the Reverend Robert Harley, who subsequently published an article on the Stanhope Demonstrator in 1879.
William Jevons and His Logic Piano
Working on a somewhat different approach was the British logician and economist William Stanley Jevons (1835-1882), who produced the earliest model of his famous Jevons’ Logic Machine in 1869. This device is notable because it was the first machine that could solve a logical problem faster than that problem could be solved without using the machine!
Jevons was an aficionado of Boolean logic, and his solution was something of a cross between a logical abacus and a piano (in fact it was sometimes referred to as a “Logic Piano”). This device, which was about 3 feet tall, consisted of keys, levers, and pulleys, along with letters that could be either visible or hidden. When the operator pressed keys representing logical operations, the appropriate letters appeared to reveal the result.
Allan Marquand and His Logic Machine
The next real advance in logic machines was made by Allan Marquand, whom we previously met in connection with his work on logic diagrams. In 1881, by means of the ingenious use of rods, levers, and springs, Marquand extended Jevons’ work to produce the Marquand Logic Machine. Like Jevons’ device, Marquand’s machine could handle only four variables, but it was smaller and significantly more intuitive to use. (Following the invention of his logic machine, Marquand abandoned logical pursuits to become a professor of art and archeology at Princeton University.)
Benjamin Burack and his Electrical Logic Machine
Things continued to develop apace. In 1936, the American psychologist Benjamin Burack from Chicago constructed what was probably the world’s first electrical logic machine. Burack’s device used light bulbs to display the logical relationships between a collection of switches, but for some reason he didn’t publish anything about his work until 1949.
In fact, the connection between Boolean algebra and circuits based on switches had been recognized as early as 1886 by an educator called Charles Pierce (1839-1914), but nothing substantial happened in this area until 1937 when a graduate student at MIT, Claude E. Shannon, submitted his master’s degree thesis, A Symbolic Analysis of Relay and Switching Circuits. In this thesis—and in a follow-up paper in 1938—Shannon showed that Boolean algebra offered an ideal technique for representing the logical operations of digital systems. Shannon had realized that the Boolean concepts of false and true could be mapped onto the binary digits 0 and 1, and that both could be easily implemented by means of electronic circuits.
Following Shannon’s paper, a substantial amount of attention was focused on developing electronic logic machines. Unfortunately, interest in special-purpose logic machines waned in the 1940s with the advent of general-purpose computers, which proved to be much more powerful and for which programs could be written to handle myriad tasks… including formal logic.
As usual, we’ve only scraped the surface of a very interesting topic. If you know of any other logical diagrammatic techniques or machines, it would be great if you could share them with the rest of us in the comments below.

Leave a Reply
You must be logged in to post a comment.