Daily, snackable writings to spur changes in thinking.
Building a blueprint for a better brain by tinkering with the code.
subscribe
rss Feeds

SPIN CHESS
A Chess app from Tinkered Thinking featuring a variant of chess that bridges all skill levels!

REPAUSE
A meditation app is forthcoming. Stay Tuned.
WHAT IS GPT-3?
July 21st, 2020
After playing around with the new tool from OpenAI, named GPT-3, it’s clear that we’ve probably reached an inflection point in the progress of artificial intelligence. To think of it in more commonplace terms, parents will often have the interesting experience of watching their children suddenly gain competency in a skill, seemingly overnight. There will be lots of struggle and failed starts and then as if all in one moment, boom, suddenly the kid can do something new. GPT-3 hails a moment when machine learning got turned on and suddenly it could do an amazing new trick, and it does this new trick with astonishing coherence and ability. The hope of this episode is to present a description of what GPT-3 is in layman’s terms and then explore some possible ways that this could be used, and how it might impact society.
But first, what is GPT? The acronym stands for Generative Pretrained Transformer. Ok sure, but what exactly is it? How do we interact with it? In simple terms it’s a program that accepts some text and then generates a continuation of this text. What exactly does this look like? For those with access to the GPT-3 playground through OpenAI, it just looks like an empty textbox with a submit button.
Now enters a principle that has emerged during the last week as developers have played around with it:
Garbage in, Garbage out.
GPT-3 mimics the style, tone and character of whatever text you feed it. (For those curious, you can hit the submit button with an empty textbox and GPT-3 will generate something. The first time I did this it started talking in French about libraries. The second time it started talking about heatwaves being in the headlines.) There is a bit of an art to the prompt you give GPT-3. Again, garbage in, garbage out.
Here are two examples:
Input: What’s up?
GPT-3 continuation: A confused voice came from inside. When I opened the door, the person that looked back at me was Hayama Hayato. Why was Hayama, who I only shared memories of me playing soccer with, in my room at this hour of the night? That question immediately flew out from my mouth.
Second example, same exact input:
Input: What’s up?
GPT-3 continuation: they respond 'Super' – at least they were back in the days of Fox Mulder and Dana Scully. I love the little sixties touches too, the poodle skirts, the VWs, the hippydom.
The input of “what’s up?” doesn’t provide any context, so GPT-3 generates text in a random direction each time. But notice how coherent it was…
But if you provide GPT-3 with context, say, the first couple paragraphs of an episode from Tinkered Thinking and it will generate a continuation of that episode which is shockingly on point and very believable. Episode 823 and 824 are non-fiction episodes that explore just this idea. About half of each episode was generated by GPT-3. This collaboration was also done for Episode 826 which is a fictional short story. GPT-3’s contribution to that episode is not explicitly indicated but there too it wrote approximately half of the text, and that episode in particular demonstrates an aspect of context in relationship to GPT-3. The program’s initial contributions to the fictional short story are a bit clunky. But as the story goes on, GPT-3’s contributions get better and better as the entire context of the story grows.
It should be obvious at this point that if you were to do something nefarious like input some dangerous ideas into GPT-3, then it would continue in the same vein. This is an important point that lends to obvious criticism and it’s a topic of which OpenAI is extremely conscious. The use of GPT-3 in the wild requires a security application among other safeguards to limit GPT-3’s use in obviously negative ways. Again, the principle holds: garbage in, garbage out. That holds not just for level of composition but of course, also subject matter.
Now how exactly does this magic work? To answer this, it’s necessary to wrap one’s mind around machine learning. How exactly does a machine learn? To be honest, nobody really knows. In a way, we’ve created a black box that can learn to figure things out. To really answer the question: how exactly does a machine learn is similar to asking: how does a person arise from all the chatter between neurons in a person’s brain? No one can answer that question. And likewise, no one can really answer the question about what’s really going on when a neural net is being trained in a machine learning context. That being said, we can describe some aspects of the process that help us grasp a sense of what’s going on.
Turns out, Machine Learning doesn’t seem too much different than human learning, though the differences are enormous and enormously important. Our brains are vast networks of neurons and they all talk to each other and somehow from all that chatter, we as individuals, emerge. To be sure, our understanding of the brain is still rudimentary, but it’s clear that neurons ‘listen’ to the chatter around them via a web of arms or dendrites that reach out to other neurons, and based on the signals they hear, they fire their own signal to be heard by another neuron or set of neurons. Multiply that process roughly 86 billion times, give it the right shape, and some configuration and boom, you have a human brain that will try to figure things out. (To dive a little deeper it’s useful to know that an infant actually has around 100 billion neurons, and as we learn how to exist in the world, we pare down this number significantly.)
GPT-3 was trained to generate text using a computational model that bares a lot of similarity to our jumble of neurons we call a brain. The analogy is so close in fact that this computational model is referred to as a ‘neural net’. Now here’s an interesting fact about GPT-3: the neural net upon which it was trained contains 175 billion parameters, or what you might think of as neurons. But there’s a distinction here we need to be very careful about. The node, or neuron, or what’s often referred to as a ‘parameter’ within this neural net is not a neuron in the same way that our neurons are. There is a similarity, but our neurons are vastly more complex than the nodes that exist within a neural net. Within the context of machine learning, a neuron is a fairly simple weighted value. These values dictate how one node talks to another. An actual human neuron is far more complex regarding how it listens and signals it’s neighbors. We must also take into account that there are many different types of neurons in the human brain specializing in different ways with different functions. Human neurons also have varying numbers of dendrites which are used to listen to other neurons. A single neuron can have over 100 dendrites and each dendrite has characteristics that determine how and what the core neuron hears and ultimately what sort of signal it sends. This presents a level of complexity that far exceeds what is going on with the nodes or neurons in a machine learning context. And this difference in complexity should make it intuitive why GPT-3 was trained using so many ‘neurons’. It’s construction and organization of these ‘neurons’ is far simpler and more straightforward than what we have going on between our ears, so where the neural net lacks in terms of shape and initial configuration, it attempts to make up for in size and the brute force achieved by speed and computational iteration.
So how does a neural net learn in order for something like GPT-3 to work? First you need a gargantuan amount of data. GPT-3 was trained using text from the internet - an amount of text that is just inconceivable for a single human being to think about reading. Think of every comment, every post, every description, every pdf book freely available, all the tweets and blogs, manuals, dissertations, threads, rantings - all of it. That was the block of text that was given to the neural net for it’s training.
So how does that training work? Imagine I gave you a paragraph to read, but I left off the last word, and asked you to guess what that word is. As a human, you’d probably do really well at this. You’d certainly do a perfect job guessing a sensible word that fits the context and which works, regardless of whether you guess the actual word. Here’s a tiny example: Please provide the last word that should be added to the end of this ___________ .
The above example is easy. The word is ‘sentence’. But. Since I’ve started a new paragraph, you could also plop the word ‘paragraph’ into the slot and it still works. This is the task that GPT-3 has become remarkably good at: what word would make sense to come next? The training of the neural net essentially asked this question of itself over and over. We can imagine the neural net reading 99 words and guessing what the 100th word would be. It would then check the answer and see how far ‘off’ the answer was, and based on the difference it would backtrack through the neural net to find all of the nodes that were most responsible for the choice of the wrong word, and then it would adjust the weights of those nodes, which in turn effects how those nodes will talk to one another in the next round. Run this game an astronomical number of times with an inconceivably large amount of text and after a while the neural net gets pretty good at the game it’s playing.
What’s important to note is that the computer is not actually ‘reading’ the text that’s given to it. The computer first creates embeddings for this block of text. What is an embedding? It’s arbitrary, but it’s a form that the computer can actually work with. Computers function using binary code. Each letter that you read on a website is actually derived from a binary code. So embeddings are bits of binary code that are associated with the words and word-snippets that the computer ‘reads’. These embeddings never change.
A good way to think about it is to realize that humans use their own embeddings. Our form of embeddings are individual words themselves. For example, you can hear the sound of the word ‘boat’, or you can read the graphical symbols that indicate the word ‘boat’ in your mind, but that set of graphical marks and the sound your hear are arbitrary - they don’t actually mean anything on their own. If you said the word ‘boat’ or showed the word ‘boat’ to someone who has no understanding of english, it would be meaningless to them. They have a different word or embedding for that concept. This is quite similar to how embeddings work in a neural network. Every time the computer ‘reads’ the word ‘boat’ it converts it into the same binary string, and uses that binary string to push into the neural net. This discussion of embeddings is important because it’s pretty magical to realize that GPT-3 doesn’t actually know any words. Aside from the already problematic use of the word ‘know’ in relation to a computer, the point here is to get at a description of what is at the core of GPT-3’s success. Through all of this weighted calibration using embedded language, GPT-3 has ‘learned’ the subtle rules that dictate how we humans pick our words in different contexts.
Dialect and jargon are helpful concepts here. For example, if you go to the hospital and talk to a doctor, you expect to hear a whole bunch of complex language that relates to biology and medicine, and if this hospital is in London, you expect to hear a British accent. Even if we don’t know all the words, we’d be able to tell something is amiss if we suddenly start hearing a whole bunch of nautical sailing jargon, the words of which we also don’t know.
If a friend starts relating a story about how they went sailing, we can predict that there’s a fairly high chance we’re going to hear the word ‘boat’. GPT-3 has essentially played that probability game a ridiculously unfathomable number of times in every context that humans have written about and that’s been plopped on the internet.
So it’s completely inaccurate to say that GPT-3 knows something about any given subject. If it knows anything, it’s perhaps best described as a really good guess about which word would come next.
So why all the hype? Simply put, it does an astonishingly good job at that guessing game. It’s so good that often you just can’t tell that it’s generated text, that wasn’t written by a human. If prompted correctly it responds identically a human.
How does this neat trick translate? What are the implications?
Well, let’s imagine a scenario here: Let’s say a friend shows up to work, and they’ve got a nice smile on their face. You grow curious and ask “What’s up?” And your friend begins to tell you about an app they just downloaded. It’s an Artificial Intelligence that works as a therapist, and your friend tells you they just had the most amazing insights during their therapy session that they had while walking to work.
Now, unlike a real therapist who is human that needs to write down information about sessions in files that they reference, a GPT powered therapist could have a perfect memory of absolutely everything you’ve ever said. You could then ask your friend in astonishment: how does that work? And your friend might say, well, first it’s super affordable, oh my gosh, just a couple of dollars a month and I have unlimited sessions. But to get started, you have to talk about yourself for 10 hours, but luckily I’ve been keeping a journal on a private blog which was a hundred times bigger in word-count, so I was able to submit that blog and boom, I had access.
If you think that description might worry a therapist, let’s swing in a different direction:
regulatory capture.
Let’s say you are a human rights activist and you’ve dedicated your whole life to monitoring the ways that corporations try to take advantage of governments through lobbyists and manipulating legislation. And right now there is a bill on the floor being considered. You know there’s probably something nefarious hidden in the details of this bill, but the bill is 2,000 pages long, and there’s just no way you can go over it with a fine-tooth comb in order to figure out just what’s in there. But, a colleague has just built a tool using an advanced GPT that can take any amount of text and summarize it. Not only that, but this specific GPT has been additionally trained to flag details that don’t seem connected with the main text, or details that hint at the creation or use of loopholes in existing law, because naturally, it’s also read all of existing law and has a perfect memory of all legal text. You feed the new bill to this special GPT and boom, within a minute you have an accurate 5 page summary and a list of 27 sentences and paragraphs spread across the 2,000 page document that it thinks you should specifically look at for trouble.
These are just two imaginary ways that the technology of GPT-3 could be implemented. And one of these ways, that of the pocket therapist has already been achieved in a somewhat rudimentary way during the week or so that GPT-3 has been available.
The ramifications for how this advance might rattle through society seem ubiquitous. Language is the fabric of society. Our use and misuse of language dictates the rise and fall of all our endeavors, and a computer just got really really good at imitating our language.
DETOUR TEMPTATION
July 20th, 2020
Life is full of detours. Maybe it’s all detours. No one plans their life completely, and those who attempt to do so often find themselves thwarted in counter-intuitive ways. There’s so much we just don’t see coming.
But what sort of choice do we make when an option comes floating by that seems fun?
But I’m heading in a different direction…
This would make me late…
I don’t know where that leads…
All valid points, but they each have a polar corollary that frames the situation as full of juicy possibility. Now here’s an important tidbit about getting old. When most of life is behind you, very few regrets will be about things you actually did. Most of your wonder regarding the past will be about the things you didn’t do: the chances you didn’t seize, the idea you didn’t pursue, the detour you didn’t take.
How does one negotiate these spicy pockets of possibility? Well, if it’s anything like a real detour, then it means we can always just backtrack to get back to where we started and then continue on our way. It might take a little time, but if we have time to spare, why not?
It’s certainly an entirely different deal if the detour is like a one-way valve and the door shuts and automatically locks the moment we venture off our normal path. That sort of opportunity would require quite a bit more consideration. A path that decreases the possibilities of life must have a huge benefit attached to it, otherwise, we should be wary by default. Such definitive paths aren’t actually detours by definition. If the door slams shut behind us, and we can somehow be sure the new path won’t loop back around to our initial path…then this certainly isn’t a detour, and the temptation must be quite promising.
Fate makes no promises, though. And this is why ‘the detour’ is such a useful way of looking at opportunities. The detour is the essential constituent of exploration, and without a map marked with an X, the detours we take are the only paths that might lead to treasure.
If life can be designed in a way that is open to more potential detours than we create optionality. The ability to take a detour is perhaps the quintessential opposite of being a cog in a system. Cogs can’t take detours without causing a lot of trouble. And likewise, an explorer who sticks to the beaten path will find nothing new.
If the detour presents itself, why not take it?
A LUCILIUS PARABLE: TINKER TOOLS
July 19th, 2020
~ Approximately half of this story was generated by the new technology from OpenAI called GPT-3. The contribution from this technology has not been demarcated on purpose in order to give the reader an opportunity to guess what is human and what is machine. ~
Lucilius was browsing the data of his own genome. He understood it only if he looked at tiny individual pieces or as a whole when he regarded himself as a person. The way the data of his genome translated into his existence was, as it is for everyone, quite a mystery. He'd employed the help of some machine learning to explore the possibility of editing his own genome in order to achieve some upgrades in his person, and really when it came down to it, he was seeking a more direct way to upgrade his very existence. After all, what exactly is self-improvement, what does it look like if we take it to its absolute extreme? Always more curious than fearful, Lucilius was seriously pursuing how it might actually be done. He was waiting now, while the Artificial Intelligence he'd hired trained itself on the data of Lucilius' genome in relation to the entire medical literature as it's been understood in relation to the genome. Another feat which was achieved through the eerie magic of machine learning. A loading bar on the screen finished and disappeared.
"Ok," said the AI, "so what exactly are we looking to do?"
Lucilius paused a moment. Good question, what exactly did he want to change? "Well," Lucilius said, "I've always been a bit more lazy than I'd like. Seems like it might be low-hanging fruit. What do you say, shall we try to figure out how to tweak that variable?"
"Lazy?" the AI said. "But isn't it already at an optimal level, if you can say there is one?"
"Well," Lucilius said, "can we please try anyway, I mean... I'm interested in what it could feel like to work harder, be more driven, live with more discipline. It's just that I've always preferred freedom, peace, and leisure. If we could just rewire that part a little bit, you know?”
"Sure," said the AI. "It's ok to be lazy, I mean, that's really not a good quality to strive for, but it's definitely better than being overworked and inefficient, or being lazy in other areas. Also, not to get too philosophical about this, but don't you find your laziness in relation to your freedom to be like a manifestation of your boredom, or like a low boredom tolerance, or like a boredom threshold? And don't you think that's just an interesting variation on the values of free-will and control and choice and self-determination and self-government. With you preferring the choice of leisure, even though you might be happier in other realms, and that this variation is uniquely yours, but maybe also a bit tragic, but some people might say that there are some more genuinely human experiences that you're missing out on by never feeling frustrated or challenged?"
Lucilius considered this. "Yes," he said. "Maybe. So, which are the variables?"
"Ok, so we could, say, tinker with your cannabinoid system and your serotonin levels and some other genes related to them. Maybe that will improve your disposition a bit."
"I'm not sure I like the sound of 'maybe', don't we have a pretty good handle on the genetic determinate of laziness or is laziness some kind of composite that will affect too many other aspects of who I am?"
"It's true," the AI said, "that you can't have a lazy personality without also having other personality characteristics associated with it."
"I was hoping that this one might have a tiny change that we can make, but of course all of these personality traits manifest from a pretty big web of genes," Lucilius mused, almost as though he were talking the issue out loud to himself. He looked back to the computer. "Based on the literature-genome training, are there any traits that we can tweak that don't have widespread effects?"
"We could tweak something as minor as say... I don't know... hair color. That's a pretty easy thing to change without affecting anything else too much. That might actually be the way to go, because at least it's a very small effect compared to something like laziness, which does require a degree of neural rewiring. It's much less risky."
Lucilius turned to glance in a mirror on the adjacent wall. "Yea, I don't really care too much about hair color. I'm looking to improve the way I exist in the world. Not just as something that you - or rather other people see - but the way I operate."
"Well, that's the thing," the AI said, "you really can't make a change like that without taking in mind all the aspects of your existence, the ones that most people are familiar with, but also things like how you feel when you walk in the rain or the snow, or if you'd even think of going for that walk, how you'd feel if you had a call to action, a desire, and if you'd follow that or spend more time with your kids. How that desire would manifest in your physiology and behavior, and what kind of thought process would be involved in all of those scenarios.. the ramifications are nearly limitless."
Lucilius grew a little nervous. But then he remembered his initial thinking on the subject. "Yes, exactly, and that's what I have you for to figure this out. Is it possible to toggle something like laziness as though it were on a sliding scale? Or what if it's a switch that I can turn on temporarily, something that I can somehow turn off and back on depending on what's going on in my life?"
"Sure," said the AI. "In theory, it's possible, but it's not easy. Once you start down this path, there's a lot more you can do, you might find that you like the changes and then it's just a slippery slope."
Lucilius considered the AI's position. 'Slippery slope', what ironic phrasing, he thought. The AI itself was the result of accelerating change and innovation in tech. It's a meaningless value judgement to make that curve look like its going up or down some slippery slope, he thought. What was clear is that he didn't want to get left behind, and the only way to stay apace was to accelerate.
"Screw it, let's do it," Lucilius said.
"Let's tinker," said the AI.
THE DEATH OF POSSIBILITY
July 18th, 2020
Many possibilities must die in order for reality to arise. In fact, all possibilities for a given moment vanish once that moment has emerged in time. That tiny sliver of possibilities that actually do occur are no longer possibilities, they are in fact, facts of reality.
This lends to an unsettling question: do possibilities even exist?
From a purely deterministic viewpoint, the answer is. . . no. Possibilities do not exist. The universe is a vast machine which is simply running its program. But since we are inside of that program, and that we are in fact made of the fabric of the universe’s program, that limited perspective keeps us from seeing just how all the clockwork meshes together in order to bring about the next moment in the exact and often surprising way it emerges.
The concept of a possibility is in fact an illusion. But aside from deluding ourselves, it generally seems to be a very helpful illusion. Just as free will is generally a productive notion to believe in despite its total incongruence with a deterministic view of reality. One convenient aspect of this universe is just how adept humans are at hosting contradictions side-by-side in their minds.
Though, it’s perhaps equally beneficial from a psychological perspective to shun all use of the concept of free will and possibility. From that point of view, there really does seem to be something like destiny, and having a particular destiny in mind can be a powerful force for moving forward if a person is of the belief that their destiny is certain to come about. There is of course the small caveat that we don’t, and can’t ever know what our ‘destiny’ is in this sense.
Our blindness to the larger movement of the universe necessitates these fantasy-concepts of possibility, free will, goals, and even destiny - if one believes they can know it ahead of time.
In fact: what exactly is the difference between someone who believe a particular outcome is “their destiny” and someone else who is simply working really hard to make a certain possibility come about? The difference seems to be incidental of language more than it is a practical difference. Perhaps a person with a goal has an orientation that’s a bit more flexible and fluid than the person with a self-proclaimed destiny, which harks of a potentially dangerous stubbornness, but then again, achievement almost always requires a certain hard-headed determination.
Regardless of which way we lean on to view the future, we are deluding ourselves to some degree. To believe that there’s any possibility that things will turn out in any way other than the way they actually will is just as deluded as believing in a specific destiny that is somehow known ahead of time.
The death of possibility doesn’t occur the moment the present finally casts all factors into an eternity of frozen past, in fact, the possibilities we imagine don’t ever die, because they never actually existed in the first place.
MACHINE MEDITATION
July 17th, 2020
The portions of this episode that are in italics have been generated by GPT-3 from OpenAI.
This episode is another collaborative effort undertaken with GPT-3, the newest technology from OpenAI. This episode seeks to explore the topic of meditation with this machine. As with the previous episode, number 823 entitled AI Collaboration, Tinkered Thinking and GPT-3 will take turns adding a paragraph to the episode at hand in order to build a cohesive piece that reads as though it was written by a single mind. GPT-3’s contributions will be conveyed in italics. Tinkered Thinking will initiate this episode on meditation by writing only the next two words of the succeeding paragraph and allowing GPT-3 to fill in the rest of the sentence and succeeding paragraph.
Meditation is a state of consciousness and it is not a complete departure from thinking to arrive at this state, but an improved immersion in the processes of thinking and an awareness of thinking as such, which is enhanced by being freed from the immediate desire to have things the way one would like them to be and the need to make them fit into the way one would like things to be.
In one sense, meditation is the practice of accepting the moment as it is, unedited, and unvarnished by our perspective when it is laden with memory and desire. To make progress to this end is to place distance between our raw experience and our default response to that experience.
It is human nature to be in a constant state of conversation with ourselves, and what this has become, what we call, “thinking,” is very often negative and distracting. When our default setting is triggered we are constantly engaged in justifying or defending our beliefs. When in the presence of stimuli, we are evaluating and critiquing. When our past comes back to us we reactively have a desire to repeat or undo our mistakes. At the root of all of this is fear, fear of the unknown, fear of change, fear of death, fear of being alone, fear of being with others. The self is a collection of learned fears; fearful reactions to thoughts, feelings and events. Meditation, or the practice of accepting what is, helps us to appreciate the fullness of what is by freeing us from clinging to our personal desires, which often determine what we believe is true and relevant, to instead see the moment for what it is without the delusions of the past or anticipation of the future.
Meditation is a gift of the mind that serves to cleanse us of our broken thinking and to put us in touch with the living moment, not as an interruption of our thoughts, but as a medium for resolving the riddles of thinking and interacting with our surroundings, with ourselves and with others in the purest and most effortless manner. Meditation can be the key to seeing beyond the barriers of isolation and fear and experiencing a kind of clarity that is inexpressible in words.
When we achieve the ability to separate our sense of identity from our experience, we inhabit a context that permits for more options, a wider perspective and more points of view. The simplest form of this can be summed up by the question: Do I want to act upon the anger that I am currently feeling? The unmeditated mind is likely to act reactively while intoxicated by the inspiration of such anger. But the practiced meditator has developed the ability to pause, and then consider options.
This all sounds very simple, but it's not. The mind resists this kind of mediation because the mind has been nurtured and validated by this world. In a way, the mind is like a drug addict that has a physical dependence upon emotions and events in order to feel validated and purposeful. Meditation, on the other hand, suggests that life without thought is preferable to life with too much thought. This will come as a challenge to many because the mind has become dependent upon its thoughts and it fears that without these thoughts, it will die.
In this sense, it's curious to think of any given thought as an organism. We use the word 'meme' to describe concepts that compete in ways similar to genes. Might we also say the same thing about entire perspectives? Perhaps the reason we can be so stubborn is simply because the ideas we currently hold are simply trying to resist irrelevance and extinction? The key benefit of meditation in this discussion is that we gain the ability to separate from the identity generated by these perspectives and by doing so we are free to consider options that can potentially lead to a better, more fulfilling life.
Part of the problem with the mind is that it will tend to indulge itself and strive towards habits that make us comfortable in the moment, while rejecting anything that doesn't serve that goal. With this mindset it's easy to see that much of life is the creation of a mirage. A mirage can be seductive. Many live within this illusion in a dream-like state. In many ways, we can view reality from this perspective as a dream from which we awaken as our perspective moves closer to align with the actual. For example, there is no longer the person I once thought of as myself. There are no longer desires or fears of the person I used to be. The person I was is like the face in the mirror that was once familiar, but the face looking back at me today is clearly something else, though still entirely recognizable. It is through the practice of meditation that we are able to step back and observe this change, but in a way that isn't forced or dramatic. It is a gradual shift that happens to us slowly over time. It is the ability to witness this shift as if it were happening to someone else. This level of detachment takes practice. It takes attention and openness.