Michael Nielsen on visualization and making a new science

Michael Nielsen on visualization and making a new science

Michael Nielsen on visualizations, biological systems, and making a new science
72 min read
Share this post
Michael Nielsen is a quantum physicist, science writer, computer programming researcher, and modern polymath working on tools to expand human capacity to think and create. He’s previously authored pioneering quantum computing books , propelled forward the open science movement, and published research on artificial intelligence. He now researches meta-science at the Astera Institute , while writing about his many interests online .
Michael speaking at Science Online London in 2011. Image from Wikipedia
DEVON: Hello, and welcome to Tools and Craft. I'm your host, Devon Zuegel and today I'm talking with Michael Nielsen. Michael is a scientist who explores ideas and tools that help people think and create, both individually and collectively. Michael helped pioneer quantum computing and the modern open science movement, and he also has the honor of being the person who originally introduced me to the ideas of tools for thought almost a decade ago. So really, we can thank him for the existence of this entire podcast. 
Michael defies field boundaries more than anyone I can think of. He's leapt around from physics to metascience, to educational tools, to programming and beyond - and he's made tremendous contributions to each field he bumps into along the way. So Michael, thank you so much for taking the time for this conversation. 
MICHAEL: Thanks for having me on Devon. 
DEVON: One thing that you said before is that you believe far better social processes are possible in science, and that these could activate great latent potential for discovery. How have these social processes changed since you first started doing science?
MICHAEL: There's a lot more emphasis on accountability. I think that sounds like a good thing. In many ways, it is a good thing. In particular, a lot of science funders around the world do more exercises where they try to assess just how well things are going. Probably the best known example of this is in the UK, where they run what used to be called the Research Assessment Exercise , changed the name to Research Excellence Framework. They assess all the universities and university departments to try and determine the quality of their research and that determines the way they get funded. When I was in Australia, as an academic, there was a similar exercise taking place. 
It's maybe a little bit mixed. People end up gaming the system or trying to game the system. They worry so much about the impact this has on funding, that they can actually end up essentially, Goodhart’s Law , where what gets measured gradually becomes a target. People start to target these kinds of exercises, and it actually can sometimes distort what it is they do. That's not always a good thing. That's a really big change.
DEVON: And how do they measure quality?
MICHAEL: It varies from place to place. There are two common approaches. One is bibliometrics, so they will count the number of papers and citations. Those are not terrible things to be looking at. But when they become targets, they become really quite strange things to be looking at. There's an interesting paper I was just looking at earlier today, which suggests that the use of those kinds of measures in Italy since 2011, has actually perhaps led to the creation of citation rings and lots of self citation and things like that. It's not certain that that's what's going on, but at least suggestive that this is not very good for the process of science. So that's one sort of approach to this sort of very data driven approach, where you're not really looking at what's actually being done at all. You're just using this very gross sort of metadata. 
“You need to understand [science] in a way that is not immediately quantitative or scalable, before you can develop good quantitative measures.”
The other approach tends to be more panels, where individuals, usually pretty well known senior scientists, will try and assess research outputs from different people. They might ask that every person in the department has to submit their three best, five best, or however many best papers from the last three years, five years, or however many years. And they'll try and assess how important that work was, and develop some kind of aggregate sense of a department or a university. That approach is also pretty common with lots of variations between different funders. It's certainly a little bit harder to game, but there are still ways that it can be gamed.
DEVON: Let's assume that a scientist does not want to commit fraud, but they do want to craft their research agenda to get as high of a score on this as possible, maybe maximize the number of citations they can get. What are the characteristics of research agendas or lines of inquiry that will tend to get more citations and tend to map to this score more highly?
MICHAEL: There's at least two things. Joining very large collaborations is often a way of increasing this for a couple of reasons. One is you may end up as an author on a very large number of different papers, where you're just sort of part of the collaboration that is producing it. Maybe you're working on a detector in the particle physics experiment. And that detector is used in a lot of different experiments. As a result, you end up as an author on a lot of different papers. What's more, there will be quite legitimately a lot of self citation amongst those papers. 
The other thing that it will certainly tend to do is drive people to work on things which are extremely fashionable, which have a lot of people already working on them and there's some potential for you to have a high impact in this kind of way. It's significantly more difficult to carve out a career, doing something where you're the only person or there's only one or two other people in the world who are currently interested  in that topic, you're just not going to look so good on a bunch of these measures. When people are very aware of that kind of thing, it tends to suppress such work.
DEVON: Hmm. That's really interesting. And collaboration that produces many papers, maybe each of those papers is very good and of a very high quality, but they're closer in the ideas space to each other than papers on completely different topics. So it seems a little odd for that to be counted as the same number of citations, even if it's sort of the same cluster of concepts.
“One of the reasons why we regard an idea is tremendously important, is because it was a priori, so unlikely… so there's some funny, interesting tension between the importance of an idea and how a priori plausible it is”
MICHAEL: Yeah, absolutely. It's just sort of a funny thing. It's using citations for a purpose they're not designed for at all. The person who started the Science Citation Index , and sort of all interest in citation counting, Eugene Garfield , he wasn't thinking of it as a method for impact assessment at all. Well, that's a little too strong. He was very slightly interested in it, but that wasn't his main interest. He was interested in tracking the lineage of ideas. In particular, he was really interested in enabling people to track down errors in the scientific record, so quite a different sort of purpose. Citations themselves, they're not meant to be intrinsically a measure of importance. It's just saying, we were influenced in this way by that paper. 
Arrangement of the Science Citation Index. Image from the Illinois Digital Environment for Access to Learning and Scholarship
DEVON: It also might have been a more effective measure initially, before people caught on that it was going to be used that way. The idea of taking a snapshot of a system is very different from using a metric for a system that is going to be an input to the system itself later on, because now you change the system by observing it. 
If you had to design a quantitative metric for the quality of science, how would you do it? Or would you reject the concept entirely?
MICHAEL: I think what I probably do, if you put a gun to my head and said I had to do it, is recruit a lot of different teams, made up of very different types of people. There'd be a lot of historians of science, sociologists of science, philosophers of science - people who are interested in the qualitative evolution of ideas and the details to try and understand things in that very qualitative way. You need to understand it in a way that is not immediately quantitative or scalable, before you can develop good quantitative measures. Scientists are not making their judgments individually just by citation counting, or something like that. You try to understand the evolution of ideas in your own area and figure out what is actually intrinsically important, what should you be working on? What kinds of ideas are actually important for the future of the field? And people, you know, there's a fair amount of heterogeneity, different people have different opinions and it's not necessarily clear that you can actually entirely eliminate that heterogeneity. I believe it is probably true that those differences of opinion are actually very generative and necessary.  If you take that point of view seriously, then no metric is ever going to capture all of that. So the attempt to say what is most important is actually answering the wrong question, what you should be trying to do instead is generate the portfolio in different directions, some of which may be, in some sense, inconsistent. 
Visualization of massive bodies warping space-time. Image from LIGO Lab
Let me give you an example. Quantum Gravity is a really sort of famous example, in physics, there's this long standing problem resolving an inconsistency between the theories of quantum mechanics and the theory of general relativity, our best theory of gravity. There are a number of different schools of thought about how best to do this. If you insist on just picking out something as the best, you will then concentrate all your attention on one particular approach. If instead you say,  just allow for at least some period of time, different approaches to flourish and not try and determine which of the two, three, five or 10 different approaches is best. I think, usually, that's a significantly better way. 
It’s something I like about Silicon Valley actually. If you have two competitors in a space, let's say you talk about, say, Netflix and Blockbuster, sort of two classic competitors. The fact is for a long time, although they're competing for the same market, in some respects, they have their own internal infrastructure, capital, and momentum. And you don't have a situation where employees at Blockbuster are conducting the performance evaluations for people at Netflix, and determining who gets promoted and these kinds of things. Yet, the situation in science often resembles that a little bit, where somebody in one school of thought is actually doing the peer review for somebody in a completely different school of thought. I think it's actually good to be able to have those silos exist independently for some period, extended period of time. Ultimately, you do want to be able to sort of make a resolution of which is the better approach. But in the meantime, giving one veto power over the other just seems kind of hopeless to me. Certainly, you know, I'm quite certain that the CEO of Blockbuster would have given very negative performance review, the people at Netflix for a long time.
DEVON: As soon as you made that analogy, I was like, Oh, my goodness, that is exactly what's going on. The imagery that pops to mind of the silos is of the Galapagos Islands . On these islands, different species were able to evolve independent of each other and you end up with much greater diversity.
“Computers would begin to be based on evolutionary processes and other ideas from biology. We would accept this because we get so much power from them, but we give up some of our control.”
How do you keep those silos, while also allowing some of the cross pollination of ideas that is also so important?
MICHAEL: I mean, the cross pollination of ideas is independent to some extent of the mutual evaluation of the ideas. And to be clear, some mutual evaluation is great. But you don't want that to be sort of the only factor that determines people's continued existence.
One of the ideas that people have talked about and explored that is becoming a popular area of investigation is the idea of randomizing grant funding. So basically, somebody applies with a project idea. Instead of having a peer review, where you essentially try and rank all the proposals and only find the “best”. Instead, there's kind of a basic sanity check round of peer review, so that people with crank proposals are being rejected. Then after that, there's simply a random selection made and whoever's lottery numbers come up, get funded. The benefit of doing this kind of thing is that you're not suppressing ideas, you're not putting the Blockbuster employees in charge of the performance evaluation of the Netflix employees. You're just relying on individual scientists to make the best judgment about what they think is the best possible idea, subject to that basic checkup, sanity check. I think the first place that did this in the world was the New Zealand Health Research Council , I think it's called, and a few other grant agencies have done small pilot trials since then, but it's certainly not a widely used approach.
DEVON: I suppose if you as a funding agency knew the relative importance of different research ahead of time, you would have already done science that is sort of solving the problem.
MICHAEL: Yeah, I mean, it's such a funny thing, where one of the reasons why we regard an idea is tremendously important, is because it was a priori, so unlikely. I know one of my favorite examples. One of my heroes, the biologist Lynn Margulis , she's the person who proposed the incredible, almost ludicrous idea of symbiosis. She was opposed for a long time but to my understanding, this is now accepted by biologists. So, there's some funny, interesting tension between the importance of an idea and how a priori plausible it is. She's one of these people who just has tremendous force and sort of willingness to persist even when everybody else is telling her over and over that she said that she's wrong.
Lynn Margulis accepting her National Medal of Science Award. Image from The New York Times
Something also really interesting about biology is that we regard both single celled systems and multicellular organisms, both as forms of life. And that's incredibly interesting that we call we unite both of those. Why don't we regard an animal with, you know, 1000 cells? Why don't we regard a thought as 1000 separate living beings? It's interesting that we don't do that, like what's sort of the boundary between what we consider an organism and and a cell is, at least to me, not entirely obvious.
“I always feel like JavaScript seems a little bit more, almost more biological. ”
DEVON: I wonder now if we're missing things, because when we put things in categories, it's often so that we can think about them and make generalizations about the things in the category. So now it makes me wonder, what are we missing by blobbing them all together? And what might we learn if we were to split them apart?
MICHAEL: Yeah, one of my favorite little facts is that apparently in linguistics, the subject of what is a word is somewhat controversial . There's sort of different schools of thought on what should be considered a word or not. As a physicist, you get really used to having often very clean categories and some notion of sort of correct categories. At first, it's really frustrating that there doesn't seem to be these quick clean, conceptual cleavages. It’s really interesting.
DEVON: I often feel that way, moving into different realms from programming. Programming is a little messier, maybe than physics, but it still has a lot of things that are literally binary in the sense of they're a boolean value. And when things are digital, there's a sense that this code produces the results I'm looking for, or it does not. Whereas moving into other realms more like design or politics. Often I'll ask someone a question that I think must have a simple answer. And it's always like, well, it depends. And sometimes it might even depend on what the person involved had for lunch that day. So there's no clean answer.
MICHAEL: I just feel like JavaScript is kind of a little bit of an exception in something. It's so surprising, I often can't remember how stuff is done. Because, there's all these weird exceptions and things are done in strange ways. If you want to delete a DOM element , you need to find the parent and then delete the child. This is crazy as a design, whereas something like Python seems much more like, they tried to make a language that was pretty consistent. I always feel like JavaScript seems a little bit more, almost more biological. There's no real uniformity or consistency, it evolved to be whatever it is.
DEVON: That is a great way to put it. And I see that as its greatest strength and its greatest weakness.
MICHAEL: I mean, it's a beautiful language in some really interesting ways. I don't know if you've ever read Kevin Kelly 's book, Out of Control .
DEVON: No, I haven't
MICHAEL: He wrote this great book in the early 90s and at the time, I thought, “Oh, this is good”, but it wasn't like making a really deep impression on me. The thesis is that human beings have gotten really good at designing and building systems, which they understand really well and where they try and get very close, tight control of all the different elements. Sort of like Le Corbusier , the modernist city planner, lots of straight lines and this kind of stuff. 
Le Courbusier's plan for Ville Radieuse, a contemporary city for 3 million people. Image from the Artists Rights Society
Kelly's thesis was that we were going to move our machines to much more biological kinds of systems - we would essentially trade off power against the ability to control. Computers would begin to be based on evolutionary processes and other ideas from biology, and we would accept this because we would get so much power from them. But we would give up some of our expectations. You're sort of seeing this, these systems that formerly were under control, or getting out of control.
DEVON: It reminds me of conversations I've had with a lot of friends who previously ran fairly small companies and then scaled them up, or they are deciding whether or not they want to go for being a larger organization. And when it's just you and two or three other employees, it's very easy to see what's going on and have visibility and control the situation. But there's only so much that four people can get done. Whereas if you're running a 10,000 person organization, you have very little control over what's going on. But you get a lot more power. 
It also reminds me of a lot of artificial intelligence discourse, where so much around being able to understand what an AI is doing is interpretability. It is such a key first step to being able to control it - something that I think a lot of people working on AI are concerned about, because the capabilities are far beyond what we understand at the moment.
MICHAEL: Just to go back to your mutual friend who is a CEO, the comment was that it is often harder for them to know what's going on inside the company than it is for almost anybody else. I think when you're in a very senior position, although it gives you like this bird's eye view, there are certain things that the person who is the underling might be uncomfortable sharing with their boss or their boss's boss, or their boss's boss's boss's boss's boss. Just for sort of natural reasons, you don't want to complain that you don't like the lunch menu, or that you worry a lot about your relationship to that person. So it creates a really interesting opacity to the person at the top of the organization. In some sense, the CEO has less visibility into some parts of the organization than just a random new hire who can go and find out what people really think about certain issues.
DEVON: It's that observability effect again. It's similar to the citations like, as soon as someone's looking at something, their behavior will change, or at least mask itself so that it doesn't get read in a way that is unfavorable, or out of control for that particular person.
MICHAEL: Yeah, that seems exactly right.
“I've noticed over and over again, that many sorts of guilty pleasures like that, ultimately end up becoming, five years or 10 years later, the creative projects that I work on.”
DEVON: You'd said something about biology being one of the fields that you aren't so familiar with. I would go so far as to say it's one of the few fields that you're not familiar with considering how many different places that you've worked in science? As you've gone from different fields to different fields? What are signs that you read that help you see that you maybe can contribute something to a field you hadn't previously studied?
MICHAEL: There are two things that go on. One, sometimes I just get curious about the field. And, yeah, there's no particular sort of opportunity to contribute. I might just pay attention to it for some period of time, talk with people about it, read about it. So I'm not working in the field at that point. What I want is just one insight that seems worth building on and like it might actually potentially be important. It's a very local kind of thing, it's not anything more than just having one little creative thing that feels like it should be done. Often actually, it feels like a little tiny creative thing, a project that should take a week and that takes three months, six months, or even two years. It's sort of unfolding, I do a lot of things like that. I've never seriously worked on AI, but every once in a while, I'll just take a week or something because there'll be some little question I have that I can't find the answer to in any paper. So I'll just play around and try to get a little bit of insight into it, and maybe write up some notes. Yeah, you sort of just repeat that often enough. Sometimes things start to seem very interesting in some particular direction and maybe actually start to develop some expertise. I don't know, I don't have a sort of a conscious theory of how this happens.
DEVON: When was the last time that you had a little project like that, that ended up expanding to take more of your headspace than you anticipated? 
MICHAEL: Yeah. A project with Kanjun Q , that's an essay that we thought might take a couple of months to write about? Well, actually, we didn't even know what we did what I now think it was about, we had no idea that it was going to be about that, at first, it's just going to be some thoughts on science funding and how it might be done better. And it's turned into something completely different. It's taken 18 months, and many 1000s of hours to do. It started with some simple ideas and turned out to be about the extent to which the scientific system, the discovery system that we have, is able to learn and change and update itself. That system does a really great job of updating its ideas - lots of people can contribute to that. But if you're talking about the social processes and institutions of power by which it operates, those only change relatively slowly. That's the question that ultimately, we got interested in and started to think about.
Screenshot from Michael's treatise on metascience with Kanjun Qiu. Image from Michael Nielsen
DEVON: I remember you talking about some of these questions many, many, many years ago. So it sounds like something you've been thinking about and chewing on in the back of your head for many years - is that typical of the types of problems that ended up grabbing your attention for long periods of times where you've had something working in the back of your head for a while, or sometimes it comes out of the blue, and it's something you've never really thought about before.
MICHAEL: Both happen. In this particular case? About 15 years ago, I took a few years to work as an open science advocate, and to develop a bunch of ideas around open science. And those are all sort of very concrete instantiations of this problem, how do you change and update the social processes or science. I didn't conceive of them in that kind of abstract form then, but later came back and realized, “oh I've seen a lot of problems with these very specific forms”. They're actually instances of a much more general pattern. In this particular case, it was kind of revisiting something from a long time ago. 
One pattern I really like. Actually, maybe this is just an excuse for self indulgence. I think most people have, you know, various guilty pleasures, things that they enjoy doing, or reading or whatnot, that they feel that they really shouldn't. They should be doing hard work on the actual serious projects. I've noticed over and over again, that many sorts of guilty pleasures like that, ultimately end up becoming, five years or 10 years later, the creative projects that I work on. I was a quantum physicist, I wasn't supposed to be thinking about open science. But I just got really interested in the way Linux kernel development was being done and the way Wikipedia was being constructed and then a whole bunch of other strange questions. This sounded very highfalutin, there were also many other sorts of less highfalutin things, but that's what turned ultimately into the interest in open science. This pattern just sort of happens over and over again.
Chart of Linux Kernel version history. Image from Wikipedia
“We get interested in things, for reasons that often we can't articulate. In fact, it's the fact that we can't articulate them that we get interested in... There's some interesting new structure in the world that is actually grabbing our attention.”
DEVON: Yeah, that's been a pattern in my own work as well. Although it's it's quite different types of work, but the kinds of things that I find myself working on thinking about today are the types of things that I was blogging, thinking about, and researching, in my off hours three to six years ago, which I always thought, well, I could never work on that. 
MICHAEL: Paul Buchheit had this really nice heuristic he followed when it started investing. He would sometimes turn some founders down. They would pitch  and he would think “that's obviously not going to work”. And then he would keep thinking about their proposal over and over in the days after. Eventually, he'd realize that whenever that was the case, he should go back and invest. His unconscious mind, I guess, was trying to tell him something about the proposal. That's a very general pattern. We get interested in things, for reasons that often we can't articulate. In fact, it's the fact that we can't articulate them that we get interested in. It's the sense of, you know, there's some interesting new structure in the world that we haven't seen before that is actually grabbing our attention. But of course, once you start to master that, it becomes a tool that you can use in other pursuits. Anyway, that's my way of justifying many wasted hours on YouTube and other places.
DEVON: I think it's very compelling. I think I'm gonna be thinking about this over the next few days, in fact.
MICHAEL: Yeah, urbanism. So yeah, one of your great interests is very much that for me. We both know and love the work of Jane Jacobs , and many people like that. I had no work-related reason for reading Jane Jacobs, that's for sure, but she's one of the people who influenced me the most.  Just the way she thinks about building complex systems , what it means to do well, and how we should do it. That influences absolutely everything else I do, but I wasn't thinking that when I read her work, it was just that this is really interesting for some reason I don't entirely understand.
DEVON: Now that you say that, I see a lot of resonance between the way you think about science as an ecosystem, and the way she talks about cities as ecosystems and how they can remain healthy and strong, and vibrant. 
MICHAEL: Yeah. It's another example of the use of decentralized knowledge to improve society. That very much so influences the way I think about science funding. You have sort of the eyes on the street in Jane Jacobs, which provides safety, security, and so many other things. In science, you can either have very centralized funders making all the decisions, or you can try and devolve a lot of trust out to individual scientists, and trust that individual scientists may actually have a much better idea of how to spend their talents, than some centralized kind of vetoing power. There's a structural similarity between those arguments. There's many, many ways that that kind of thing has influenced my thinking about science funding.
Author and activist Jane Jacobs at a community meeting in Greenwich Village’s Washington Square Park in 1963. Image from Fred McDarrah.
DEVON: The beauty of a truly powerful analogy is that it can predict things that you hadn't considered yet. Instead of being that one to one mapping, it's many to many mapping. Once you observe something about one system, you realize, “oh, maybe this is true about the other one as well”. So have you, have you had that experience with comparing cities to science?
MICHAEL: Well, actually, I'm just having it right now. In fact, I've never really made this particular sort of type of analogy before. I'm just wondering, you know, what the analogy of safety and eyes on the street is in science? Yeah, maybe there's some analogy to be made to the process by which we keep scientific results reliable. 
Jane Jacobs makes this lovely point that you want city block sizes to be very small. The smaller they are, the more rapidly you can get mixing, because you have a cross strait so people from other states can come in and instead of having a one dimensional system, you actually start to get access to, to a full two dimensional grid. I'm wondering what the analog of small block sizes is in science. Actually, maybe it's got something to do with the size of the organizational units you use to do science. Maybe if they're relatively small and they're put in serious contact with each other on a relatively frequent basis, there's kind of an interdisciplinary confab where two or three of these units get together and talk to each other. 
DEVON: On the safety one, it makes me think about policing and how communities that have the highest rates of safety and the lowest rates of crime, typically not because they have tons of police on the streets. It tends to be because there's social and cultural norms that keep people from harming each other. Also, lots of other factors like wealth and opportunities, and whether people have options besides robbing each other, and so on. 
I have never been a scientist, but I could imagine that it's much more motivating for people to do great science and have honest results, if it's because they want to impress their fellow scientists, and they want to legitimately find an interesting contribution. Whereas if there’s a national body that makes sure that your science is good, that's probably a lot less personally motivating.
“I certainly had the experience of starting to really understand mathematics as a teenager, and really giving up video games as a result… Why would I be trying to play video games, when I could be doing something that was 1000 times more rewarding?”
MICHAEL: That's certainly true. But I think there's an even more fundamental thing, which is just the enjoyment of doing good work and understanding. 
I certainly had the experience of starting to really understand mathematics as a teenager, and really giving up video games as a result, because mathematics was so much more interesting. Then a little bit later, science was so much more interesting. Like, why would I be trying to play video games, when I could be doing something that was 1000 times more rewarding? Just intrinsically, I didn't need anybody else. Mathematics really has this incredible clarity in that way.
I remember first understanding Euclid’s proof , that there are infinitely many primes, and the idea that the square root of two is irrational - it's just the most incredibly beautiful thing. Then you start to be given little problems of your own and you start to have the same kind of shock of understanding. That's just so rewarding. So I think there's a lot of that as well, which is independent of impressing your peers, which is not to say that's not also important. I think for many people, discovery is already such a motivating factor. I find I meet a surprising number of people who have heard about the replication crisis in psychology or things like this. They think that scientists need to be policed, or that somehow the incentives aren't right. Sometimes they don't understand what an enormous intrinsic incentive there is. 
DEVON: On the point of intrinsic motivation, how widespread do you think that is in the population of scientists? Like if you were to pluck a random scientist out of a hat? Do you think that that would apply to the person more likely than not?
MICHAEL: Certainly, the scientists I've met in my life, I think the great majority of them, if they wanted to become wealthy or powerful, or high status, that many other things that they could have done that would have afforded them greater opportunity to do that. I think most of them just absolutely adored science. That's not always true, actually it tends to be that they either absolutely adored science as kids, or else there's sort of a little bit of luck or chance. They were in the right lab at the right time and all of a sudden they realized, “Oh, this is great”. So they forwent those other possibilities. 
With that said, it is difficult to escape some type of careerism. People like to eat. Most people, all other things being equal, prefer to have a higher salary. There are these very sad cases of people committing fraud or something very dubious. Very often, what's going on is people who just become too invested in some false notion of success. They're disconnected from what they're really supposed to be doing. It's sort of sad individually, and of course, pretty terrible for our society.
DEVON: I could even imagine a situation where that's not inconsistent with some intrinsic motivation. I can reflect on times where, let's say, I'm having a conversation with somebody about a topic I'm deeply interested in. I actually want to learn from them really, really badly. But then sometimes I'll have a moment where I think “If I asked this question, I will look stupid, it will lower my status in their eyes, because they will think 'she doesn't know that already?'”. I always try very hard to push through that feeling because I value information, but I can feel the tug of “maybe maybe I should hold back, maybe I shouldn't ask this question". I don't think that that is inconsistent with me wanting to know the answer. It's just another feeling that is layered on top. Those two things, while they end up having conflicting results, they can exist, they can certainly coexist.
In something you wrote , you quoted Paul Dirac . They said that in the 1920s, when quantum mechanics was discovered, it was a period in which it was very easy for any second rate physicists to do first rate work. The UN said history suggests that the early days of new scientific fields are often golden ages, with fundamental questions about the world being answered quickly and easily. And this feels to me like one of those, you know, when you see things, but I'm really curious, like what to you? Does it mean to create a new field?
MICHAEL: Oh, yeah, that's a really interesting question. What is a field that is a really interesting one, and really complicated. And as far as I know, there's not. Okay, so like, one possible set of associations you might have is, to some deep set of related ideas. For example, the Maxwell Lorenz equations are used to describe electromagnetic phenomena, this is an incredibly deep set of ideas. You can spend all of your days just studying the consequences of these equations, and understanding all kinds of electrical and magnetic phenomena. So in that kind of approach, to think about what a field is, a field is somehow a social structure that grows up around a particular deep set of ideas.
Diagram of Maxwell’s interpretation of electromagnetic fields. Image from the Journal of Modern Physics
Now, something like physics, it's not a single deep set of ideas. It's actually a whole lot of sort of loosely related, when there's quantum mechanics is condensed matter physics, there's astrophysics, there's hydrodynamics. It's a loose agglomeration. So I've started with deep ideas, but then you also have these kinds of political structures that sit on top of them, you know, is it a field? Is it something which has a named department at universities? You've made it entirely a political thing. 
But actually, the time at which you start to get program officers and program managers at major funders is often a very important time of transition. In a field, it will, you know, before that happens, there's not really any opportunity for a set of ideas to start to grow. I was involved a fair amount, in the relatively early days of quantum computing. It was really interesting to watch as people started to get the ability to do things like run conferences, or submit to journals. It sounds crazy, but it was actually quite difficult to know where to publish, because there wasn't really a sort of a publication home. There are different transitions that take place in the life of a field, but I do think, at the core, if there's not a deep set of ideas, there can't possibly be a field. 
The reason I've expounded at such an enormous length here, there are two interesting things that can happen. I like to think of Potemkin fields, which I won't try to name any particular examples. But it's when, you know, sort of money and interest, tries to declare that something is a field, but there's no deep set of ideas. You can spend as much money as you'd like, if no deep set of ideas is discovered, you'll have plenty of activity. You might have journals, you might have all kinds of things happening, but there will never really be any, any substance to it. 
This affects to some extent, or can affect interdisciplinary work, where maybe you have two fields, both of which are grounded in deep sets of ideas. But just declaring something to be at the intersection doesn't mean that there's sort of interesting, deep ideas to be discovered that sometimes there is actually I think quantum computing basically occurred by people mashing up quantum mechanics with computer science. But sometimes, interdisciplinary work can sort of flounder a bit, because there'll be people trying to work at the intersection of two fields. 
DEVON: For these Potemkin fields, I like that phrase a lot. 
MICHAEL: You might be able to make the case that computing in the 19th century was an example of a Potemkin field. So people like Charles Babbage attempted to invent the field of computer science. They actually got a fair amount of support for it, but the time wasn't quite right. They were asking great fundamental questions, and were able to make a little bit of progress, but the time wasn't right for the field to exist. So they weren't really able to get to the incredible set of ideas that Von Neumann and others had in the 1930s and 1940s, with the invention of the transistor and so on.
John Von Neumann by the stored-program computer at the Institute for Advanced Study. Image from Getty
DEVON: What would you consider the fieldeist field like the archetypal field of a field?
MICHAEL: Well, you know, physics was very successful early, particularly classical mechanics, Newton's Newtonian mechanics. So it's very tempting to say, “oh, this is the prototype for a field”. But part of what's exciting is when things work in very different ways. I don't know much about AI, but when I hear some of the criticisms that are made by skeptics, I maybe hear a little bit that they're trying to force it into an old mold. So they want systems that they can really understand, like, you know, how they operate. If they're talking about things like language models, they want to be able to understand the relationship of the model to things like the different parts of speech, nouns, verbs, grammar, and so on. And we just don't have that. We don't have a principled way of understanding the way any of the big foundational models are operating. Maybe some of the old standards actually shouldn't be applied. 
There's usually some back and forth interplay between tinkering and fooling around with stuff. That [tinkering can] sometimes be improved by a detailed scientific understanding of what's going on. Maybe in the case of AI, you don't actually need the deep understanding step quite as much because we can do the experiments so quickly. I think the skeptic says, “You're never going to have successful artificial intelligence, if you don't stop and understand what these systems are doing. You know, look at the history of the way we've improved technologies in the past, we always needed to understand how they operate”. The counterpoint to that is well, yeah, that's true. But actually, we're in a different situation. Today, our ability to test our tinkering is just much greater than it has been in the past, we can try many, many, many more possibilities automatically. So we don't need really strong theoretical explanations to rule out incorrect lines of investigation, we can instead just try a trillion or a trillion trillion trillion different things and rely on our ability to recognize when something is working, rather than derive from first principles why it's working.
DEVON: It sounds like you're describing a sort of feeling towards understanding as opposed to thinking towards understanding.
MICHAEL: A little bit. Yeah, I mean, very much this intuitive approach, where you don't necessarily systematically understand why things are improving. Feeling is not a not a bad way of putting it. But you can't go from like, the steam engine to a modern Tesla without having a lot of scientific explanation somewhere in between. In some sense, we're trying to do that on the AI side without that sort of detailed understanding in between time. So it's easy to see why you might be skeptical. But you just couldn't test trillion trillion trillions of intermediate technologies in the past. 
DEVON: I think what I'm imagining is if technologies we could operate are a tree, and you have all these different choices that you could make, similar to the choices you could make in chess, the human mind has to build models that make it more clear which paths to go down because we just cannot compute all of the different possible paths. But with an AI with a computer, we can run them down all the paths and see, “is there anything at the end?” That's interesting. It can come back and tell us so we won’t need to have as much of a model of understanding about why that path is the right one, because we've tried them all. We just pick the one that has the result that we want.
MICHAEL: Yeah, that's, you've just said much more clearly what I was trying to say than how I said it.
DEVON: No, makes a lot of sense. I had never considered that before. I don't follow AI nearly as closely as you do. Is this something that you're seeing in the behavior of how AI researchers are doing their work?
MICHAEL: It's just when I talk to AI researchers versus some of the critics of the AI researchers, I see that dichotomy a lot. It is possible, I'm unfairly characterizing some or all of these people. But when I see people like David, or to some extent, Gary Marquez, certainly Noam Chomsky, critiquing approaches to AI, that seems to me, like a big part of what they're saying is that we should have detailed understanding, telling us, you know, which, which in your, in your way of describing it, which of those branches not to go down? You know, which ones are the correct ones to go down. And they're not very happy with a situation where people essentially say, let's just go down all of them, and rely on the ability to recognize at the end, whereas the thrust of the AI researchers seems to be they're much happier, just just trying a very large number of different things.
Ian Goodfellow in interview at EmTech Digital. Image from MIT Technology Review
So yeah, a really interesting question and answer with the AI researcher Ian Goodfellow . Somebody asks a question , it's along the lines of “Do you feel upset when one of your experiments doesn't work out?” And he says, “oh, no, not at all. I adopt what I consider a high throughput approach to experimentation. In fact, while running while doing this question and answer session, I've said a whole bunch of different experiments running and seeing the results”. So for him, it's just the sort of very cheap thing where he's able to keep asking questions and keep getting answers very, very, very quickly enough so that he can do it while actually participating in the Q&A. I really liked that terminology, sort of a high throughput approach to experimentation. It seems to me like, in many ways, a fundamental shift in how we think about science.
Something that friends who work in AI labs, commercial, industrial AI labs now do tell me that one thing that they have noticed that is a difference is there is much more emphasis put on having access to really first rate engineering resources. So in the past, some poor grad student had to write all the code and, you know, arrange all the infrastructure. Now, there will be a team that just specializes in making sure all the infrastructure is working extremely well, being able to rapidly scale up experiments and do these kinds of things. 
“For certain problems, they actually had to flip pieces of paper over the monitor to cut the kids off. Instead, they actually had to stop and think - sometimes speed is your enemy.”
DEVON: How do you think physics experiments would be different, or physics in general, if physicists had more of this mindset of “let's just throw as many experiments on the wall as possible and see what sticks".
MICHAEL: There's often intrinsic timescales in physical systems that make it hard. In my one published experiment, I'm a theoretical physicist by one published experiment, we're using a particular molecule to do an NMR experiment. And we needed to wait roughly three minutes at the end of every cycle, simply for the molecule to relax back to its original ground state. When we're sort of just setting up we would typically wait like 50, 60, 70 seconds. So it wasn't really being reset properly, but it was mostly being reset. That was just for us to kind of do some initial rough calibration. But later on, when we're actually doing the real experiment, there was just this really annoying three minute wait. Everything would have been so much easier, if we could have eliminated that. If we'd had multiple machines in parallel, that also would have been very profound for us. But I can remember that the cost of the machine was on the order of $10 million, so having multiple machines was out. 
Example of NMR Spectrometer machine. Image from Oregon State University
I don't know what the consequences would have been. I instinctively feel that it would have completely transformed the process for us, not always, not entirely for the better - we did a lot of good thinking and those three minute intervals. Basically, when doing an experimental run, you'd see the results. In the three minutes, while we were waiting, we'd write up a few lines about what we just saw. Sometimes, of course, we'd spend much more time if something very important just happened. But most of the time, it was just a few minutes of writing. There were a lot of good thoughts that were jotted down there. Experimentalists make comments to the effect of more thinking and less work is sometimes the right approach to problems, but it's not always the easiest thing to do. 
There's another interesting example. I remember who I heard this from, possibly from Alan Kay , possibly from somebody who was at his school in Los Angeles, in the 70s or 80s. They've got a bunch of kids who are some of the first users of computers. For certain problems, they actually had to flip pieces of paper over the monitor to cut the kids off from the monitor. So the kids couldn't just keep randomly trying calculations to solve some problem. Instead, they actually had to stop and think, sometimes speed is your enemy. Actually, that's a fun question, I think about cases when really dramatically speeding something up, like the ability to iterate, has actually resulted in a sort of secular slowdown. Like the overall process has maybe been burned by it. Can you think of any examples like that?
DEVON: I would say debugging often has this effect, in programming, where if I get into a rhythm on debugging, and I'm just, I'm just trying all sorts of different things. This is something many, many people experience where as soon as you step away from the computer to go to the restroom, or grab a coffee, or go to sleep, that's when the solution sort of hits your brain. I think the fact that you can, you have so many ways to test something at your fingertips can often stop you from stepping away and just letting the problem sink into your subconscious. 
MICHAEL: Going for a walk is often the best debugger.
DEVON: Right, right. Luckily, there are certain basic human needs that you have to fulfill that sort of force on you, like “Oh, I haven't eaten in 12 hours, maybe I should stop trying to find this bug that is making me so frustrated”. And then you go have a sandwich and you think, “Oh, I solved that”. On the other hand, those hours of debugging, poking at the system, getting lots of different inputs into your brain, are helping you build a model. That subconscious process would not be able to do what you want, if you hadn't done a lot of that. 
Children playing with the Xerox Alto in 1973; each cost $70,000, and only 2,000 were produced. Image from Xerox Archives
As good as slowing things down can be, sometimes I think that there are plenty of processes in the world that slow things down already. And it's much harder to speed things up. So I'll tend to bias towards tools that will make me faster. There's just plenty of things that are going to make you stop and you know, give you time to think, but not so many that help you get interesting, relevant inputs more quickly.
MICHAEL: I wonder, the use of libraries in computing seems closely related, where some particular libraries will become so canonical. I'm thinking of the LINPACK library and in science, which is used to do linear algebra. Maybe there are people who like debugging all the time, but I also wonder if important parts of the code may just end up being like the sort of civilization we neglected. Like they become like infrastructure that everybody uses, but nobody ever thinks about it. 
It is a really interesting long term trend, I think in our civilization some group or a person spends an enormous amount of effort developing [the infrastructure]. If they're really good, then they just sort of sit there. And nobody really has to think about them for a long time. And we're gradually building more and more layers of that. So it's really good from an immediate speed perspective, you just want to be able to import NumPy or whatever. Over the long term, it creates sort of an interesting failure point. For catastrophic failure, if there's a bug in one of those things in some important way . It can potentially have sort of interesting system wide effects.
DEVON: This is the stuff of my nightmares. I’m not going to sleep tonight. Thanks, Michael.
MICHAEL: I don't know. I'm sorry. I flew over the main distribution center of, I guess, supermarkets in Toronto once. And it was just interesting, kind of looking down from the plane. And thinking, you, when you've got a centralized system like that, that is so important, like an essential function of a city, what happens when that starts to fail? You know, just how brittle are the systems? I wonder about it. 
I gave the example of LINPACK. I'm sure that LINPACK is used in things like the discovery by the Large Hadron Collider of the Higgs Boson, or the discovery of gravitational waves by LIGO and probably in a zillion other important discoveries. It's just a sort of a fun game to play to think about. Yeah, how would you detect subtle software bugs in the output from those experiments? I'm quite confident those teams know a lot about what they're doing. But at least as a science fictional sort of scenario, it's fun to think about the possibility of bugs very deep inside software, actually really causing us to come to erroneous conclusions about the way the world works. Maybe the way in which food gets distributed, or electricity is the electrical grid stabilized or something like that.
DEVON: There's all those stories of biology labs using Excel, and then like, genes get parsed as dates via Excel, instead of genes. Then all these results are way off as a result. I hear stories like this pop up pretty frequently and those are just the ones that we hear about, I'm sure that there are many, many more that we are not aware of. 
I think this is where the importance of viewing problems in different ways becomes so important through different lenses, so that you can sanity check the results. I mean, you know, in math class, I remember in elementary school, they said to solve the problem two different ways. And that way, if you made a mistake in one of the paths, it will show up and you'll see that there's a different result. I would imagine that with some of these problems in physics, it's probably pretty difficult to show it in one way, let alone in two ways.
MICHAEL: So actually, the further the Higgs Boson is, there are at least two detectors at the LHC. And so you can in principle, sort of run not entirely independently, like this suddenly gonna be a lot of common infrastructure still, like they'll still all be using whatever, you know, Intel or AMD. So there's still the possibility and it's not like they're completely uncorrelated, but there's quite a bit of independence for LIGO. They had a team whose job was to infect it to inject fake signals essentially. Yeah, it was to act as sort of a Chaos Monkey in the system, which I think is just great. Having that kind of adversarial group built into the design of the experiment is so cool. It also must have been so much fun. I mean, imagine you're part of a scientific group and you're hired actually to sneak into the lab at night and do nasty things, kind of essentially. It's the role that they have, and you want to be resilient against that kind of behavior.
DEVON: It sounds like white hat penetration testing and software systems. But in intellectual systems. That's fun, I would definitely watch a heist movie with that.
MICHAEL: That's such a great idea. That's such a great idea.
DEVON: Or at least like a sci-fi novel heist movie, I think I think that would be fun.
MICHAEL: I'll bet some actually have. A great thing about science fiction is that it seems like any idea you or I could possibly have, some sci-fi author has at least written a short story about it. There's so imaginative,
DEVON: We'll have to dig one up and link it in the show notes if we can find one. That'd be fun. Besides intellectual chaos monkeys, what are some of the most successful tools for thought that you've seen in physics?
MICHAEL: Language, mathematics, symbols? Yeah, these are all things we take for granted. But of course, they're just amazingly important . Things like Arabic numerals, super easy to take for granted, even actually, just the idea that place matters for numerals is incredibly deep. The fact that if I consider the number 21, 2 has a very different meaning in 21, than when it's just 2 alone. We don't think about that, but it's actually quite a different symbol. In some sense, the context has changed its meaning. So I think you're probably asking about much more recent things, though, is that right?
DEVON: That is true, but that was also a good answer. What are some recent tools for thought developed in physics, say, in the last 100 years?
MICHAEL: Just just narrowing down to a century? Yeah. Not even not even giving me 100,000 years?
DEVON: Yea.
MICHAEL: Certainly, Mathematica has had an enormous impact on physics. So to have things like MATLAB and NumPy , and things like that. I have used those systems, but I'm not a master in the way some people are. I had a student, Henry Haselgrove . What Henry could do with MATLAB was astonishing. We would have a conversation about what I thought of as a very conceptual, theoretical kind of a question. Then, he would just spend a few minutes in MATLAB, and even though he was doing computations with particular matrices, he was able to get conceptual insight about abstract mathematics, which I thought was very interesting. It was much more rapid than I could possibly do, being restricted to just a completely different mode of thinking. 
Graphical User Interface for Mathematica. Image from VNC
Thought that it's really interesting to work with people who have that kind of capacity to generate very concrete calculations with specific numbers, but to draw interesting conclusions about very abstract conceptual questions. Physics has certainly changed. To the point where computational and numerical methods have really become a third way of understanding so many systems. You just can't analytically solve a lot of problems. You know, if you want to understand what the gravitational wave signature of colliding black holes is going to look like, you need to do some pretty heavy duty numerical calculation. That will tell you then what sort of smoking gun signatures to look for. 
In some sense, just that ability to do simulation has transformed, not just physics, but all of science. It's like you're able to take a system specification, which is very broad. And then answer questions about specific behaviors under specific circumstances which absent simulation, would just be completely and utterly inaccessible to you. You can't solve the problem of figuring out what kind of gravitational wave signature will be provided by colliding black holes. You can't do that. In theory, the problem is too complicated. You could do it experimentally, if you knew that you had two black holes nearby. But we don't know that we have two black holes nearby, we're trying to figure out if what we're seeing is, in fact, two black holes. So it's important then to have this extra method, this method of simulation, which is able to say, “Oh, two black holes colliding would look like this in gravitational wave signature”. So then you study a whole bunch of sort of events like that maybe you didn't see in your simulations about black holes and how they behave. That then can be used as input to other experiments. I don't know if that's before you asked that particular example.
How is it transforming science in that, in that particular instance? It's enabling you to make inferences about systems where you're not actually. In the experiment, you're not actually sure what the constituent systems are and you're instead inferring it from the outcome of some theory plus numerical simulation. So that's a new ability in science and a pretty significant one. I’m thinking in real time here
I think somehow, like, you can certainly do the thing where you just try simulating lots and lots of different possible systems that you think might be out there somewhere, you know, neutron stars colliding neutron stars, and a black hole colliding other sorts of many other possibilities. You can simulate all of them. And then you can look in the data to see are there signatures in the data?  That's been the process. I do know that. They did do early simulations of some of these important sorts of plausible classes of astrophysical phenomena. And then they actually just had to wait and see what actually showed up and in the data, there was no guarantee that they were going to see black holes colliding. 
DEVON: What are some areas of physics where you think that we could benefit from better tools for thought where our thinking is kind of hazy because we don't have maps that help us really understand what's going on?
Grant Sanderson's math animations. Image from 3Blue1Brown channel.
MICHAEL: It is certainly true, and I have played a lot with tools, which are just meant to represent particular individual systems. As an example, I'll just use my friend, Grant Sanderson ’s YouTube channel, 3Blue1Brown . It's not a tool, but at least makes the point clear. He does these beautiful animated 3D graphics, which show systems from illustrating phenomena and mathematics, sometimes from physics, sometimes from other areas, but most often formed from mathematics. You're just able to build intuition, often relatively easily, by seeing these kinds of animated drawings that you wouldn't have in any other way. It's extremely striking. I mean, it's a very simple example, and super common, but in many ways, I feel like it's a pity in some sense that it's not more routine. There's quite a startup cost to doing it. To the extent that we can make it easier and easier to do those kinds of things, it's certainly a good intuition generator.
“Visual processing system is so powerful. It operates in parallel, whereas the, you know, symbolic manipulation I was doing before… it's very sort of serialized.”
DEVON: It's really difficult to create the quality of graphics that Grant does for his videos, I do think that there are relatively lower bars that people still don't make it over. Just even drawing a very simple line graph can often be so elucidating. I can't even count the number of times that in a company, I spent an hour or two drawing a visual that explained my thinking and thinking “This won't help that much. This is kind of a waste of time. I should have been doing real work”. And then I share it with other people on my team and they go, “Oh, now I understand”. This makes sense. Sometimes an entire team will form around that diagram and solve a problem. It's always so valuable to do it. I always then kick myself and think, why don't I spend more of my time doing this? Because suddenly, all of these people understand something that they previously didn't. And my own understanding is better. Because once I drew it out, I realized, “oh, this thing doesn't quite add up” or “this thing in my head actually doesn't map to paper very well”. This either means there's something wrong, or it means that there's some way I'm representing it that doesn't quite capture what matters. But yeah, these things are always so valuable. I think it seems that most people underrate them, myself included. Why is that?
MICHAEL: So myself included as well, sort of it's funny like that. The question, why did you know different representations of the same ideas help so much? It's pretty clear that lots of people have that experience. I enjoy Venkatesh Rao , the writer. He loves to draw two by twos for absolutely everything. Yeah, they make you think about things in a slightly different way. Yeah, there's a cheap answer, I think there's a better answer. The cheap answer is when you arrange something in a new representation, it does make it sometimes easier to see other connections, which had formally escaped you, so you see, “oh, this might be related to this other thing” and such in such a way, maybe that I think that the better answer. 
Some years ago, I developed a little prototype. It was for studying the motion of physical systems in one dimension. And the thing that I did in the prototype was that I found a way of representing the conservation of energy, so that you could just see conservation of energy directly in the visual representation provided by the prototype. Even though I've been doing physics for most of my life, 30 odd years, I applied conservation of energy, I don't know how many 1000s or 10s of 1000s of times before, I found just being able to see it directly in this visual representation of the system completely changed the way I related to it. It was no longer an algebraic manipulation that I was doing. Instead, I could just see what to expect. That simple change really made a difference to how I thought about it.
One thing, I mean, it's just because of course, you know, a visual processing system is so powerful. You know, it operates in parallel, whereas the, you know, symbolic manipulation I was doing before, to think about conservation of energy before that, it's very sort of serialized. So it's m

Images Powered by Shutterstock