Architects of Intelligence
Page 54
CEO, THE ALLEN INSTITUTE FOR ARTIFICIAL INTELLIGENCE
Oren Etzioni is the CEO of the Allen Institute for Artificial Intelligence, an independent organization established by Microsoft co-founder Paul Allen and dedicated to conducting high-impact research in artificial intelligence for the common good. Oren oversees a number of research initiatives, perhaps most notably Project Mosaic, a $125 million effort to build common sense into an artificial intelligence system—something that is generally considered to be one of the most difficult challenges in AI.
MARTIN FORD: Project Mosaic sounds very interesting. Could you tell me about that and the other projects that you’re working on at the Allen Institute?
OREN ETZIONI: Project Mosaic is focused on endowing computers with common sense. A lot of the AI systems that humans have built, to date, are very good at narrow tasks. For example, humans have built AI systems that can play Go very well, but the room may be on fire, and the AI won’t notice. These AI systems completely lack common sense, and that’s something that we’re trying to address with Mosaic.
Our over-arching mission at the Allen Institute for Artificial Intelligence is AI for the common good. We’re investigating how you can use artificial intelligence to make the world a better place. Some of that is through basic research, while the rest has more of an engineering flavor.
A great example of this is a project called Semantic Scholar. In the Semantic Scholar project, we’re looking at the problem of scientific search and scientific hypothesis generation. Because scientists are inundated with more and more publications, we realize that scientists, just like all of us when we’re experiencing information overload, really need help in cutting through that clutter; and that’s what Semantic Scholar does. It uses machine learning and natural language processing, along with various AI techniques, to help scientists figure out what they want to read and how to locate results within papers.
MARTIN FORD: Does Mosaic involve symbolic logic? I know there was an older project called Cyc that was a very labor-intensive process, where people would try to write down all the logical rules, such as how objects related, and I think it became kind of unwieldy. Is that the kind of thing you’re doing with Mosaic?
OREN ETZIONI: The problem with the Cyc project is that, over 35 years in, it’s really been a struggle for them, for exactly the reasons you said. But in our case, we’re hoping to leverage more modern AI techniques—crowdsourcing, natural language processing, machine learning, and machine vision—in order to acquire knowledge in a different way.
With Mosaic, we’re also starting with a very different point of view. Cyc started, if you will, inside out, where they said, “OK. We’re going to build this repository of common-sense knowledge and do logical reasoning on top of it.” Now, what we said in response is, “We’re going to start by defining a benchmark, where we assess the common-sense abilities of any program.” That benchmark then allows us to measure how much common sense a program has, and once we’ve defined that benchmark (which is not a trivial undertaking) we’ll then build it and be able to measure our progress empirically and experimentally, which is something that Cyc was not able to do.
MARTIN FORD: So, you’re planning to create some kind of objective test that can be used for common sense?
OREN ETZIONI: Exactly! Just the way the Turing test was meant to be a test for artificial intelligence or IQ, we’re going to have a test for common sense for AI.
MARTIN FORD: You’ve also worked on systems that attempt to pass college examinations in biology or other subjects. Is that one of the things you’re continuing to focus on?
OREN ETZIONI: One of Paul Allen’s visionary and motivating examples, which he’s investigated in various ways even prior to the Allen Institute for AI, was the idea of a program that could read a chapter in a textbook and then answer the questions in the back of that book. So, we formulated a related problem by saying, “Let’s take standardized tests, and see to what extent we can build programs that score well on these standardized tests.” And that’s been part of our Aristo project in the context of science, and part of our Euclid project in the context of math problems.
For us it is very natural to start working on a problem by defining a benchmark task, and then continually improving performance on it. So, we’ve done that in these different areas.
MARTIN FORD: How is that progressing? Have you had successes there?
OREN ETZIONI: I would say the results have been mixed, to be frank. I would say that we’re state of the art in both science and math tests. In the case of science, we ran a Kaggle competition, where we released the questions, and several thousand teams from all over the world joined. With this, we wanted to see whether we were missing anything, and we found that in fact our technology did quite a bit better than anything else out there, at least who participated in the test.
In the sense of being state of the art and having that be a focus for research, and publishing a series of papers and datasets, I think it’s been very positive. What’s negative is that our ability on these tests is still quite limited. We find that, when you have the full test, we’re getting something like a D, not a very stellar grade. This is because these problems are quite hard, and often they also involve vision and natural language. But we also realized that a key problem that was blocking us was actually the lack of common sense. So, that’s one of the things that led us to Project Mosaic.
What’s really interesting here is that there’s something I like to call the AI paradox, where things that are really hard for people—like playing World Championship-level Go—are quite easy for machines. On the other hand, there are things that are easy for a person to do, for example if you look at a question like, “Would an elephant fit through a doorway?”, while most people can answer that question almost instantaneously, machines will struggle. What’s easy for one is hard for the other, and vice versa. That is what I call the AI paradox.
Now, the standardized test writers, they want to take a particular concept like photosynthesis, or gravity, and have the student apply that concept in a particular context, so that they demonstrate their understanding. It turned out that representing something like photosynthesis, at a 6th grade level, and representing that to the machine is really quite easy, so we have an easy time doing that. But where the machine struggles is when it’s time to applying the concept in a particular situation that requires language understanding and common-sense reasoning.
MARTIN FORD: So, you think your work on Mosaic could accelerate progress in other areas, by providing a foundation of common-sense understanding?
OREN ETZIONI: Yes. I mean, a typical question is. “If you have a plant in a dark room and you move it nearer the window, will the plant’s leaves grow faster, slower or at the same rate?” A person can look at that question and understand that if you move a plant nearer to the window then there’s more light, and that more light means the photosynthesis proceeds faster, and so the leaves are likely to grow faster. But it turns out that the computer really struggles with this—because the AI doesn’t necessarily understand what you mean when you say, “What happens when you move a plant nearer the window.”
These are some examples that indicate what led us to Project Mosaic, and what some of our struggles have been with things like Aristo and Euclid over the years.
MARTIN FORD: What led you to work in AI, and how did you end up working at the Allen Institute?
OREN ETZIONI: My foray into AI really started in high school when I read the book, Gödel, Escher, Bach: An Eternal Golden Braid. This book explored the themes of logician Kurt Gödel, the artist M. C. Escher, and the composer Johann Sebastian Bach, and expounded many of the concepts that are relatable to AI, such as mathematics and intelligence. This is where my fascination with AI began.
I then went to Harvard, for college, where they were just starting AI classes when I was a sophomore. So, I took my first AI class, and I was completely fascinated. They were not doing much in the way of AI a
t the time but, just a short subway ride away, I found myself at the MIT AI Lab and I remember that Marvin Minsky, the co-founder of MIT’s AI lab, was teaching. And actually Douglas Hofstadter, the author of Gödel, Escher, Bach, was a visiting professor, so I attended Douglas’s seminar and became even more enchanted with the field of AI.
I got a part-time job as a programmer at the MIT AI Lab and for somebody who was just starting their career, I was, as they say, over the moon. As a result, I decided to go to graduate school to study AI. Graduate school for me was at Carnegie Mellon University where I worked with Tom Mitchell, who is one of the founding fathers of the field of machine learning.
The next step in my career was when I became a faculty member at the University of Washington, where I studied many topics in AI. At the same time got involved in a number of AI-based start-ups, which I found to be very exciting. All of this resulted in me joining the Allen Institute of AI, and more specifically in 2013 Paul Allen’s team reached out to me saying that they wanted me to launch an Institute for AI. So, in January 2014 we launched the Allen Institute for AI. Now fast forward to 2018 and here we are today.
MARTIN FORD: As the leader of one of Paul Allen’s institutes, I assume you have a lot of contact with him. What would you say about his motivation and vision for the Allen Institute of AI?
OREN ETZIONI: I’m really lucky in that I’ve had a lot of contact with Paul over the years. When I was first contemplating this position, I read Paul’s book Idea Man, which gave me a sense of both his intellect and his vision. While reading Paul’s book, I realized that he’s really operating in the tradition of the Medicis. He’s a scientific philanthropist and has signed the Giving Pledge that was started by both Bill and Melinda Gates and Warren Buffett, where he’s publicly dedicated most of his wealth to philanthropy. What drives him here is that he’s been fascinated by AI and the questions around how we can imbue semantics and an understanding of text in machines since the 1970s.
Over the years, Paul and I have had many conversations and email exchanges, and Paul continues to help shape the vision of the institute, not just in terms of the financial support but in terms of the project choices, and the direction of the institute. Paul is still very much hands-on.
MARTIN FORD: Paul has also founded the Allen Institute of Brain Science. Given that the fields are related, is there some synergy between the two organizations? Do you collaborate or share information with the brain science researchers?
OREN ETZIONI: Yes, that’s correct. So, way back in 2003, Paul started the Allen Institute of Brain Science. In our corner of the Allen Institutes, we call ourselves “AI2,” partly because it’s a bit kind of tongue-in-cheek as we’re the Allen Institute of AI but also because we’re the second Allen Institute.
But going back to Paul’s scientific philanthropy, his strategy is to create a series of these Allen Institutes. There’s a very close exchange of information between us all. But the methodologies that we use are really quite different, in that the Institute of Brain Science is really looking at the physical structure of the brain, while here at AI2 we’re adopting a rather more classical-AI methodology for building software.
MARTIN FORD: So, at AI2 you’re not necessarily trying to build AI by reverse-engineering the brain, you’re actually taking more of a design approach, where you’re building an architecture that’s inspired by human intelligence?
OREN ETZIONI: That is exactly right. When we wanted to figure out flight, we ended up with airplanes, and now we’ve developed Boeing 747s, which are very different than birds in several ways. There are some of us within the AI field who think that it is quite likely that our artificial intelligence will be implemented very differently than human intelligence.
MARTIN FORD: There’s enormous attention currently being given to deep learning and to neural networks. How do you feel about that? Do you think it’s overhyped? Is deep learning likely to be the primary path forward in AI, or just one part of the story?
OREN ETZIONI: I guess my answer would be all of the above. There have been some very impressive achievements with deep learning, and we see that in machine translation, speech recognition, object detection, and facial recognition. When you have a lot of labeled data, and you have a lot of computer power, these models are great.
But at the same time, I do think that deep learning is overhyped because some people say that it’s really putting us on a clear path towards artificial intelligence, possibly general artificial intelligence, and maybe even superintelligence. And there’s this sense that that’s all just around the corner. It reminds me of the metaphor of a kid who climbs up to the top of the tree and points at the moon, saying, “I’m on my way to the moon.”
I think that in fact, we really have a long way to go and there are many unsolved problems. In that sense, deep learning is very much overhyped. I think the reality is that deep learning, and neural networks are particularly nice tools in our toolbox, but it’s a tool that still leaves us with a number of problems like reasoning, background knowledge, common sense, and many others largely unsolved.
MARTIN FORD: I do get the sense from talking to some other people, that they have great faith in machine learning as the way forward. The idea seems to be that if we just have enough data, and we get better at learning—especially in areas like unsupervised learning—then common-sense reasoning will emerge organically. It sounds like you would not agree with that.
OREN ETZIONI: The notion of “emergent intelligence” was actually a term that Douglas Hofstadter, the cognitive scientist, talked about back in the day. Nowadays people talk about it in various contexts, with consciousness, and with common sense, but that’s really not what we’ve seen. We do find that people, including myself, have all kinds of speculations about the future, but as a scientist, I like to base my conclusions on the specific data that we’ve seen. And what we’ve seen is people using deep learning as high-capacity statistical models. High capacity is just some jargon that means that the model keeps getting better and better the more data you throw at it.
Statistical models that at their core are based on matrices of numbers being multiplied, and added, and subtracted, and so on. They are a long way from something where you can see common sense or consciousness emerging. My feeling is that there’s no data to support these claims and if such data appears, I’ll be very excited, but I haven’t seen it yet.
MARTIN FORD: What projects would you point to, in addition to what you’re working on, that are really at the forefront? What are the most exciting things happening in AI? The places you’d look for the next big developments. Would that be AlphaGo, or are there other things going on?
OREN ETZIONI: Well, I think DeepMind is where some of the most exciting work is currently taking place.
I’m actually more excited about what they called AlphaZero, than AlphaGo, and so the fact that they were able to achieve excellent performance without hand-labeled examples, I think is quite exciting. At the same time, I think everybody in the community agrees that when you’re dealing with board games, it’s black and white, there’s an evaluation function, it’s a very limited realm. So, I would look to current work on robotics, and work on natural language processing to see some excitement. And I think that there’s also some work in fields called “transfer learning,” where people are trying to map from one task to the other.
I think Geoffrey Hinton is trying to develop a different approach to deep learning. I think at AI2, where we have 80 people who are looking at how do you put together the symbolic type of approaches and knowledge with the deep learning paradigm, I think that’s also very exciting.
There’s also “zero-shot learning,” where people are trying to build programs that can learn when they see something even for the first time. And there is “one-shot learning” where a program sees a single example, and they’re able to do things. I think that’s exciting. Brenden Lake who’s an assistant professor of Psychology and Data Science at NYU, is doing some work along
those lines.
Tom Mitchell’s work, with lifelong learning at CMU, is also very interesting—they’re trying to build a system that looks more like a person: it doesn’t just run through a dataset and build a model and then it’s done. Instead, it continually operates and continually tries to learn, and then learn based on that, over a longer extended period of time.
MARTIN FORD: I know there’s an emerging technique called “curriculum learning,” where you start with easier things and then move to the harder things, in the same way a human student would.
OREN ETZIONI: Exactly. But if we just take a step back for a minute here, we can see that AI is a field that’s rife with bombastic and overly grandiose misnomers. In the beginning, the field was called “artificial intelligence,” and to me, that’s not the best name. Then there’s “human learning” and “machine learning,” both of which sound very grandiose but actually, the set of techniques they use are often very limited. All these terms that we just talked about—and curricular learning is a great example—refer to approaches where we’re simply trying to extend a relatively limited set of statistical techniques, and to start to take on more of the characteristics of human learning.
MARTIN FORD: Let’s talk about the path toward artificial general intelligence. Do you believe it is achievable, and if so, do you think it’s inevitable that we will ultimately get to AGI?
OREN ETZIONI: Yes. I’m a materialist, so I don’t believe there’s anything in our brain other than atoms and consequently, I think that thought is a form of computation and so I think that it’s quite likely that over some period of time we’ll figure out how to do it in a machine.
I do recognize that maybe we’re just not smart enough to do that, even with the help of a computer, but my intuition is that we will likely achieve AGI. As for the time line though, we’re very far from AGI because there are so many problems that need to be solved that we haven’t even been able to define appropriately for the machine.