Chris McKinstry Replies: Telecopes, AI And More
News | Posted by timothy on Monday July 10, @11:00AM
from the international-man-of-mystery dept.
A few weeks ago you asked the multi-talented Chris McKinstry questions, about the telescope projects he's involved with (ESO's Very Large Telescope -- VLT -- and the OverWhelmingly Large telescope -- OWL), about his project to synthesize AI by collecting a database of answers to questions common and obscure, and about the possibilities of discovering extraterrestrial life. Read what he has to say on everything from humans leaving the solar system to telescopes staying here on Earth.
by Dungeon Dweller
I have an active interest in artificial intelligence. I study it as part of my major, and hope to do research in it in the future. As a young man coming up in the world, I am hoping to enter into research eventually, am entering into research at my university (WVU).
Your project reminds me of several projects/theories that have been discussed before. In the psychological debate, your system depends entirely upon nurture, it would seem. I like that kind of system and research. I do have a few questions.
I know that I posed a lot of questions, but several could be answered in combination, I just didn't want to ask 2 questions at the same time.
- What separates this from other projects in the field?
- Where did you draw your inspiration for this project?
- What kind of support staff do you recommend to an individual who has never led research before, but would like to? (I ask this of many of my professors who conduct research)
- Where are you getting the bulk of your input for this project?
- What do you hope to learn from this project?
- At what time will you consider this project a success?
There are three primary features of the MindPixel Digital Mind Modeling Project (also known by GAC -- for Generic Artificial Consciousness -- which is public interface to the project) that distinguish it from other large scale knowledge projects such as CYC.
- The first phase is a completely public, internet based effort. All the data it will be collecting will come from average people, with no specific training in AI or psychology. It is like seti@home in many respects, except that we're not after your CPU's cycles, but rather your humanness. We're actually seeking to extract the entire content of an average person's mind bit by literal bit from millions of different internet users. We're not trying to write the algorithm for consciousness, but rather create the world's most rigorous fitness test (a Dawkinsian continuous variable) and get it into the hands of researchers who will attempt to make systems that will learn or evolve into consciousness by feeding back against this fitness test. Not only will we be collecting consensus fact, but also consensus emotion. (When the project is fully operational, in addition to collecting information about each MindPixel's truth or falsity, we will also collect emotional data based on Mehrabian's PAD model of emotion.)
- The second phase of the project involves releasing the data collected to the scientific community and providing those researchers with some funds (generated by advertising to the people supplying the data) to conduct their research. As a side note, Jeff Elman's page contrains information about recurrent neural networks that are very good at processing just the kind of data that this project will collect and distribute. Specifically his 1990 article, Finding Structure in Time (PDF) is one of the most important neural network papers ever written; it strongly influenced me.
- Finally, the project is a meritocracy. People will gain voting rights that will give them a say in every aspect of how the project is run, from data collection and use to the distribution of data and research funds, based entirely on the amount of data they have contributed to the project. The more work you do, the stronger your voice becomes.
My primary inspiration for the project comes from observation: I observed that computers are stupid and know nothing of human existence. I concluded a very long time ago that either we had to write a "magic" program that was able to go out in the world and learn like a human child, or we just had to sit down and type in ALL the data. When I was studying psychology in the late 80's I wanted to begin to gnaw the bullet and start getting people to type in ALL the data. It was my plan then to get people to enter data as part of an intro psych course, or get the university to allow me ask people for data when they logged on to the university's computer system. I was never able to get permission for either and the idea sat on the shelf until I downloaded my first copy of NCSA's Mosaic in 1994. I saw in following my first hyperlink, a different path.
I decided to collect my data via the internet. But, the problem was, that I needed to think of a standard format for the data; some way of representing human knowledge that an average person could learn quickly. That idea didn't come to me until I was preparing an entry for the 1995 Loebner Prize. Jackie, my program, was a stimulus response creature. You would ask her a full text question, and she scan her database for a canned full text response. My idea for the Loebner competition was to have her talk to a lot of people a get a lot of canned responses (at the time I was consulting for a large insurance company and brought Jackie to work everyday where she could talk to my colleagues) As well, I stored the responses in a number of different ways: phonetically using soundex, again with all the words in each stimulus sorted alphabetically, and also with a primitive concept token system. So, if there was no direct match, she would look for a phonetic match or sorted or conceptual match. Essentially I was breaking down each stimulus and standardizing it like a Fourier transform breaks down a waveform.
Then suddenly Hugh Loebner changed the rules. No longer was passing a text based Turing Test good enough for him. Now he would only award his prize if the system passed a full audio/video Inquisition. I hit the roof! Hell, there were tens of thousands of people on the planet that couldn't pass that kind of test! Anyone blind or deaf are just two obvious examples. I withdrew Jackie in a loud protest, stating that intelligence didn't depend on the bandwidth of the communication channel; intelligence could be communicated with one bit! If you locked a person in a box I could detect them with a series of yes/no questions and nothing more. And there all of a sudden, I had my answer (and a quick paper - The Minimum Intelligent Signal Test - An Objective Turing Test in Canadian Artificial Intelligence, issue 41.) There was a minimum intelligent signal, and it was just one bit. I would store my model of the human mind in binary propositions. I would make a digital model of the mind.
I realized within minutes that a giant database of these propositions could be used to train a neural net to mimic a conscious, thinking, feeling human being! I thought, maybe I'm missing something obvious. So, I emailed Marvin Minsky and asked him if he thought it would be possible to train a neural network into something resembling human using a database of binary propositions. He replied quickly saying "Yes, it is possible, but the training corpus would have to be enormous." The moment I finished reading that email, I knew I would spend the rest of my life building and validating the most enormous corpus I could.
Support staff! I recommend using the entire planet as support staff! Seriously, don't even dream about it. Almost every researcher I know works on their own or with a handful of collaborators. When you're a big cheese you might get a student or two, but other than that you'll get nothing more than shared use of a departmental secretary. You'll definitely be writing all your own code for a very long time.
I can't tell you that yet because at the time I wrote this, the project was not yet online (should be now though.) What I can tell you is that in 1995 I did try to collect this same data, using a web based form that sent an email back to me. I managed to collect some 450,000 items. This time, I expect to collect more and higher quality data and I expect it to come from a wide cross section of the internet public. I should also note MindPixels will be collected in multiple languages, which opens up the future prospect of mapping the sampled human languages to each other concept by concept. It will be very interesting to see exactly how an artificial consciousness trained in English differs at the conceptual level from one trained in say, Spanish.
I hope to learn what the human conceptual network looks like. I hope that in a few years I will be able to access a map of all the concepts in the head of an average person or to have learned why I can't.
I will consider the project a complete success when the cover of Science announces that for the first time in history there exists an artificial system that has passed a scientifically strong form of the Turing Test known as the Minimum Intelligent Signal Test.
2) How do you guys do it?
With exptremely high magnification, how in heck do you keep the telescope still enough to take photos?
The slightest movement ought to mean millions of miles so thoes pesky little earthquakes should be a problem. Not to mention how you guys move the telescope accurately?
Chris: You're quite right about the system being very sensitive; if I walk on the azimuth platform of a VLT telescope while we're observing, I will destroy the observation. For normal tracking we use a software system called Tpoint written by a well known telescope genius named Pat Wallace (Pat has a wonderful and detailed article about telescope pointing that anyone seriously interested in telescope pointing should read); the same system is in use on telescopes all over the world. Basically what we do is build a pointing model for each of our telescopes. This involves pointing each telescope to a number of different points uniformly covering the sky. At each sample point, we observe a guide star and record how it moves from the center of the field over about one minute of tracking time. After we have collected enough data, we build a computer model of the telescope's tracking error. Then we basically run the model backwards into the telescope control system and thus apply corrections that try to cancel out the tracking errors of the telescope. This of course can't correct for any unusual vibrations, we rely on normal guide star tracking and hydraulic isolation of the telescope for that. And baring a large earthquake, Tpoint, automatic guide star corrections and the isolation work pretty well (In the event of a large earthquake, there are giant airbags that inflate to protect the mirror from damage.)
3) How can we help?
You probably know about SETI At Home, which lets people volunteer spare CPU time to processing radio-telescope data, in a (so far vain) attempt to find extra-terrestrial intelligence. Is there a similar way that we can help process some of the data that you gather?
As a simple example, one could compute the differences between a sequence of pictures of the same portion of the sky, looking for anomalies like giant asterioids on their way to wiping us all out.
Chris: seti@home is one of the most impressive demonstrations of how the world of science has changed. There are now over 2 million average people working together for a common scientific goal. I just wish they sold advertising to raise funds for other worthy (meritocratically determined) projects. It really bugs me that my Pentium III 450 which has done over 7000 hours of seti@home processing since last June, hasn't shown me a single science supporting ad. What a waste!
Now as for your idea of doing the same thing in optical wavelengths, I think in it there is a great deal of merit. There are a whole pile on new survey telescopes coming online soon that will be useful for just what you proposed. And if you read ahead to my answer to question 11, you'll see I do think it is a problem we have to pay attention to. (As well, I know of at least two virtual telescope projects; the NRC's National Virtual Telescope. See NVO White Paper (PDF) and ESO's ASTROVIRTEL which seek to allow data mining of previously collected telescope data.
In general, I think the future will see a lot more distributed processing projects doing useful science. The question remains whether or not it is more cost effective to build supercomputers for critical projects or harness the CPU's of private citizens, and I think the answer will need to be determined on a case by case basis. As well, there will be some projects (my own for example) where the CPU cycles are incidental; where what we want to harvest is not your electricity and capital equipment, but actually your humanity.
4) Division between Science and Spirituality
I am continuously frustrated that people's general perception seems to be that science and art, spirituality, and so forth are divided by an uncrossable schism. People feel the need to pit science against spirituality; logic against intuition. It is a rare thing that people accept the idea that these are different ways of approaching the same reality. As a dreamer and artist as well as a respected scientist, what do you say to people who doubt that scientists can be spiritual and artistic people?
Chris: Science for me at least, is concerned with the external, the measurable; while art is concerned with the internal and immeasurable. Every scientist knows measurement can only go so far; that nature at its most fundamental is immeasurable. Unfortunately many scientists turn away from what they can't measure (and conversely, many artists turn from measurement) instead of finding some way, any way to express it. It is this turning away or fear of the immeasurable (or many artist's converse fear of reduction to measurement) that creates doubt; that separates science from art. The scientist can learn that one does not become any less of a scientist for attempting to express the inexpressible or attempting to measure the immeasurable, just as the artist can learn that because we are neurons and our neurons atoms, doesn't mean we are any less human.
5) CCD or what?
What kind of imaging does a telescope of this scale use? Is it an OWLCCD or something else? What kind of resolution? And how far away would an object need to be before the resolution becomes a shortcoming?
Chris: I actually can't answer this question. I am only aware of one discussion regarding instrumentation for the OWL and I haven't read it yet. See FROM ISAAC TO GOLIATH, OR BETTER NOT!? INFRARED INSTRUMENTATION CONCEPTS FOR 100M CLASS TELESCOPES (PDF) on the OWL project page.
6) Yeah, they're big ...
But what do you do with them?
What kind of work do the telescopes at your facility generally do? Do local astronomers get to come in and do research or are the scopes reserved for some large project?
Chris: There is a very wide spectrum of observing programs for the VTL; from the study of comets in our solar system to the detection and measurement of objects on the edge of the observable universe. The VTL operates in two primary modes: visitor and service. In visitor mode, scientists actually travel to Chile and execute their observing program interactively with the support of operations personnel like myself who know the telescope and control system intimately and staff astronomers that know the instruments and science. Visitor mode is best utilized when the program contains interactive components, for example when what the observer does next depends on the results of what he has just completed. In service mode, observers don't travel to Chile but instead submit observing programs that don't have a large interactive component. Service programs are executed by staff astronomers and the data is automatically returned to the observer upon completion. Service mode is much like the old batch mode of mainframe computers. In both service and visitor modes, the programs that get time are determined by an observing program committee made up of scientists from all over the world based on scientific merit. And yes, a portion of the time (I believe it is 10%) automatically goes to Chilean astronomers in exchange for Chile's donation of the land for the project.
7) How parellelizable?
How parallelizable is the problem of micro-adjusting small portions of a large deformable mirror to correct for atmospheric distortion?
I remember a Scientific American article stating that you'd have to devote a top-of-the-line Cray to continuously recalculate the deformations needed given data from the guide star, or laser simulated guide star. If this problem is highly parallelizable, you may be able to get away with _much_ cheaper hardware.
I'm sure the idea has occured to you, but I want to know what your thoughts are on it.
Chris: My experience with deformable mirrors is entirely practical and I'm really not qualified to comment on the theory behind them. However, speaking from a practical standpoint, the VLT's 450 force actuators (150 per operating telescope) are each activated about 1000 times per night, night after night almost without error (7 non-critical electronic failures up to May of this year). I see no obvious reason why it shouldn't scale smoothly to 130 or 150 meters with current computer technology. And we certainly don't have any supercomputers doing the deformation calculations.
8) Why single-mirror?
I could have been mis-reading the article, but it seemed to me as though the idea was to build a single-mirror system. On the other hand, in radio astronomy, and in the insect world, arrays are considered the norm. Is there some advantage that a single mirror gives that cannot be duplicated using multiple smaller mirrors? (Simpler optics is an obvious one, paradoxically. :) Or is this (at least in part) NerdTrek III: The Search for Sponsors, where a record-setting single telescope is going to get more interest than a comparable array?
(A supplementary question, to go along with this. Let's say, for the sake of argument, that optical arrays are practical. Do you see any possibility of optical astronomers adopting the same line as radio astronomers, in trying to build an effective 1Km+ optical telescope, using an array?)
Chris: Actually, it isn't a single mirror. It is "filled aperture" telescope. The aperture is filled with many smaller mirrors, just like Keck. And as for optical arrays (interferometers), the VLT (called VLTI in this mode) will be the first real large scale test of such a system. But that stage of the project is still a few years away. In short, we'll have to wait and see how effective it is before we even consider giant optical interferometers.
What's the biggest hurdle to hop over in getting funding for projects like OWL? And how did you pull it off?
Chris: The biggest hurdle for getting funding for projects like OWL, is getting funding for construction! Construction of OWL hasn't been funded, so nothing has really been pulled off, yet. But, if the public really wants projects like this to go ahead, then they need to be active about it. If you're in Europe, write your representatives and mention OWL by name and direct them to the OWL project page. If you're not in Europe, urge your representatives to find some way to participate in this project or projects like it.
10) Terrestrial Optical Telescopes
What are the benefits of having an Earth-bound, optical telescope? Or rather, what can a larger optical telescope find better from Earth that we can't already find on other wavelengths and from other venues (i.e. The Hubble)?
If there are no advantages here, is it more cost-effective, or what?
Chris: What you should actually ask is what advantage does a space based telescope have over a ground based telescope? The only thing you gain from being in space for an optical telescope is better image quality due to lack of atmospheric turbulence. By for every other measure (maintenance, support, materials, etc.) being in space is much, much more expensive and limited. Which is why the Hubble and it's 2.4 meter primary cost a number of times more than the projected cost of of the 100 meter OWL. Recent advances in computer technology (adaptive and active optics) have greatly reduced the advantage that being in space provides at optical wavelengths. For some non-optical telescopes (x-ray, IR, gamma ray) there will always be an advantage to being in orbit.
11) might as well ask it now..
by Blue Lang
I noticed in your 'fave books' section that you have the blind watchmaker, et al.
so, with an eye towards dawkins' views on evolution, what's your personal take on the probability (not the possibility) of humans locating extraterrestrial life without going outside the solar system?
Chris: Actually I'm quite pessimistic about the prospects of us locating ETL, AND independently about leaving the solar system. My main reason for this is that I doubt our civilization can last long enough. Not that I think we're going to self-destruct, but rather I think that nature is going to do it for us. It is my opinion that it is much more PROBABLE that we are driven into or close to extinction by an asteroid collision, than it is we will detect ETL or travel outside the solar system. This is one of the major reasons I strongly support construction of self-supporting Lunar and Martian colonies (and sky survey telescopes!) I just don't like us having all our eggs in the one basket called Earth. Having said all that, if we survive, I am confident we will eventually detect ETL, and that we will leave the solar system.
< Fling:Anonymous Protocol Suite | IIS Vs. Apache and Netcraft Results >