麻豆淫院


Machines that learn better

Machines that learn better
Cameron Freer, left, an instructor in pure mathematics; and Daniel Roy, right, a PhD student in the Department of Electrical Engineering and Computer Science. Photo: Jason Dorfman/CSAIL

(麻豆淫院Org.com) -- In the last 20 years or so, many of the key advances in artificial-intelligence research have come courtesy of machine learning, in which computers learn how to make predictions by looking for patterns in large collections of training data. A new approach called probabilistic programming makes it much easier to build machine-learning systems, but it鈥檚 useful for a relatively narrow set of problems. Now, MIT researchers have discovered how to extend the approach to a much larger class of problems, with implications for subjects as diverse as cognitive science, financial analysis and epidemiology.

Historically, building a machine-learning system capable of learning a new task would take a graduate student somewhere between a few weeks and several months, says Daniel Roy, a PhD student in the Department of Electrical Engineering and who along with Cameron Freer, an instructor in pure mathematics, led the new research. A handful of new, experimental, probabilistic programming languages 鈥 one of which, Church, was developed at MIT 鈥 promise to cut that time down to a matter of hours.

At the heart of each of these new languages is a so-called inference algorithm, which instructs a machine-learning system how to draw conclusions from the data it鈥檚 presented. The generality of the inference algorithm is what confers the languages鈥 power: The same algorithm has to be able to guide a system that鈥檚 learning how to recognize objects in digital images, or filter spam, or recommend DVDs based on past rentals, or whatever else an artificial-intelligence program may be called upon to do.

The inference algorithms currently used in probabilistic programming are great at handling discrete data but struggle with continuous data. For an idea of what that distinction means, consider three people of different heights. Their rank ordering, from tallest to shortest, is discrete: Each of them must be first, second, or third on the list. But their absolute heights are continuous. If the tallest person is 5 feet 10 inches tall, and the shortest is 5 feet 8 inches, you can鈥檛 conclude that the third person is 5 feet 9 inches: He or she could be 5 feet 8.5 inches, or 5 feet 9.6302 inches or an infinite number of other possibilities.

Designers of probabilistic programming languages are thus avidly interested in whether it鈥檚 possible to design a general-purpose inference algorithm that can handle continuous data. Unfortunately, the answer appears to be no: In a , Freer, Roy, and Nate Ackerman of the University of California, Berkeley, mathematically demonstrate that there are certain types of statistical problems involving continuous data that no general-purpose algorithm could solve.

But there鈥檚 good news as well: Last week, at the International Conference on Artificial Intelligence and Statistics, Roy presented a paper in which he and Freer not only demonstrate that there are large classes of problems involving continuous data that are susceptible to a general solution but also describe an inference algorithm that can handle them. A probabilistic that implemented the algorithm would enable the rapid development of a much larger variety of machine-learning systems. It would, for instance, enable systems to better employ an analytic tool called the P贸lya tree, which has been used to model stock prices, disease outbreaks, medical diagnoses, census data, and weather systems, among other things.

鈥淭he field of probabilistic programming is fairly new, and people have started coming up with probabilistic programs, but Dan and Cameron are really filling the theoretical gaps,鈥 says Zoubin Ghahramani, professor of information engineering at the University of Cambridge. The hope, Ghahramani says, 鈥渋s that their theoretical underpinnings will make the effort to come up with probabilistic programming languages much more solidly grounded.鈥

Chung-chieh Shan, a computer scientist at Rutgers who specializes in models of linguistic behavior, says that the MIT researchers鈥 work could be especially useful for artificial-intelligence systems whose future behavior is dependent on their past behavior. For instance, a system designed to understand spoken language might have to determine words鈥 parts of speech. If, in some context, it notices that a word tends to be used in an uncommon way 鈥 for instance, 鈥渕an鈥 is frequently used as a verb instead of a noun 鈥 then, going forward, it should have greater confidence in assigning that word its unusual interpretation.

Often, Shan explains, treating problems as having such 鈥渟erial dependency鈥 makes them easier to describe. But it also makes their solutions harder to calculate, because it requires keeping track of an ever-growing catalogue of past behaviors and revising future behaviors accordingly. Freer and Roy鈥檚 algorithm, he says, provides a way to convert problems that have serial dependency into problems that don鈥檛, which makes them easier to solve. 鈥淎 lot of models would call for this kind of picture,鈥 Shan says. Roy and Freer鈥檚 work 鈥渋s narrowing this gap between the intuitive description and the efficient implementation.鈥

While Freer and Roy鈥檚 algorithm is guaranteed to provide an answer to a range of previously intractable problems, Shan says, 鈥渢here鈥檚 a difference between coming up with the right algorithm and implementing it so that it runs fast enough on an actual computer.鈥 Roy and Freer agree, however, which is why they haven鈥檛 yet incorporated their algorithm into Church. 鈥淚t鈥檚 fairly clear that within the set of models that our algorithm can handle, there are some that could be arbitrarily slow,鈥 Roy says. 鈥淪o now we have to study additional structure. We know that it鈥檚 possible. But when is it efficient?鈥

Citation: Machines that learn better (2010, May 18) retrieved 17 May 2025 from /news/2010-05-machines.html
This document is subject to copyright. Apart from any fair dealing for the purpose of private study or research, no part may be reproduced without the written permission. The content is provided for information purposes only.

Explore further

A Grand Unified Theory of Artificial Intelligence

0 shares

Feedback to editors