忍者ブログ
Technical News
[1]  [2]  [3]  [4]  [5]  [6]  [7]  [8]  [9]  [10
×

[PR]上記の広告は3ヶ月以上新規記事投稿のないブログに表示されています。新しい記事を書く事で広告が消えます。

Quantum Teleporting, Yes; the Rest Is Movie Magic

In a battle waged with popcorn, floodlights, chalk and star power, science and art squared off at the
Massachusetts Institute of Technology one night last month.

MIND FLIGHT Hayden Christensen is a superhero on the move in “Jumper.”

On one side of a vaunted cultural divide were Doug Liman, director of the coming movie “Jumper,” about a young man who discovers he can transport himself anywhere he wants just by thinking about it, and Hayden Christensen, the film’s star.

On the other were a pair of the institute’s physics professors, Edward Farhi and Max Tegmark, experts on the type of physics the movie was purporting to portray, who had been enlisted to view a few scenes from it and talk about science.

In the middle were hundreds of M.I.T. students who had waited for hours to jam into a giant lecture hall known as Room 26-100 and who proved that future scientists and engineers could be just as rowdy and star-struck as the crowds outside the MTV studios in Times Square.

“I guess I wasn’t expecting such a lively group,” Mr. Christensen said.

The evening was the brainchild of Warren Betts, a veteran Hollywood publicist who has helped promote a number of movies with scientific or technological themes, including “Apollo 13.” Mr. Betts said he had gotten excited after a Caltech physicist told him that teleportation was actually an accomplished fact in the quirky realm of quantum physics.

Mr. Betts arranged for clips from the movie, scheduled for a Feb. 14 release, to be shown, and then inveigled Dr. Farhi, an expert on quantum computers, and Dr. Tegmark, a cosmologist, to participate in a panel discussion. They agreed, as long as they could talk about real physics.

“What do I know about movie production?” asked Dr. Farhi, calling himself “clueless.” He said, “If the students learn something, it’s fine, I’m happy.”

The corridor outside M.I.T.’s venerable lecture hall was transformed for the occasion into a red carpet — sans the actual red carpet — lined with television cameras and reporters. At the appointed hour, Mr. Christensen, who played the young Anakin Skywalker in “Star Wars Episode II: Attack of the Clones,” and “Star Wars Episode III: Revenge of the Sith,” began to proceed slowly down the line.

Mr. Liman, the director, meanwhile, confessed to being nervous. “We’re about to see a couple of M.I.T. professors rip me to shreds,” he said. “I hope they appreciate that I tried to respect the physics of the planet we live on.”

Mr. Liman, who directed “The Bourne Identity,” and “Mr. and Mrs. Smith,” said he had been a “physics prodigy” in high school, which had gotten him into Brown University despite a checkered adolescence. He never took a physics class in college, however. “Being good at it made it a little boring,” he said.

He said he had fallen in love with the “Jumper” script — adapted by David S. Goyer, Jim Uhls and Simon Kinberg from a series of young adult novels by Steven Gould — because of its honesty. The first thing the new superhero does with his powers is rob a bank. “The story was as honest as it could be,” Mr. Liman said.

He said he had spent a lot time trying to figure what teleportation would actually look like, never mind what causes it. If a body suddenly disappeared, for example, there would be a rush of air into the vacuum left behind.

Physics, Mr. Liman said, is more connected to filmmaking than one might expect. “I liked problem solving,” he said. “A film,” he added, “is one big problem.”

An hour later, Dr. Farhi and Dr. Tegmark, true to their words, let the air out of the “Jumper” balloon.

In real experiments recently, Dr. Farhi told the movie fans, physicists had managed to “teleport” a single elementary particle, a photon, which transmits light, about one and a half miles, “a little less exotic than what you see in the movie.”

What is actually teleported in these experiments, he explained, is not the particle itself but all the quantum information about the particle.

To accomplish this is no small matter. Among other things, the teleporters have to create a pair of so-called entangled particles, which maintain a kind of spooky correlation even though they are separated by light years. Both of them exist in a kind of quantum fog of possibility until one or the other is observed. Measuring one particle instantly affects its separated-at-birth twin no matter how far away. If one is found to be spinning clockwise, for example, the other will be found to be spinning counter clockwise.

In order to use this magic to “teleport” a third particle, Dr. Farhi emphasized, you have to send a conventional signal between the entangled twins, and that takes time, according to Einstein. “You cannot get that thing over there faster than the speed of light,” Dr. Farhi said, to cheers from the crowd.

The real lure, he said, is not transportation, but secure communication. If anybody eavesdrops on the teleportation signal, the whole thing doesn’t work, Dr. Farhi said. Another use is in quantum computing, which would exploit the ability of quantum bits of information to have different values, both one and zero, at the same time to perform certain calculations, like factoring large prime numbers, much faster than ordinary computers.

As Dr. Tegmark said, “Nobody can hack your credit card, and then you can build a quantum computer and hack everybody else’s card.”

One student asked the physicists if they rolled their eyes at the scientific miscues in movies. That was too much like work, protested Dr. Farhi, who said he was more interested in the acting and the characters. Dr. Tegmark said that even inaccurate science fiction movies could inspire scientists to think. You could see something that you think is impossible, he said, but that might start you thinking. “Why is that impossible? It can trigger a train of thought,” he said.

“The hard part of science is finding the right questions,” Dr. Tegmark said.

Asked if science mattered, Mr. Liman said that he always tried to get to know the reality behind a film, but that it was not always so easy. One professor he approached for advice about “Jumper” threw him out of his office, he said.

He went on to describe his attempts to portray the teleportation jumps realistically. Wind would rush to fill the vacuum left by the departing body, he said, and papers would fly around.

“Yeah,” Dr. Tegmark said.

Under some conditions moisture would condense out of the air into clouds.

The physicists nodded. “In any other place, I would sound very scientific,” Mr. Liman said, to laughter and applause.

By now the divide between the two cultures was getting as fuzzy and blurred as some quantum fog.

Dr. Tegmark asked what scientists could do to help the movie makers.

“Watch ‘Jumper,’ ” Mr. Christensen answered, “and then get to work and figure out how to do it.”

PR

New Electronics Promise Wireless at Warp Speed

Company uses nanoscale metals to build faster radios to wirelessly process video and other massive data files

 
PHIAR'D UP: Phiar's metal-insulator technology stacks metals and insulators at nanoscale thicknesses to enhance the performance and cut the costs of wireless networks, introducing a simpler, less expensive manufacturing process compared to silicon-based semiconductors.

 
METAL SANDWICH: Phiar's basic metal-double-insulator-metal (MIIM) diode includes sputter-deposited thin films of metals and their oxides that create quantum wells at the insulators' interface.

Wireless networking technology will one day deliver high-definition video content and other large data files via the airwaves far faster than that information can be now be delivered over wired systems. But it will take major advances in the electronics that drive computer and radio-frequency systems to create such a high-powered wireless highway.

One of the most basic examples of such a system is a laptop computer equipped with a radio for wireless connectivity. The computer's performance has generally been improved through upgrades in digital semiconductor performance: shrinking the size of the semiconductor's transistors to ramp up transaction speed, packing more of them onto the chip to increase processing power, and even substituting silicon with compounds such as gallium arsenide or indium phosphide, which allow electrons to move at a higher velocity.

The key to squeezing higher performance out of the radio side of the equation, according to one company, is using metal-insulator components. "We are potentially at another stepping point, where instead of solid-state semiconductor electronics, we will have metal-insulator electronics," says Garret Moddel, chief technology officer and chairman of Phiar Corporation in Boulder, Colo.

Moddel has good reason to believe this, given that his company builds diodes, radio-frequency (RF) detectors and RF receivers using metal-insulator technology.

Although Phiar's technology will not be commercially available until next year, the company's approach is expected to enhance the performance and cut the costs of wireless networks by introducing a simpler, less expensive manufacturing process. The company does this by using stacks of metals and insulators at nanoscale thicknesses—tens of angstroms, a unit of length equal to one ten-billionth of a meter; the space between atoms is generally two or three angstroms—to create high-frequency, up to three terahertz. (A terahertz is a trillion hertz, a thousand times speedier than a gigahertz-level processor, but slower than an optical network.)

"We're bridging the performance between photonics and electronics," says Adam Rentschler, Phiar's director of business development.

Conventional semiconductors are built using silicon-based substrates (the material upon which semiconductor devices are fabricated), but metal-insulator electronics can be made atop less pricey glass, metal or plastic substrates. Phiar's approach is to place two metal layers on either side of a double layer of insulation. When voltage is applied, electrons tunnel through the insulator layers with the help of a "quantum well" that forms between the two insulators.

Phiar will not specify which metals it uses, it is a trade secret, but says they are amorphous rather than crystalline, like silicon. This means these metals can be layered on top of a variety of other substances, including standard complementary metal–oxide semiconductor (CMOS) circuitry. As such, Phiar's metal-insulator diodes or other electrical components could be combined with semiconductors on the same microchip. Among other potential options: using metal-insulator radios to replace the copper chip-to-chip interconnect wires on today's printer circuit boards, eliminating one of computing's worst performance bottlenecks. In the more distant future, metal-insulator devices could even replace the digital transistors within a semiconductor.

Phiar's technology is expected to hit the market through a series of partnerships. Phiar and Motorola, Inc., last year signed a joint development agreement that could make Phiar's metal-insulator electronics an integral part of the 60-gigahertz mobile wireless high-definition multimedia interfaces and imaging technologies that Motorola is developing. Motorola has successfully incorporated Phiar's metal-insulator diodes into a 60-gigahertz prototype system and demonstrated multigigabit-per-second data rates.

The ability to operate at 60 gigahertz—where wavelengths are only a few millimeters—is crucial to the wireless personal area networks (PANs) that Motorola and other tech vendors are looking to offer. Whereas Wi-Fi signals—which operate at frequencies no greater than 5.8 gigahertz—can penetrate walls, glass and other barriers, enabling one wireless access point to serve an entire household, 60-gigahertz signals are more easily contained. This means a household could have one 60-gigahertz network in one room and an entirely different 60-gigahertz setup in the next room (hence the name "personal area" network). Operating in the wider bandwidth 60-gigahertz range also increases the speed of Wi-Fi data transfers.

"The seminal point about moving to higher frequencies is that there is more bandwidth available, which means it's faster," Moddel says. This translates into content that is delivered faster over wireless networks, which is important because the demand for large data files and video content is on the rise.

Phiar's technology is optimal for building radios because metal-insulator electronics can detect higher frequency signals than silicon-based semiconductors can. "Phiar's technology vastly bumps up the speed at which you can detect an incoming signal," Ian Lao, a senior analyst with In-Stat, a technology research firm based in Scottsdale, Ariz., and a division of tech publisher Reed Elsevier, Inc. This has many implications in the development of new sensors and telecommunications equipment that will be able to quickly detect and accurately read wireless signals.

This could be a useful component in a high-speed communications network, "the kind of thing that a wireless company would want for their base stations," says Michael Kozicki, an Arizona State University professor of electrical engineering and director of the school's Center for Applied Nanoionics. Phiar technology could become a central part of imaging systems such as airport security devices that rely on radiation moving at terahertz speeds to scan passengers for contraband, penetrating materials otherwise opaque to visible or infrared radiation.

Motorola and Phiar's approach is expected to face competition from a number of companies eager to take wireless communications beyond Wi-Fi speeds. IBM and Taiwanese semiconductor maker Mediatek, Inc., are developing their own very fast chip sets that can wirelessly transmit a full-length, high-definition movie to and from a home PC, handheld device, retail kiosk or television set. IBM and Mediatek plan to ultimately develop ways to wirelessly connect high-definition TVs to set-top boxes using millimeter-wave radio technology, which takes advantage of the highest frequency portion of the radio spectrum where massive amounts of information can be sent quickly.

The companies will integrate IBM's millimeter-wavelength radio chips, antenna and package technology with Mediatek's digital base band and video processing chips. IBM demonstrated a prototype packaged chip set as small as a dime to wirelessly transmit uncompressed HD video two years ago.

"Don't sell your stock in all of those silicon companies because this is not a replacement for silicon in any way, shape or form," Kozicki says. Although Phiar's metal-insulator technology could become a crucial component of future high-frequency imaging systems or communications networking equipment, it is not meant to perform the kind of processing that silicon does so well.

Super Tuesday: Markets Predict Outcome Better Than Polls

Internet-based financial markets appear to forecast elections better than polls do. They also probe how well the next George Clooney drama will do at the box office and how bad the next flu season will be.

In late March 1988 three economists from the University of Iowa were nursing beers at a local hangout in Iowa City, when conversation turned to the news of the day. Jesse Jackson had captured 55 percent of the votes in the Michigan Democratic caucuses, an outcome that the polls had failed to intimate. The ensuing grumbling about the unreliability of polls sparked the germ of an idea. At the time, experimental economics—in which economic theory is tested by observing the behavior of groups, usually in a classroom setting—had just come into vogue, which prompted the three drinking partners to deliberate about whether a market might do better than the polls.

A market in political candidates would serve as a novel way to test an economic theory asserting that all information about a security is reflected in its price. For a stock or other financial security, the price summarizes, among other things, what traders know about the factors influencing whether a company will achieve its profit goals in the coming quarter or whether sales may plummet. Instead of recruiting students to imitate “buyers” or “sellers” of goods and services, as in other economics experiments, participants in this election market would trade contracts that would provide payoffs depending on what percentage of the vote George H. W. Bush, Michael Dukakis or other candidates received.

If the efficient-market hypothesis, as the theory relating to securities is known, applied to contracts on political candidates as well as shares of General Electric, it might serve as a tool for discerning who was leading or trailing during a political campaign. Maybe an election market could have foretold Jackson’s win. Those beer-fueled musings appear to have produced one of the most notable successes in experimental economics—and have blossomed into a subdiscipline devoted to studying prediction markets that allow investing or betting (pick the term you like best) not just on elections but on the future of climate change, movie box-office receipts and the next U.S. military incursion.

Make Your Best Bet
When the three academics—George R. Neumann, Robert Forsythe and Forrest Nelson—sought support from the university, the dean of its business college, a free-market advocate, could not contain his enthusiasm. On the other hand, the dean of the college of arts and sciences, a political scientist, characterized the proposal as “the stupidest thing he had ever heard of,” Neumann recalls. “At best, it would be a shadow of the polls,” he was told.

With the business school dean onboard, the three pressed forward. They wanted to use real money as an incentive for participants to take the exercise seriously. But they needed permission to allow students and faculty to gamble legally on campus. The university’s general counsel resisted, but Iowa’s state attorney general let the real-money market go ahead under a state law that permits office-betting pools.

The World Wide Web was still a glint in the eye of Tim Berners-Lee when the Iowa Political Stock Market opened on June 1, 1988. Nearly 200 students and faculty members began buying contracts on George H. W. Bush, Dukakis and others using the relatively primitive tools of the pre-Web Internet. A Bush or Dukakis contract was bought or sold in a futures market, the same type in which Iowa hog farmers trade pork bellies. Instead of pigs, however, the investors in the Iowa Political Stock Market were trading contracts on the share of the vote that a candidate would receive on Election Day.

Up until the morning of the election, traders carried out their transactions, although a rule stipulated that no one could invest more than $500. Taking a simplified example, a Bush contract in the vote-share market paid $0.53, corresponding to Bush’s 53 percent of the vote, and a Dukakis contract paid $0.45, tied to the Democrat’s popular vote percentage. If you had bought a Bush security at $0.50 before the market closed the morning of the election, you would have made a gain of $0.03.

To the three economists, finding out who won or lost money—or the election—was less important than whether this exercise answered the question posed in the barroom: Would the expected share of the votes represented by the market’s closing prices on Election Day match the actual share the candidates obtained more closely than the polls would? The experiment worked. The final market price corresponded to Bush’s and Dukakis’s market shares better than Gallup, Harris, CBS/New York Times and three other major polls.

In 1992 the Iowa Political Stock Market was redubbed the Iowa Electronic Markets (IEM), and trading was opened to anyone from Dubuque to Beijing who could come up with the requisite minimum of $5. The Commodity Futures Trading Commission (CFTC) had granted the University of Iowa an exemption from regulation because the IEM is mainly run for research purposes (only minor sums are transacted).

The election exchange has continued to beat the polls consistently for presidential elections and at times has prevailed in congressional and international races. A paper being prepared for publication by several Iowa professors compares the performance of the IEM as a predictor of presidential elections from 1988 to 2004 with 964 polls over that same period and shows that the market was closer to the outcome of an election 74 percent of the time. The market, moreover, does better than the polls at predicting the outcome not just around Election Day but as long as 100 days before.

The IEM will never be the New York Stock Exchange. But even with the CFTC trading restrictions, it has flourished. The number of contracts traded expanded from 15,286 in 1988 (a dollar volume of $8,123) to 339,222 ($46,237) in the 2004 elections. And another IEM market that furnishes a payoff only to those who picked the winner of an election had even more activity in 2004 (1,106,722 contracts totaling $327,385). Television commentators have recognized this new barometer of voter sentiment by sometimes mentioning market prices in the months running up to an election. The IEM’s status has risen among those who contribute to the incessant blog-based chatter that has become a cornerstone of contemporary political discourse. And after a spike in trading during the 2004 election, the IEM office received e-mails charging that über-financier George Soros was trying to manipulate the market to create a bandwagon effect for Democratic presidential candidate John Kerry, an assertion for which there was never any proof.

The How and Why
The IEM continues to serve not only as a forecasting tool but as an energizing environment for students to learn about markets and, perhaps most important, as a testing ground for experimental economists to probe theories of how and why markets appear to make accurate predictions. Its track record provides arguably the best empirical evidence to date to justify the case for prediction markets. But when researchers have tried to backtrack, looking for theories of why markets serve as effective means of forecasting, straightforward answers have not been forthcoming. Some of these analyses have even called into question the basic assumption that a market does a good job of foretelling what lies ahead.

At first, the idea that a market can prophesize the outcome of an election does not seem particularly startling. After all, the chairman of the Federal Reserve or the chief economist at Goldman Sachs will routinely look at the price of stocks or commodities as a guide to making forecasts about the economy, and the futures market for orange juice concentrate predicts Florida weather better than the National Weather Service does.

Developers of the IEM and other prediction markets contrast a poll with a market by saying that the latter takes a reading not of whom people are going to vote for but of whom they think will win—and cash wagered indicates the strength of those beliefs. You might have voted for Kerry in the 2004 election because you opposed the Iraq War, but after watching news shows and talking to neighbors, you may have decided that George W. Bush was going to win. When putting money down, you might have picked Bush.

The question, though, of how one individual’s belief—that IBM’s stock will rise or that a Bush will be elected—gets combined with those of every other trader and then translated into a price that is an accurate predictor continues to provoke heated debates in the research community. Economic theoreticians have yet to understand precisely why this novel means of forecasting elections should work better than well-tested social science methods.

On close inspection, the characteristics of IEM traders would drive a statistician batty. Early on it became clear that the traders are by no means a representative sampling of the population at large, the prerequisite for any poll. And a survey of them in the 2004 presidential election market underscored the point: most were found to be well educated, affluent, white, male Republicans who tended to have a high opinion of their own political insight into the face-off between Bush and Kerry, a grouping that does not fit the definition of a well-designed sample. In about one in five transactions, traders had no personal opinions or beliefs at all about the Swift Boat smear campaign or prisoners being held in Guantánamo. Rather those buying or selling were “robots”—automated trading programs that buy and sell when the software perceives that a security is too high or low. Automated programs routinely execute trades on Wall Street. And IEM election market researchers are still plumbing what a machine’s trading patterns add to the market’s ability to deduce the outcome of an election.

As early as the aftermath of the 1988 presidential race, the Iowa team began to probe deeply into why the IEM seems to predict election outcomes with such precision. Discounting pure luck and the possibility that traders somehow constitute a representative sample of the population, the team analyzed trading patterns and found a select group of “marginal traders” who would buy and sell actively when the share price was not valued properly. This group might have bought, say, Bush securities if the price was way under what the members thought was the likely percentage of votes the Republican would attract.

These traders were the Warren Buffetts of the 1988 race, investing an average of $56, twice the level of less active participants who might have simply bought and held contracts for the candidate they liked best, without making a careful judgment about that candidate’s prospects. The wallflowers would typically make nothing from their trades, whereas marginal traders took home 9.6 percent returns (a whopping $5.38; the reason such small sums act as an incentive to traders—or the use of play money in other markets—is also closely studied).

The identification of marginal traders, described in a 1992 paper in the American Economic Review, has sometimes elicited phone calls from Wall Street types interested in new insight into the perennial question of the traits of a person who can beat the market. Other than noting that most of those investing are male, the Iowa researchers have not succeeded in identifying more specific qualities of this special class of trader.

One possibility is that they do not exist. James Surowiecki, a New Yorker columnist who wrote The Wisdom of Crowds, a book first published in 2004 that brought attention to prediction markets and other novel means of group decision making, thinks that the marginal trader is a myth. No individual or subgroup in a market has the financial wherewithal to sway prices in the way the marginal-trader hypothesis suggests—an opinion that is echoed by some economists.

Just a Word Argument
Perhaps the most incisive critique of prediction markets has come from Charles F. Manski, an econometrician at Northwestern University whose academic research focuses on how people assign probabilities to future events, such as the possibility that they might lose their job. Manski started wondering a few years ago about the theoretical basis for statements made repeatedly in the popular press that markets can predict an election better than polls and experts can.

Advocates of prediction markets often invoke Austrian-born economist Friedrich Hayek, who argued in 1945 that prices aggregate information held by a group—“dispersed bits of incomplete and frequently contradictory knowledge which all the separate individuals possess.” That knowledge is combined into a price that expresses the relative desirability of a commodity or public sentiment at a given moment, whether it be a pork belly or a candidate for the U.S. presidency. Manski went back to Hayek’s original work to examine the quantitative underpinnings of his ideas. No hard numbers supported the notion of the collective wisdom of crowds. “It’s a very loose argument,” he says. “There’s no theory in the modern sense of the word. It’s just a word argument.”

So Manski set out to explore whether he could build a mathematical model that would confirm Hayek’s notion of the market as an information aggregation mechanism and, secondarily, bolster the empirical findings taken from the IEM. Manski created a model of a diverse group of traders using the IEM’s winner-take-all market in which a trader buys a contract for a candidate that pays $1 for a victory and nothing for a loss. If the market worked in accordance with the way that proponents of prediction markets have interpreted Hayek, the price would represent the average, or mean, value of traders’ belief that a particular candidate would win. A Kerry contract selling for $0.49 would mean that there would be a 49 percent probability that Kerry would win.

But Manski’s model did not confirm this conjecture. In many instances, the mean did not necessarily coincide with the price and could even diverge sharply, a finding suggesting that the market would not serve as a particularly accurate prediction tool. If, for instance, the price was $0.50, the mean of traders’ beliefs could be anything from a 25 to 75 percent chance that Kerry would win. Manski remarks that even if the price and the mean were the same, it would not be certain that the mean would correspond to a reasonable probability of a candidate’s chances.

Manski is a respected economist, and his finding caused a minor furor because it appeared to contradict an emerging consensus about the value of these markets for making predictions about anything from elections to public policy. But two subsequent papers offered a way to reconcile the dispute. They also compared prices with the mean but factored in a variable called risk aversion—which measures how traders react to uncertainty in the market. In the revised model, said by the authors to offer a more realistic scenario, the price and the mean were about the same, which seemed to confirm that a price is, indeed, a good measure of a probability.

But the debate has never been resolved, and exactly how the markets achieve success remains unclear. Manski, for his part, suspects that his critics’ models do not account for all the actual ways prediction markets operate in the real world. “There isn’t going to be a simple inter­pretation of that market price that always works as a prediction,” he observes. “It really depends on the beliefs and the attitudes toward risk of those trading.” Manski also remains unsatisfied with the IEM’s proponents’ reliance on its record of consistently besting the polls. “Comparison to the polls is not the best comparison,” he says. “Everyone knows there are all kinds of problems with the polls, and they’re just one piece of information.” In fact, Manski notes, IEM traders may be taking the polls into account as one of many factors in making decisions about when to buy or sell.

Oft-cited statistics about election markets beating the polls have come under scrutiny from other quarters. A 2005 analysis by political scientists Robert S. Erikson of Columbia University and Christopher Wlezien of Temple University insisted that polls and election markets do not serve the same functions and so do not merit direct comparison. The authors contended that the polls identify vote preferences on the day each poll is taken, whereas the IEM market prices forecast what is to happen on the day of the election. In their analysis, they made a series of mathematical adjustments to the polls, which they then found to be more accurate in projecting Election Day outcomes than both the IEM’s vote-share and winner-take-all markets.

Controversy again ensued. One dissenter, Justin Wolfers, an economist at the Wharton School of the University of Pennsylvania who has done extensive analyses of prediction markets, criticized Erikson and Wlezien’s results, saying that their study only compared a few elections and polls. Wolfers also objects because the 2005 analysis “adjusts polls but doesn’t make a corresponding adjustment of prediction markets.”

The Triumph of the Market
It will take years to put these debates to rest. In spite of persistent wrangling, the IEM has inspired formation of other prediction markets, many of them outside an academic setting. On the Hollywood Stock Exchange, traders speculate on box-office sales for new movies. NewsFutures trades in current events. Some markets allow traders to buy and sell securities on the prospects for new ideas or technologies. Without the CFTC exemption accorded to the IEM, other U.S. markets use virtual play money on the Internet. In Ireland, which lacks similar restrictions, TradeSports and Intrade, both part of the same company, accept real cash for trading on sports, elections or other events. Intrade, for instance, provides a contract that will furnish a payoff if the U.S. or Israel executes an air strike against Iran by March 31. Another contract will provide recompense if the U.S. economy goes into recession during 2008.

The place accorded markets in U.S. society, along with the revolution in new forms of information sharing afforded by the World Wide Web, has meant that prediction markets are now being increasingly adopted as innovative decision-making tools in both government and private institutions. The ardor for market-based answers can at times border on the hyperbolic. Robin Hanson, a professor of economics at George Mason University, has advocated that if trading patterns on prediction markets suggest that implementation of a particular policy will cause the economy to grow and unemployment to shrink, then policy officials should, by fiat, adopt that policy—an interest rate cut or a public works project, perhaps. Hanson reasons that the collective information held by traders is superior to the analyses that can be marshaled by a panel of economists or other experts. Hanson has even proposed a form of government called futarchy, based on policy-making markets.

Such utopian leanings have sometimes led advocates to push too far too fast. Several years ago the Defense Advanced Research Projects Agency (­DARPA) began planning for a project called the Policy Analysis Market, which would have allowed investors to trade on geopolitical events, not unlike the Intrade Iran contract, including assassinations, wars and the next al-Qaeda attack. If the market—for which Hanson was an adviser—bid up a contract that would pay off if a terrorist attack occurred, the Department of Homeland Security might then decide to raise the threat condition status from yellow to red.

Or so went the rationale. The idea of a “terrorist futures market” repulsed many in Washington, and the market died quickly, even forcing the resignation of DARPA head John Poin­dexter (but not before TradeSports launched a market to speculate on the prospects of his ouster). Senator Barbara Boxer of California fumed when she learned about the Policy Analysis Market: “There is something very sick about it.”

But not everyone experienced the same distaste. Some argued that a prediction market able to serve as an efficient intelligence-gathering mechanism just might avert a pending crisis. Writing in the Washington Post, Wolfers and his colleague Eric Zitzewitz speculated that a contract on whether Niger had made a sale of uranium to Saddam Hussein would have been trading at low levels in early 2003, reflecting the actual intelligence consensus that the transaction never occurred and thereby undercutting one of the Bush administration’s rationales for going to war in Iraq.

The attacks on the Policy Analysis Market ultimately doomed the project, although the hoopla managed to boost public awareness of prediction markets. DARPA’s project became an informal tutorial that broadened public awareness of prediction markets. “It actually took the DARPA thing to get people’s attention,” comments Joyce Berg, a professor of accounting and IEM’s interim director.

New types of markets intended to assist in formulating government or internal corporate decision making have continued to emerge. Here again the University of Iowa has been a leader. Its markets for predicting influenza outbreaks serve as an example. In one, which ran for seven months, beginning in mid-September 2004, an IEM spinoff sold influenza futures contracts to a set of 62 health care professionals in Iowa to predict influenza activity for each week of the flu season. If a contract for the third week of January accurately forecast flu prevalence—gauged by a Centers for Disease Control and Prevention scale (ranked as no activity, sporadic, local, regional or widespread)—it would pay $1. The market accurately predicted the beginning, the peak and the end of the influenza season two to four weeks ahead of the CDC reports on influenza activity.

“Prediction markets will never replace traditional surveillance systems, but they may provide an efficient and relatively inexpensive source of information to supplement existing disease surveillance systems,” says Philip M. Polgreen, a physician and professor at the University of Iowa’s Carver College of Medicine, who helped to run the market. The university has more recently begun a market, in collaboration with Pro-MED mail, an electronic disease-reporting system, that is intended to predict events related to the H5N1 “bird flu” virus.

Attracted by the markets’ apparent soothsaying powers, companies such as Hewlett-Packard (HP), Google and Microsoft have established internal markets that allow employees to trade on the prospect of meeting a quarterly sales goal or a deadline for release of a new software product. As in other types of prediction markets, traders frequently seem to do better than the internal forecasts do.

HP has refined the running of prediction markets to make them effective for groups that might be too small to make accurate predictions. Before a market is launched, HP gauges the expertise level of participants and their attitude toward risk—factors that are then used to mathematically adjust the predictions made when participants place their bets on some future outcome. “Our mechanism basically distills the wisdom of the crowd from a very small group,” says Bernardo Huberman, director of the social computing laboratory at HP. This filtering process achieves better results than does a market alone or the predictions of the most knowledgeable members of the group.

The burgeoning interest in prediction markets evokes the prepoll era of the early 20th century, when betting on election results was ubiquitous. Newspapers would routinely run stories on the odds for a particular candidate, reports that often proved to be surprisingly prescient. In that sense, prediction markets may truly hark back to the future. “My long-run prediction is that newspapers in 2020 will look like newspapers in 1920,” Wharton School’s Wolfers says. If that happens, the wisdom of crowds will have arrived at a juncture that truly rivals the musings of the most seasoned pundits.

Lasers Make Other Metals Look Like Gold 

 
From left, aluminum turned a gold color, titanium turned to blue, and platinum turned gold. 

All that glitters golden is not gold. It could be aluminum. Or tungsten. Or another metal of Chunlei Guo’s choosing.

In a feat of optical alchemy, Dr. Guo, a professor of optics at the University of Rochester, and Anatoliy Y. Vorobyev, a postdoctoral researcher, use ultrashort laser bursts to pockmark the surface of a metal in a way that is not perceptible to the touch — it still feels smooth to the finger — but that alters how the metal absorbs and reflects light.

The result is that pure aluminum looks like gold, and the appearance is literally skin deep.

“I cannot tell it’s not gold,” Dr. Guo said. “It looks very pretty.”

Dr. Guo and Dr. Vorobyev reported their findings in the journal Applied Physics Letters published online Thursday.

The golden aluminum follows work a little more than year ago where Drs. Guo and Vorobyev reported that they could make gold and other metals look black — indeed a black that is blacker than the usual black, sucking up almost all light that impinged upon it.

The laser bursts — each lasting only about 60 millionths of a billionth of a second — melt and vaporize metal atoms near the surface, which then reassemble in minuscule structures including pits, spheres and rods that are a fraction of a millionth of a meter in size.

By changing the length, strength and number of pulses, the researchers found they could vary the resulting color.

In some cases, the change causes the structures to absorb a range of colors so that they cannot be seen. But the colors that are not absorbed are still reflected, and thus visible, resulting in gold aluminum or dark blue tungsten.

In other cases, the laser pulses create a periodic array of structures that cause the reflected light to interact and interfere with itself, producing an iridescent, shimmering rainbow — much like some butterfly wings, Dr. Guo said.

Dr. Guo imagines a kaleidoscope of potential uses, from the practical (a reflective filter) to the whimsical (etching the family photograph onto a metal refrigerator door, for instance). Another possibility is custom colors for bicycles or cars, without the need for any paint.

“It’s pretty robust, because it’s right on the metal, and it’s not going to peel off,” Dr. Guo said.

He cannot yet make all metals into all colors but says he believes that it is only a matter of trial-and-error to find the right recipe for each permutation.

With his black metal finding, Dr. Guo suggested the possibility of black gold rings. He was surprised when jewelers started calling. “They are actually indeed interested in making colored jewelry,” he said.

In the new article, he suggests a blue gold ring, perhaps a blue to match the eyes of a fiancé. 

Megavoltage CT Imaging Unlocks Fossil Mysteries

University of Wisconsin professor tests the proficiency of cancer-care computerized tomography on geologic finds

 
BURIED AMMONOID: As viewed through megavoltage computerized tomography (CT), a 370-million-year-old ammonoid from the Devonian period, an extinct cephalopod relative of squids, octopuses and the chambered nautilus. Ammonoids died out around the same time as dinosaurs, about 65 million years ago, at the Cretaceous–Tertiary (K–T) extinction event. The TomoTherapy machine's four-million-electron-volt x-rays are energetic enough to penetrate rock, as this image reveals.

 
ANOTHER AMMONOID?: Although this specimen was obtained under the pretense of being another Devonian ammonoid, the CT scan reveals smooth and uniformly concave septa, or dividing walls, that separate the individual chambers, whereas ammonoids typically had slightly ruffled septa. Additionally, the scan shows a siphuncle, a tube that allowed organisms to adjust buoyancy by increasing or decreasing the amount of gas within the chambers of its shell. The siphuncle was located on the shell periphery in ammonoids but pierced the center of the septum in nautiloids, indicating that this fossil is also around 370 million years old, but a nautiloid.

 
LIMESTONE SAMPLE BEFORE: Is this ordinary round chunk of limestone solid or hollow? Density measurements showed it to be a bit light, suggesting that it might be hollow (a geode). It is well known, however, that limestone comes in three distinct densities. Thus, the density alone cannot prove whether this is solid or hollow.

 
LIMESTONE SAMPLE AFTER: Megavoltage CT scanning reveals that this rock is indeed a geode.

 
SANDSTONE MYSTERY: This megavoltage CT image of a slab of sandstone reveals some imperfections within the rock.

 
SANDSTONE MYSTERY SOLVED: Upon chipping sandstone away, a shark tooth was discovered within the stony matrix.

Using a novel radiotherapy technology called helical tomotherapy—in essence, the marriage of a computerized tomography (CT) scanner and a radiotherapy linear accelerator—James Welsh, associate professor of medical physics and human oncology at the University of Wisconsin School of Medicine and Public Health, and a group of colleagues have created images of fossil specimens of various types and ages.

TomoTherapy, Inc.'s Hi-Art radiation machine, developed at the University of Wisconsin–Madison, is designed to treat cancer patients with intensity-modulated radiation therapy (IMRT). The CT component provides images of tumor position right before each daily treatment session so that doctors can target malignancies while avoiding healthy cells. The Hi-Art device is designed to rotate around a patient, delivering narrow, potent and precise doses of radiation to tumors without harming healthy tissue nearby—a process that takes 15 to 25 minutes. "I was fortunate to be the first MD to ever treat a patient with this technology," Welsh says. "I suspect these are the world's first megavoltage CT images of rocks and fossils from such a unit."

The use of image-guided IMRT technology has shown promise in treating cancer, and TomoTherapy is expected to soon have competitors. Sharp Grossmont Hospital in La Mesa, Calif., offers patients access to a Hi-Art system, but there are only about 30 medical facilities in the country using helical tomotherapy. This is expected to change after Varian Medical Systems, Inc., in Palo Alto, Calif., gets Food and Drug Administration approval to sell its RapidArc IMRT system. Elekta, AB, in Sweden also offers comparable image-guided radiation therapy technology.

Research is currently being conducted on other possible uses for these devices. "I plan to obtain more images and continue a scientific evaluation of the potential of this clinical machine for paleontological and geological purposes," Welsh says. "I expect a peer-reviewed manuscript will be ready by midyear."

Calendar
03 2024/04 05
S M T W T F S
1 2 3 4 5 6
7 8 9 10 11 12 13
14 15 16 17 18 19 20
21 22 23 24 25 26 27
28 29 30
Timepiece
タグホイヤー フォーミュラー1 ドリームキャンペーン
Blog Plus
SEO / RSS
Podcast
by PODCAST-BP
New TB
Bar Code
Data Retrieval
Oldest Articles
(09/30)
(09/30)
(09/30)
(09/30)
(09/30)
Photo Index
Reference
Latina




RSS Reader
無料RSSブログパーツ
Misc.
◆BBS


◆Chat


◆Micro TV


Maps



顔文字教室




Copyright © Tech All Rights Reserved.
Powered by NinjaBlog
Graphics by 写真素材Kun * Material by Gingham * Template by Kaie
忍者ブログ [PR]