Skip to main content

Computer scientists at Stanford University and Google have created technology that can track time down to 100-billionths of a second. It could be just what Wall Street is looking for.

System engineers at Nasdaq, the New York-based stock exchange, recently began testing an algorithm and software they hope can synchronize a giant network of computers with that nanosecond precision. They say they have built a prototype and are in the process of deploying a bigger version.

For an exchange such as Nasdaq, such refinement is essential to accurately order the millions of trades that are placed on its computer systems every second.

Open this photo in gallery:

Driven by the needs of the financial industry, researchers are working on ways to manage vast computer systems down to 100 billionths of a second.MINH UONG/The New York Times News Service

“The financial industry has easily become the most obsessed with time,” said Balaji Prabhakar, a Stanford electrical engineer who is one of the designers of the new synchronization system.

Because the orders are placed from locations around the world, they frequently arrive at the exchange’s computers out of sequence. The new system allows each computer to time stamp an order when it takes place.

As a result, the trades can be sorted and executed in correct sequence. In a networked marketplace, this precision is necessary not only to prevent illicit trading on advance information known as “front-running,” but also to ensure the fair placement of orders.

The importance of technical advances in measuring time was underscored by European regulations that went into effect in January and that require financial institutions to synchronize time-stamped trades with microsecond accuracy.

Being able to trade at the nanosecond level is vital to Nasdaq. Two years ago, it unveiled the Nasdaq Financial Framework, a software system that it has envisioned eventually trading everything from stocks and bonds to fish and car-sharing rides.

The new synchronization system will make it possible for Nasdaq to offer “pop-up” electronic markets on short notice anywhere in the world, Mr. Prabhakar said. He cited the World Cup as a hypothetical example of a short-term electronic marketplace.

“There are tickets needed, housing, people will need transportation,” he said. “Think of an electronic market almost like a massive flea market hosted by Nasdaq software.”

To go from trading equities to managing all sorts of financial transactions will require more than an order of magnitude speed-up in the company’s networks of computers. It will be possible only if all of the exchange’s computers agree on time with nanosecond accuracy.

A generation ago, computing usually took place in a single mainframe or personal computer. Now it is routinely spread across thousands of independent processors in machines that can be separated by a few metres or entire continents.

Chip designers have long struggled to maintain the precise timing needed to order mathematical operations inside individual computing chips. And synchronizing these vast ensembles of them has become the limiting factor in the speed and processing power of what Google describes as “planetary-scale” computers.

“It’s kind of mind-boggling,” said Peter Hochschild, a Google software engineer who specializes in the challenges associated with spreading software and data across networked computers. “Inside a processor, an enormous amount of stuff happens in a billionth of a second.”

A billionth of a second is roughly the time it takes light to travel one foot. It has long been viewed as a crucial measure in computing. In the 1960s, computing pioneer Grace Murray Hopper would hand out 11.8-inch lengths of wire to illustrate how designing smaller electronic parts would create faster computers.

Distance has become even more significant as software has begun to escape the boundaries of individual computers and make its way into the cloud – the web of giant computer data centres that have come to blanket the planet.

They are near dams to take advantage of cheap hydroelectric power and in cold climates to save on cooling costs. Microsoft has even begun submerging them in the ocean to take advantage of power generated by tidal surges.

Because software and data are no longer in the same place, correctly calculating the order of the events that may be separated by short or long distances has become the dominant factor in the speed with which data can be processed.

“So much of our expectation about computing being correct depends essentially on knowing this order,” said Krishna Palem, a theoretical computer scientist at Rice University.

In the world of cloud computing, entire databases are scattered among different computers and data centres.

That has created tremendous challenges for the designers of electronic commerce systems. The new software synchronization standard under which Nasdaq’s system would work, known as Huygens, is intended to replace a 33-year-old Network Time Protocol, or NTP, as well as more expensive approaches that have relied on atomic clocks and global positioning satellites.

Huygens, named for Dutch physicist Christiaan Huygens, who invented the pendulum clock in 1656, uses machine-learning techniques to synchronize a network of computers to within 100-billionths of a second. In contrast, the NTP standard can synchronize computers no more accurately than a millisecond, or one one-thousandth of a second.

To ensure that buyers and sellers are treated fairly, Nasdaq has for decades looked for ways to ensure that trades are processed in the order they are placed.

While building a network for Nasdaq in the 1990s, Brian Reid, a computer scientist at Digital Equipment Corp., experimented by coiling large rolls of cables of different lengths in a Massachusetts warehouse in order to insert tiny delays in the time it took data to travel in a network to make sure that messages were delivered fairly. He then employed timing information from satellites to synchronize clocks at different locations.

Google would later use this method to synchronize computers based on GPS data and atomic clocks to make sure its database system could correctly order transactions. But since the system requires super-accurate clocks and satellite receivers, it is more costly than the software-based Huygens approach.

Mr. Reid built his original system in an era when the Securities and Exchange Commission required that all stock sales be entered by humans.

“Five-millisecond accuracy in clock synchronization pleased everyone,” he said. “It took much longer than five milliseconds to press the ‘enter’ key on the big green terminals that people used.”

New York Times News Service

Interact with The Globe