Researchers at North Carolina State University have developed a new approach to software development that will allow common computer programs to run up to 20 percent faster and possibly incorporate new security measures, the university reports.
The researchers have found a way to run different parts of some programs—including, for the first time, such widely used programs as word processors and web browsers—at the same time, which makes the programs operate more efficiently.
To understand how the new technique works, first you have to understand how computers operate today.
The brain of a computer chip is its central processing unit, or “core.” Computing technology has advanced to the point where it’s now common to have between four and eight cores on each chip.
But for a program to use these cores, it has to be broken down into separate “threads,” so each core can execute a different part of the program simultaneously. The process of breaking down a program into threads is called parallelization, and it allows computers to run programs very quickly.
But some programs are difficult to parallelize, including word processors and web browsers. These programs operate much like a flow chart, with certain program elements dependent on the outcome of others. These programs can only use one core at a time, minimizing the benefit of multi-core chips.
Now, NC State researchers have developed a technique that allows these hard-to-parallelize applications to run in parallel, by using nontraditional approaches to break programs into threads.
Every computer program consists of multiple steps. The program will perform a computation, then perform a memory-management function—which prepares memory storage to contain data or frees up memory storage that is currently in use.
The program repeats these steps over and over again, in a cycle. For hard-to-parallelize programs, both of these steps traditionally have been performed in a single core.
“We’ve removed the memory-management step from the process, running it as a separate thread,” says Yan Solihin, an associate professor of electrical and computer engineering at NC State. Solihin is the director of this research project and co-author of a paper describing the research. Under this approach, he said, the computation thread and memory-management thread are executing simultaneously, allowing the computer program to operate more efficiently.
“By running the memory-management functions on a separate thread, these hard-to-parallelize programs can operate approximately 20 percent faster,” Solihin said. “This also opens the door to development of new memory-management functions that could identify anomalies in program behavior, or perform additional security checks. Previously, these functions would have been unduly time-consuming, slowing down the speed of the overall program.”
Using the new technique, when a memory-management function needs to be performed, “the computational thread notifies the memory-management thread—effectively telling it to allocate data storage and to notify the computational thread of where the storage space is located,” said Devesh Tiwari, a Ph.D. student at NC State and lead author of the paper.
“By the same token, when the computational thread no longer needs certain data, it informs the memory-management thread that the relevant storage space can be freed.”
The paper, “MMT: Exploiting Fine-Grained Parallelism in Dynamic Memory Management,” will be presented April 21 at the IEEE International Parallel and Distributed Processing Symposium in Atlanta.
Link:
Memory Management Thread research
- ‘Buyer’s remorse’ dogging Common Core rollout - October 30, 2014
- Calif. law targets social media monitoring of students - October 2, 2014
- Elementary world language instruction - September 25, 2014