wrote the paper with his current Ph.D. students, Mahesh Nanjundappa and Bijoy A. Jose, also of Virginia Tech, and a past Ph.D. advisee Hiren D. Patel who is now an ECE assistant professor at the University of Waterloo in Canada.
Shukla and his collaborators said that they were able to demonstrate how to speed up the simulation performance of certain SystemC based hardware models “by exploiting the high degree of parallelism afforded by today’s general purpose graphic processor units (GPGPU).” These units have multiple core processors capable of very high computation and data throughput. When parallelism is applied, it means that the processor units can run various parts of the simulations simultaneously, and not just as a single sequence of computations. Their experiments were carried out on an NVIDIA Tesla 870 with 256 processing cores. This equipment was donated to Shukla’s lab by NVIDIA during fall 2008.
Shukla said their preliminary experiments showed they were able to speed up SystemC based simulation by factors of 30 to 100 times that of previous performances.
They named their simulation infrastructure SCGPSim. The Air Force Office of Scientific Research and the National Science Foundation helped support this research.
In the past, Shukla said, “significant effort was aimed at improving the performance of SystemC simulations, but little had been directed at making them operate in parallel. And none of the attempts were ever targeted at a massively parallel platform such as a general purpose graphic processor unit.”
Another aspect of their work was the use of a specific programming model called Compute Unified Device Architecture (CUDA). It is an extension to the C software language that “exploits the processing power of graphic processor units to solve complex compute-intensive problems efficiently,” Shukla explained. “High performance is achieved by launching a number of threads and making each thread execute a part of the application in parallel.”
The CUDA execution model differs from the more commonly known central processing unit (CPU) based execution in terms of how the threads are scheduled. With CUDA, it is possible to have all of the threads execute simultaneously on separate processor cores and intermittently converge on the same path, thus increasing the efficiency.
The work at Virginia Tech was conducted in the Formal Engineering Research with Models, Abstractions and Transformations (FERMAT) Laboratory, founded by Shukla in 2002. Its focus is in designing, analyzing and predicting performance of electronic systems, particularly systems embedded in automated systems. http://www.fermat.ece.vt.edu/
“Speeding up simulation of complex hardware models is extremely important for semiconductor electronics industry to producer newer and newer products in shorter times, thus improving the quality of computing and consumer electronics products faster. If such models can be simulated 10 times faster, then if validating a model took 10 days in the past, now it would take one day. This is why faster simulation performance probably attracted the attention of the ASP-DAC ’10 awards committee.” Shukla said.
ASP-DAC is one of the three conferences sponsored by IEEE Circuits and Systems Society, and ACM Special Interests Group on Design Automation, on the topic of electronics design automation. These three conferences are held every year in the US (DAC) , in Europe (DATE) and in the Asia-pacific region (ASP-DAC).
Virginia Tech’s College of Engineering is internationally recognized for its excellence in 14 engineering disciplines and computer science. As the nation’s third largest producer of engineers with baccalaureate degrees, undergraduates benefit from an innovative curriculum that provides a hands-on, minds-on approach to engineering education. It complements classroom instruction with two unique design-and-build facilities and a strong Cooperative Education Program. With more than 50 research centers and numerous laboratories, the college offers its 2,000 graduate students opportunities in advanced fields of study, including biomedical engineering, state-of-the-art microelectronics, and nanotechnology. http://www.eng.vt.edu/main/index.php
Lynn A. Nystrom | Newswise Science News
NASA CubeSat to test miniaturized weather satellite technology
10.11.2017 | NASA/Goddard Space Flight Center
New approach uses light instead of robots to assemble electronic components
08.11.2017 | The Optical Society
The formation of stars in distant galaxies is still largely unexplored. For the first time, astron-omers at the University of Geneva have now been able to closely observe a star system six billion light-years away. In doing so, they are confirming earlier simulations made by the University of Zurich. One special effect is made possible by the multiple reflections of images that run through the cosmos like a snake.
Today, astronomers have a pretty accurate idea of how stars were formed in the recent cosmic past. But do these laws also apply to older galaxies? For around a...
Just because someone is smart and well-motivated doesn't mean he or she can learn the visual skills needed to excel at tasks like matching fingerprints, interpreting medical X-rays, keeping track of aircraft on radar displays or forensic face matching.
That is the implication of a new study which shows for the first time that there is a broad range of differences in people's visual ability and that these...
Computer Tomography (CT) is a standard procedure in hospitals, but so far, the technology has not been suitable for imaging extremely small objects. In PNAS, a team from the Technical University of Munich (TUM) describes a Nano-CT device that creates three-dimensional x-ray images at resolutions up to 100 nanometers. The first test application: Together with colleagues from the University of Kassel and Helmholtz-Zentrum Geesthacht the researchers analyzed the locomotory system of a velvet worm.
During a CT analysis, the object under investigation is x-rayed and a detector measures the respective amount of radiation absorbed from various angles....
The quantum world is fragile; error correction codes are needed to protect the information stored in a quantum object from the deteriorating effects of noise. Quantum physicists in Innsbruck have developed a protocol to pass quantum information between differently encoded building blocks of a future quantum computer, such as processors and memories. Scientists may use this protocol in the future to build a data bus for quantum computers. The researchers have published their work in the journal Nature Communications.
Future quantum computers will be able to solve problems where conventional computers fail today. We are still far away from any large-scale implementation,...
Pillared graphene would transfer heat better if the theoretical material had a few asymmetric junctions that caused wrinkles, according to Rice University...
15.11.2017 | Event News
15.11.2017 | Event News
30.10.2017 | Event News
20.11.2017 | Earth Sciences
20.11.2017 | Earth Sciences
20.11.2017 | Life Sciences