HomeNewsIndia NewsParallel programming made easy by chips created at MIT

    Parallel programming made easy by chips created at MIT

    Parallel programming Computer chips have stopped getting faster: The regular performance improvements we’ve come to expect are now the result of chip makers’ adding more cores, or processing units, to their chips, rather than increasing their clock speed.

    In theory, doubling the number of cores doubles the chip’s efficiency, but splitting up computations so that they run efficiently in parallel isn’t easy. On the other hand, say a trio of computer scientists from MIT, Israel’s Technion, and Microsoft Research, neither is it as hard as had been feared.

    Commercial software developers writing programs for multi core chips frequently use so-called “lock-free” parallel algorithms, which are relatively easy to generate from standard sequential code. In fact, in many cases the conversion can be done automatically.

    Yet lock-free algorithms don’t come with very satisfying theoretical guarantees: All they promise is that at least one core will make progress on its computational task in a fixed span of time. But if they don’t exceed that standard, they squander all the additional computational power that multiple cores provide.

    In recent years, theoretical computer scientists have demonstrated ingenious alternatives called “wait-free” algorithms, which guarantee that all cores will make progress in a fixed span of time. But deriving them from sequential code is extremely complicated, and commercial developers have largely neglected them.

    In a paper to be presented at the Association for Computing Machinery’s Annual Symposium on the Theory of Computing in May, Nir Shavit, a professor in MIT’s Department of Electrical Engineering and Computer Science; his former student Dan Alistarh, who’s now at Microsoft Research; and Keren Censor-Hillel of the Technion demonstrate a new analytic technique suggesting that, in a wide range of real-world cases, lock-free algorithms actually give wait-free performance.

    “In practice, programmers program as if everything is wait-free,” Shavit says. “This is a kind of mystery. What we are exposing in the paper is this little-talked-about intuition that programmers have about how [chip] schedulers work, that they are actually benevolent.”

    The researchers’ key insight was that the chip’s performance as a whole could be characterized more simply than the performance of the individual cores. That’s because the allocation of different “threads,” or chunks of code executed in parallel, is symmetric. “It doesn’t matter whether thread 1 is in state A and thread 2 is in state B or if you just swap the states around,” says Alistarh, who contributed to the work while at MIT. “What we noticed is that by coalescing symmetric states, you can simplify this a lot.”

    In a real chip, the allocation of threads to cores is “a complex interplay of latencies and scheduling policies,” Alistarh says. In practice, however, the decisions arrived at through that complex interplay end up looking a lot like randomness. So the researchers modeled the scheduling of threads as a process that has at least a little randomness in it: At any time, there’s some probability that a new thread will be initiated on any given core.

    The researchers found that even with a random scheduler, a wide range of lock-free algorithms offered performance guarantees that were as good as those offered by wait-free algorithms.

    ELE Times Bureau
    ELE Times Bureauhttps://www.eletimes.ai/
    ELE Times provides a comprehensive global coverage of Electronics, Technology and the Market. In addition to providing in depth articles, ELE Times attracts the industry’s largest, qualified and highly engaged audiences, who appreciate our timely, relevant content and popular formats. ELE Times helps you build awareness, drive traffic, communicate your offerings to right audience, generate leads and sell your products better.

    Related News

    Must Read

    How Ethernet accelerates the move to software-defined vehicles

    By: Madison Ecklund, systems manager, Body Electronics & Lighting A...

    Intel Foundry Advances Chip Power Delivery with Next-Generation Capacitor Technology

    Courtesy: Intel Highlights Intel and Intel Foundry researchers demonstrated three...

    Keysight Partners with 3dB Labs to Enable Interoperable Signal Monitoring and Analysis

    Keysight Technologies announced the integration of its spectrum analysers...

    India to Boost Local Chip-making with a 1 Trillion Rupees Funding

    India is expected to unveil a funding of more...

    Qualitas Semiconductor Picks Anritsu’s Vector Network Analyzer for High-Speed Interconnect Signal Integrity Verification

    Qualitas Semiconductor Co., Ltd., a leading developer specialising in...

    The Tomorrow for AI and India’s edge advantage

    Courtesy: Qualcomm Artificial intelligence is entering its next chapter, one...

    Posifa Technologies Introduces PVC4001-C MEMS Pirani Vacuum Transducer for Wide-Range Vacuum Measurement

    Posifa Technologies has introduced its new PVC4001-C MEMS Pirani vacuum...

    STMicroelectronics to support AI infrastructure demand with high-volume production of its industry-leading silicon photonics platform

    STMicroelectronics is now entering high-volume production for its state-of-the-art...

    Impact of AI on Computing and the Criticality of Testing

    Courtesy: Teradyne Artificial intelligence (AI) is transforming industries, enhancing our...