US Energy dept starts handing out cash for exaflop superputer quest

Need to get the 'leccy bill down below that of a city

The Department of Energy is continuing to dole out cash to pay for some of the basic research that needs to be done if the United States is going to field exascale-class supercomputers by 2020 or so. This time around, Nvidia and Intel have taken down some contracts, and El Reg hears that Big Blue is getting some funding as well.

IBM is in its quiet period just before putting out its financial results for the second quarter, and that could be why the company has not said anything about its awards under the FastForward portion of the DOE's Extreme-Scale Computing Research and Development program, which has become the main vehicle for the creation of systems with 1,000 petaflops of number-crunching performance and 1,000 petabytes of data store.

As El Reg explained in detail last week, the first phase of the exascale project is called FastForward, and it has a nominal amount of funding that is designed to get some primary research done that will lead to exascale systems. We have contacted the DOE a number of times to get more details about the budgeting for the three phases of the Extreme-Scale project, but thus far DOE is keeping a tight lip about the plans. The development of exascale systems is expected to cost hundreds of millions if not billions of dollars, so you can understand why DOE doesn't want to talk about spending that much dough in an election year. Moreover, while the Extreme-Scale project is funded in part by the DOE's Office of Science, which does basic research, the National Nuclear Security Administration part of the DOE, which manages the American nuclear weapons stockpile, is also not only kicking in funds but is also expected to be the first big beneficiary of exascale systems. Ever more powerful supercomputing is needed to simulate old nukes and confirm they'll still work, or to design new ones: the problems are particularly intricate as real-world explosions are forbidden under the Nuclear Test Ban Treaty.

The Extreme-Scale project has an initial research phase, called FastForward, and which has an intentionally somewhat vague statement of work since it is dedicated to some of the primary research in processor, memory, and storage and I/O technologies that will be required in exascale systems. The engineering challenges are daunting. Extrapolating current technologies with Moore's Law will be woefully insufficient to meet the power envelope, cost (both capital and operating), and reliability requirements of an exascale system, so real engineering breakthroughs must be made to get a machine that might only require 20 megawatts of juice to do an exaflops of work against an exabyte of data.

Last week, AMD announced it received two contracts worth $15.6m to do research on memory subsystems and interconnects relating to memory as well as exploring the use of hybrid CPU-GPU architectures to boost the performance of HPC machines. Whamcloud, a company that offers commercial support for the open source Lustre parallel clustered file system, also took down a contract, rumored to be worth $8m, in conjunction with partners HDF Group (which has application I/O expertise), EMC (which is bringing in experts for system I/O and I/O aggregation), and Cray (which is doing scale-out testing of the big bad storage systems).

Nvidia's Bill Dally, the head of the computer science department at Stanford University who took a two-year sabbatical to be chief scientist and head of Nvidia Research, came back to Nvidia in March in those roles and is spearheading the $12.4m project that Nvidia is doing as part of the FastForward phase of the Extreme-Scale DOE project. The research will be based on the earlier "Echelon" system that Nvidia conceived as part of the US Defense Advanced Research Project Agency's Ubiquitous High Performance Computing program announced in March 2010 and now superseded by the DOE effort.

Applied Micro X-Gene ARM block diagram

Block diagram of Nvidia's Echelon exascale system (click to enlarge)

Echelon proposed creating a special processor with four double-precision fused multiply-add (DFMA) units and some Level 0 instruction and data cache for those units and their main registers and operand registers. This block would be able to do about 20 gigaflops of peak number crunching. You pack eight of these into a single unit with a switch linking them to a shared Level 1 cache memory and comprising a "streaming multiprocessor" or SM. A single chip would have 128 of these SMs, rated at a combined 20.5 teraflops and would include plus 1,024 banks of Level 2 static RAM cache (measuring 256KB each for a total of 256GB of cache), plus memory controllers, network interface cards, and eight of what Nvidia called latency processors.

An Echelon node multichip module would have eight of these linked together on a single substrate linking out to 3D stacked DRAM memory (through 1.4TB/sec of memory bandwidth) with 150GB/sec of network bandwidth coming out of the interconnect. There are also ports on the node for direct link to non-volatile storage memory. A node module would be rated at 160 teraflops peak and would have 12.8TB/sec of memory bandwidth at 2TB of main memory. A cabinet of Echelon iron would have 16 modules offering a total of 2.56 petaflops and would consume about 38 kilowatts, according to a presentation put together by Dally for the DARPA work, and 400 cabinets would get you to an exaflop at around 15 megawatts.

This sounds like mission accomplished, right? Wrong. For one thing, Cray was on the Echelon team and its "Aries" interconnect, which will probably be commercialized as Dragonfly, was an integral part of the Nvidia design. So was the high-radix router module, a component of the Aries interconnect presumably, that Nvidia was proposing to put in its Echelon chip. Now Intel owns Cray's Aries interconnect as well as the patents behind the interconnect biz and the key employees (minus former Crayon Steve Scott, now Tesla CTO at Nvidia). And you can bet that Intel is in no mood to share with Nvidia – particularly when it wants its Xeon Phi x86-based coprocessors to be used instead of Nvidia GPUs in petaflops and exaflops systems.

In a blog post, Dally confirmed that the FastForward contract with the DOE would leverage the work done on Echelon, and just to put all the engineering into perspective, Dally said in theory an exascale system using today's CPU technology would take about 2 gigawatts – yeah, 1,000 megawatts – to power and cool, and that this was equivalent to the entire output of the Hoover Dam. If you moved to the forthcoming Tesla K20 GPU coprocessors, coming later this year, to do the math, you are down to maybe 150 megawatts to juice up an exascale box.

"Achieving this level of efficiency will require extraordinary innovation on a number of fronts," Dally wrote. "However, we firmly believe that heterogeneous computing offers the best approach to get there."

That's another way of saying that you need to minimize the use of traditional CPUs, which are optimized for single-thread performance, and go massively parallel with modest processors, like those in the GPU.

Intel Federal, the unit of the chip maker established last September to chase big HPC and big data deals at the US government, has taken down two FastForward contracts worth a combined $19m.

At the same time, Mark Seager, former head of HPC systems at LLNL, was tapped by Intel to be CTO for the chipmaker's HPC Ecosystems group. Seager was in charge of the Advanced Simulation and Computing Initiative (ASCI) program at LLNL, and was responsible for ASCI Blue Pacific, ASCI White, ASCI Purple, and BlueGene/L, which were all based on IBM's Power servers and proprietary interconnects. Intel wants to own this business, and has made no secret about it. Intel will be working on processor projects, featuring future Xeon and Xeon Phi designs, as well as on interconnects. The expectation is that Intel will perfect its near threshold voltage chippery, demonstrated with the "Claremont" Pentium chip, as well as the 3D memory stacking technology (called Hybrid Cube Memory, or HCM) that Intel is developing with Micron Technology, Samsung Electronics, and others. But the specifics of the work Intel is doing were not detailed.

"High-performance computing is a transformative technology that will allow current and future generations of scientists and engineers to develop breakthrough advancements to address our most pressing societal issues," said David Patterson, president of Intel Federal, in a statement. "This is a great example of how public-private partnerships will significantly help move high performance computing forward and push the boundaries of innovation." ®

Broader topics

Other stories you might like

  • Monero-mining botnet targets Windows, Linux web servers
    Sysrv-K malware infects unpatched tin, Microsoft warns

    The latest variant of the Sysrv botnet malware is menacing Windows and Linux systems with an expanded list of vulnerabilities to exploit, according to Microsoft.

    The strain, which Microsoft's Security Intelligence team calls Sysrv-K, scans the internet for web servers that have security holes, such as path traversal, remote file disclosure, and arbitrary file download bugs, that can be exploited to infect the machines.

    The vulnerabilities, all of which have patches available, include flaws in WordPress plugins such as the recently uncovered remote code execution hole in the Spring Cloud Gateway software tracked as CVE-2022-22947 that Uncle Sam's CISA warned of this week.

    Continue reading
  • Red Hat Kubernetes security report finds people are the problem
    Puny human brains baffled by K8s complexity, leading to blunder fears

    Kubernetes, despite being widely regarded as an important technology by IT leaders, continues to pose problems for those deploying it. And the problem, apparently, is us.

    The open source container orchestration software, being used or evaluated by 96 per cent of organizations surveyed [PDF] last year by the Cloud Native Computing Foundation, has a reputation for complexity.

    Witness the sarcasm: "Kubernetes is so easy to use that a company devoted solely to troubleshooting issues with it has raised $67 million," quipped Corey Quinn, chief cloud economist at IT consultancy The Duckbill Group, in a Twitter post on Monday referencing investment in a startup called Komodor. And the consequences of the software's complication can be seen in the difficulties reported by those using it.

    Continue reading
  • Infosys skips government meeting – and collecting government taxes
    Tax portal wobbles, again

    Services giant Infosys has had a difficult week, with one of its flagship projects wobbling and India's government continuing to pressure it over labor practices.

    The wobbly projext is India's portal for filing Goods and Services Tax returns. According to India's Central Board of Indirect Taxes and Customs (CBIC), the IT services giant reported a "technical glitch" that meant auto-populated forms weren't ready for taxpayers. The company was directed to fix it and CBIC was faced with extending due dates for tax payments.

    Continue reading
  • Google keeps legacy G Suite alive and free for personal use

    Google has quietly dropped its demand that users of its free G Suite legacy edition cough up to continue enjoying custom email domains and cloudy productivity tools.

    This story starts in 2006 with the launch of “Google Apps for Your Domain”, a bundle of services that included email, a calendar, Google Talk, and a website building tool. Beta users were offered the service at no cost, complete with the ability to use a custom domain if users let Google handle their MX record.

    The service evolved over the years and added more services, and in 2020 Google rebranded its online productivity offering as “Workspace”. Beta users got most of the updated offerings at no cost.

    Continue reading
  • GNU Compiler Collection adds support for China's LoongArch CPU family
    MIPS...ish is on the march in the Middle Kingdom

    Version 12.1 of the GNU Compiler Collection (GCC) was released this month, and among its many changes is support for China's LoongArch processor architecture.

    The announcement of the release is here; the LoongArch port was accepted as recently as March.

    China's Academy of Sciences developed a family of MIPS-compatible microprocessors in the early 2000s. In 2010 the tech was spun out into a company callled Loongson Technology which today markets silicon under the brand "Godson". The company bills itself as working to develop technology that secures China and underpins its ability to innovate, a reflection of Beijing's believe that home-grown CPU architectures are critical to the nation's future.

    Continue reading

Biting the hand that feeds IT © 1998–2022