Ever-growing volumes of data mean computational storage is becoming crucial for HPC, say boffins at Dell's tech chinwag

Data at rest should remain at rest

Dell believes that technologies such as computational storage will soon play a part in high-performance computing (HPC) in response to ever-growing volumes of data. It doesn't see the general-purpose CPU disappearing any time soon, but says it will be complemented with specialised processors for specific tasks, with composability seen as both an opportunity and a problem.

Dell Technologies holds regular online sessions for its HPC Community, and the latest saw experts from the company discussing how HPC and mainstream enterprise IT feed off each other and where this may lead to, with advances in one sector often being picked up by the other.

Most modern HPC systems are now built from clusters of commodity servers, for example, while the use of GPUs for accelerating complex workloads started in HPC research labs and is now filtering into enterprise data centres.

Onur Celebioglu, Dell senior director of Engineering for HPC and Emerging Workloads, said that expanding volumes of data were a big part of HPC today, and this has led to the rise of specialised silicon for dedicated tasks like AI and machine learning.

"One of the ways to analyse data and get intelligence out of all that data is by utilising statistical analysis techniques and AI techniques. We're seeing that trend continuing, and AI is one of the great ways to be able to get meaning out of the data, so many businesses are investing in the opportunity to get that intelligence faster and quicker," he said.

When trying to optimise for just one thing such as deep learning or training, it is possible to fine-tune the architecture to get the best performance per watt using a special ASIC, for example, so Celebioglu sees a lot of investment going into that to get ever-faster analytics. But that does not take away the role of the CPU, since HPC covers a diverse range of problems that need to be solved, and tackling all these needs the ability to handle general purpose processing.

"When you are doing general-purpose HPC, you will be running a wide variety of algorithms, and different applications. If you look at molecular dynamics the algorithms there are very different than what's used for genomics or oil and gas, so you need a general purpose machine like a processor to cater for that wide variety of applications," he said.

When it comes to new technologies, the discussion turned to composability, techniques that allow for a bunch of systems to be viewed as a pool of resources at the rack level (or even larger). This aims to enable any application to run, even if it would require more memory or other resources than would fit inside a single node, by composing a big enough system with resources from other nodes.

A key technology here is set to be Compute Express Link (CXL), said Garima Kochhar, Distinguished Engineer at Dell. "I think that HPC will be the first adopters of CXL, and once we figure that out it will be extremely important, useful and relevant for mainstream IT as well as for the optimised infrastructure," she said.

However, she warned that there would likely be performance trade-offs in creating larger systems from distributed bits and pieces this way.

"I do think that a big piece of this composable thing is going to be a conversation between latency and any overhead versus the value that you get from being able to disaggregate," Kochhar said. She added that it could still prove useful for areas of HPC where having a larger memory space was more critical than latency, and the technology would also improve over time.

Celebioglu said that growing volumes of data would call for broader use of technology such as computational storage, which allows data to be processed without incurring many of the penalties in moving huge amounts of data around.

"The growth in data and how we move data, how do you handle that if you have very large volumes of data? That's going to continue being a challenge, not just for HPC, but for traditional IT as well," he said. Some of the challenges can be alleviated by building faster networks and giving systems larger memories, but the sheer volume of data will still risk creating a bottleneck.

"If we can analyse data where it is, that's going to be one way to shift the paradigm, and I think computational memory, computational storage technologies are going to start playing a bigger role, both in HPC and general IT," he opined.

Computational storage actually covers a range of architectural models, but the most familiar features a CPU or ASIC embedded into the controller of an SSD, allowing for high-bandwidth, low-latency access to the data stored there.

Jimmy Pike, a senior VP and senior fellow at Dell, agreed, saying that even as far back as the days of Gene Amdahl there was a saying that data at rest should remain at rest.

"What he was talking about is if you're going to do analysis on data, for god's sake do it where the data is rather than moving it around all over the place," Pike said.

Another topic touched on was SmartNICs, with Celebioglu expecting to see these enhanced network adapters cropping up more and more to offload some of the functions that are taking valuable compute resources from the system, such as security functions and serving as a point for remote management control, for bare metal provisioning, as an example.

This is an area where hyperscalers such as AWS are already ahead, with the Nitro cards inside the servers fulfilling similar functions to SmartNICs. VMware has also floated plans to use SmartNICs inside the physical hosts for virtualized infrastructure, taking on functions such as firewalls at first, but with an eye to the potential to offload software-defined network tasks as well.

Finally, on the subject of cloud versus on-premises for HPC infrastructure, Pike pointed out that the ecosystem was developing pretty much along the lines of enterprise cloud services.

"Do I think the cloud guys will continue to invest and drive HPC? You bet. Do I think that the on-prem version of that, where people are going to own and drive their own HPC will continue? You bet. Do I think there will be HPC stuff for rent, which is what the cloud is about? There is no doubt about it," he said.

Both Dell and HPE have developed HPC as-a-service via their own brands, APEX and GreenLake. ®

Broader topics

Other stories you might like

  • Lenovo halves its ThinkPad workstation range
    Two becomes one as ThinkPad P16 stands alone and HX replaces mobile Xeon

    Lenovo has halved its range of portable workstations.

    The Chinese PC giant this week announced the ThinkPad P16. The loved-by-some ThinkPad P15 and P17 are to be retired, The Register has confirmed.

    The P16 machine runs Intel 12th Gen HX CPUs, but only up to the i7 models – so maxes out at 14 cores and 4.8GHz clock speed. The laptop is certified to run Red Hat Enterprise Linux, and can ship with that, Ubuntu, and Windows 11 or 10. The latter is pre-installed as a downgrade right under Windows 11.

    Continue reading
  • US won’t prosecute ‘good faith’ security researchers under CFAA
    Well, that clears things up? Maybe not.

    The US Justice Department has directed prosecutors not to charge "good-faith security researchers" with violating the Computer Fraud and Abuse Act (CFAA) if their reasons for hacking are ethical — things like bug hunting, responsible vulnerability disclosure, or above-board penetration testing.

    Good-faith, according to the policy [PDF], means using a computer "solely for purposes of good-faith testing, investigation, and/or correction of a security flaw or vulnerability."

    Additionally, this activity must be "carried out in a manner designed to avoid any harm to individuals or the public, and where the information derived from the activity is used primarily to promote the security or safety of the class of devices, machines, or online services to which the accessed computer belongs, or those who use such devices, machines, or online services."

    Continue reading
  • Intel plans immersion lab to chill its power-hungry chips
    AI chips are sucking down 600W+ and the solution could be to drown them.

    Intel this week unveiled a $700 million sustainability initiative to try innovative liquid and immersion cooling technologies to the datacenter.

    The project will see Intel construct a 200,000-square-foot "mega lab" approximately 20 miles west of Portland at its Hillsboro campus, where the chipmaker will qualify, test, and demo its expansive — and power hungry — datacenter portfolio using a variety of cooling tech.

    Alongside the lab, the x86 giant unveiled an open reference design for immersion cooling systems for its chips that is being developed by Intel Taiwan. The chip giant is hoping to bring other Taiwanese manufacturers into the fold and it'll then be rolled out globally.

    Continue reading

Biting the hand that feeds IT © 1998–2022