Microsoft brings Hadoop option to SQL Server

Open-source crunchware SQOOPs to conquer


Microsoft customers running SQL Server are getting a taste of really big data processing through an injection of Hadoop.

The company has released early code that will let Microsoft customers plug the open-source Java architecture from Doug Cutting into SQL Server 2008 R2, SQL Server Parallel Data Warehouse for huge data warehouses, as well as the next version of Microsoft's database, which is codenamed Denali.

Hadoop was built by Cutting, who was inspired by Google's MapReduce. It is becoming something of an industry standard for processing huge amounts of data on clustered servers thanks to the fact that its code is open. Hadoop has also been adopted by top-tier web properties including Amazon, Facebook and Twitter.

The industry thinking is that Hadoop can trickle down to customers outside the rarefied circles of serious number-crunchers, where it is used to understand the changing minutiae of millions of users' likes and status updates in order to change services in response. The aim is for Hadoop to find its feet in more mainstream IT.

Microsoft's Research unit has been working on something that sounds remarkably similar to Hadoop, called Dryad, since about 2006. Earlier this year the plan was to "productise" Dryad through integration with SQL Server and its Windows Azure cloud. There have been no updates from Microsoft, but it seems Dryad must now compete for the affections of big-data lovers on SQL Server.

The Microsoft connectors are called Hadoop Connector for SQL Server Parallel Data Warehouse and Hadoop Connector for SQL Server and are available as Community Technology Previews (CTPs).

The connectors are two-way, letting you move data backwards and forwards between Hadoop and Microsoft's database servers

Microsoft said the connectors would let its customers analyse unstructured data in Hadoop and then pull that back into the SQL Server environments for analysis.

Both connectors use SQL to Hadoop (SQOOP) to transfer the data "efficiently" between the Hadoop File System (HDFS) and Microsoft's relational databases. The Parallel Data Warehouse uses PDW Bulk Load/Extract tool for fast import and export of data.

SQL Server PDW customers can get the Hadoop connector from Microsoft while users of the regular SQL Server 2008 R2 can get the code for Hadoop Connector for SQL Server here. ®


Other stories you might like

  • 381,000-plus Kubernetes API servers 'exposed to internet'
    Firewall isn't a made-up word from the Hackers movie, people

    A large number of servers running the Kubernetes API have been left exposed to the internet, which is not great: they're potentially vulnerable to abuse.

    Nonprofit security organization The Shadowserver Foundation recently scanned 454,729 systems hosting the popular open-source platform for managing and orchestrating containers, finding that more than 381,645 – or about 84 percent – are accessible via the internet to varying degrees thus providing a cracked door into a corporate network.

    "While this does not mean that these instances are fully open or vulnerable to an attack, it is likely that this level of access was not intended and these instances are an unnecessarily exposed attack surface," Shadowserver's team stressed in a write-up. "They also allow for information leakage on version and build."

    Continue reading
  • A peek into Gigabyte's GPU Arm for AI, HPC shops
    High-performance platform choices are going beyond the ubiquitous x86 standard

    Arm-based servers continue to gain momentum with Gigabyte Technology introducing a system based on Ampere's Altra processors paired with Nvidia A100 GPUs, aimed at demanding workloads such as AI training and high-performance compute (HPC) applications.

    The G492-PD0 runs either an Ampere Altra or Altra Max processor, the latter delivering 128 64-bit cores that are compatible with the Armv8.2 architecture.

    It supports 16 DDR4 DIMM slots, which would be enough space for up to 4TB of memory if all slots were filled with 256GB memory modules. The chassis also has space for no fewer than eight Nvidia A100 GPUs, which would make for a costly but very powerful system for those workloads that benefit from GPU acceleration.

    Continue reading
  • GitLab version 15 goes big on visibility and observability
    GitOps fans can take a spin on the free tier for pull-based deployment

    One-stop DevOps shop GitLab has announced version 15 of its platform, hot on the heels of pull-based GitOps turning up on the platform's free tier.

    Version 15.0 marks the arrival of GitLab's next major iteration and attention this time around has turned to visibility and observability – hardly surprising considering the acquisition of OpsTrace as 2021 drew to a close, as well as workflow automation, security and compliance.

    GitLab puts out monthly releases –  hitting 15.1 on June 22 –  and we spoke to the company's senior director of Product, Kenny Johnston, at the recent Kubecon EU event, about what will be added to version 15 as time goes by. During a chat with the company's senior director of Product, Kenny Johnston, at the recent Kubecon EU event, The Register was told that this was more where dollars were being invested into the product.

    Continue reading

Biting the hand that feeds IT © 1998–2022