The controversy between open source and pure faction

Source: Internet
Author: User
Keywords Large data Idc hadoop mapreduce
Tags analysis apache data distributed distributed file system file framework hadoop

IDC released its latest report in Monday, which said that the software market revenues associated with Hadoop and MapReduce programming frameworks for large data analysis would surge from $77 million trillion in 2011 to $812.8 million in 2016, with an annual compound growth rate of 60.2% per cent.

Hadoop is an open source implementation of the MapReduce framework, hosted by the Apache Software Foundation, which has a number of supporting software projects, including the Hadoop Distributed File System (HDFS) and the Pig programming language.

MapReduce and Hadoop can process data across a large number of nodes, based on the principle of decomposing large amounts of data and then splitting them into parallel chunks. They are closely related to the industry buzzword "big data", which refers mainly to the increasing amount of information generated by websites, social media, sensors, and other sources, especially unstructured information.

In general, Hadoop has secured support from Business Analytics and database vendors in recent years, and these vendors have started to provide Hadoop products and services.

While the use cases and advantages of Hadoop and MapReduce, "have produced too many fantastic and overwhelming amounts of exaggeration, there is no doubt that it does provide a relatively low-cost approach to tapping into significant commercial value from very large scattered data," Carl Olofson and Dan Vesset, IDC analyst, wrote in the report.

The IDC report believes that the Hadoop-mapreduce "ecosystem" of explosive growth in the question should be justified.

This year, "the mainstream it world leader adopters will shift from ' proof-of-concept ' to real value," the report wrote.

But the report also points out that the lack of qualified talent will limit the technology's development over the next 2-3 years.

Next year we'll see a "debate between open source and pure pie," the pure faction thinks that the kernel of Hadoop must be developed purely on the Apache project code, "IDC said. However, most IT organizations will use commercial and open-source components in their hadoop environments.

"There is also competition between open source vendors and copies of their cloned source code, which may force the latter to reduce their licensing fees, which could result in a reduction in software revenues on the market." ”

Related Article

Contact Us

The content source of this page is from Internet, which doesn't represent Alibaba Cloud's opinion; products and services mentioned on that page don't have any relationship with Alibaba Cloud. If the content of the page makes you feel confusing, please write us an email, we will handle the problem within 5 days after receiving your email.

If you find any instances of plagiarism from the community, please send an email to: info-contact@alibabacloud.com and provide relevant evidence. A staff member will contact you within 5 working days.

A Free Trial That Lets You Build Big!

Start building with 50+ products and up to 12 months usage for Elastic Compute Service

  • Sales Support

    1 on 1 presale consultation

  • After-Sales Support

    24/7 Technical Support 6 Free Tickets per Quarter Faster Response

  • Alibaba Cloud offers highly flexible support services tailored to meet your exact needs.