Apache Hadoop and MapReduce attract a large number of large data analysis experts and business intelligence experts. However, a wide range of Hadoop decentralized file systems, or the ability to write or execute mapreduce in the Java language, requires truly rigorous software development techniques. Apache Hive will be the only solution. The Apache Software Foundation Engineering Hive's database component, is also based on the cloud Hadoop ecosystem, provides the context based query statement called Hive query statement. This set of ...
Pig is a Yahoo donated project to Apache and is currently in the Apache incubator, but the basic functionality is already available. Today I would like to introduce you to this useful pig.pig is Sql-like language, is built on the mapreduce of an advanced query language, Some operations are compiled into the MapReduce model's map and reduce, and users can define their own capabilities. Yahoo Grid Computing department developed another clone of Google's project: Sawzall. Supported operations ...
Zhang Fubo: The following part of the forum is mainly four guests, talk about cloud practice. Beijing First Letter Group is the Beijing government's integration company, mainly responsible for the capital window of the construction, they are also in the domestic, in the government industry earlier in a company, as the first letter Group Technical Support Center General Manager Zhang Ninglai for us to do the report. Zhang: Good afternoon, we have just introduced, I am from Beijing First Letter Development Co., Ltd., I bring today is the result of our practice in cloud computing technology these years. Today is mainly divided into three parts, we mainly do is the field of e-government applications, we are mainly ...
Forbes: hadoop--Big Data tools you have to understand now Apache Hadoop has become the driving force behind the development of the big data industry. Techniques such as hive and pig are often mentioned, but they all have functions and why they need strange names (such as Oozie,zookeeper, Flume). Hadoop has brought in cheap processing of large data (large data volumes are usually 10-100GB or more, with a variety of data types, including structured, unstructured, etc.) capabilities. But this is the same as before ...
In the past few years, the use of Apache Spark has increased at an alarming rate, usually as a successor to the MapReduce, which can support thousands of-node-scale cluster deployments. In the memory data processing, the Apache spark is more efficient than the mapreduce has been widely recognized, but when the amount of data is far beyond memory capacity, we also hear some organizations in the spark use of trouble. Therefore, with the spark community, we put a lot of energy to do spark stability, scalability, performance, etc...
February 28, 2012, the world's leading cloud computing services provider hundred major cloud computing event one by one "mobile office 100 will cloud Computing conference 2012" held in the Centennial Auditorium of Peking University. The conference to "mobile office" as the theme, to explore the inspirational, cloud computing leading applications, partners, etc. as the main content. "The future enterprise is not in the factory, not in the office, but in the clouds." "Hundred meeting CEO Jianhua thinks, the modern enterprise is changing quietly, the enterprise is more and more is not a static entity, it is closer to a group or many group relations, is a real name, the division of labor detail ..."
The content source of this page is from Internet, which doesn't represent Alibaba Cloud's opinion;
products and services mentioned on that page don't have any relationship with Alibaba Cloud. If the
content of the page makes you feel confusing, please write us an email, we will handle the problem
within 5 days after receiving your email.
If you find any instances of plagiarism from the community, please send an email to:
info-contact@alibabacloud.com
and provide relevant evidence. A staff member will contact you within 5 working days.