Big data is changing our lives, influencing our way of thinking and solving problems, in order to adapt to the trend of the Times, the organization must learn to use data to speak, if sitting on a large number of data but helpless or indifferent, that and no data is the same. However, in the analysis of data, complete self-creation is undesirable, because there are a large number of experience can be followed and learned to save a lot of time and cost. Recently, Orionx.net's co-founder, Shahin Khan, published an article about his team's experience and discipline from the big data, Internet of things, and cloud computing markets.
Big data becomes possible because it retains enough data, so don't delete it anyway, because you don't know when it will be used and what legal risk it will be to delete it. The cost of preserving data is low, and if anything happens in the future, you can also find evidence from that data.
Most data collection efforts focus on ongoing activities, but once you know how to use the data, the willingness to get more data increases.
There are few medium-sized big data systems, and once the idea of a project proves to be promising, it will quickly grow and hatch new projects as it grows rapidly.
Unused data is an idle asset that is likely to result in a depreciation of value. If you consider big data as a workflow, you must flow data to the most valuable places.
Most of the big data scenarios are valuable and worth the effort, but their causal relationships are complex, data incomplete, and user biases unavoidable.
There is a lot of data, but most of it is useless, and only a few are worth it. The more data collected, the more obvious this phenomenon, which means that irrelevant data is growing at a much higher rate than the relevant data.
Synthesis is required after the analysis is complete, which requires the introduction of machine learning and cognitive algorithms.
Data is an asset, although it can appreciate, but most of the time as new data replaces old data, the value of historical data will be lower because its relevance will become worse. So you must know the "interest rate" of the data and know how fast it will depreciate.
The quality of data directly affects the quality of data mining.
The greater the amount of data, the more difficult it is to find valuable information, the complexity of the data, the irrational motives and the ignorance may result in invalid conclusions, but on the other hand, the more data there is, the more evidence is available to support the hypothesis, and, in a completely scientific way, sometimes the approval rate is even closer to 100%.
In the new media age, interesting but superficial content is much more than insightful content, value mining needs to have a deep understanding of the data, but it takes time.
If you have 200 rows of data, you can use a spreadsheet, but if you have 2 billion rows of data, you must use HPC. In addition, as the volume of data grows, mathematical and scientific knowledge is needed to transform the data into models.
Big Data Big Law