- Project plans
- Project activities
- Legislation and standards
- Industry context
Last edited 31 Oct 2017
Top big data tools used to store and analyze data
Big data is a phrase used for a collection of data sets so big and complex that it is difficult to process using traditional applications/tools. Due to the variety of information that it encompasses, big data consistently brings several challenges relating to its volume and complexity.
A recent survey claims that 80% of the data generated in the world are unstructured. One question is how these unstructured information can be structured, before we try to understand and capture the most important data. Another challenge is how we could store it. Listed below are the top tools utilised to store and analyse big data.
 Apache Hadoop
Apache Hadoop is a java-based free software framework that can effectively store great deal of information in a cluster. This frame runs in parallel on a cluster and has an ability to enable us to process data across all nodes. Hadoop Distributed File System (HDFS) is the storage system of Hadoop which splits big information and distribute across several nodes in a cluster. This also replicates data in a bunch thus providing high availability.
 Microsoft HDInsight
While the traditional SQL can be effectively utilised to handle large quantity of structured data, we want NoSQL (Not Just SQL) to deal with unstructured data. NoSQL databases store unstructured information with no particular schema.
NoSQL gives better performance in storing massive number of data. There are lots of open-source NoSQL DBs available to analyse big data.
This supports SQL-like query option HiveSQL (HSQL) to get big data. This may be primarily used for its data-mining function.
This works on top of SQL Server 2012 Parallel Data Warehouse (PDW) and is used to get data stored in PDW. PDW is a data-warehousing appliance built for processing any quantity of relational data and provides an integration with Hadoop allowing the additional provision of non-relational information.
Lots of people are comfortable doing data analytics, therefore, the users may even connect data stored in Hadoop using Excel 2013. The Power View feature of Excel 2013 can be used to easily summarise the information. Similarly, Microsoft's HDInsight enables us to connect to big data stored in Azure Cloud using a power query option.
Facebook has developed and recently open-sourced its Query engine (SQL-on-Hadoop) called Presto which is built to manage petabytes of information. Unlike Hive, Presto doesn't depend on MapReduce technique and can quickly retrieve information.
 Related articles on Designing Buildings Wiki
Featured articles and news
Read our introductory article to the completion date in construction contracts.
Almost 90% of freight in London is moved by road. The River Thames could add much needed extra capacity.
National Infrastructure Commission warn that large infrastructure projects are at risk of falling behind.
The quality of Cambridge owes as much to its open spaces as to its architectural uniqueness.
If events occur that cause the completion of the works to be delayed then these may be compensation events.
BSRIA's new Building MOTs Scheme is designed to provide guidance on the next steps after compliance.
At an ICE discussion, the focus was on delivering a Northern Infrastructure Strategy based on opportunity for all.
The Considerate Constructors Scheme officially launch the new Ultra Site status for contractors and supply chains.
The risk of specification errors in the cladding sector is "worryingly high" after Grenfell, according to major distributor.
Here is our outline work plan for a private sector design and build project.