Enterprise technology professionals talk a lot about Big Data in its various forms, discussing how to best approach what is seen as a growing resource for the generation of insights. Data volumes will continue to grow exponentially as new data sources and systems come online, delivering upon the promise for a deeper understanding of what makes business move forward. Because Big Data remains such a popular buzzword, many view data in only this form, the piles and piles of stored data that data analysts and scientists must dig into in order to learn about their surroundings. So the focus remains on data volumes, ignoring what may be an even more crucial matter, data velocity.
It’s true that data is fast before it’s big. The speed at which data moves is another necessary consideration for modern data-driven organizations. It seems only natural that business should intertwine big and fast data to achieve an overarching scheme with which to make decisions, both in the moment and historically. As data continues to pile up, new sources open as relevant for collection in many verticals, making the practice of keeping up with the expansion burdensome.
Big Data is certainly valuable. No doubt there is an incredible amount of knowledge that can be derived from historical data stores. However, in world increasingly being flooded with self-service technology, fast data is what drives real-time decision making. When we think about the types of data that users want to analyze on a continual basis, 21st century data sources come to mind. While still expanding, these sources are largely made up of data streaming to and from business applications, sensor networks, on social media, and involving financial transactions. Modern sources of data are proliferating at a rate beyond which even the most forward-thinking technology can comprehend.
For these reasons, many enterprise companies have turned away from legacy storage and collection technologies in favor of more agile, open source frameworks, like Hadoop, Spark and those developed by the Apache Software Foundation. Fast data provides the modern digital enterprise with a unique opportunity to analyze data on the fly while its ingested. This allows for mass collection of data for later use, but with the caveat that newer feature enhancements provider users with the tools they need to act on an event-driven basis.
Though Big Data has been top-of-mind in the technology world for years now, it’s probably correct to assume that enterprises are just scratching the surface as to what’s possible with data at this scale. The benefits of utilizing fast data, or data as it streams in real-time, are numerous. If common practice evolves to a point where this becomes the norm enterprise-wide, access to “fresh” data could very well change the game. This data has intrinsic value, since it paints a picture of what is happening at a particular moment in time. Everyone knows that the quicker data can be acted upon, the more likely it is to have long-standing business impact.
The Big Data movement was largely drive by the demand for scale in the volume and variety of data, leading to an evolutionary new era in the enterprise and a different approach to Data Management. It seems obvious that the next step would be to make use of fast data, once and for all processing data at the speed of insight and conquering the entirety of the data spectrum.
Latest posts by Timothy King (see all)
- 5 Data Quality Tools Vendors to Watch in 2018 - October 19, 2017
- Five Questions to Ask When Evaluating Data Management Tools - October 17, 2017
- GoodData Adds Governance to Enterprise Insights Platform - October 12, 2017