10 Big Data Implementation Best Practices
Big data is still relatively new with many organizations, and its significance in business processes and outcome has been changing every day. Here are some of the key best practices that implementation teams need to increase the chances of success.
1. Gather business requirements before gathering data. Begin big data implementations by first gathering, analyzing and understanding the business requirements; this is the first and most essential step in the big data analytics process. If you take away nothing else, remember this: Align big data projects with specific business goals.
2. “Implementing big data is a business decision not IT.” This is a wonderful quote that wraps up one of the most important best practices for implementing big data. Analytics solutions are most successful when approached from a business perspective and not from the IT/Engineering end. IT needs to get away from the model of “Build it and they will come” to “Solutions that fit defined business needs.”
3. Use Agile and Iterative Approach to Implementation. Typically, big data projects start with a specific use-case and data set. Over the course of implementations, we have observed that organization needs evolve as they understand the data – once they touch and feel and start harnessing its potential value. Use agile and iterative implementation techniques that deliver quick solutions based on current needs instead of a big bang application development. When it comes to the practicalities of big data analytics, the best practice is to start small by identifying specific, high-value opportunities, while not losing site of the big picture. We achieve these objectives with our big data framework: Think Big, Act Small.
4. Evaluate data requirements. Whether a business is ready for big data analytics or not, carrying out a full evaluation of data coming into a business and how it can best be used to the business’s advantage is advised. This process usually requires input from your business stakeholders. Together we analyze what data needs to be retained, managed and made accessible, and what data can be discarded.
5. Ease skills shortage with standards and governance. Since big data has so much potential, there’s a growing shortage of professionals who can manage and mine information. Short of offering huge signing bonuses, the best way to overcome potential skills issues is standardizing big data efforts within an IT governance program.
6. Optimize knowledge transfer with a center of excellence. Establishing a Center of Excellence (CoE) to share solution knowledge, plan artifacts and ensure oversight for projects can help minimize mistakes. Whether big data is a new or expanding investment, the soft and hard costs can be shared across the enterprise. Another benefit from the CoE approach is that it will continue to drive the big data and overall information architecture maturity in a more structured and systematical way.
7. Embrace and plan your sandbox for prototype and performance. Allow data scientists to construct their data experiments and prototypes using their preferred languages and programming environments. Then, after a successful proof of concept, systematically reprogram and/or reconfigure these implementations with an “IT turn-over team.” Sometimes, it may be difficult to even know what you are looking for, because the technology is often breaking new ground and achieving results that were previously labeled “can’t be done.”
8. Align with the cloud operating model. Analytical sandboxes should be created on-demand and resource management needs to have a control of the entire data flow, from pre-processing, integration, in-database summarization, post-processing, and analytical modeling. A well planned private and public cloud provisioning and security strategy plays an integral role in supporting these changing requirements. The advantage of a public cloud is that it can be provisioned and scaled up instantly. In those cases where the sensitivity of the data allows quick in-and-out prototyping, this can be very effective.
9. Associate big data with enterprise data: To unleash the value of big data, it needs to be associated with enterprise application data. Enterprises should establish new capabilities and leverage their prior investments in infrastructure, platform, business intelligence and data warehouses, rather than throwing them away. Investing in integration capabilities can enable knowledge workers to correlate different types and sources of data, to make associations, and to make meaningful discoveries.
10. Embed analytics and decision-making using intelligence into operational workflow/routine. For analytics to be a competitive advantage, organizations need to make “analytics” the way they do business; analytics needs to be a part of the corporate culture. Nowadays, the competitive advantage of data-driven organizations is no longer just a good ally, but a “must have” and a “must do.” The range of analytical capabilities emerging with big data and the fact that businesses can be modeled and forecasted is becoming a common practice Analytics need not be left to silos of teams, but rather made a part of the day-to-day operational function of front-end staff.