How to get the enterprise data strategy 'right' (Part I) : The roadmap to reliable data integration

There is immense untapped potential in the data universe 

There is data flowing relentlessly resulting from the ever expansive information exchange between its creators and consumers .Organizations are more or less baffled by the spooky fluctuations in businesses processes,workflow and enterprise routines that is created by this massive flux of data in the conduct of businesses.Indeed its not a thing of coincidence that the lifecycle of every business processes can be punctuated as a function of data.However, how to qualify, evaluate and improve the multifarious and heterogeneous data for a comprehensive data management and governance strategy is the essential question.

Organizations mostly take a very broad strategic approach to data management.To gain access to specific business information as opposed to unrefined bulk data streaming through API interfaces, there needs to be robust data formatting and integrity assessments.A more prudent way is to take a tactical approach to data integration, wherein the goal is to build a target data ecosystem and then establish cross-departmental links .This implies that each functional silo of business, focuses on its specific user-value to streamline.

A good starting point is asking how well conducted and managed is your ETL pipeline. Rather than over-focussing on best tools (choices like redshift vs kafka) to modify data based on advanced AI-driven cloud infrastructures, you should first get your basics dialled aka deeper understanding your business challenges and conforming to the best practices of ETL.This will only establish the right conducive environment for big data systems to take charge of producing quality analytics. While detecting anomalies its in best interest to devise more expressive integrity constraints and automate creation of dynamic master data catalogues.A few of the most critical criteria to consider are -

1.Data maintenance scoping from request for new features, broken data fixing, addition of new connections etc

2. Understanding transformation needs and keeping it user-oriented( in this case the various business functions)

3. Scale the ETL process for future anticipated changes and needs. The problem with combating these challenges is fundamentally rooted in highly volatile and changing nature of input data, changing requirements and business rules,.

Now having done the ground work, we need to transition towards the objective of setting up the control landscape for data management.Here in this series, I analyse the data pipeline through 3 data-qualifying conceptsLineage and provenance , metadata consistency, integrity and interoperability (based on conceptualization of specific ontologies).

I shall discuss lineage and provenance and its paramount impact on data management and comprehensive data strategy in the next blog post in the same series.

Stay tuned!

Comments

Popular posts from this blog

Becoming a Leader : A growth based maturity model

Value-added data-asset lifecycle management (Part II) : Enterprise capability development

Blockchain, multi-agents and dynamic learning in supply chains