The Changing Role and Importance of Data Engineering

Over the last decade, the role of the data engineer has evolved drastically. Industry shifts such as the emergence of scalable processing and cloud data warehouses mean that today’s data is collected more frequently, is far less siloed and more accessible to everyone across an organisation than ever before. The role of the engineer has become integral, as these systems require increasingly sophisticated infrastructure.

This growth in significance, responsibilities and available tools means that today it can sometimes be difficult to define the data engineer role, and many companies actually differ in what they expect from their engineering talent. This creates a confusing landscape for data engineers — how can you tell what a prospective employer will require of you or how to categorise their own data engineering capabilities when the field now encompasses a broad swathe of work?

With this in mind, we have gathered our top pieces of advice for data engineers to help them keep pace with an ever-changing role, charting how the position has changed and how to work alongside data science teams to deploy the best standards of practice.

Data engineering, then and now

Ten years ago, data engineering consisted of data warehousing, business intelligence, and ETL (tooling to move data from place to place). There was limited development of pipelines for analytical models, but for the majority of industries these often remained in the ‘proof of concept’ phase due to no clear path to production.

In order to remain competitive, let alone stay ahead of the game, today’s companies are eager to invest in analytics and productionise models. The sheer size and quantity of data has also increased, giving scalability added importance.

With the changing attitudes and priorities of businesses, there are a couple of key focuses for today’s data engineers to thrive in this modern landscape:

Best practices in the software development lifecycle (SDLC) — These include proper use of version control, release management, and automated devOps pipelines.

Information security — With the ever-increasing threat of hacks, data engineers need to understand cloud security best practices and be vigilant in handling data. This includes managing data privacy in an evolving regulatory landscape (e.g. GDPR, etc).

Data architecture principles — These have always been important in data engineering, and include separation of concerns, degree of logical groupings, traceability, lineage, and well-defined data models.

Business domain knowledge — Domain expertise is increasingly required to draw insights from data.

Three trends that changed the role of the Data Engineer

Alongside the above areas that are key for all data engineers to understand, we believe it is vital for them to be aware of the industry trends that have helped shape the role over the last decade. Understanding these developments in analytics will help engineers explain the value of their work in conversations with organisations, who may not be aware of these revolutionary advancements.

The rise of cloud

Cloud has finally reached a tipping point where even institutions such as finance and government, that have historically shied away, are embracing it. In the last 4 years alone, the market for cloud computing has doubled from ~$114B to ~$236B. Amazon Web Services has led the market over the past several years (currently at 33% market share) but Microsoft Azure (13%) and Google Cloud Platform (6%) are catching up.

The expansion of open source

Data engineering used to be dominated by closed-source, proprietary tools. Now we are seeing a growth of open source tools and, in many cases, a preference for these tools in data organizations. Open source libraries such as Spark and Tensorflow have become widespread and many organisations are seeking to minimise vendor or product lock-in.

The growth of data in scale

Companies simply have more data at their disposal now more than ever before, which makes it more important for data engineers to understand how to scale. More than 90% of the world’s data was created in the last few years. Data engineers need proficiency in tools that can help to quickly organize and assess this massive amount of data.

Organisation archetypes

Before working closely with an organisation, data engineers need to assess the existing infrastructure for analytics to see what level of onboarding will be required. For example, how you approach a complex project with a company inexperienced with data science and data engineering will be completely different from an organisation with an existing analytics platform.

Clients often fit one of four archetypes in regard to their data engineering capabilities. These are:

  1. Starter — No data science or data engineering capabilities, very limited infrastructure for analytics
  2. Siloed — Pockets of analytics capabilities, but uncoordinated and with messy infrastructure
  3. Potential — Has a functioning analytics centre of excellence (CoE) but team needs to build more capability or is missing process
  4. North Star — Mature data organisation, working analytics platform

In all of these archetypes, data engineering plays a critical role; it is often the make or break factor in organisations achieving their North Star in analytics. It is widely cited that 80% of data science consists of data engineering. However, even in companies that see analytics as a competitive differentiator and have it on their agenda, we often see ineffective internal data organisation. This leads to a conflict between upper management, who wish to see value from analytics, and the reality on the ground, mired in a challenging and technical data environment.

Unlocking Analytics through Data Engineering

For analytics teams to work efficiently with organisations and convince them to invest in data engineering, no matter the existing analytics capability, it is important to show the long term value that engineering talent can bring to the table. Data engineers can “unlock” data science and analytics in an organisation, as well as build well curated, accessible data foundations.

Data engineering and data science should work hand in hand.

In short, companies must consciously invest in developing their data engineering capability in order to have a truly successful analytics program. This includes setting a solid foundation with data governance — identifying gaps and quality issues while improving data collection. The organisations that thrive in the years ahead will be those that not only acknowledge the challenges of developing and productionising models, but actively invest in engineer talent in order to maximise value from data.

Source : QuantumBlack

One comment

Leave a Reply

Fill in your details below or click an icon to log in: Logo

You are commenting using your account. Log Out /  Change )

Twitter picture

You are commenting using your Twitter account. Log Out /  Change )

Facebook photo

You are commenting using your Facebook account. Log Out /  Change )

Connecting to %s