data infrastructure – Dataconomy https://dataconomy.ru Bridging the gap between technology and business Wed, 16 Oct 2019 16:33:09 +0000 en-US hourly 1 https://dataconomy.ru/wp-content/uploads/2022/12/DC-logo-emblem_multicolor-75x75.png data infrastructure – Dataconomy https://dataconomy.ru 32 32 A Data Scientist’s relationship with building Predictive Models https://dataconomy.ru/2019/10/16/a-data-scientists-relationship-with-building-predictive-models/ https://dataconomy.ru/2019/10/16/a-data-scientists-relationship-with-building-predictive-models/#comments Wed, 16 Oct 2019 18:00:44 +0000 https://dataconomy.ru/?p=20671 If you’re a Data Scientist, you’ve likely spent months earnestly developing and then deploying a single predictive model. The truth is that once your model is built – that’s only half the battle won. A quarter of a Data Scientist’s working life often goes something like this: You met with business stakeholders to scope the […]]]>

If you’re a Data Scientist, you’ve likely spent months earnestly developing and then deploying a single predictive model. The truth is that once your model is built – that’s only half the battle won.

A quarter of a Data Scientist’s working life often goes something like this: You met with business stakeholders to scope the model and what it should do. You gathered, ingested, explored, and prepped the data. You iteratively built, tested, and tweaked the model. And then — just when you finally hit the AUC (Area Under the Curve)  threshold you had been targeting — you shared it with business stakeholders, and chances are, it wasn’t exactly what they had in mind.

So, you started the process over again. And finally, after countless iterations and reviews, your model was ready for production.

From there, you worked with the Engineering or the IT team to operationalize the model — whether that meant building an app, integrating into another system, or serving insights to business decision-makers through a chart or graph. Chances are, your code had to be rewritten in another programming language to meet the requirements of your production environment. But you worked through it and — ta-da! — your model is running.

Organizations that invest in Data Science can and should expect that a lot of time and energy will be dedicated to a single model before it even starts to impact the business. But then what? What happens to a model once it’s been deployed, whether serving insights to humans or triggering automated workflows that directly impact end customers?  

Organizations’ management of models once they’re in production is critical to maximizing their impact.

Most Data Scientists today would say the core of their job is building a model. Team’s incentive structures often reflect this, with one Data Scientist saying, “I get paid for what I build this year, not maintaining what I built last year.”

Once a model has been deployed in production, its ownership transfers to either business IT or data science management. But too often, those tasked with managing production models are not equipped to keep close tabs on how they’re using key resources or maintain visibility to their models in production. The default is a “set it and forget it” mentality. This is dangerous and severely limits the impact of an organization’s data science efforts.

A Data Scientist’s relationship with building Predictive Models

In the context of a broader model management framework, we refer to the pillar that allows an organization to keep a finger on the pulse of the activity, cost, and impact of each model as “model governance.”

Governance is key to any mission-critical system, however governing a growing, complex systems of models is particularly difficult for a few reasons:

  • Rapidly evolving toolkits: Models use computationally-intensive algorithms that benefit from scalable compute and specialized hardware like GPUs, and they leverage packages from a vibrant and constantly innovating ecosystem. Data Scientists need extremely agile technology infrastructure to accelerate research. Most enterprise IT organizations are accustomed to provisioning new servers in a flexible and automated manner leveraging Continuous Integration – Continuous Deployment (CI/CD) processes, but Data Scientists aren’t used to following CI/CD processes or including DevOps in the model building cycle until they are ready to deploy. When IT engineers can’t respond to an overwhelming volume of requests immediately before a critical deployment, data scientists seek to create their own shadow IT to support their models.

  • Research-based development: The process to develop models is different from well-established approaches to software engineering or data management. Data Science is research — it’s experimental, iterative, and exploratory. You might try dozens or hundreds of ideas before getting something that works. In software development, such false starts and dead ends are not preserved. When you make a mistake, it’s a bug. In Data Science, a failure can be the genesis of the next breakthrough. IT organizations that presume their source control and database access systems are sufficient will fail to capture critical metadata and documentation.  
  • Probabilistic behavior: Unlike software, which implements a specification, models prescribe action based on a probabilistic assessment. Statistician George Box captured the difference well saying, “All models are wrong but some are useful.” Models have no “correct” answer — they just have better or worse answers once they’re live in the real world. And while nobody needs to “retrain” software, models should change as the world changes around them. Organizations need to plan for rapid iteration and establish tight feedback loops with stakeholders to mitigate the risk of model drift.

Given these unique characteristics of models, existing processes and technology for managing them are often insufficient, leaving organizations vulnerable to inefficiencies and risks. The result? Long cycle times to deploy a revolving door of “buggy” models that don’t communicate with the rest of the technology ecosystem and don’t reflect the current state of the world. This directly threatens the business’ competitive advantage that could have been achieved through effective model management. This is why a new capability of model governance is critical.

What is model governance?

Model governance is the Data Science management function that provides visibility into Data Science projects, models, and their underlying infrastructure. High-functioning Data Science organizations implement model governance to keep tabs on models that are in development and running in production to ensure they’re doing what they’re supposed to be doing, and impacting the business as they should.

While governance sounds antithetical to the experimental ideals of data science, it is required to make sure the Data Science team is delivering business value and mitigating risk that can undermine the transformative potential of models.

How will model governance change your world?

  • Data Science leaders gain real-time transparency into the aggregate model portfolio and impact, rather than wondering exactly how many models are in-development and bemoaning perpetually outdated model inventories. Transparency into models can also help Data Science leaders quickly identify and address model bias before it presents a problem.

  • Data Scientists — especially those early in their careers — can clearly see how their work is used (and potentially misused). They no longer underestimate the risks of models or wonder how their work fits into the broader organization.

  • IT establishes an alignment with Data Science, and both groups gain granular knowledge of where key resources are used and how they can be used more efficiently. Clashes between the two groups stemming from difficulties forecasting compute resources and software, leading to missed budgets or wasted resources, are minimized or eliminated entirely. Concerns over security of the data used in models are also alleviated with complete provenance provided by model governance.

  • Infrastructure teams get real-time mapping of their model graph, encompassing all of the dependencies and linkages across critical system artifacts. They no longer have to deal with CACE (change anything, change everything) problems and the unknown risk to downstream models and systems.

Who is ultimately responsible for this?

Multiple stakeholders across the business should be involved to ensure model governance is successful — from Data Science practitioners to IT to line of business stakeholders to compliance teams.  The Data Science leader who is tasked with growing Data Science as a function within most companies is responsible for establishing and enforcing a model governance policy.

How do I get started?

Investing in a holistic model management strategy — that emphasizes model governance — leads to maximizing the impact of Data Science across the organization. As you think about model governance, in particular, you should first tackle challenges around model visibility. Here are some important tasks that can get you started on the right path:

  • Build and keep an inventory of all models that are in production.
  • Identify the production models that haven’t been reviewed and/or updated in a long time. What’s considered a “long time” is relative to your business and the unique purpose of each model, but generally speaking, three months is a good benchmark. Pay particular attention to those models operating in situations that may have changed significantly since the model was built.
  • Get stakeholders across the business involved, and work together to agree on a feedback mechanism that can be standardized to streamline improvements to production models moving forward.
  • Maintain an audit trail of all models in production and how they were built. Whenever changes are made to a model, track that in your audit log. This is a best practice for knowledge management generally and is required if you operate in a regulated industry.
  • Keep track of not just the models themselves and their code derivations, but also the artifacts associated with them — such as charts, graphs, interesting insights, or even feedback provided along the way from stakeholders.
  • Consider investing in a data science platform that can streamline and automate many of these tasks.  

Final Thoughts

Building models continues to be a critical element of a Data Scientist’s job. However, for companies looking to rapidly scale their organizations and build a competitive advantage with Data Science, model governance should be at the top of their mind. It’s better to build it proactively than wait until you are responding to a crisis. Not only does model governance mitigate downside risks, but it also helps your organization become increasingly productive as the organization grows.

]]>
https://dataconomy.ru/2019/10/16/a-data-scientists-relationship-with-building-predictive-models/feed/ 4
GDPR and the skills gap that could cost you €20million https://dataconomy.ru/2017/08/23/gdpr-skills-gap/ https://dataconomy.ru/2017/08/23/gdpr-skills-gap/#comments Wed, 23 Aug 2017 08:00:04 +0000 https://dataconomy.ru/?p=18291 It’s not new news that Europe is suffering from a near-chronic skills gap. It’s been going on for a while now, with industry experts and government bodies all scratching their heads over how to solve it. The problem is about to get a whole lot worse, as the soon-to-be-enforced General Data Protection Regulation (GDPR) will […]]]>

It’s not new news that Europe is suffering from a near-chronic skills gap. It’s been going on for a while now, with industry experts and government bodies all scratching their heads over how to solve it. The problem is about to get a whole lot worse, as the soon-to-be-enforced General Data Protection Regulation (GDPR) will turn the skills gap into a chasm.

GDPR breaches can cost up to €20million

Under GDPR, your business could be one mistake away from a breach that could cost you up to €20million or 4% of your global revenue. A fine-worthy breach includes data hacks, loss or misuse of data. ‘Misuse’ covers a number of sins that employees lacking in the right technical knowledge risk committing every day.

To make the issue even worse, to prepare for GDPR itself you’re going to need staff that are adequately trained in data protection, data management and GDPR compliance. As GDPR affects nearly every company in the EU (and those that do business with EU citizens), people who have knowledge of all of the above are going to be in extremely high demand.

Data literacy training from within

For some companies, it will make sense to nurture and develop staff from within. A quick search online brings up several GDPR training courses you can send your IT team and other technical staff on so they can get clued up on GDPR and its requirements. Some businesses will need a dedicated Data Protection Officer (DPO). Again, because of the vast number of businesses affected, DPOs are going to be in high demand. One solution for smaller businesses that cannot afford to fight for a DPO is to appoint a third-party who can act like one.

There is the further issue of your staff potentially leaking customer data, misusing it or storing it incorrectly. As part of your GDPR preparations, you will have to ensure all staff are aware of GDPR, its implications and what GDPR-compliance looks like. You’ll have to go into detail over what constitutes a breach, as well as put in place policies on bring-your-own equipment and data governance that all staff will have to be trained in.

There’s no one-size-fits-all type of training that will speak to all your employees. Therefore, you should consider holding a few different training sessions with your employees based on how tech literate they are and how clued up they are on GDPR. You’ll also have to schedule in regular refresher sessions in case anything changes and to really ensure compliance and include GDPR in induction sessions for new employees. The emphasis here is to do several different levels of workshop, however, a fresh faced graduate who has grown up surrounded by email, social media and smartphones is going to handle GDPR readiness very differently to someone who isn’t quite as confident with technology.

Organiztations should focus on solid data infrastructures

Between setting up employee training and finding yourself a DPO, it’s very easy to forget about the main preparations for GDPR readiness. That is, getting your data infrastructure up to standard as well. Privacy by design will become the default approach, where you hold the minimum amount of data for the task you need to carry out. Likewise, you’ll need to carry out a data audit to ensure all your data is stored correctly and securely, is easily transferable when requested and has all the required consent.

When beginning your GDPR preparations, you should firstly take a long hard look at your data infrastructure before training your staff. However, you don’t want a chicken and egg situation on your hands if your staff aren’t skilled enough to audit your infrastructure in the first place. This rings especially true for smaller businesses and start-ups. Again, to save yourself the hassle of trying to hire someone, it’s worth considering third parties who can audit your systems for you.

Conclusion

There’s a significant amount of work to be put in before the May 2018 deadline when GDPR is enforced. The skills gap only makes this mountain harder to climb. It makes sense to begin your preparations now before the impending deadline and skills gap makes it a sellers’ market. Invest in the right people now and it’ll pay off in the long run – and at the very least save your business from that €20 million fine.

Like this article? Subscribe to our weekly newsletter to never miss out!

]]>
https://dataconomy.ru/2017/08/23/gdpr-skills-gap/feed/ 1