Our Manifesto

three dotsthree dots

Enterprises deserve better data science

The data science industry is at an inflection point. We are surrounded by high-velocity innovation and new AI technologies that have the potential to create entirely new paradigms for how we work, communicate, and even make decisions.

This momentum is certainly a testament to the incredible talent and ambition in our industry. But as we push forward, we must acknowledge that the long-term success of enterprise data science is held back by critical challenges–be it the technology-first mindset that tempts business leaders to replace proven processes with AI tools that promise magic but ultimately deliver opacity, or spiraling pay-as-you-go costs that hamper economies of scale and all-in strategies that lead to vendor lock-in.

In light of these industry trends, we would like to suggest a different way forward for enterprise data science; one that turns data science into the industrial-grade practice it deserves to be and one that empowers enterprises to own their data science and ultimately achieve return on the money invested.

5 challenges holding back industrial-grade data science

Let’s be brutally honest with ourselves: the practice of industrial-grade data science has not yet achieved its full potential. If we, the data science industry, want to realize the long-term success of enterprise data science, we must ambitiously tackle the challenges that we face. Consider the following five which we at Probabl believe are critical.

1. The rising tide of technoleogy-first thinking
New technologies have the potential to create new paradigms and AI tooling momentum tells enterprises that legacy applications and processes must be replaced because AI will surpass human creativity and productivity. When innovation in data science doesn’t allow you to understand and reuse your existing experiments and models, it creates technical debt and amplifies costs.

2. The pay-as-you-go trap
On-demand pricing has become the norm. Pay for compute. Pay for GPUs. Pay for tokens. Costs spiral out of control. Budget forecasting becomes impossible. Scaling your business no longer creates economies of scale, it creates uncontrollable OPEX expansion. When you give your suppliers open-ended access to your bank account, your expansion generates their profits, not yours.

3. All-in strategies create lock-in
Cloud-only, GPU-only or AI-only sound like modern and decisive strategies. They create strategic dependencies that contradict long-term value creation. When you lose autonomy, you lose freedom of movement, and your infrastructure decisions become vendor lock-in.

4. Data science has not reached the industrial maturity it deserves
Machine learning models rarely make it to production. Experiments are lost when team members leave, and reproducibility remains an aspiration rather than a standard. The discipline has grown in adoption but not always in rigor. Practitioners still reinvent wheels, lack shared quality standards, and operate without the engineering discipline that data science deserves. When most data science work never delivers business value because it can't scale beyond notebooks and proof-of-concepts, it's the lack of industrial-grade practice.

5. Scientific thinking has been forgotten
Data science is not software engineering. It requires a different discipline. Adopting new data science technologies should not undermine peer review, explainability, and ultimately trust. It should not create opacity and remove your control over business-critical systems. Because methodology matters, because statistical rigor matters, because explainability and understanding of your models matters, you should not rush to replace scientific discipline with automated tools that promise magic but deliver opacity.

Another way forward: Bringing the science of data to the world

To tackle these challenges, we must take a pragmatic shift to the practice of data science. At Probabl, we advocate firmly for an approach that is built on the following principles.

A. Transparency and explainability lead to ownership, trust, and impact
When you understand and can see how your models work, you can improve and trust them. Trust enables confidence in your decisions and accountability in your results. Understanding drives business value and competitive advantage.

B. Composability leads to agility and independence
We believe in agility and independence. By choosing tools that are modular and plug into your existing stack, you retain the freedom to adapt to change and choose the best tool for each specific use case. This ensures you control your destiny and pay the right price rather than being forced into a walled garden or vendor lock-in.

C. Reusability leads to economies of scale
Innovation should not mean that your existing investments become obsolete. When past experiments and models are treated as building blocks, you can build on experience and create true long-term value.

D. Science first
Data science was born from the scientific method–hypothesis, experimentation, measurement, and peer review. These foundations are precisely why data science creates value for enterprises. When science comes first, you start with the problem, not the tool. You validate before you deploy. You question before you trust. Methodology should drive tooling, not the other way around. At Probabl, we advocate for starting with the problem rather than the tool, validating before you deploy, and ensuring methodology drives your tooling, not the other way around.

By returning to these principles, we can move away from automated tools that promise magic but deliver opacity, and return to the rigor and strategic autonomy that business-critical systems require.