The FINANCIAL — After decades of false starts and unfulfilled expectations, artificial intelligence has now gone mainstream. Companies are successfully applying AI to a wide variety of current and novel processes, products, and services, and this success comes none too soon; AI is essential for business to address the ballooning complexity brought on by digitization and big data.
Visionary executives have started to imagine the potential for implementing AI throughout their companies. For example, BCG estimates that AI could help the top ten banks generate an additional $150 billion to $220 billion in annual operating earnings.
As pioneers across industries strive to reap these rewards by scaling up AI, however, they are stumbling against what we call the “AI paradox”: it is deceptively easy to launch AI pilots and achieve powerful results. But it is fiendishly hard to move toward “AI@scale.” All sorts of problems arise, threatening to undercut the AI revolution at its inception, according to BCG.
The paradox is easy to explain but hard to resolve. AI forces business executives to deal simultaneously with technology infrastructure as well as more traditional business issues. The core challenge is a tightening Gordian knot. Typical IT systems consist of data input, a tool, and data output. These systems are relatively easy to modularize, encapsulate, and scale. But AI systems are not so simple. AI algorithms learn by ingesting data—the training data is an integral part of the AI tool and the overall system. This entanglement is manageable during pilots and isolated uses but becomes exponentially more difficult to address as AI systems interact and build upon one another.
What sounds like a mere technical issue has multifaceted implications and challenges for companies. For example, vendor management becomes both more strategic and more complex. For the foreseeable future, vendors will play a large role in deploying AI because they have hired so much AI talent and have even promised to help with entanglement. Companies need to prepare for working with AI vendors in ways that do not put their data at risk or create long-term dependencies but instead strengthen competitive advantage.
People challenges also loom large for AI@scale. On one hand, companies confront a scarcity of AI data scientists and systems engineers; on the other hand, current employees are concerned about interacting with machines and even about losing their jobs, according to BCG.
Also, the organizational demands of AI require a delicate three-dimensional balancing act. Data governance, core AI expertise, and system management should be centralized. The development of use cases, learning, and training should occur in business units or functions and be managed by agile, cross-functional teams. Finally, AI action remains decentralized—in the marketplace, on the shop floor, or in the field.
Below, we provide a roadmap to systematically resolve the AI paradox. The AI@scale program is a full transformation to a new operating model.
A SYSTEMATIC APPROACH TO AI SCALE
While strategy and AI use cases are not the focus of this publication, they are critical to AI@scale transformations. So that is where we startcale
Strategy for AI Scale
Strategy setting in an era of disruption is daunting, and AI is the epitome of disruption. So while it makes sense to initiate an AI strategy within or close to the current business model, companies must also take into account the evolution of value pools in the wider market as other players embrace AI.
The first step in that process involves studying the potential and overall effect of AI use cases for both internal processes and external offerings. When we ran such an analysis for the top ten banks worldwide, it generated the estimate (referenced earlier) of up to $220 billion in additional annual operating earnings with AI.
HOW AI COULD FATTEN THE BOTTOM LINE
The second step—understanding the evolution of value pools—is equally important. The rest of the world is not standing still, passively watching these rewards being realized. Other companies both within and outside a company’s traditional industry are focusing on how AI can help them attack value pools.
In a prior publication, we illustrated potential value shifts in the health care industry among biopharma companies, insurers, providers, medtech companies, new entrants from the technology industry, and consumers. (See Putting Artificial Intelligence to Work, BCG report, September 2017.) A foreshadowing of such shifts occurred in January 2018 when the stocks of major retail pharmacy chains and insurers dropped by up to 5% relative to the S&P 500 after Amazon, JPMorgan Chase, and Berkshire Hathaway announced their entry into health care. Financial services, media, retail, as well as mobility and logistics are other industries susceptible to major AI-triggered disruption, according to BCG.
Two warnings are in order in setting strategy.
It’s not all about amassing data. While critical for the success of AI, internal data is not the Holy Grail. For one thing, companies increasingly will need to negotiate access to data outside their walls—from suppliers, partners, and customers, for example. Also, new approaches are emerging that avoid the need for massive troves of data. Generative adversarial networks (GANs), for instance, are virtual learning environments in which algorithms train one another to solve problems—such as creating and discriminating false pictures of faces. “Digital twins,” meanwhile, are virtual replicas of physical assets that enable algorithms to optimize end-to-end operations, for example, without the same real-world data requirements of traditional AI.
It’s also not enough to be agile and fast. Flexibility and speed are important, but they are not excuses to engage in a random walk through the AI wilds. To achieve competitive advantage, companies need to point in a specific direction. As research we have conducted with the MIT Sloan Management Review shows, the gap between companies believing in the transformational potential of AI and those actually having an AI strategy in place is stunning.
Identifying, Prioritizing, and Deploying AI Use Cases
Companies that have successfully piloted several AI use cases can get seduced into thinking that the move toward AI@scale is not that hard. Use cases can be powerful levers to boost performance. And they can help companies adopt agile test-and-learn methods; build transparency and cybersecurity measures into AI systems; and satisfy internal and regulatory requirements, according to BCG.
But isolated use cases provide limited guidance for the challenges of navigating through the AI paradox. They can sputter and grind to a halt when interacting at scale unless companies transform their operating models.
As a consequence, companies should initially prune the number of use cases, focusing on the most promising. (See “How to Address AI Use Cases.”) There is peril—not safety—in numbers when scaling up AI use cases and pilots.
HOW TO ADDRESS AI USE CASES
PREPARING FOR AI SCALE
Against that backdrop of strategy, priority setting, and piloting, let’s explore how companies can plan for a successful AI@scale transformation.
Creating a Robust AI Architecture
IT is often treated as a standalone function that is brought in to implement decisions already made by business leaders. This approach does not work with AI. If executives want to deploy machine intelligence to solve essential business problems, they must prioritize machine architecture equally to “people architecture” or traditional organizational issues. Otherwise, the entanglement concerns described earlier will disrupt their plans.
As companies move beyond the pilot phase, AI algorithms systemically interact with one another by learning from data, as explained earlier. Researchers at Google raised the entanglement issue several years ago, leading many prominent AI users to develop comprehensive machine-learning architectures. For example, Uber’s Michelangelo, in the words of the company, is “an end-to-end system that enables users across the company to easily build and operate machine learning systems at scale.” Likewise, a leading European online fashion retailer has built an AI platform that allows individual AI systems to work together and reuse components.
Vendors such as Amazon, Google, IBM, and Microsoft have all now launched or announced platform offerings to AI users. Whether organizations buy one of these emerging platforms from a vendor or build their own, companies that want to deploy AI at scale need a rigorous and consistent framework and system to manage, document, and monitor workflow from data input to final action. In addition to addressing issues such as entanglement and reusability, for example, companies need to monitor the security of these systems as well as the integrity and compliance of automated actions generated by AI.
Moreover, companies need to ensure that they have the storage, computing, and bandwidth to handle multiple AI engines and their timely actions. Owing to its flexibility, the cloud is often a preferred option to address these needs. In industrial automation and Internet of Things networks, however, latency and bandwidth constraints may prevent the cloud from serving as a complete solution. These settings require novel structures, such as edge computing, in which part of the processing power is kept closer to the action in the periphery, according to BCG.
Ultimately, the intelligent algorithms are a relatively small part at the core of the overall system. Consequently, many companies expanding their AI operations need only a small number of data scientists and AI experts at the core but require a large number of data and systems engineers with AI experience to ensure the performance and resilience of the pipeline and peripheral systems. For instance, with experienced data engineers and architects, a leading oil and gas company has built its own integrated-operations center, leveraging AI to enable predictive maintenance and optimize production. The projected efficiency gains amount to several billion dollars over the next few years.
Structuring an Ecosystem
The question is not if companies will work with vendors and partners but how. At least for the short term, it is their only way to access necessary AI talent and capabilities. Even companies with strong AI backgrounds rely to some extent on vendors.
With AI, however, these relationships could jeopardize core sources of competitive advantage for the companies. Vendors need to train their AI tools, often using sensitive client data.
Companies can work with AI vendors in many ways, from outsourcing an entire process to buying selected services, seeking help in building in-house solutions, or training internal staff. Executives should view these options in light of two questions:
How valuable is the process or offering to our future success?
How strong is our ownership, control, or access to high-quality, unique data, relative to the AI vendor’s?
By analyzing the AI landscape this way, companies will discover that their AI efforts land in one of four quadrants that each requires a different strategy:
Commodities (low-value potential, low differentiated data access). Manage vendor relationships to reduce costs and improve performance in support functions and elsewhere.
Danger Zones (high-value potential, low differentiated data access). Develop or acquire competitive data sources to turn danger zones into gold mines.
Hidden Opportunities (low-value potential, high differentiated data access). Leverage technical expertise of vendors to generate quick wins and insights.
Gold Mines (high-value potential, high differentiated data access). Capitalize on superior data access in strategically relevant areas and develop AI independently.
Beyond the critical issues of trust and confidentiality, the overall commercial offerings of vendors—including their newly launched platforms—differ widely in capabilities, integration abilities, and ease of use. So companies need sufficiently solid AI capabilities to select and manage these relationships. Many business leaders ultimately choose to hire an independent advisor to assess the tradeoffs while they build up the internal capability to bring work in-house.
Developing People, Skills, and Processes
Companies moving toward AI@scale should carefully consider how the transformation will affect their workforce through the creation of skills shortages, elimination of jobs, or both. This complex topic requires intensive strategic workforce planning, retraining, and process re-engineering to accommodate a human-machine world.
The most immediate need is critical technical skills. While many companies focus on hiring data scientists, they are encountering a greater shortage of people who have both business skills and an understanding of AI, as well as systems and data engineers, as mentioned earlier.
Longer term, companies need to understand the fundamental and nuanced shifts in the workplace that occur as humans and machines work side by side. When AI is focused on the augmentation of employees in their current roles, changes to the workforce and processes are relatively modest. Indeed, the introduction of augmentation in many internal processes and service functions often improves employees’ experience. In financial services, for example, the use of AI is reducing the burden, while enhancing the quality of many processes.
These changes become more extreme as processes start being reimagined for true human-machine collaboration. In radiology, doctors and hospitals are already starting to prepare their processes and business models for a time when regulators accept machines as a second opinion—and ultimately a first opinion—for diagnostics, with a human in the loop to correct potential errors.
Finally, in what we call “hyperlearning,” machines are capable of teaching themselves in a virtual environment with little or no human intervention and without relying on massive data input from the real world. As discussed earlier, GANs and digital twins are such environments. It is conceivable that algorithms will not need human training or even real-world data in these environments in the near future. This evolution has already occurred in the relatively restricted environment of chess but also, for instance, in flight simulators for drones and end-to-end supply chain optimization.
As companies move along this continuum, they will need to address new and novel change management and reskilling demands. Renault has already embarked on this path with its digital transformation and with the creation of its Digital Hub, a large center dedicated to the development of digital products, training and innovation, according to BCG.
Corporate Center. Certain types of expertise and governance should be centralized. These include excellence hubs that house experts in many of the (horizontal) AI building blocks—such as machine vision, natural-language processing, and some general frontier topics in machine learning—and support AI initiatives and vendor assessment throughout the company.
Data is the raw material of AI, but it also contains some of the company’s most sensitive information. A world-class data governance function, which sets data permissions across the organization, is critical in an AI world, to ensure both competitive advantage and regulatory compliance.
Finally, data architects, who manage global topics such as industrialization frameworks and platforms as well as vendors, should be centralized along with cybersecurity experts.
Business Units and Functions. As a general rule, cross-functional agile teams situated within business units or functions should be responsible for the development of AI-enabled processes, products, and services. These teams should include both AI experts and topic specialists. Roving “SWAT teams,” including HR experts and change specialists, should also be available to help decentralized units affected by AI use cases with implementation, training, and other issues.
Field Level. The employees responsible for managing the processes and actions modified by AI are, of course, decentralized, located close to the market, shop floor, or field. These units will need to understand the new tools and how they affect processes and skill requirements and will carry the principal burden of change. The SWAT teams will be available to bring these people up to speed on how to leverage the new technology and assist with the transformation.
PULLING TOGETHER THE TRANSFORMATION PROGRAM
An AI scale transformation should occur through a series of top-down and bottom-up actions to create alignment, buy-in, and follow-through. This has ensured the successful industrialization of AI across companies and their value chains:
AI Ambition and Maturity Assessment. This top-down step establishes the overall context of the transformation and helps prevent the company from pursuing disconnected AI pilots. The maturity assessment is typically based on a combination of surveys and interviews.
Evaluation of AI Initiatives and the Operating Model. This bottom-up step provides a baseline of current AI initiatives. It should include goals, business cases, accountabilities, work streams, and milestones in addition to an analysis of data management, algorithms, performance metrics, and cybersecurity. A review of the current AI operating model should also be conducted at this stage.
Priority Setting and Gap Analysis. The next top-down step prioritizes AI initiatives, focusing on easy wins and unicorns. This step also identifies the required changes to the operating model.
Discussion about this post