I write about national security, especially its business dimensions.
Aug 14, 2023, 11:02am EDT
On August 11, the editors of Aviation Week & Space Technology posted a podcast with the provocative title, “Why AmazonAMZN Could Be The Next Big Defense Prime.” The discussion wasn’t so much about Amazon as about how software-driven projects increasingly are shaping military modernization plans.
AvWeek’s Chief Technology Editor, Graham Warwick, noted in the exchange that the Air Force’s next-generation fighter will be largely defined by its software, without which the aircraft would be unable to meet its performance requirements.
Some observers believe that the evolution of combat aircraft is inexorably progressing towards a future in which human pilots will no longer be part of the design, and software will enable every facet of aircraft operations—presumably with speed and precision that no human operator could match.
It is easy to overstate the degree to which software is eclipsing hardware in current military plans. What good is the source code without the plane? However, the Pentagon’s latest software modernization strategy, released last year, expresses what looks to be the conventional wisdom in asserting “software increasingly defines military capabilities.”
Viewed from this perspective, the Pentagon’s recent embrace of artificial intelligence is just the latest chapter in a long-running trend that is concentrating military power in algorithms and code rather than human hands.
Pentagon officials insist that life-and-death decisions will never be turned over to machines, but the reality is that if adversaries like China follow the same course, at some point the only alternative to automating warfare may be to accept defeat.
Here are five reasons why software increasingly dominates the thinking of military visionaries, sometimes to the exclusion of traditional industrial (and human) processes.
Software enhances the performance of hardware. The electronic content of warfighting systems has been growing for generations, and with the advent of the digital revolution, that content increasingly includes high-power computers that run applications software for onboard systems.
The result has been huge gains in functionality. Don’t take my word for it, just compare the performance of your current iPhone with the cellphone you used ten years ago. The incredible versatility concentrated in this compact device is largely enabled by software, and supported by a network that is itself software-driven.
A similar dynamic applies to military technology. The 75 improvements that will be integrated in the next round of F-35 fighter upgrades will depend on agile software running on more powerful computers. The same is true of ongoing upgrades to the Aegis combat system on destroyers—without the underlying software, they would be literally impossible.
Software takes less time to develop. Today’s military software can contain millions of lines of code, but it is easier to develop and field than hardware. It typically consists of modular building blocks constructed according to open architecture principles, and much of the code is itself generated using software. In other words, the generation of software is automated in a way that construction of fighters or warships cannot be.
A senior shipbuilding executive involved in the construction of warships once remarked to me that the state of play in his yards reflected decisions made by Congress seven years earlier. That tells you something about how long it takes to build complex military hardware. The Air Force’s F-35 became operational 15 years after the contract was awarded.
Software generally is developed and fielded according to more compressed timelines. In fact, all the steps from design to development to testing to installation can be accomplished in a fraction of the time required for new hardware. So, the acquisition system naturally defaults to software as the preferred way of upgrading capabilities.
Software is less expensive to implement. Developing and producing new military hardware involves big investments in capital equipment and the creation of articulated supply chains. A skilled workforce must be trained to integrate components unique to a particular program.
Such challenges are not unheard of in generating software, but they usually require far less financial resources to overcome. One reason is that coding of software often is fungible across diverse applications and industries—hence AvWeek’s notion that Amazon skills might be useful in advancing military capabilities. The Pentagon’s search for commercial technologies relevant to warfighting is grounded largely in leveraging private-sector software skills for new uses.
Software can take the place of costly personnel. Replacing humans with software may raise ethical concerns for the warfighting profession, but it potentially has big budgetary benefits. Retired Major General Arnold Punaro, a legendary Washington insider, figures that the fully loaded cost of a single soldier in the All-Volunteer Force is $400,000 annually. Even at that steep price, the services are having trouble attracting new recruits.
Many military jobs can be performed more cost-effectively by using software in imaginative ways. That is particularly true with the advent of artificial intelligence programs using deep-learning processes. With the federal government spending a trillion dollars more than it takes in each year, the fiscal appeal of substituting software for people will become increasingly attractive—and not just in the military.
Software lowers barriers to entry. Policymakers frequently complain that high barriers to entry in the defense industry limit options for introducing new products and processes. Greater reliance on software potentially ameliorates this problem, because there are hundreds of successful commercial software firms that can apply their skills to military tasks. Even when they perform as subcontractors to traditional primes, such firms can stimulate the adoption of new ideas.
The above five considerations barely scratch the surface of reasons why software is eclipsing hardware in military technology plans. As Grahan Warwick points out in the August 11 podcast, even when the subject is hardware, the underlying processes (like prototyping) are increasingly software-driven. The digital revolution is transforming the technological landscape, and agile software has become the coin of the realm.
Several agencies partnered to release the first federal recommendations for organizations to begin upgrading their networks and systems to quantum cryptography-resilient security schemes.
Federal agencies are leading the charge to usher in the shift to post-quantum cryptography standards, releasing an authoritative factsheet Monday surrounding PQC standards and the impact of quantum information technologies.
The Quantum-Readiness: Migration to Post-Quantum Cryptography fact sheet — co-released by the Cybersecurity and Infrastructure Security Agency, the National Security Agency and the National Institutes of Standards and Technology — lays out a roadmap for public and private entities to use as quantum computing technologies continue to advance and potentially threaten the standard cryptographic schemes that safeguards modern digital infrastructure.
Among the recommendations included in the roadmap is an emphasis on close communication between organizations and technology vendors — a reflection of the Biden administration’s goal of fostering better partnerships between the public and private sectors.
“It is imperative for all organizations, especially critical infrastructure, to begin preparing now for migration to post-quantum cryptography,” said CISA Director Jen Easterly in a statement. “CISA will continue to work with our federal and industry partners to unify and drive efforts to address threats posed by quantum computing. Our collective aim is to ensure that public and private sector organizations have the resources and capabilities necessary to effectively prepare and manage this transition.”
Upgrading today’s cryptography to be able to withstand a potential attack from a fault-tolerant quantum computer is a long overhaul. CISA and NIST recommend in their new document to begin by having firms analyze which parts of their network systems and assets rely on quantum-vulnerable cryptography; that is, which network components create and validate security measures like digital signatures.
“Having an inventory of quantum-vulnerable systems and assets enables an organization to begin the quantum risk assessment processes, demonstrating the prioritization of migration,” the document says.
The agencies prioritize technology vendors’ role in facilitating migration efforts. Given that PQC migration will involve software and occasional firmware updates, the document prompts vendors to chart their own timelines for PQC migration efforts for successful product integration.
“The authoring agencies also urge organizations to proactively plan for necessary changes to existing and future contracts,” the document says. “Considerations should be in place ensuring that new products will be delivered with PQC built-in, and older products will be upgraded with PQC to meet transition timelines.”
This new guidance supports the timelineimposed by a Biden administration memorandum, which requests that government agencies modernize their networks to PQC standards by the year 2035.
“Post-quantum cryptography is about proactively developing and building capabilities to secure critical information and systems from being compromised through the use of quantum computers,” said Rob Joyce, director of NSA Cybersecurity. “The transition to a secured quantum computing era is a long-term intensive community effort that will require extensive collaboration between government and industry. The key is to be on this journey today and not wait until the last minute.”
Too many companies bet on having a cut-throat, high-pressure, take-no-prisoners culture to drive their financial success.
But a large and growing body of research on positive organizational psychology demonstrates that not only is a cut-throat environment harmful to productivity over time, but that a positive environment will lead to dramatic benefits for employers, employees, and the bottom line.
Although there’s an assumption that stress and pressure push employees to perform more, better, and faster, what cutthroat organizations fail to recognize is the hidden costs incurred.
First, health care expenditures at high-pressure companies are nearly 50% greater than at other organizations. The American Psychological Association estimates that more than $500 billion is siphoned off from the U.S. economy because of workplace stress, and 550 million workdays are lost each year due to stress on the job. Sixty percent to 80% of workplace accidents are attributed to stress, and it’s estimated that more than 80% of doctor visits are due to stress. Workplace stress has been linked to health problems ranging from metabolic syndrome to cardiovascular disease and mortality.
The stress of belonging to hierarchies itself is linked to disease and death. One study showed that, the lower someone’s rank in a hierarchy, the higher their chances of cardiovascular disease and death from heart attacks. In a large-scale study of over 3,000 employees conducted by Anna Nyberg at the Karolinska Institute, results showed a strong link between leadership behavior and heart disease in employees. Stress-producing bosses are literally bad for the heart.
Second is the cost of disengagement. While a cut-throat environment and a culture of fear can ensure engagement (and sometimes even excitement) for some time, research suggests that the inevitable stress it creates will likely lead to disengagement over the long term. Engagement in work — which is associated with feeling valued, secure, supported, and respected — is generally negatively associated with a high-stress, cut-throat culture.
And disengagement is costly. In studies by the Queens School of Business and by the Gallup Organization, disengaged workers had 37% higher absenteeism, 49% more accidents, and 60% more errors and defects. In organizations with low employee engagement scores, they experienced 18% lower productivity, 16% lower profitability, 37% lower job growth, and 65% lower share price over time. Importantly, businesses with highly engaged employees enjoyed 100% more job applications.
Lack of loyalty is a third cost. Research shows that workplace stress leads to an increase of almost 50% in voluntary turnover. People go on the job market, decline promotions, or resign. And the turnover costs associated with recruiting, training, lowered productivity, lost expertise, and so forth, are significant. The Center for American Progress estimatesthat replacing a single employee costs approximately 20% of that employee’s salary.
Wellbeing comes from one place, and one place only — a positive culture.
Creating a positive and healthy culture for your team rests on a few major principles. Our own research (see here and here) on the qualities of a positive workplace culture boils down to six essential characteristics:
Caring for, being interested in, and maintaining responsibility for colleagues as friends.
Providing support for one another, including offering kindness and compassion when others are struggling.
Avoiding blame and forgive mistakes.
Inspiring one another at work.
Emphasizing the meaningfulness of the work.
Treating one another with respect, gratitude, trust, and integrity.
As a boss, how can you foster these principles? The research points to four steps to try:
1. Foster social connections. A large number of empirical studies confirm that positive social connections at work produce highly desirable results. For example, people get sick less often, recover twice as fast from surgery, experience less depression, learn faster and remember longer, tolerate pain and discomfort better, display more mental acuity, and perform better on the job. Conversely, research by Sarah Pressman at the University of California, Irvine, found that the probability of dying early is 20% higher for obese people, 30% higher for excessive drinkers, 50% higher for smokers, but a whopping 70% higher for people with poor social relationships. Toxic, stress-filled workplaces affect social relationships and, consequently, life expectancy.
2. Show empathy. As a boss, you have a huge impact on how your employees feel. A telling brain-imaging study found that, when employees recalled a boss that had been unkind or un-empathic, they showed increased activation in areas of the brain associated with avoidance and negative emotion while the opposite was true when they recalled an empathic boss. Moreover, Jane Dutton and her colleagues in the CompassionLab at the University of Michigan suggest that leaders who demonstrate compassion toward employees foster individual and collective resilience in challenging times.
3. Go out of your way to help.Ever had a manager or mentor who took a lot of trouble to help you when he or she did not have to? Chances are you have remained loyal to that person to this day. Jonathan Haidt at New York University’s Stern School of Business shows in his research that when leaders are not just fair but self-sacrificing, their employees are actually moved and inspired to become more loyal and committed themselves. As a consequence, they are more likely to go out of their way to be helpful and friendly to other employees, thus creating a self-reinforcing cycle. Daan Van Knippenberg of Rotterdam School of Management shows that employees of self-sacrificing leaders are more cooperative because they trust their leaders more. They are also more productive and see their leaders as more effective and charismatic.
4. Encourage people to talk to you – especially about their problems. Not surprisingly, trusting that the leader has your best interests at heart improves employee performance. Employees feel safe rather than fearful and, as research by Amy Edmondson of Harvard demonstrates in her work on psychological safety, a culture of safety i.e. in which leaders are inclusive, humble, and encourage their staff to speak up or ask for help, leads to better learning and performance outcomes. Rather than creating a culture of fear of negative consequences, feeling safe in the workplace helps encourage the spirit of experimentation so critical for innovation. Kamal Birdi of Sheffield University has shownthat empowerment, when coupled with good training and teamwork, leads to superior performance outcomes whereas a range of efficient manufacturing and operations practices do not.
When you know a leader is committed to operating from a set of values based on interpersonal kindness, he or she sets the tone for the entire organization. In Give and Take, Wharton professor Adam Grant demonstrates that leader kindness and generosity are strong predictors of team and organizational effectiveness. Whereas harsh work climates are linked to poorer employee health, the opposite is true of positive work climates where employees tend to have lower heart rates and blood pressure as well as a stronger immune systems. A positive work climate also leads to a positive workplace culture which, again, boosts commitment, engagement, and performance. Happier employees make for not only a more congenial workplace but for improved customer service. As a consequence, a happy and caring culture at work not only improves employee well-being and productivity but also improved client health outcomes and satisfaction.
In sum, a positive workplace is more successful over time because it increases positive emotions and well-being. This, in turn, improves people’s relationships with each other and amplifies their abilities and their creativity. It buffers against negative experiences such as stress, thus improving employees’ ability to bounce back from challenges and difficulties while bolstering their health. And, it attracts employees, making them more loyal to the leader and to the organization as well as bringing out their best strengths. When organizations develop positive, virtuous cultures they achieve significantly higher levels of organizational effectiveness — including financial performance, customer satisfaction, productivity, and employee engagement.
Editor’s note : Due to a typo, this article initially misstated the number of workdays lost due to stress each year. That number is estimated at 550 million, not 550 billion. The sentence has been corrected.
Measuring, tracking, and benchmarking developer productivity has long been considered a black box. It doesn’t have to be that way.
Measuring, tracking, and benchmarking developer productivity has long been considered a black box. It doesn’t have to be that way.
Compared with other critical business functions such as sales or customer operations, software development is perennially undermeasured. The long-held belief by many in tech is that it’s not possible to do it correctly—and that, in any case, only trained engineers are knowledgeable enough to assess the performance of their peers. Yet that status quo is no longer sustainable. Now that most companies are becoming (to one degree or another) software companies, regardless of industry, leaders need to know they are deploying their most valuable talent as successfully as possible.
Sidebar
About the authors
This article is a collaborative effort by Chandra Gnanasambandam, Martin Harrysson, Alharith Hussin, Jason Keovichit, and Shivam Srivastava, representing views from McKinsey’s Digital and Technology, Media & Telecommunications Practices.
There is no denying that measuring developer productivity is difficult. Other functions can be measured reasonably well, some even with just a single metric; whereas in software development, the link between inputs and outputs is considerably less clear. Software development is also highly collaborative, complex, and creative work and requires different metrics for different levels (such as systems, teams, and individuals). What’s more, even if there is genuine commitment to track productivity properly, traditional metrics can require systems and software that are set up to allow more nuanced and comprehensive measurement. For some standard metrics, entire tech stacks and development pipelines need to be reconfigured to enable tracking, and putting in place the necessary instruments and tools to yield meaningful insights can require significant, long-term investment. Furthermore, the landscape of software development is changing quickly as generative AI tools such as CopilotX and ChatGPT have the potential to enable developers to complete tasks up to two times faster.
To help overcome these challenges and make this critical task more feasible, we developed an approach to measuring software developer productivity that is easier to deploy with surveys or existing data (such as in backlog management tools). In so doing, we built on the foundation of existing productivity metrics that industry leaders have developed over the years, with an eye toward revealing opportunities for performance improvements.
This new approach has been implemented at nearly 20 tech, finance, and pharmaceutical companies, and the initial results are promising. They include the following improvements:
20 to 30 percent reduction in customer-reported product defects
20 percent improvement in employee experience scores
60-percentage-point improvement in customer satisfaction ratings
Leveraging productivity insights
With access to richer productivity data and insights, leaders can begin to answer pressing questions about the software engineering talent they fought so hard to attract and retain, such as the following:
What are the impediments to the engineers working at their best level?
How much does culture and organization affect their ability to produce their best work?
How do we know if we’re using their time on activities that truly drive value?
How can we know if we have all the software engineering talent we need?
Understanding the foundations
To use a sufficiently nuanced system of measuring developer productivity, it’s essential to understand the three types of metrics that need to be tracked: those at the system level, the team level, and the individual level. Unlike a function such as sales, where a system-level metric of dollars earned or deals closed could be used to measure the work of both teams and individuals, software development is collaborative in a distinctive way that requires different lenses. For instance, while deployment frequency is a perfectly good metric to assess systems or teams, it depends on all team members doing their respective tasks and is, therefore, not a useful way to track individual performance.
Another critical dimension to recognize is what the various metrics do and do not tell you. For example, measuring deployment frequency or lead time for changes can give you a clear view of certain outcomes, but not of whether an engineering organization is optimized. And while metrics such as story points completed or interruptions can help determine optimization, they require more investigation to identify improvements that might be beneficial.
In building our set of metrics, we looked to expand on the two sets of metrics already developed by the software industry. The first is DORA metrics, named for Google’s DevOps research and assessment team. These are the closest the tech sector has to a standard, and they are great at measuring outcomes. When a DORA metric returns a subpar outcome, it is a signal to investigate what has gone wrong, which can often involve protracted sleuthing. For example, if a metric such as deployment frequency increases or decreases, there can be multiple causes. Determining what they are and how to resolve them is often not straightforward.
The second set of industry-developed measurements is SPACE metrics (satisfaction and well-being, performance, activity, communication and collaboration, and efficiency and flow), which GitHub and Microsoft Research developed to augment DORA metrics. By adopting an individual lens, particularly around developer well-being, SPACE metrics are great at clarifying whether an engineering organization is optimized. For example, an increase in interruptions that developers experience indicates a need for optimization.
On top of these already powerful metrics, our approach seeks to identify what can be done to improve how products are delivered and what those improvements are worth, without the need for heavy instrumentation. Complementing DORA and SPACE metrics with opportunity-focused metrics can create an end-to-end view of software developer productivity (Exhibit 1).
These opportunity-focused productivity metrics use a few different lenses to generate a nuanced view of the complex range of activities involved with software product development.
Inner/outer loop time spent. To identify specific areas for improvement, it’s helpful to think of the activities involved in software development as being arranged in two loops (Exhibit 2). An inner loop comprises activities directly related to creating the product: coding, building, and unit testing. An outer loop comprises other tasks developers must do to push their code to production: integration, integration testing, releasing, and deployment. From both a productivity and personal-experience standpoint, maximizing the amount of time developers spend in the inner loop is desirable: building products directly generates value and is what most developers are excited to do. Outer-loop activities are seen by most developers as necessary but generally unsatisfying chores. Putting time into better tooling and automation for the outer loop allows developers to spend more time on inner-loop activities.
Top tech companies aim for developers to spend up to 70 percent of their time doing inner-loop activities. For example, one company that had previously completed a successful agile transformation learned that its developers, instead of coding, were spending too much time on low-value-added tasks such as provisioning infrastructure, running manual unit tests, and managing test data. Armed with that insight, it launched a series of new tools and automation projects to help with those tasks across the software development life cycle.
Developer Velocity Index benchmark. The Developer Velocity Index (DVI) is a survey that measures an enterprise’s technology, working practices, and organizational enablement and benchmarks them against peers. This comparison helps unearth specific areas of opportunity, whether in backlog management, testing, or security and compliance.1 For example, one company, well known for its technological prowess and all-star developers, sought to define standard working practices more thoughtfully for cross-team collaboration after discovering a high amount of dissatisfaction, rework, and inefficiency reported by developers.
Contribution analysis. Assessing contributions by individuals to a team’s backlog (starting with data from backlog management tools such as Jira, and normalizing data using a proprietary algorithm to account for nuances) can help surface trends that inhibit the optimization of that team’s capacity. This kind of insight can enable team leaders to manage clear expectations for output and improve performance as a result. Additionally, it can help identify opportunities for individual upskilling or training and rethinking role distribution within a team (for instance, if a quality assurance tester has enough work to do). For example, one company found that its most talented developers were spending excessive time on noncoding activities such as design sessions or managing interdependencies across teams. In response, the company changed its operating model and clarified roles and responsibilities to enable those highest-value developers to do what they do best: code. Another company, after discovering relatively low contribution from developers new to the organization, reexamined their onboarding and personal mentorship program.
Talent capability score. Based on industry standard capability maps, this score is a summary of the individual knowledge, skills, and abilities of a specific organization. Ideally, organizations should aspire to a “diamond” distribution of proficiency, with the majority of developers in the middle range of competency.2 This can surface coaching and upskilling opportunities, and in extreme cases call for a rethinking of talent strategy. For example, one company found a higher concentration of their developers in the “novice” capability than was ideal. They deployed personalized learning journeys based on specific gaps and were able to move 30 percent of their developers to the next level of expertise within six months.
Avoiding metrics missteps
As valuable as it can be, developer productivity data can be damaging to organizations if used incorrectly, so it’s important to avoid certain pitfalls. In our work we see two main types of missteps occur: misuse of metrics and failing to move past old mindsets.
Misuse is most common when companies try to employ overly simple measurements, such as lines of code produced, or number of code commits (when developers submit their code to a version control system). Not only do such simple metrics fail to generate truly useful insights, they can have unintended consequences, such as leaders making inappropriate trade-offs. For example, optimizing for lead time or deployment frequency can allow quality to suffer. Focusing on a single metric or too simple a collection of metrics can also easily incentivize poor practices; in the case of measuring commits, for instance, developers may submit smaller changes more frequently as they seek to game the system.
To truly benefit from measuring productivity, leaders and developers alike need to move past the outdated notion that leaders “cannot” understand the intricacies of software engineering, or that engineering is too complex to measure. The importance of engineering talent to a company’s success, and the fierce competition for developer talent in recent years, underscores the need to acknowledge that software development, like so many other things, requires measurement to be improved. Further, attracting and retaining top software development talent depends in large part on providing a workplace and tools that allow engineers to do their best work and encourages their creativity. Measuring productivity at a system level enables employers to see hidden friction points that impede that work and creativity.
Getting started
The mechanics of building a developer productivity initiative can seem daunting, but there is no time like the present to begin to lay the groundwork. The factors driving the need to elevate the conversation about software developer productivity to C-level leaders outweigh the impediments to doing so.
The increase in remote work and its popularity among developers is one overriding factor. Developers have long worked in agile teams, collaborating in the same physical space, and some technology leaders believe that kind of in-person teamwork is essential to the job. However, the digital tools that are so central to their work made it easy to switch to remote work during the pandemic lockdowns, and as in most sectors, this shift is hard to undo. As remote and hybrid working increasingly becomes the norm, organizations will need to rely on broad, objective measurements to maintain confidence in these new working arrangements and ensure they are steadily improving the function that could easily determine their future success or failure. The fact that the markets are now putting greater emphasis on efficient growth and ROI only makes it more important than ever to know how they can optimize the performance of their highly valued engineering talent.
Another key driver of this need for greater visibility is the rapid advances in AI-enabled tooling, especially large-language models such as generative AI. These are already rapidly changing the way work is done, which means that measuring software developers’ productivity is only a first step to understanding how these valuable resources are deployed.
But as critical as developer productivity is becoming, companies shouldn’t feel they have to embark on a massive, dramatic overhaul almost overnight. Instead, they can begin the process with a number of key steps:
Learn the basics. All C-suite leaders who are not engineers or who have been in management for a long time will need a primer on the software development process and how it is evolving.
Assess your systems. Because developer productivity has not typically been measured at the level needed to identify improvement opportunities, most companies’ tech stacks will require potentially extensive reconfiguration. For example, to measure test coverage (the extent to which areas of code have been adequately tested), a development team needs to equip their codebase with a tool that can track code executed during a test run.
Build a plan. As with most analytics initiatives, getting lost in mountains of data is a risk. It’s important to start with one area that you know will result in a clear path to improvement, such as identifying friction points and bottlenecks. Be explicit about the scope of such a plan, as even the best approaches, no matter how comprehensive, will not be a silver bullet.
Remember that measuring productivity is contextual. The point is to look at an entire system and understand how it can work better by improving the development environment at the system, team, or individual level.
No matter the specific approach, measuring productivity should ideally create transparency and insights into key improvement areas. Only then can organizations build specific initiatives to drive impact for both developer productivity and experience—impact that will benefit both those individuals and the company as a whole.
ABOUT THE AUTHOR(S)
Chandra Gnanasambandam and Martin Harrysson are senior partners in McKinsey’s Bay Area office, where Alharith Hussin and Shivam Srivastava are partners; and Jason Keovichit is an associate partner in the New York office.
The authors wish to thank Pedro Garcia, Diana Rodriguez, and Jeremy Schneider for their contributions to this article.
Tech corporation IBM has unveiled a new “prototype” of an analog AI chip that works like a human brain and performs complex computations in various deep neural networks (DNN) tasks.
The chip promises more. IBM says the state-of-the-art chip can make artificial intelligence remarkably efficient and less battery-draining for computers and smartphones.
Introducing the chip in a paperpublished by IBM Research, the company said: “The fully integrated chip features 64 AIMC cores interconnected via an on-chip communication network. It also implements the digital activation functions and additional processing involved in individual convolutional layers and long short-term memory units.”
Reinventing ways in which AI is computed
The new AI chip is developed in IBM’s Albany NanoTech Complex and comprises 64 analog in-memory compute cores. By borrowing key features of how neural networks run in biological brains, IBM explains that it has embedded the chip with compact, time-based analog-to-digital converters in each tile or core to transition between the analog and digital worlds.
Each tile (or core) is also integrated with lightweight digital processing units that perform simple nonlinear neuronal activation functions and scaling operations, explained IBM in a blog published on August 10.
A replacement for current digital chips?
In the future, IBM’s prototype chip could replace the current chips powering heavy AI applications in computers and phones. “A global digital processing unit is integrated into the middle of the chip that implements more complex operations that are critical for the execution of certain types of neural networks,” further said the blog.
With more and more foundation models and generative AI tools entering the market, the performance and energy efficiency of traditional computing methods that these models run on are at a testing limit.
IBM wants to bridge that gap. The company says that many of the chips being developed today have a split in their memory and processing units, thus slowing down computation. “This means the AI models are typically stored in a discrete memory location, and computational tasks require constantly shuffling data between the memory and processing units.”
Speaking to BBC, Thanos Vasilopoulos, a scientist based at IBM’s research lab in Switzerland, compared the human brain to traditional computers and said that the former “is able to achieve remarkable performance while consuming little power.”
He said that the superior energy efficiency (of the IBM chip) would mean “large and more complex workloads could be executed in low power or battery-constrained environments”, for example, cars, mobile phones, and cameras.
“Additionally, cloud providers will be able to use these chips to reduce energy costs and their carbon footprint,” he added.
Attributable to Deputy Secretary of Defense Kathleen Hicks:
The Department of Defense must meet the urgency of today’s threats and tomorrow’s challenges with innovation in all portfolios — including how we build and execute our budget. This is critical not only to maintain the trust of the American taxpayer, but also to ensure that DoD can rapidly transition, integrate, and deliver cutting-edge capabilities to the warfighter at speed and scale.
It’s no secret that DoD’s resource allocation process was born in the industrial age, and the Department has undertaken significant reforms to improve it. The recommendations made in the Commission on Planning, Programming, Budgeting, and Execution (PPBE) Reform’s Interim Report released today will further these efforts. I am directing the Department to adopt all actions that can be implemented now, as recommended by the Commission and within its purview. We look forward to working with Congress on all other proposed recommendations included in this interim report.
The PPBE Reform Commission’s work is immensely important to assisting Congress and DoD in the nation’s efforts to stay ahead of the pacing challenge, ensuring our agility in fielding combat credible forces at speed and scale. I stand ready to provide continued DoD cooperation with the Commission and to receive its final report in March 2024.
September/October 2023 Published on August 16, 2023
It’s 2035, and artificial intelligence is everywhere. AI systems run hospitals, operate airlines, and battle each other in the courtroom. Productivity has spiked to unprecedented levels, and countless previously unimaginable businesses have scaled at blistering speed, generating immense advances in well-being. New products, cures, and innovations hit the market daily, as science and technology kick into overdrive. And yet the world is growing both more unpredictable and more fragile, as terrorists find new ways to menace societies with intelligent, evolving cyberweapons and white-collar workers lose their jobs en masse.
Just a year ago, that scenario would have seemed purely fictional; today, it seems nearly inevitable. Generative AIsystems can already write more clearly and persuasively than most humans and can produce original images, art, and even computer code based on simple language prompts. And generative AI is only the tip of the iceberg. Its arrival marks a Big Bang moment, the beginning of a world-changing technological revolution that will remake politics, economies, and societies.
Like past technological waves, AI will pair extraordinary growth and opportunity with immense disruption and risk. But unlike previous waves, it will also initiate a seismic shift in the structure and balance of global power as it threatens the status of nation-states as the world’s primary geopolitical actors. Whether they admit it or not, AI’s creators are themselves geopolitical actors, and their sovereignty over AI further entrenches the emerging “technopolar” order—one in which technology companies wield the kind of power in their domains once reserved for nation-states. For the past decade, big technology firms have effectively become independent, sovereign actors in the digital realms they have created. AI accelerates this trend and extends it far beyond the digital world. The technology’s complexity and the speed of its advancement will make it almost impossible for governments to make relevant rules at a reasonable pace. If governments do not catch up soon, it is possible they never will.
Thankfully, policymakers around the world have begun to wake up to the challenges posed by AI and wrestle with how to govern it. In May 2023, the G-7 launched the “Hiroshima AI process,” a forum devoted to harmonizing AI governance. In June, the European Parliament passed a draft of the EU’s AI Act, the first comprehensive attempt by the European Union to erect safeguards around the AI industry. And in July, UN Secretary-General Antonio Guterres called for the establishment of a global AI regulatory watchdog. Meanwhile, in the United States, politicians on both sides of the aisle are calling for regulatory action. But many agree with Ted Cruz, the Republican senator from Texas, who concluded in June that Congress “doesn’t know what the hell it’s doing.”
Unfortunately, too much of the debate about AI governance remains trapped in a dangerous false dilemma: leverage artificial intelligence to expand national power or stifle it to avoid its risks. Even those who accurately diagnose the problem are trying to solve it by shoehorning AI into existing or historical governance frameworks. Yet AI cannot be governed like any previous technology, and it is already shifting traditional notions of geopolitical power.
The challenge is clear: to design a new governance framework fit for this unique technology. If global governance of AI is to become possible, the international system must move past traditional conceptions of sovereignty and welcome technology companies to the table. These actors may not derive legitimacy from a social contract, democracy, or the provision of public goods, but without them, effective AI governance will not stand a chance. This is one example of how the international community will need to rethink basic assumptions about the geopolitical order. But it is not the only one.
A challenge as unusual and pressing as AI demands an original solution. Before policymakers can begin to hash out an appropriate regulatory structure, they will need to agree on basic principles for how to govern AI. For starters, any governance framework will need to be precautionary, agile, inclusive, impermeable, and targeted. Building on these principles, policymakers should create at least three overlapping governance regimes: one for establishing facts and advising governments on the risks posed by AI, one for preventing an all-out arms race between them, and one for managing the disruptive forces of a technologyunlike anything the world has seen.
Like it or not, 2035 is coming. Whether it is defined by the positive advances enabled by AI or the negative disruptions caused by it depends on what policymakers do now.
FASTER, HIGHER, STRONGER
AI is different—different from other technologies and different in its effect on power. It does not just pose policy challenges; its hyper-evolutionary nature also makes solving those challenges progressively harder. That is the AI power paradox.
The pace of progress is staggering. Take Moore’s Law, which has successfully predicted the doubling of computing power every two years. The new wave of AI makes that rate of progress seem quaint. When OpenAI launched its first large language model, known as GPT-1, in 2018, it had 117 million parameters—a measure of the system’s scale and complexity. Five years later, the company’s fourth-generation model, GPT-4, is thought to have over a trillion. The amount of computation used to train the most powerful AI models has increased by a factor of ten every year for the last ten years. Put another way, today’s most advanced AI models—also known as “frontier” models—use five billiontimes the computing power of cutting-edge models from a decade ago. Processing that once took weeks now happens in seconds. Models that can handle tens of trillions of parameters are coming in the next couple of years. “Brain scale” models with more than 100 trillion parameters—roughly the number of synapses in the human brain—will be viable within five years.
With each new order of magnitude, unexpected capabilities emerge. Few predicted that training on raw text would enable large language models to produce coherent, novel, and even creative sentences. Fewer still expected language models to be able to compose music or solve scientific problems, as some now can. Soon, AI developers will likely succeed in creating systems with self-improving capabilities—a critical juncture in the trajectory of this technology that should give everyone pause.
AI models are also doing more with less. Yesterday’s cutting-edge capabilities are running on smaller, cheaper, and more accessible systems today. Just three years after OpenAI released GPT-3, open-source teams have created models capable of the same level of performance that are less than one-sixtieth of its size—that is, 60 times cheaper to run in production, entirely free, and available to everyone on the Internet. Future large language models will probably follow this efficiency trajectory, becoming available in open-source form just two or three years after leading AI labs spend hundreds of millions of dollars developing them.
As with any software or code, AI algorithms are much easier and cheaper to copy and share (or steal) than physical assets. Proliferation risks are obvious. Meta’s powerful Llama-1 large language model, for instance, leaked to the Internet within days of debuting in March. Although the most powerful models still require sophisticated hardware to work, midrange versions can run on computers that can be rented for a few dollars an hour. Soon, such models will run on smartphones. No technology this powerful has become so accessible, so widely, so quickly.
AI also differs from older technologies in that almost all of it can be characterized as “dual use”—having both military and civilian applications. Many systems are inherently general, and indeed, generality is the primary goal of many AI companies. They want their applications to help as many people in as many ways as possible. But the same systems that drive cars can drive tanks. An AI application built to diagnose diseases might be able to create—and weaponize—a new one. The boundaries between the safely civilian and the militarily destructive are inherently blurred, which partly explains why the United States has restricted the export of the most advanced semiconductors to China.
All this plays out on a global field: once released, AI models can and will be everywhere. And it will take just one malign or “breakout” model to wreak havoc. For that reason, regulating AI cannot be done in a patchwork manner. There is little use in regulating AI in some countries if it remains unregulated in others. Because AI can proliferate so easily, its governance can have no gaps.
What is more, the damage AI might do has no obvious cap, even as the incentives to build it (and the benefits of doing so) continue to grow. AI could be used to generate and spread toxic misinformation, eroding social trust and democracy; to surveil, manipulate, and subdue citizens, undermining individual and collective freedom; or to create powerful digital or physical weapons that threaten human lives. AI could also destroy millions of jobs, worsening existing inequalities and creating new ones; entrench discriminatory patterns and distort decision-making by amplifying bad information feedback loops; or spark unintended and uncontrollable military escalations that lead to war.
Nor is the time frame clear for the biggest risks. Online misinformation is an obvious short-term threat, just as autonomous warfare seems plausible in the medium term. Farther out on the horizon lurks the promise of artificial general intelligence, the still uncertain point where AI exceeds human performance at any given task, and the (admittedly speculative) peril that AGI could become self-directed, self-replicating, and self-improving beyond human control. All these dangers need to be factored into governance architecture from the outset.
AI is not the first technology with some of these potent characteristics, but it is the first to combine them all. AI systems are not like cars or airplanes, which are built on hardware amenable to incremental improvements and whose most costly failures come in the form of individual accidents. They are not like chemical or nuclear weapons, which are difficult and expensive to develop and store, let alone secretly share or deploy. As their enormous benefits become self-evident, AI systems will only grow bigger, better, cheaper, and more ubiquitous. They will even become capable of quasi autonomy—able to achieve concrete goals with minimal human oversight—and, potentially, of self-improvement. Any one of these features would challenge traditional governance models; all of them together render these models hopelessly inadequate.
TOO POWERFUL TO PAUSE
As if that were not enough, by shifting the structure and balance of global power, AI complicates the very political context in which it is governed. AI is not just software development as usual; it is an entirely new means of projecting power. In some cases, it will upend existing authorities; in others, it will entrench them. Moreover, its advancement is being propelled by irresistible incentives: every nation, corporation, and individual will want some version of it.
Within countries, AI will empower those who wield it to surveil, deceive, and even control populations—supercharging the collection and commercial use of personal data in democracies and sharpening the tools of repression authoritarian governments use to subdue their societies. Across countries, AI will be the focus of intense geopolitical competition. Whether for its repressive capabilities, economic potential, or military advantage, AI supremacy will be a strategic objective of every government with the resources to compete. The least imaginative strategies will pump money into homegrown AI champions or attempt to build and control supercomputers and algorithms. More nuanced strategies will foster specific competitive advantages, as France seeks to do by directly supporting AI startups; the United Kingdom, by capitalizing on its world-class universities and venture capital ecosystem; and the EU, by shaping the global conversation on regulation and norms.
The vast majority of countries have neither the money nor the technological know-how to compete for AI leadership. Their access to frontier AI will instead be determined by their relationships with a handful of already rich and powerful corporations and states. This dependence threatens to aggravate current geopolitical power imbalances. The most powerful governments will vie to control the world’s most valuable resource while, once again, countries in the global South will be left behind. This is not to say that only the richest will benefit from the AI revolution. Like the Internet and smartphones, AI will proliferate without respect for borders, as will the productivity gains it unleashes. And like energy and green technology, AI will benefit many countries that do not control it, including those that contribute to producing AI inputs such as semiconductors.
At the other end of the geopolitical spectrum, however, the competition for AI supremacy will be fierce. At the end of the Cold War, powerful countries might have cooperated to allay one another’s fears and arrest a potentially destabilizing technological arms race. But today’s tense geopolitical environment makes such cooperation much harder. AI is not just another tool or weapon that can bring prestige, power, or wealth. It has the potential to enable a significant military and economic advantage over adversaries. Rightly or wrongly, the two players that matter most—Chinaand the United States—both see AI development as a zero-sum game that will give the winner a decisive strategic edge in the decades to come.
China and the United States both see AI development as a zero-sum game.
From the vantage point of Washington and Beijing, the risk that the other side will gain an edge in AI is greater than any theoretical risk the technology might pose to society or to their own domestic political authority. For that reason, both the U.S. and Chinese governments are pouring immense resources into developing AI capabilities while working to deprive each other of the inputs needed for next-generation breakthroughs. (So far, the United States has been far more successful than China in doing the latter, especially with its export controls on advanced semiconductors.) This zero-sum dynamic—and the lack of trust on both sides—means that Beijing and Washington are focused on accelerating AI development, rather than slowing it down. In their view, a “pause” in development to assess risks, as some AI industry leaders have called for, would amount to foolish unilateral disarmament.
But this perspective assumes that states can assert and maintain at least some control over AI. This may be the case in China, which has integrated its tech companies into the fabric of the state. Yet in the West and elsewhere, AI is more likely to undermine state power than to bolster it. Outside China, a handful of large, specialist AI companies currently control every aspect of this new technological wave: what AI models can do, who can access them, how they can be used, and where they can be deployed. And because these companies jealously guard their computing power and algorithms, they alone understand (most of) what they are creating and (most of) what those creations can do. These few firms may retain their advantage for the foreseeable future—or they may be eclipsed by a raft of smaller players as low barriers to entry, open-source development, and near-zero marginal costs lead to uncontrolled proliferation of AI. Either way, the AI revolution will take place outside government.
To a limited degree, some of these challenges resemble those of earlier digital technologies. Internet platforms, social media, and even devices such as smartphones all operate, to some extent, within sandboxes controlled by their creators. When governments have summoned the political will, they have been able to implement regulatory regimes for these technologies, such as the EU’s General Data Protection Regulation, Digital Markets Act, and Digital Services Act. But such regulation took a decade or more to materialize in the EU, and it still has not fully materialized in the United States. AI moves far too quickly for policymakers to respond at their usual pace. Moreover, social media and other older digital technologies do not help create themselves, and the commercial and strategic interests driving them never dovetailed in quite the same way: Twitter and TikTok are powerful, but few think they could transform the global economy.
This all means that at least for the next few years, AI’s trajectory will be largely determined by the decisions of a handful of private businesses, regardless of what policymakers in Brussels or Washington do. In other words, technologists, not policymakers or bureaucrats, will exercise authority over a force that could profoundly alter both the power of nation-states and how they relate to each other. That makes the challenge of governing AI unlike anything governments have faced before, a regulatory balancing act more delicate—and more high stakes—than any policymakers have attempted.
MOVING TARGET, EVOLVING WEAPON
Governments are already behind the curve. Most proposals for governing AI treat it as a conventional problem amenable to the state-centric solutions of the twentieth century: compromises over rules hashed out by political leaders sitting around a table. But that will not work for AI.
Regulatory efforts to date are in their infancy and still inadequate. The EU’s AI Act is the most ambitious attempt to govern AI in any jurisdiction, but it will apply in full only beginning in 2026, by which time AI models will have advanced beyond recognition. The United Kingdom hasproposedan even looser, voluntary approach to regulating AI, but it lacks the teeth to be effective. Neither initiative attempts to govern AI development and deployment at the global level—something that will be necessary for AI governance to succeed. And while voluntary pledges to respect AI safety guidelines, such as those made in July by seven leading AI developers, including Inflection AI, led by one of us (Suleyman), are welcome, they are no substitute for legally binding national and international regulation.
Advocates for international-level agreements to tame AI tend to reach for the model of nuclear arms control. But AI systems are not only infinitely easier to develop, steal, and copy than nuclear weapons; they are controlled by private companies, not governments. As the new generation of AI models diffuses faster than ever, the nuclear comparison looks ever more out of date. Even if governments can successfully control access to the materials needed to build the most advanced models—as the Biden administration is attempting to do by preventing China from acquiring advanced chips—they can do little to stop the proliferation of those models once they are trained and therefore require far fewer chips to operate.
For global AI governance to work, it must be tailored to the specific nature of the technology, the challenges it poses, and the structure and balance of power in which it operates. But because the evolution, uses, risks, and rewards of AI are unpredictable, AI governance cannot be fully specified at the outset—or at any point in time, for that matter. It must be as innovative and evolutionary as the technology it seeks to govern, sharing some of the characteristics that make AI such a powerful force in the first place. That means starting from scratch, rethinking and rebuilding a new regulatory framework from the ground up.
The overarching goal of any global AI regulatory architecture should be to identify and mitigate risks to global stability without choking off AI innovation and the opportunities that flow from it. Call this approach “technoprudentialism,” a mandate rather like the macroprudential role played by global financial institutions such as the Financial Stability Board, the Bank of International Settlements, and the International Monetary Fund.Their objective is to identify and mitigate risks to global financial stability without jeopardizing economic growth.
A technoprudential mandate would work similarly, necessitating the creation of institutional mechanisms to address the various aspects of AI that could threaten geopolitical stability. These mechanisms, in turn, would be guided by common principles that are both tailored to AI’s unique features and reflect the new technological balance of power that has put tech companies in the driver’s seat. These principles would help policymakers draw up more granular regulatory frameworks to govern AI as it evolves and becomes a more pervasive force.
The first and perhaps most vital principle for AI governance is precaution.As the term implies, technoprudentialism is at its core guided by the precautionary credo: first, do no harm. Maximally constraining AI would mean forgoing its life-altering upsides, but maximally liberating it would mean risking all its potentially catastrophic downsides. In other words, the risk-reward profile for AI is asymmetric. Given the radical uncertainty about the scale and irreversibility of some of AI’s potential harms, AI governance must aim to prevent these risks before they materialize rather than mitigate them after the fact. This is especially important because AI could weaken democracy in some countries and make it harder for them to enact regulations. Moreover, the burden of proving an AI system is safe above some reasonable threshold should rest on the developer and owner; it should not be solely up to governments to deal with problems once they arise.
AI governance must also be agile so that it can adapt and correct course as AI evolves and improves itself. Public institutions often calcify to the point of being unable to adapt to change. And in the case of AI, the sheer velocity of technological progress will quickly overwhelm the ability of existing governance structures to catch up and keep up. This does not mean that AI governance should adopt the “move fast and break things” ethos of Silicon Valley, but it should more closely mirror the nature of the technology it seeks to contain.
In addition to being precautionary and agile, AI governance must be inclusive, inviting the participation of all actors needed to regulate AI in practice. That means AI governance cannot be exclusively state centered, since governments neither understand nor control AI. Private technology companies may lack sovereignty in the traditional sense, but they wield real—even sovereign—power and agency in the digital spaces they have created and effectively govern. These nonstate actors should not be granted the same rights and privileges as states, which are internationally recognized as acting on behalf of their citizens. But they should be parties to international summits and signatories to any agreements on AI.
Such a broadening of governance is necessary because any regulatory structure that excludes the real agents of AI power is doomed to fail. In previous waves of tech regulation, companies were often afforded so much leeway that they overstepped, leading policymakers and regulators to react harshly to their excesses. But this dynamic benefited neither tech companies nor the public. Inviting AI developers to participate in the rule-making process from the outset would help establish a more collaborative culture of AI governance, reducing the need to rein in these companies after the fact with costly and adversarial regulation.
AI is a problem of the global commons, not just the preserve of two superpowers.
Tech companies should not always have a say; some aspects of AI governance are best left to governments, and it goes without saying that states should always retain final veto power over policy decisions. Governments must also guard against regulatory capture to ensure that tech companies do not use their influence within political systems to advance their interests at the expense of the public good. But an inclusive, multistakeholder governance model would ensure that the actors who will determine the fate of AI are involved in—and bound by—the rule-making processes.In addition to governments (especially but not limited to China and the United States) and tech companies (especially but not limited to the Big Tech players), scientists, ethicists, trade unions, civil society organizations, and other voices with knowledge of, power over, or a stake in AI outcomes should have a seat at the table. The Partnership on AI—a nonprofit group that convenes a range of large tech companies, research institutions, charities, and civil society organizations to promote responsible AI use—is a good example of the kind of mixed, inclusive forum that is needed.
AI governance must also be as impermeable as possible. Unlike climate change mitigation, where success will be determined by the sum of all individual efforts, AI safety is determined by the lowest common denominator: a single breakout algorithm could cause untold damage. Because global AI governance is only as good as the worst-governed country, company, or technology, it must be watertight everywhere—with entry easy enough to compel participation and exit costly enough to deter noncompliance. A single loophole, weak link, or rogue defector will open the door to widespread leakage, bad actors, or a regulatory race to the bottom.
In addition to covering the entire globe, AI governance must cover the entire supply chain—from manufacturing to hardware, software to services, and providers to users. This means technoprudential regulation and oversight along every node of the AI value chain, from AI chip production to data collection, model training to end use, and across the entire stack of technologies used in a given application. Such impermeability will ensure there are no regulatory gray areas to exploit.
Finally, AI governance will need to be targeted, rather than one-size-fits-all. Because AI is a general-purpose technology, it poses multidimensional threats. A single governance tool is not sufficient to address the various sources of AI risk. In practice, determining which tools are appropriate to target which risks will require developing a living and breathing taxonomy of all the possible effects AI could have—and how each can best be governed. For example, AI will be evolutionary in some applications, exacerbating current problems such as privacy violations, and revolutionary in others, creating entirely new harms. Sometimes, the best place to intervene will be where data is being collected. Other times, it will be the point at which advanced chips are sold—ensuring they do not fall into the wrong hands. Dealing with disinformation and misinformation will require different tools than dealing with the risks of AGI and other uncertain technologies with potentially existential ramifications. A light regulatory touch and voluntary guidance will work in some cases; in others, governments will need to strictly enforce compliance.
All of this requires deep understanding and up-to-date knowledge of the technologies in question. Regulators and other authorities will need oversight of and access to key AI models. In effect, they will need an audit system that can not only track capabilities at a distance but also directly access core technologies, which in turn will require the right talent. Only such measures can ensure that new AI applications are proactively assessed, both for obvious risks and for potentially disruptive second- and third-order consequences. Targeted governance, in other words, must be well-informed governance.
THE TECHNOPRUDENTIAL IMPERATIVE
Built atop these principles should be a minimum of three AI governance regimes, each with different mandates, levers, and participants. All will have to be novel in design, but each could look for inspiration to existing arrangements for addressing other global challenges—namely, climate change, arms proliferation, and financial stability.
The first regime would focus on fact-finding and would take the form of a global scientific body to objectively advise governments and international bodies on questions as basic as what AI is and what kinds of policy challenges it poses. If no one can agree on the definition of AI or the possible scope of its harms, effective policymaking will be impossible. Here, climate change is instructive. To create a baseline of shared knowledge for climate negotiations, the United Nations established the Intergovernmental Panel on Climate Change and gave it a simple mandate: provide policymakers with “regular assessments of the scientific basis of climate change, its impacts and future risks, and options for adaptation and mitigation.” AI needs a similar body to regularly evaluate the state of AI, impartially assess its risks and potential impacts, forecast scenarios, and consider technical policy solutions to protect the global public interest. Like the IPCC, this body would have a global imprimatur and scientific (and geopolitical) independence. And its reports could inform multilateral and multistakeholder negotiations on AI, just as the IPCC’s reports inform UN climate negotiations.
The world also needs a way to manage tensions between the major AI powers and prevent the proliferation of dangerous advanced AI systems. The most important international relationship in AI is the one between the United States and China. Cooperation between the two rivals is difficult to achieve under the best of circumstances. But in the context of heightened geopolitical competition, an uncontrolled AI race could doom all hope of forging an international consensus on AI governance. One area where Washington and Beijing may find it advantageous to work together is in slowing the proliferation of powerful systems that could imperil the authority of nation-states. At the extreme, the threat of uncontrolled, self-replicating AGIs—should they be invented in the years to come—would provide strong incentives to coordinate on safety and containment.
On all these fronts, Washington and Beijing should aim to create areas of commonality and even guardrails proposed and policed by a third party. Here, the monitoring and verification approaches often found in arms control regimes might be applied to the most important AI inputs, specifically those related to computing hardware, including advanced semiconductors and data centers. Regulating key chokepoints helped contain a dangerous arms race during the Cold War, and it could help contain a potentially even more dangerous AI race now.
Few powerful constituencies favor containing AI.
But since much of AI is already decentralized, it is a problem of the global commons rather than the preserve of two superpowers. The devolved nature of AI development and core characteristics of the technology, such as open-source proliferation, increase the likelihood that it will be weaponized by cybercriminals, state-sponsored actors, and lone wolves. That is why the world needs a third AI governance regime that can react when dangerous disruptions occur. For models, policymakers might look to the approach financial authorities have used to maintain global financial stability. The Financial Stability Board, composed of central bankers, ministries of finance, and supervisory and regulatory authorities from around the world, works to prevent global financial instability by assessing systemic vulnerabilities and coordinating the necessary actions to address them among national and international authorities. A similarly technocratic body for AI risk—call it the Geotechnology Stability Board—could work to maintain geopolitical stability amid rapid AI-driven change. Supported by national regulatory authorities and international standard-setting bodies, it would pool expertise and resources to preempt or respond to AI-related crises, reducing the risk of contagion. But it would also engage directly with the private sector, recognizing that key multinational technology actors play a critical role in maintaining geopolitical stability, just as systemically important banks do in maintaining financial stability.
Such a body, with authority rooted in government support, would be well positioned to prevent global tech players from engaging in regulatory arbitrage or hiding behind corporate domiciles. Recognizing that some technology companies are systemically important does not mean stifling start-ups or emerging innovators. On the contrary, creating a single, direct line from a global governance body to these tech behemoths would enhance the effectiveness of regulatory enforcement and crisis management—both of which benefit the whole ecosystem.
A regime designed to maintain geotechnological stability would also fill a dangerous void in the current regulatory landscape: responsibility for governing open-source AI. Some level of online censorship will be necessary. If someone uploads an extremely dangerous model, this body must have the clear authority—and ability—to take it down or direct national authorities to do so. This is another area for potential bilateral cooperation. China and the United States should want to work together to embed safety constraints in open-source software—for example, by limiting the extent to which models can instruct users on how to develop chemical or biological weapons or create pandemic pathogens. In addition, there may be room for Beijing and Washington to cooperate on global antiproliferation efforts, including through the use of interventionist cybertools.
Each of these regimes would have to operate universally, enjoying the buy-in of all major AI players. The regimes would need to be specialized enough to cope with real AI systems and dynamic enough to keep updating their knowledge of AI as it evolves. Working together, these institutions could take a decisive step toward technoprudential management of the emerging AI world. But they are by no means the only institutions that will be needed. Other regulatory mechanisms, such as “know your customer” transparency standards, licensing requirements, safety testing protocols, and product registration and approval processes, will need to be applied to AI in the next few years. The key across all these ideas will be to create flexible, multifaceted governance institutions that are not constrained by tradition or lack of imagination—after all, technologists will not be constrained by those things.
PROMOTE THE BEST, PREVENT THE WORST
None of these solutions will be easy to implement. Despite all the buzz and chatter coming from world leaders about the need to regulate AI, there is still a lack of political will to do so. Right now, few powerful constituencies favor containing AI—and all incentives point toward continued inaction. But designed well, an AI governance regime of the kind described here could suit all interested parties, enshrining principles and structures that promote the best in AI while preventing the worst. The alternative—uncontained AI—would not just pose unacceptable risks to global stability; it would also be bad for business and run counter to every country’s national interest.
A strong AI governance regime would both mitigate the societal risks posed by AI and ease tensions between China and the United States by reducing the extent to which AI is an arena—and a tool—of geopolitical competition. And such a regime would achieve something even more profound and long-lasting: it would establish a model for how to address other disruptive, emerging technologies. AI may be a unique catalyst for change, but it is by no means the last disruptive technology humanity will face. Quantum computing, biotechnology, nanotechnology, and robotics also have the potential to fundamentally reshape the world. Successfully governing AI will help the world successfully govern those technologies as well.
The twenty-first century will throw up few challenges as daunting or opportunities as promising as those presented by AI. In the last century, policymakers began to build a global governance architecture that, they hoped, would be equal to the tasks of the age. Now, they must build a new governance architecture to contain and harness the most formidable, and potentially defining, force of this era. The year 2035 is just around the corner. There is no time to waste.
1. As artificial intelligence and machine learning continue to expand, the ability to engage in nuanced decision-making can help humans stay employable.
Amid nonstop technological change, leaders must commit to a program of efficient, continuous, rapid learning, said Vanessa Tanicien, director of client success at LifeLabs Learning, a management training consultancy.
“Machines are going to do a lot for us in the future, but one of the things that they will not be able to do are some of those human-flavored skills,” Tanicien said at the recent EmTech Next conference. “Take the time to double down on learning, because that’s what is going to keep us different [from] the machines.”
Tanicien highlighted two skills that are essential to continuous learning:
• Extraction
- the process of applying learning to diverse situations to understand root causes, events, and results and to identify systemic problems.
• Transfer
- intentionally moving knowledge from short- to long-term memory and then applying the new learning to other business situations.
2. It’s often said that economic advancement comes down to who you know. That might include relatively distant connections.
A recent paper co-authored by MIT Sloan professor Dean Eckles and Eaman Jahani, PhD ’21, details a positive correlation between evidence of economic well-being and a greater number of long ties — people who you know but with whom you share no other mutual connections. This could include a childhood neighbor, a former colleague, or a college roommate.
By examining public interactions among all Facebook users in the U.S. over a six-month period, the researchers found that people who move out of state, attend college out of their home state, or transfer to another high school tend to have more long ties and are more likely to be better off financially as a result.
Maintaining relationships with long ties takes work — and the disruption from moving can be difficult — but it results in a more diverse network and increased access to economic opportunities, the researchers found.
“If a close friend has valuable information, it’s likely the same valuable information that a lot of mutual friends have as well,” Eckles said. “Long ties give you access to information and opportunities you wouldn’t have otherwise.”
3. Would planting a trillion trees help to mitigate global warming? The idea has been touted by House Speaker Kevin McCarthy and other Republican leaders.
But new analysis by MIT Sloan professor John Sterman and Andrew P. Jones, executive director of the nonprofit Climate Interactive, indicates that the plan wouldn’t work, partly because of the long lag time for trees to reach maturity and absorb large amounts of carbon.
As reported in The Washington Post, the pair used the En-ROADS global climate simulator, developed by Climate Interactive and the MIT Sloan Sustainability Initiative, to determine that planting a trillion trees would prevent only 0.15 degrees Celsius (0.27 Fahrenheit) of warming by 2100.
The simulator also indicated that planting a trillion trees would sequester only 6% of the carbon dioxide that the world needs to avoid emitting by 2050 to meet the goal of the Paris climate accord.
“Trees are great,” Sterman told the Post. “But the reality is very simple: You can plant a trillion trees, and even if they all survived, which wouldn’t happen, it just wouldn’t make that much difference to the climate.”
+ IDEAS THAT MATTER
How the industrial metaverse helps solve real-world problems.
The industrial metaverse enables more resilient supply chains and localized manufacturing through the use of existing tools such as digital twins.
While the metaverse has been most closely associated with elevated shopping and entertainment experiences for consumers, the industrial sector has pragmatic designs on using it to improve collaboration, optimize operations, and train manufacturing talent.
The suite of tools for constructing the industrial metaverse includes digital twins, artificial intelligence, extended reality, blockchain, 5G connectivity, and cloud and edge computing.
At a recent event hosted by MIT Technology Review, experts discussed how the industrial metaverse links real and digital worlds to enable collaboration while solving real-world business problems. Potential use cases include:
• Improved design and engineering.
Team members from different departments or locations can work in a collaborative setting. Realistic simulations allow for more extensive testing and validation. •
Virtual commissioning and plant design.
By using immersive digital twins, manufacturers can design shop floors in the metaverse, enabling them to detect and correct errors without disrupting ongoing production or incurring unnecessary investment risks. • Enhanced operations. Through simulations and real-time data collection, manufacturers can garner insights to optimize equipment, minimize downtime, and predict and prevent failures.
• Upskilling of employees.
The metaverse gives employees remote access to expert skills and virtual training regardless of where they are physically located.
Governor Kathy Hochul today announced New York’s first-ever statewide cybersecurity strategyaimed at protecting the State’s digital infrastructure from today’s cyber threats. The Strategy articulates, for the first-time, a set of high-level objectives for cybersecurity and resilience across New York. It clarifies agency roles and responsibilities, outlines how existing and planned initiatives and investments knit together into a unified approach, and reiterates the State’s commitment to providing services, advice, and assistance to county and local governments. New York State’s cybersecurity strategy provides public and private stakeholders with a roadmap for cyber risk mitigation and outlines a plan to protect critical infrastructure, networks, data, and technology systems.
“Our interconnected world demands an interconnected defense leveraging every resource available,” said Governor Hochul. “This strategy sets forth a nation-leading blueprint to ensure New York State stands ready and resilient in the face of cyber threats.”
The strategy unifies New York’s cybersecurity services in order to safeguard critical infrastructure, personal information and digital assets from malicious actors. It also provides a framework to align the actions and resources of both private and public stakeholders, including county and other local governments.
New York’s cybersecurity strategy is not just about protecting our digital assets. It is about ensuring the safety and security of all New Yorkers and maintaining our ability to function and thrive in the digital age. From the State employees who deliver digital services to the residents who access and rely on them, the strength of the State’s cyber defense impacts all New Yorkers. This strategy highlights the Governor’s commitment to cybersecurity, not just for State Government systems but for New Yorkers everywhere, as a core responsibility of the State.
Governor Hochul announced her commitment to bolster New York State’s centralized cybersecurity during this year’s State of the State address. The historic $90 million investment for cybersecurity included in the Fiscal Year 2024 Budget made $30 million in shared services funding available to assist local governments in strengthening their own defenses against cyber threats. This initiative signaled a new and stronger collaboration between the state and its local governments on this critical and evolving issue. A part of this strategy includes providing $500 million to enhance New York State’s healthcare information technology, primarily cybersecurity infrastructure, as well as $7.4 million to expand the New York State Police’s Cyber Analysis Unit, Computer Crimes Unit and Internet Crimes Against Children Center.
The state’s comprehensive cybersecurity strategy is defined by three central principles: Unification, Resilience and Preparedness. When taken together, New York State can lean on these tenets to present a unified and more resilient defense against new and more sophisticated cyber threats; preventing the vast majority of attacks but also isolating, controlling and mitigating potential threats; and preparing, adapting and always being ready for the cyber challenges of the future. This strategy offers a blueprint for cybersecurity stakeholders across New York, from State agencies to local governments, to understand how they fit into a larger plan. The blueprint provides objectives, lines of effort, and a commitment from the Governor that they can use when doing future planning and program design.
Governor Hochul also signed legislation to expand New York’s technology talent pool and provide funding to help ensure that New York-based employers are able to hire and retain necessary cybersecurity personnel. Governor Hochul acknowledged the importance of strengthening this sector during today’s announcement, which was attended by recent graduates of the MTA’s Operational Technology Cybersecurity Program.
This strategy sets forth a nation-leading blueprint to ensure New York State stands ready and resilient in the face of cyber threats.”
Governor Hochul
New York State Chief Cyber Officer Colin Ahern said, “Our vision for unification, resilience, and preparedness addresses the critical need for advanced resources and expertise across the state. We’re ensuring that every New Yorker is equally protected from digital threats.”
Acting National Cyber Director KembaWalden said, “The President’s National Cybersecurity Strategy articulates an affirmative vision for a defensible, resilient, and values-aligned digital ecosystem that benefits all Americans and enables our grandest national ambitions. The strategy calls for bold action to rebalance the responsibility for managing cyber risk onto those entities that are most capable of keeping us all safe, and shifting incentives to drive coordinated investments in long-term resilience. The New York strategy similarly articulates a fundamentally affirmative vision for cyberspace—that is, it is not simply reactive to threat actor behavior—and advances policy in areas such as: (1) public-private operational collaboration, (2) regulation of critical infrastructure, (3) cyber education and workforce development, and (4) IT modernization.”
New York State Division of Homeland Security and Emergency Services Commissioner Jackie Bray said,“Cyber threats are more prevalent than ever as the world relies on digital technology. New York’s strategy positions us strongly in preparing for and responding to attacks. Thanks to Governor Hochul’s vision and historic investment in cybersecurity, New York continues to lead the way.”
MTA Chair and CEO Janno Lieber said,“In recent years, MTA has made cybersecurity one of our top priorities. It is essential to ensure we can deliver service for the 6 million-plus commuters and travelers who depend on us every day.”
Acting New York State Chief Information Officer Jennifer Lorenz said, “Thanks to Governor Hochul, cybersecurity is a priority in New York State – backed up by real funding, real human resources and a real strategy designed to protect the state’s assets from intrusion and attack. The Office of Information Technology Services is proud to be a partner in carrying out this important mission, and know it will result in cyber defenses that are better, stronger and more agile.”
New York State Police Acting Superintendent Steven A. Nigrelli said,“Law enforcement is constantly challenged to keep pace with evolving online technologies that are utilized by criminals to steal from the public or gain access to sensitive information. We want to remind the public that just like you should be aware of your physical surroundings, you should be aware of your online presence too. The State Police will continue to work with our partners in safeguarding our cybersecurity networks.”
Department of Financial Services Superintendent Adrienne A. Harris said, “The Department of Financial Services is a pioneer in recognizing the risk of cyberattacks and the importance of adapting to new threats and more sophisticated cyber criminals. The Department’s first-in-the-nation requirements to protect DFS-regulatedbanks, insurance companies, financial services institutions and the consumers they serve stand as a model and I applaud Governor Hochul for advancing a comprehensive plan to address these issues across government and other critical industries.”
Chair and Chief Executive Officer of the Public Service Commission Rory M. Christian said, “New York is a hub for significant financial, governmental, manufacturing, and transportation infrastructure that has higher than normal risk of cyberattack for either criminal or geopolitical reasons. With her drive and determination, Governor Hochul’s leadership will help ensure that New York has the best cybersecurity protection available.”
SUNY Chancellor John B. King, Jr. said, “SUNY campuses are leading in cybersecurity education, both in preparing our students for jobs, and in providing cutting-edge laboratories for faculty, staff, and student research. Our students are in high demand for jobs, and we are thankful for the Governor’s investment to help us educate more New Yorkers in this field, and, via the Digital Transformation Fund in the 2024 enacted budget, protect our campuses from cyberattacks. We applaud Governor Hochul for leading this comprehensive statewide strategy.”
New York State Health Commissioner Dr. James McDonald said, “Under Governor Hochul’s leadership, New York State has significantly enhanced its cyber defenses, which is critically important to our health care system. This first-ever statewide cybersecurity strategy builds on these achievements by helping protect critical systems from cyber threats. The Department of Health looks forward to continuing to work closely with our agency partners on ensuring New York’s hospitals and health care facilities stay secure.”
Contact the Governor’s Press Office
Contact us by phone:
Albany: (518) 474 – 8418 New York City: (212) 681 – 4640
WASHINGTON — Director of National Intelligence Avril Haines on Thursday released the 2023 National Intelligence Strategy (NIS), focusing on “strategic competition” with China and Russia across the economic, political and military spheres — and calling on the Intelligence Communityto up its game on everything from information warfare to supply chain control to rapid adoption of emerging technologies.
“The United States faces an increasingly complex and interconnected threat environment characterized by strategic competition between the United States, the People’s Republic of China (PRC), and the Russian Federation, felt perhaps most immediately in Russia’s ongoing aggression in Ukraine. In addition to states, sub-national and non-state actors—from multinational corporations to transnational social movements—are increasingly able to create influence, compete for information, and secure or deny political and security outcomes, which provides opportunities for new partnerships as well as new challenges to U.S. interests,” Haines writes in a forward to the document.
“In addition, shared global challenges, including climate change, human and health security, as well as emerging and disruptive technological advances, are converging in ways that produce significant consequences that are often difficult to predict,” she added.
The NIS “is a foundational document for the IC and reflects the input of leaders from each of the 18 intelligence elements, as it directs the operations, investments, and priorities of the collective,” explains the Office of the Director of National Intelligence (ODNI) in a press release announcing the new strategy.
Besides the usual three-letter IC agiencis, those 18 organizations include the National Reconnaissance Office, the National Geospatial-Intelligence Agency and the intelligence arms of the military service, with the most recent addition being the Space Force in early 2021.
The document lays out six goals for the interagency community:
Position the IC for Intensifying Strategic Competition: This includes improving the “ability to provide timely and accurate insights into competitor intentions, capabilities, and actions by strengthening capabilities in language, technical, and cultural expertise and harnessing open source, ‘big data,’ artificial intelligence, and advanced analytics.”
Recruit, Develop, and Retain a Talented and Diverse Workforce that Operates as a United Community: The IC needs an increasingly technical and diverse workforce, the document says. “The Community must overcome long-standing cultural, structural, bureaucratic, technical, and security challenges to reimagine and deliver the IC workforce of the future.”
Deliver Interoperable and Innovative Solutions at Scale: To do so, the strategy says, the IC must establish “unified IC procurement authorities, centralized solicitation systems, and a Community-wide contracting system, all bolstered by automation tools. A Community-wide, data-centric approach based on common standards is crucial to realizing the full promise of new capabilities.”
Diversify, Expand, and Strengthen Partnerships: “Even as we continue to invest in existing partnerships like those with our Five-Eyes partnersand forge new ones, the evolving set of challenges — from cyberattacks and climate change to pandemics and foreign malign influence — also require investing in new and more diverse partnerships, especially with non-state and sub-national actors. From companies to cities to civil society organizations, these actors’ ideas, innovations, resources, and actions increasingly shape our societal, technological, and economic futures.”
Expand IC Capabilities and Expertise on Transnational Challenges: Such challenges, the NIS explains, include “more frequent and intense crises due to the effects of climate change, narcotics trafficking, financial crises, supply chain disruptions, corruption, new and recurring diseases, and emerging and disruptive technologies” that in turn are piquing security crises such as civil unrest and migration.
Enhance Resilience: This includes increasing the IC’s role in protecting critical infrastructure to improve early warning that can allow more robust “recovery and response,” as well as “expanding its role in understanding threats and vulnerabilities to supply chains and helping to mitigate threats to government and industry partners’ infrastructure.”