The European parliament has voted to take steps to regulate AI technology such as ChatGPT.
The proposed rules would require such systems to indicate that content was AI-generated.
Talks with EU member states will shape the precise wording of the legislation known as the AI Act.
The European Union has taken the first steps towards regulating artificial intelligence, with its parliament backing a ban on the technology for biometric surveillance, emotion recognition, and predictive policing.
Europe will also seek to require systems such as ChatGPT to indicate that content was generated by AI, and deem AI systems used to influence voters to be “high-risk.”
The rules “aim to promote the uptake of human-centric and trustworthy AI and protect the health, safety, fundamental rights and democracy from its harmful effects,” per a press release from the European parliament on Wednesday.
The measures were passed by 499 votes to 28, with 93 abstentions. Talks will begin with EU member states on the precise wording of the legislation known as the AI Act.
The rules aim to ensure that AI developed and used in Europe complied with EU rights and values including human oversight, safety, privacy, transparency, non-discrimination, and social and environmental wellbeing.
Sam Altman, the CEO of ChatGPT maker OpenAI, met several European leaders in May to discuss the potential impact of AI on society. He said he had no plans to pull ChatGPT out of Europe despite saying proposed EU rules to govern AI could backfire.
As Insider previously reported, one of Altman’s main concerns with the EU’s proposed law centers on its definition of “high-risk” systems and could ensnare ChatGPT.
European parliament members voted to deem AI systems that “pose significant harm to people’s health, safety, fundamental rights or the environment” as high-risk, per the press release. AI systems used to influence voters and the outcome of elections, and recommender systems used by social-media platforms are also on the “high-risk” list.
Co-rapporteur Brando Benifei of Italy said Europe had devised a “concrete response” to the potential dangers posed by AI. “We want AI’s positive potential for creativity and productivity to be harnessed but we will also fight to protect our position and counter dangers to our democracies and freedoms,” he said.
Co-rapporteur Dragos Tudorache of Romania said Europe’s AI Act would shape the development and governance of AI globally and ensure it was used in line with the “European values of democracy, fundamental rights, and the rule of law.”
Collaboration is taking over the workplace. According to data collected by the authors over the past two decades, the time spent by managers and employees in collaborative activities has ballooned by 50% or more. There is much to applaud about these developments—but when consumption of a valuable resource spikes that dramatically, it should also give us pause.
At many companies, people spend around 80% of their time in meetings or answering colleagues’ requests, leaving little time for all the critical work they must complete on their own. What’s more, research the authors have done across more than 300 organizations shows that the apportionment of collaborative work is often extremely lopsided. In most cases, 20% to 35% of value-added collaborations come from only 3% to 5% of employees. The avalanche of demands for input or advice, access to resources, or sometimes just presence in a meeting causes performance to suffer. Employees take assignments home, and soon burnout and turnover become real risks.
Leaders must start to manage collaboration more effectively in two ways: (1) by mapping the supply and demand in their organizations and redistributing the work more evenly among employees, and (2) by incentivizing people to collaborate more efficiently.
Collaboration is taking over the workplace.
As business becomes increasingly global and cross-functional, silos are breaking down, connectivity is increasing, and teamwork is seen as a key to organizational success. According to data we have collected over the past two decades, the time spent by managers and employees in collaborative activities has ballooned by 50% or more.
Certainly, we find much to applaud in these developments. However, when consumption of a valuable resource spikes that dramatically, it should also give us pause. Consider a typical week in your own organization. How much time do people spend in meetings, on the phone, and responding to e-mails? At many companies the proportion hovers around 80%, leaving employees little time for all the critical work they must complete on their own. Performance suffers as they are buried under an avalanche of requests for input or advice, access to resources, or attendance at a meeting. They take assignments home, and soon, according to a large body of evidence on stress, burnout and turnover become real risks.
What’s more, research we’ve done across more than 300 organizations shows that the distribution of collaborative work is often extremely lopsided. In most cases, 20% to 35% of value-added collaborations come from only 3% to 5% of employees. As people become known for being both capable and willing to help, they are drawn into projects and roles of growing importance. Their giving mindset and desire to help others quickly enhances their performance and reputation. As a recent study led by Ning Li, of the University of Iowa, shows, a single “extra miler”—an employee who frequently contributes beyond the scope of his or her role—can drive team performance more than all the other members combined.
But this “escalating citizenship,” as the University of Oklahoma professor Mark Bolino calls it, only further fuels the demands placed on top collaborators. We find that what starts as a virtuous cycle soon turns vicious. Soon helpful employees become institutional bottlenecks: Work doesn’t progress until they’ve weighed in. Worse, they are so overtaxed that they’re no longer personally effective. And more often than not, the volume and diversity of work they do to benefit others goes unnoticed, because the requests are coming from other units, varied offices, or even multiple companies. In fact, when we use network analysis to identify the strongest collaborators in organizations, leaders are typically surprised by at least half the names on their lists. In our quest to reap the rewards of collaboration, we have inadvertently created open markets for it without recognizing the costs. What can leaders do to manage these demands more effectively?
Precious Personal Resources
First, it’s important to distinguish among the three types of “collaborative resources” that individual employees invest in others to create value: informational, social, and personal. Informational resources are knowledge and skills—expertise that can be recorded and passed on. Social resources involve one’s awareness, access, and position in a network, which can be used to help colleagues better collaborate with one another. Personal resources include one’s own time and energy.
These three resource types are not equally efficient. Informational and social resources can be shared—often in a single exchange—without depleting the collaborator’s supply. That is, when I offer you knowledge or network awareness, I also retain it for my own use. But an individual employee’s time and energy are finite, so each request to participate in or approve decisions for a project leaves less available for that person’s own work.
Up to a third of value-added collaborations come from only 3% to 5% of employees.
Unfortunately, personal resources are often the default demand when people want to collaborate. Instead of asking for specific informational or social resources—or better yet, searching in existing repositories such as reports or knowledge libraries—people ask for hands-on assistance they may not even need. An exchange that might have taken five minutes or less turns into a 30-minute calendar invite that strains personal resources on both sides of the request.
Consider a case study from a blue-chip professional services firm. When we helped the organization map the demands facing a group of its key employees, we found that the top collaborator—let’s call him Vernell—had 95 connections based on incoming requests. But only 18% of the requesters said they needed more personal access to him to achieve their business goals; the rest were content with the informational and social resources he was providing. The second most connected person was Sharon, with 89 people in her network, but her situation was markedly different, and more dangerous, because 40% of them wanted more time with her—a significantly greater draw on her personal resources.
We find that as the percentage of requesters seeking more access moves beyond about 25, it hinders the performance of both the individual and the group and becomes a strong predictor of voluntary turnover. As well-regarded collaborators are overloaded with demands, they may find that no good deed goes unpunished.
The exhibit “In Demand, Yet Disengaged,” reflecting data on business unit line leaders across a sample of 20 organizations, illustrates the problem. People at the top center and right of the chart—that is, those seen as the best sources of information and in highest demand as collaborators in their companies—have the lowest engagement and career satisfaction scores, as represented by the size of their bubbles. Our research shows that this ultimately results in their either leaving their organizations (taking valuable knowledge and network resources with them) or staying and spreading their growing apathy to their colleagues.
Leaders can solve this problem in two ways: by streamlining and redistributing responsibilities for collaboration and by rewarding effective contributions.
Redistributing the Work
Any effort to increase your organization’s collaborative efficiency should start with an understanding of the existing supply and demand. Employee surveys, electronic communications tracking, and internal systems such as 360-degree feedback and CRM programs can provide valuable data on the volume, type, origin, and destination of requests, as can more in-depth network analyses and tools. For example, Do.com monitors calendars and provides daily and weekly reports to both individual employees and managers about time spent in meetings versus on solo work. The idea is to identify the people most at risk for collaborative overload. Once that’s been done, you can focus on three levers:
Encourage behavioral change.
Show the most active and overburdened helpers how to filter and prioritize requests; give them permission to say no (or to allocate only half the time requested); and encourage them to make an introduction to someone else when the request doesn’t draw on their own unique contributions. The latest version of the team-collaboration software Basecamp now offers a notification “snooze button” that encourages employees to set stronger boundaries around their incoming information flow. It’s also worth suggesting that when they do invest personal resources, it be in value-added activities that they find energizing rather than exhausting. In studying employees at one Fortune 500 technology company, we found that although 60% wanted to spend less time responding to ad hoc collaboration requests, 40% wanted to spend more time training, coaching, and mentoring. After their contributions were shifted to those activities, employees were less prone to stress and disengagement.
To stem the tide of incoming requests, help seekers, too, must change their behavior. Resetting norms regarding when and how to initiate e-mail requests or meeting invitations can free up a great deal of wasted time. As a step in this direction, managers at Dropbox eliminated all recurring meetings for a two-week period. That forced employees to reassess the necessity of those gatherings and, after the hiatus, helped them become more vigilant about their calendars and making sure each meeting had an owner and an agenda. Rebecca Hinds and Bob Sutton, of Stanford, found that although the company tripled the number of employees at its headquarters over the next two years, its meetings were shorter and more productive.
In addition, requests for time-sapping reviews and approvals can be reduced in many risk-averse cultures by encouraging people to take courageous action on decisions they should be making themselves, rather than constantly checking with leaders or stakeholders.
Leverage technology and physical space to make informational and social resources more accessible and transparent.
Relevant technical tools include Slack and Salesforce.com’s Chatter, with their open discussion threads on various work topics; and Syndio and VoloMetrix (recently acquired by Microsoft), which help individuals assess networks and make informed decisions about collaborative activities. Also rethink desk or office placement. A study led by the Boston University assistant professor Stine Grodal documented the detrimental effects of team meetings and e-mails on the development and maintenance of productive helping relationships. When possible, managers should colocate highly interdependent employees to facilitate brief and impromptu face-to-face collaborations, resulting in a more efficient exchange of resources.
Consider structural changes.
Can you shift decision rights to more-appropriate people in the network? It may seem obvious that support staff or lower-level managers should be authorized to approve small capital expenditures, travel, and some HR activities, but in many organizations they aren’t. Also consider whether you can create a buffer against demands for collaboration. Many hospitals now assign each unit or floor a nurse preceptor, who has no patient care responsibilities and is therefore available to respond to requests as they emerge. The result, according to research that one of us (Adam Grant) conducted with David Hofmann and Zhike Lei, is fewer bottlenecks and quicker connections between nurses and the right experts. Other types of organizations might also benefit from designating “utility players”—which could lessen demand for the busiest employees—and possibly rotating the role among team members while freeing up personal resources by reducing people’s workloads.
Rewarding Effective Collaboration
We typically see an overlap of only about 50% between the top collaborative contributors in an organization and those employees deemed to be the top performers. As we’ve explained, many helpers underperform because they’re overwhelmed; that’s why managers should aim to redistribute work. But we also find that roughly 20% of organizational “stars” don’t help; they hit their numbers (and earn kudos for it) but don’t amplify the success of their colleagues. In these cases, as the former Goldman Sachs and GE chief learning officer Steve Kerr once wrote, leaders are hoping for A (collaboration) while rewarding B (individual achievement). They must instead learn how to spot and reward people who do both.
Why Women Bear More of the Burden
The lion’s share of collaborative work tends to …
Consider professional basketball, hockey, and soccer teams. They don’t just measure goals; they also track assists. Organizations should do the same, using tools such as network analysis, peer recognition programs, and value-added performance metrics. We helped one life sciences company use these tools to assess its workforce during a multibillion-dollar acquisition. Because the deal involved consolidating facilities around the world and relocating many employees, management was worried about losing talent. A well-known consultancy had recommended retention bonuses for leaders. But this approach failed to consider those very influential employees deep in the acquired company who had broad impact but no formal authority. Network analytics allowed the company to pinpoint those people and distribute bonuses more fairly.
Efficient sharing of informational, social, and personal resources should also be a prerequisite for positive reviews, promotions, and pay raises. At one investment bank, employees’ annual performance reviews include feedback from a diverse group of colleagues, and only those people who are rated as strong collaborators (that is, able to cross-sell and provide unique customer value to transactions) are considered for the best promotions, bonuses, and retention plans. Corning, the glass and ceramics manufacturer, uses similar metrics to decide which of its scientists and engineers will be named fellows—a high honor that guarantees a job and a lab for life. One criterion is to be the first author on a patent that generates at least $100 million in revenue. But another is whether the candidate has worked as a supporting author on colleagues’ patents. Corning grants status and power to those who strike a healthy balance between individual accomplishment and collaborative contribution. (Disclosure: Adam Grant has done consulting work for Corning.)
Collaboration is indeed the answer to many of today’s most pressing business challenges. But more isn’t always better. Leaders must learn to recognize, promote, and efficiently distribute the right kinds of collaborative work, or their teams and top talent will bear the costs of too much demand for too little supply. In fact, we believe that the time may have come for organizations to hire chief collaboration officers. By creating a senior executive position dedicated to collaboration, leaders can send a clear signal about the importance of managing teamwork thoughtfully and provide the resources necessary to do it effectively. That might reduce the odds that the whole becomes far less than the sum of its parts.
A version of this article appeared in the January–February 2016 issue (pp.74–79) of Harvard Business Review.
Reb Rebele is a researcher for Wharton People Analytics and teaches in the Master of Applied Positive Psychology (MAPP) program at the University of Pennsylvania.
The agency’s CHIPS Research and Development Office aims to advance U.S. research and development efforts and “ensure America’s global leadership” in the semiconductor sector, officials said.
The National Institute of Standards and Technology has compiled a new team of leaders to help spearhead federal research and development efforts that aim to boost U.S. leadership in semiconductor manufacturing.
NIST Director Laura Locascio said Tuesday that the heads of the CHIPS Research and Development Office “will propel CHIPS for America and the nation’s semiconductor sector forward” by overseeing four new programs dedicated to advancing U.S. semiconductor technology and innovation.
The office was established after Congress passed the Creating Helpful Incentives to Produce Semiconductors — or CHIPS — Act last year, which included over $50 billion for domestic semiconductor research and development.
Lora Weiss, senior vice president for research at Penn State University, will serve as director of the office. Weiss, who also serves as president of the Penn State Research Foundation, oversees research efforts across the university’s 12 academic colleges, as well as seven interdisciplinary research institutes and a university-affiliated research center for the Navy.
Eric Lin, former director of the NIST Material Measurement Laboratory, will serve as deputy director, and Neil Alderoty, former executive administrator of the laboratory, will serve as executive officer.
The CHIPS Research and Development Office is tasked with managing four integrated semiconductor programs, including the National Semiconductor Technology Center and the National Advanced Packaging Manufacturing Program.
The office will also oversee up to three new manufacturing institutes focused on semiconductor technologies, as well as the CHIPS Metrology research and development program, which conducts measurement science to help develop new materials and production methods for U.S. semiconductors.
Richard-Duane Chambers will take on the role of associate director for integration and policy after serving as a lead staffer on the Senate Committee on Commerce, Science and Transportation’s subcommittee on space and science. Maria Dowell will assume the role of director of the CHIPS Metrology Research and Development program, after previously serving as director of the NIST Communications Technology Laboratory.
Secretary of Commerce Gina Raimondo said the new research and development programs “will ensure America’s global leadership by creating a robust semiconductor R&D ecosystem.
“These leaders bring exactly the depth and breadth of organizational, programmatic and technical leadership experience that CHIPS needs to stand up new, transformational R&D programs,” Raimondo said in a statement.
The U.S. currently produces an estimated 10% of the world’s supply of semiconductors, while East Asia is responsible for more than 75% of global production. In addition to the nearly $50 billion included in the CHIPS Act for research and development, the bill also featured a $10 billion investment in regional innovation and technology hubs nationwide, as well as funding to support science, technology, engineering and mathematics education and workforce development activities.
“Full-stack” data scientist means mastery of machine learning, statistics, and analytics. Today’s fashion in data science favors flashy sophistication with a dash of sci-fi, making AI and machine learning the darlings of the job market. Alternative challengers for the alpha spot come from statistics, thanks to a century-long reputation for rigor and mathematical superiority. What about analysts?Whereas excellence in statistics is about rigor and excellence in machine learning is about performance, excellence in analytics is all about speed. Analysts are your best bet for coming up with those hypotheses in the first place. As analysts mature, they’ll begin to get the hang of judging what’s important in addition to what’s interesting, allowing decision-makers to step away from the middleman role. Of the three breeds, analysts are the most likely heirs to the decision throne.
The top trophy hire in data science is elusive, and it’s no surprise: a “full-stack” data scientist has mastery of machine learning, statistics, and analytics. When teams can’t get their hands on a three-in-one polymath, they set their sights on luring the most impressive prize among the single-origin specialists. Which of those skills gets the pedestal?
Today’s fashion in data science favors flashy sophistication with a dash of sci-fi, making AI and machine learning the darlings of the job market. Alternative challengers for the alpha spot come from statistics, thanks to a century-long reputation for rigor and mathematical superiority. What about analysts?
Analytics as a second-class citizen
If your primary skill is analytics (or data-mining or business intelligence), chances are that your self-confidence has taken a beating as machine learning and statistics have become prized within companies, the job market, and the media.
But what the uninitiated rarely grasp is that the three professions under the data science umbrella are completely different from one another. They may use some of the same methods and equations, but that’s where the similarity ends. Far from being a lesser version of the other data science breeds, good analysts are a prerequisite for effectiveness in your data endeavors. It’s dangerous to have them quit on you, but that’s exactly what they’ll do if you under-appreciate them.
Instead of asking an analyst to develop their statistics or machine learning skills, consider encouraging them to seek the heights of their own discipline first. In data science, excellence in one area beats mediocrity in two. So, let’s examine what it means to be truly excellent in each of the data science disciplines, what value they bring, and which personality traits are required to survive each job. Doing so will help explain why analysts are valuable, and how organizations should use them.
Excellence in statistics: rigor
Statisticians are specialists in coming to conclusions beyond your data safely — they are your best protection against fooling yourself in an uncertain world. To them, inferring something sloppily is a greater sin than leaving your mind a blank slate, so expect a good statistician to put the brakes on your exuberance. They care deeply about whether the methods applied are right for the problem and they agonize over which inferences are valid from the information at hand.
The result? A perspective that helps leaders make important decisions in a risk-controlled manner. In other words, they use data to minimize the chance that you’ll come to an unwise conclusion.
Excellence in machine learning: performance
You might be an applied machine learning/AI engineer if your response to “I bet you couldn’t build a model that passes testing at 99.99999% accuracy” is “Watch me.” With the coding chops to build both prototypes and production systems that work and the stubborn resilience to fail every hour for several years if that’s what it takes, machine learning specialists know that they won’t find the perfect solution in a textbook. Instead, they’ll be engaged in a marathon of trial-and-error. Having great intuition for how long it’ll take them to try each new option is a huge plus and is more valuable than an intimate knowledge of how the algorithms work (though it’s nice to have both). Performance means more than clearing a metric — it also means reliable, scalable, and easy-to-maintain models that perform well in production. Engineering excellence is a must.
The result? A system that automates a tricky task well enough to pass your statistician’s strict testing bar and deliver the audacious performance a business leader demanded.
Wide versus deep
What the previous two roles have in common is that they both provide high-effort solutions to specific problems. If the problems they tackle aren’t worth solving, you end up wasting their time and your money. A frequent lament among business leaders is, “Our data science group is useless.” And the problem usually lies in an absence of analytics expertise.
Statisticians and machine learning engineers are narrow-and-deep workers — the shape of a rabbit hole, incidentally — so it’s really important to point them at problems that deserve the effort. If your experts are carefully solving the wrong problems, your investment in data science will suffer low returns. To ensure that you can make good use of narrow-and-deep experts, you either need to be sure you already have the right problem or you need a wide-and-shallow approach to finding one.
Excellence in analytics: speed
The best analysts are lightning-fast coders who can surf vast datasets quickly, encountering and surfacing potential insights faster than those other specialists can say “whiteboard.” Their semi-sloppy coding style baffles traditional software engineers — but leaves them in the dust. Speed is their highest virtue, closely followed by the ability to identify potentially useful gems. A mastery of visual presentation of information helps, too: beautiful and effective plots allow the mind to extract information faster, which pays off in time-to-potential-insights.
The result is that the business gets a finger on its pulse and eyes on previously-unknown unknowns. This generates the inspiration that helps decision-makers select valuable quests to send statisticians and ML engineers on, saving them from mathematically-impressive excavations of useless rabbit holes.
Sloppy nonsense or stellar storytelling?
“But,” object the statisticians, “most of their so-called insights are nonsense.” By that they mean the results of their exploration may reflect only noise. Perhaps, but there’s more to the story.
Analysts are data storytellers. Their mandate is to summarize interesting facts and to use data for inspiration. In some organizations those facts and that inspiration become input for human decision-makers. But in more sophisticated data operations, data-driven inspiration gets flagged for proper statistical follow-up.
Good analysts have unwavering respect for the one golden rule of their profession: do not come to conclusions beyond the data (and prevent your audience from doing it, too). To this end, one way to spot a good analyst is that they use softened, hedging language. For example, not “we conclude” but “we are inspired to wonder”. They also discourage leaders’ overconfidence by emphasizing a multitude of possible interpretations for every insight.
As long as analysts stick to the facts — saying only “This is what is here.” — and don’t take themselves too seriously, the worst crime they could commit is wasting someone’s time when they run it by them.
While statistical skills are required to test hypotheses, analysts are your best bet for coming up with those hypotheses in the first place. For instance, they might say something like “It’s only a correlation, but I suspect it could be driven by …” and then explain why they think that. This takes strong intuition about what might be going on beyond the data, and the communication skills to convey the options to the decision-maker, who typically calls the shots on which hypotheses (of many) are important enough to warrant a statistician’s effort. As analysts mature, they’ll begin to get the hang of judging what’s important in addition to what’s interesting, allowing decision-makers to step away from the middleman role.
Of the three breeds, analysts are the most likely heirs to the decision throne. Because subject matter expertise goes a long way towards helping you spot interesting patterns in your data faster, the best analysts are serious about familiarizing themselves with the domain. Failure to do so is a red flag. As their curiosity pushes them to develop a sense for the business, expect their output to shift from a jumble of false alarms to a sensibly-curated set of insights that decision-makers are more likely to care about.
Analytics for decision-making
To avoid wasted time, analysts should lay out the story they’re tempted to tell and poke it from several angles with follow-up investigations to see if it holds water before bringing it to decision-makers. The decision-maker should then function as a filter between exploratory data analytics and statistical rigor. If someone with decision responsibility finds the analyst’s exploration promising for a decision they have to make, they then can sign off on a statistician spending the time to do a more rigorous analysis. (This process indicates why just telling analysts to get better at statistics misses the point in an important way. Not only are the two activities separate, but another person sits in between them, meaning it’s not necessarily any more efficient for one person to do both things.)
Analytics for machine learning and AI
Machine learning specialists put a bunch of potential data inputs through algorithms, tweak the settings, and keep iterating until the right outputs are produced. While it may sound like there’s no role for analytics here, in practice a business often has far too many potential ingredients to shove into the blender all at once. One way to filter down to a promising set of inputs to try is domain expertise — ask a human with opinions about how things might work. Another way is through analytics. To use the analogy of cooking, the machine learning engineer is great at tinkering in the kitchen, but right now they’re standing in front of a huge, dark warehouse full of potential ingredients. They could either start grabbing them haphazardly and dragging them back to their kitchens, or they could send a sprinter armed with a flashlight through the warehouse first. Your analyst is the sprinter; their ability to quickly help you see and summarize what-is-here is a superpower for your process.
The dangers of under-appreciating analysts
An excellent analyst is not a shoddy version of the machine learning engineer; their coding style is optimized for speed — on purpose. Nor are they a bad statistician, since they don’t deal at all with uncertainty, they deal with facts. The primary job of the analyst is to say: “Here’s what’s in our data. It’s not my job to talk about what it means, but perhaps it will inspire the decision-maker to pursue the question with a statistician.”
If you overemphasize hiring and rewarding skills in machine learning and statistics, you’ll lose your analysts. Who will help you figure out which problems are worth solving then? You’ll be left with a group of miserable experts who keep being asked to work on useless projects or analytics tasks they didn’t sign up for. Your data will lie around useless.
When in doubt, hire analysts before other roles. Appreciate them and reward them. Encourage them to grow to the heights of their chosen career (and not someone else’s). Of the cast of characters mentioned in this story, the only ones that every business needs are decision-makers and analysts. The others you’ll only be able to use when you know exactly what you need them for. Start with analytics and be proud of your newfound ability to open your eyes to the rich and beautiful information in front of you. Data-driven inspiration is a powerful thing.
More than 80% of your health status is dependent on social determinants of health, such as where you live and work, the air you breathe, your transportation, the water you drink, and whether you use tobacco or other drugs. These factors also can—and do—impact your access to healthcare, critical resources, and opportunities for well-being.
Underprivileged and underserved populations don’t have the same access or opportunities as others in our society, and they feel these inequities acutely with poorer health and higher mortality rates. Yet, everyone feels the effects of unhealthy communities and rising healthcare costs.
Health inequity is entrenched and systemic in global healthcare. The life expectancy gap between low- and high-income countries can be as high as 18 years. Inequities account for more than $320B in annual healthcare spend in the US alone, and this is anticipated to grow to $1T by 2040. This path is repeated in many countries worldwide, which overwhelms the system and isn’t sustainable.
Just as technology has improved so many other parts of our lives, it could also advance health equity by eliminating barriers to the tools, resources, knowledge, and opportunities we all need to be as healthy as possible.
Data is at the heart of advancing health equity. People, whether as patients, clinicians, community leaders, or public health officials, need clean, usable, trustworthy data to take meaningful action. To advance health equity, we need to understand social determinants of health along with information from claims data, research, operations, and community risks. A patient’s physical environment, health-related behaviors, and economic factors make a profound difference.
Bringing all this data together in a meaningful way will be transformative. While clinical data is flowing through information exchanges, most data that impacts a person’s health is siloed and disconnected, existing outside the hospital’s electronic health record system. That’s why we’re building an open, intelligent, cloud-based healthcare platform—to connect clinical and enterprise systems for organizations and to bring in data like social determinants of health or community risk factors. We’re removing the silos and connecting disparate systems.
Just as technology has improved so many other parts of our lives, it could also advance health equity by eliminating barriers to the tools, resources, knowledge, and opportunities we all need to be as healthy as possible.
At an individual level, technology can give providers a holistic view of a patient’s health through a single, longitudinal health record. This enables them to address all aspects of a person’s condition including non-clinical factors. At a community level, technology supports leaders with a full picture of their community’s health and risk. With better information, they can more effectively reach and treat different segments of the population, direct resources, programming, and interventions equitably, and create policies that lead to health justice.
With a cloud-based health platform, we can also extend data across the healthcare ecosystem, making it available for researchers and scientists. Advancing health equity includes ensuring everyone has access to the newest and most innovative therapies available. Technology enables life sciences companies to connect with healthcare providers, their clinicians, patients, and communities to expand access to clinical trials. This results in more diverse study participants, which means researchers have more representative, complete, and comprehensive patient information to validate the safety and efficacy of treatments.
Oracle Health’s Learning Health Network includes over 100 healthcare delivery organizations dedicated to sharing deidentified data to advance clinical research. With more than 100 million patients represented, diversity has become the network’s superpower. Clinical trials run through the Learning Health Network have three times the US average of Black and Hispanic participants. Individuals, clinicians, and communities who have never had the chance to participate in clinical discovery can now do so, gaining access to leading therapeutics, diagnostics, and medications sooner.
Technology comes with its own challenges in equity, so we have an obligation to be responsible stewards of patient data. AI and machine learning could be game changers as we look to solve healthcare’s biggest challenges like burnout and cost. Since healthcare data represents real people and real situations, we need to be careful that any technology relying on this data doesn’t further exacerbate the biases and inequities that exist today.
For example, healthcare providers have been using clinical calculators for decades to estimate risk and predict outcomes. Yet, many of these calculators include race as a data element, which isn’t evidence-based and can cause harm to segments of the population. Ethical AI begins with incorporating equitable and just principles into the design, development, delivery, and analysis of products, solutions, and services. Plus, in partnership across the healthcare ecosystem—especially with researchers—we need to evaluate new technologies for potential bias, eliminate it, and evaluate the results.
In short, with more data and better information, we have an incredible opportunity to create a better healthcare system that improves lives and experiences for patients, clinicians, and communities around the world regardless of economic status or geography. It will take all of us working together to achieve this. Now is our time for action.
Senior Vice President and Chief Health Officer, Oracle Health
As chief health officer, Dr. Nasim Afsar leads Oracle Health’s Health Transformation Office with a commitment to delivering healthy people, healthy workforce, and healthy business. Her team focuses on supporting clinical care, operational efficiency, and financial sustainability, leveraging big data, and advancing the future of the workforce and health equity. She also works on building the larger ecosystem of healthcare, including working closely with payers, retail, and public health agencies.
Dr. Afsar previously served as chief operating officer for University of California, Irvine (UCI) Health with the vision of delivering flawless care for patients in the region while creating the best place to work in healthcare. At UCI Health, she led inpatient and ambulatory operations, resulting in historically high ambulatory growth, inpatient volumes, surgeries, and tertiary care transfers to the institution. During COVID-19, she created a mobile field hospital and drive-thru testing centers, co-led the hospital at home program, and led a large-scale vaccination program. She also led health system contracting, spearheading new value-based products to market. As the executive for population health management, she ran a number of value-based programs, as well as the UCI Health Medicare Shared Savings Program (MSSP) accountable care organization (ACO).
Previously at University of California, Los Angeles (UCLA) Health, Dr. Afsar served as associate chief medical officer leading large-scale health system initiatives in quality, safety, and patient experience, and as chief quality officer for the Department of Medicine, overseeing population health initiatives.
She is past president of the Society of Hospital Medicine and served on its board of directors for eight years.
Japan’s early adoption of universal health coverage has attracted attention from around the world.
It is seen in may quarters as one of the foundations of an equitable society.
The key challenge is to ensure the funding and HR requirements are in place to make this approach sustainable.
Japan’s early adoption of Universal Health Coverage (UHC) has attracted worldwide attention, as it is the country with the longest healthy life expectancy in the world.
One of the reasons for this is that for more than half a century Japan has maintained a health insurance system that all permanent residents of Japan for more than three months are required to join, allowing people living in Japan to access appropriate healthcare services at a cost they can afford. This is characterized by a free-access system that allows patients to choose any healthcare provider, from small clinics to large hospitals with the latest medical facilities, and all medical services are provided at a uniform price anywhere in Japan.
In addition, the Japanese government has increased the number of medical schools, especially in rural areas, in order to increase the number of physicians under the One Prefecture, One Medical School policy approved by the Cabinet in 1973. This has also contributed to the high quality of healthcare services in the country.
With such a history and system of insured health care, Japan issued the Basic Policy for Peace and Health in 2015, and based on its own experience, has shown a commitment to strengthen the necessary support for mainstreaming universal health coverage in the international community.
At the G7 Ise-Shima Summit and G7 Kobe Health Ministers’ Meeting held in 2016, Japan became the first G7 country to set the promotion of UHC as a major theme at the summit-level meeting. Japan expressed its commitment to play a leading role in international discussions by supporting the establishment of universal health coverage in Africa, Asia, and other regions in cooperation with the international community and organizations.
Subsequently, in 2017, Japan co-hosted the high-level forum on UHC with the World Bank, the World Health Organization (WHO), and the United Nations Children’s Fund (UNICEF). Government leaders from over 30 countries, as well as representatives and experts from international organizations, gathered to discuss how to promote universal health coverage in their countries, and adopted the Tokyo Declaration on UHC, which includes a commitment to accelerate efforts to achieve UHC by 2030.
In May 2022, the Kishida administration set forth its new Global Health Strategy based on the experience of responding to the spread of COVID-19. Placing the achievement of more resilient, equitable and sustainable UHC at the centre of Japan’s international cooperation in the health sector, the strategy provides guidelines for efforts to build a global health architecture and strengthen health systems to prepare for future public health crises, including pandemics.
Corporate contribution to the realization of UHC
Ajinomoto, a Japanese food company, has been developing a project to improve infant nutrition in Ghana since 2009, working with Japan International Cooperation Agency (JICA) to promote baby nutritional supplements to reduce infant mortality due to malnutrition.
In addition, LEBER, Inc. has developed a healthcare app that can connect doctors and users anytime, anywhere, to solve the problems of people living in areas where access to healthcare is difficult. The app enables 24/7 remote doctor consultation via smartphone. With one of the largest networks of doctors in Japan, the app also functions as a physical condition management tool as well as a doctor consultation platform.
Maintaining and operating the mechanism is a challenge
Universal health coverage is attracting attention as an excellent approach, but the key issue is how to maintain and operate the system once it has been realized. Securing an operating budget and training human resources with expertise are essential to making UHC sustainable.
Even in Japan, a leading universal health coverage country, there is a growing view that the current health insurance system is at risk of collapsing in the future due to increasing medical costs associated with the rapid aging of the population and sluggish income growth caused by slow economic growth. If Japan is to further develop its healthcare functions, there is an urgent need to rebuild a sustainable healthcare system that can respond to changes in the demographic structure. To address this challenge, the Ministry of Health, Labour and Welfare has launched Health Care 2035 to discuss a paradigm shift in Japan’s healthcare, with the aim to achieve this by 2035.
Universal health coverage as the foundation for a sustainable society
It will soon be three years since the first outbreak of COVID-19. The virus has repeatedly mutated, with numerous infectious outbreaks and convergence cycles that have taken a long-term toll on the economies of countries and peoples around the world. We have become acutely aware that health issues are closely linked to economic and social issues, and the importance of Sustainable Development Goal number 3 – good health and well-being – has also been highlighted once again.
What is noteworthy about Goal 3, which is comprised of 13 targets, is that the achievement of UHC is stated as being related to all of the targets and as the foundation of the entire goal. The need for universal health coverage has become even more urgent, not only to correct the world’s increasingly pronounced health disparities, but as a foundation for economic and social stability and global health security.
Developed countries have extended assistance to people in developing countries to ensure that they have access to adequate healthcare services. On the other hand, the pandemic has also pushed the healthcare systems of developed countries to the brink of collapse. In 2023, the G7 summit is scheduled to take place in Hiroshima. In conjunction with the summit, Health Ministers’ Meeting will be held in Nagasaki in May. As the chair, Japan is expected to lead the discussion on overcoming this crises and rebuilding cooperation among countries.
Disha Jhaveri, MPH ’23, a dentist, wants to use her public health degree to make a wider impact
June 1, 2023 – Every strawberry season in Mumbai, India, Disha Jhaveri’s parents get a knock on the door from a local strawberry vendor, there to gift them fruit as thanks for their daughter, the dentist who provided him free care.
“I was buying strawberries from him on my way home one day and he was telling me he had so much tooth pain he was unable to eat,” Jhaveri said. “I said, ‘I’m a dentist. Why don’t you come to my office and let me have a look?’ I spoke to my employer, and he said I could open my chair that Sunday, when our office was closed, to treat the man at no cost. After that, most Sundays I would practice this kind of dentistry, treating local food vendors and farmers and other underserved people without charging them any fee.”
That first Sunday was about ten years ago, when Jhaveri was a new dental school graduate and was still living in Mumbai. Her de facto free clinic finished in 2018, when she moved to Boston with her husband, an infectious diseasedoctor. Nonetheless, the still grateful strawberry vendor and the rest of her Sunday patients are never far from her mind. She mentions a pair of shoes given to her by a shoemaker she treated. “I don’t wear them,” Jhaveri said. “I keep them as a reminder to myself to continue doing this kind of service.” And at home in her study hangs a gift from another grateful patient who received care that otherwise would have been inaccessible—a framed image of Lord Ganesha, the Hindu deity known as “the remover of obstacles.”
Jhaveri carried these tokens from Mumbai to Boston and then to Jackson, Mississippi, where she moved in 2021. And she found new inspiration in them as she studied for her Master of Public Health (MPH) degree at Harvard T.H. Chan School of Public Health, part of the first cohort of students in the MPH Generalist online program. Jhaveri graduated in May.
Though she may return to dentistry one day, studying public health has given her the inspiration and the tools to chart a new path toward her decade-old dream of ensuring access to health and well-being for marginalized populations.
“The vendor and the shoemaker are examples of just two people I could treat,” Jhaveri said. “But with public health, after studying for two years, I can reach thousands and thousands of people and make a meaningful impact. That’s where I’m getting pulled now. That’s what’s getting me excited.”
Connecting the classroom to the community
Newly arrived in Boston, faced with the multiyear process of obtaining an American dental license—and still wanting to make contributions to health in her home country—Jhaveri began working as a research assistant at Tufts Center for Global Public Health. There she conducted research on barriers to care among people living with tuberculosis in India. It was her mentor on the team, Dr. Ramnath Subbaraman, SM ’17, who suggested she pursue an MPH.
“Pursuing an MPH aligns with your career goals and interests, and it will be service oriented. Practicing dentistry is not the only way you can reach out to people,” Jhaveri recalls her mentor telling her.
Jhaveri had her eye on Harvard Chan School, and the School’s launch of the online MPH Generalist program in 2021 proved serendipitous, as another move—to Jackson, for her husband’s work—was imminent. When her classes began in the fall of 2021, any skepticism she and many fellow students felt about pursuing their degrees online quickly dissipated.
“We would attend Dr. Bernstein’s lecture with in-person students and could interact with each other, have live discussions,” Jhaveri said. “Laurie’s course on organizational behavior was the highlight of my final semester; it made me want to further my career in this field. And because of Professor McDonough’s lectures, I’m interested in policy. I never thought I would be, or that I would have an aptitude for it. But his engagement with us in office hours changed that. We had to extend them so many times, because we would all share our ideas and experiences and he would advise us and give us resources that weren’t on the syllabus, but that made sense for us.”
In fact, attending the School virtually was an advantage, Jhaveri said, as Jackson proved the ideal backdrop for studying public health. Her first year in Mississippi was characterized by bewilderment: She saw more fast-food restaurants and freestanding diabetes clinics than grocery stores and farmer’s markets. Cars were the only viable mode of transportation. Condoms were locked up at Walmart, purchasable only if you asked an employee. Her husband saw patients at the University of Mississippi Medical Center living with uncontrolled HIV; often they were uninsured, impoverished, and had driven hours to see him, sometimes from out of state.
“It was really weird to me,” Jhaveri said. “But I started analyzing what I was seeing around Jackson and correlating it with what we were studying—the social determinants of health—and it just clicked into place. The health status of the population, the food available, income, access to medical and dental care … everything is so interconnected. And it reminds me of the people I used to treat back home.”
Looking Ahead
“Home” is an evolving concept for Jhaveri. Whereas she once longed for the comforts of Mumbai or Boston, she now feels an allegiance to Jackson, to Mississippi, and to the Deep South.
“We as a country have progressed so much. But why have we left out certain pockets? Why is no one investing in Mississippi? The people here are part of us,” Jhaveri said. “It’s time we include them.”
For her practicum—a requirement of the MPH Generalist program, in which students tackle a public health problem through a project in a real-world setting—Jhaveri dug deep into her new community. She worked with Baby University, a nonprofit initiative based in Clarksdale, Mississippi that provides education and support to parents, in particular mothers, around how to better care for their children and themselves. Two of Jhaveri’s mentors—Bizu Gelaye, associate professor in the Department of Epidemiology, and Elizabeth Levey, assistant professor at Harvard Medical School and member of the Harvard Chan Department of Epidemiology’s GRAPE research team—advised her on her role developing tools to analyze its impact.
“The long-term goal is to be able to replicate Baby University, because everybody can’t reach Clarksdale,” Jhaveri said. “So my job was to develop an interview guide and survey for the Baby University participants, to help us understand what programming is working, what we can make better, and how to best set up our model to get funding and establish satellite campuses.”
Jhaveri also served as one of Baby University’s instructors, teaching participants about dental health for themselves and their young children. After graduation, she plans to continue her involvement. But most immediately, her plan is to take time to be with her own baby, Nivaan, who was born in February. Jhaveri said the support of her professors and peers made it possible for her to tend to herself, her pregnancy, and then her newborn son while still finishing her MPH on time—and that now, she’s looking forward to a true maternity leave.
And while she has taken the dental boards and is on the road to obtaining licensure, right now other public health projects are calling to her. Jhaveri mentions working for the University of Mississippi Medical Center’s tobacco cessation program as a potential next step, and says she also plans to share a policy proposal for introducing fluoride treatments in public schools with the help of the Medical Center’s dental school. She wrote the proposal as part of a class focused on how to draft policy proposals and briefings, and drew inspiration for it from a free pediatric health event she had organized in Mumbai, during which she helped provide fluoride treatment to 5,000 underserved school-age children.
Living in Jackson—like in Mumbai—the opportunities Jhaveri sees to make a difference in public health are endless.
“I feel I have evolved as a person. So much of what we were studying gave me the capacity to understand better and be more compassionate,” Jhaveri said. “There’s so much to do here in Mississippi. And using my experience from India and the skills and knowledge from my MPH program, there’s so much impact I can have.”
Is a new federal agency necessary to regulate AI? Maybe.
OpenAI CEO Sam Altman urged lawmakers to consider regulating AI during his Senate testimony on May 16, 2023. That recommendation raises the question of what comes next for Congress. The solutions Altman proposed – creating an AI regulatory agency and requiring licensing for companies – are interesting. But what the other experts on the same panel suggested is at least as important: requiring transparency on training data and establishing clear frameworks for AI-related risks.
Another point left unsaid was that, given the economics of building large-scale AI models, the industry may be witnessing the emergence of a new type of tech monopoly.
As a researcher who studies social media and artificial intelligence, I believe that Altman’s suggestions have highlighted important issues but don’t provide answers in and of themselves. Regulation would be helpful, but in what form? Licensing also makes sense, but for whom? And any effort to regulate the AI industry will need to account for the companies’ economic power and political sway.
AN AGENCY TO REGULATE AI?
Lawmakers and policymakers across the world have already begun to address some of the issues raised in Altman’s testimony. The European Union’s AI Act is based on a risk model that assigns AI applications to three categories of risk: unacceptable, high risk, and low or minimal risk. This categorization recognizes that tools for social scoring by governments and automated tools for hiring pose different risks than those from the use of AI in spam filters, for example.
The U.S. National Institute of Standards and Technology likewise has an AI risk management framework that was created with extensive input from multiple stakeholders, including the U.S. Chamber of Commerce and the Federation of American Scientists, as well as other business and professional associations, technology companies and think tanks.
Federal agencies such as the Equal Employment Opportunity Commission and the Federal Trade Commission have already issued guidelines on some of the risks inherent in AI. The Consumer Product Safety Commission and other agencies have a role to play as well.
Though OpenAI’s Altman suggested that companies could be licensed to release artificial intelligence technologies to the public, he clarified that he was referring to artificial general intelligence, meaning potential future AI systems with humanlike intelligence that could pose a threat to humanity. That would be akin to companies being licensed to handle other potentially dangerous technologies, like nuclear power. But licensing could have a role to play well before such a futuristic scenario comes to pass.
Algorithmic auditing would require credentialing, standards of practice and extensive training. Requiring accountability is not just a matter of licensing individuals but also requires companywide standards and practices.
Experts on AI fairness contend that issues of bias and fairness in AI cannot be addressed by technical methods alone but require more comprehensive risk mitigation practices such as adopting institutional review boards for AI. Institutional review boards in the medical field help uphold individual rights, for example.
Strengthening existing statutes on consumer safety, privacy and protection while introducing norms of algorithmic accountability would help demystify complex AI systems. It’s also important to recognize that greater data accountability and transparency may impose new restrictions on organizations.
Scholars of data privacy and AI ethics have called for “technological due process” and frameworks to recognize harms of predictive processes. The widespread use of AI-enabled decision-making in such fields as employment, insurance and health care calls for licensing and audit requirements to ensure procedural fairness and privacy safeguards.
Given the lack of transparency in the training data used by these companies, AI ethics experts Timnit Gebru, Emily Bender and others have warned that large-scale adoption of such technologies without corresponding oversight risks amplifying machine bias at a societal scale.
It is also important to acknowledge that the training data for tools such as ChatGPT includes the intellectual labor of a host of people such as Wikipedia contributors, bloggers and authors of digitized books. The economic benefits from these tools, however, accrue only to the technology corporations.
Proving technology firms’ monopoly power can be difficult, as the Department of Justice’s antitrust case against Microsoftdemonstrated. I believe that the most feasible regulatory options for Congress to address potential algorithmic harms from AI may be to strengthen disclosure requirements for AI firms and users of AI alike, to urge comprehensive adoption of AI risk assessment frameworks, and to require processes that safeguard individual data rights and privacy.
The Biden administration unveiled a docket full of more artificial intelligence regulatory efforts to promote responsible development, adoption and usage of increasingly smart systems.
The White House launched a series of new executive initiatives on fostering a culture of responsible artificial intelligence technology usage and practice within the U.S. on Tuesday, featuring a national strategic R&D plan and education objectives.
Following previous national frameworks, the three new announcements from the Biden administration act as guidelines to help codify responsible and effective AI algorithm usage, development and deployment, absent federal law.
“The federal government plays a critical role in this effort, including through smart investments in research and development (R&D) that promote responsible innovation and advance solutions to the challenges that other sectors will not address on their own,” the strategic plan executive summary reads.
Among the three announcements include a new roadmap of priority R&D areas in the AI sector for federal investments, a public request for information on how the federal government can best mitigate AI system risk, and an analysis documenting benefits and risks to AI technologies in education.
The R&D Strategic Plan, developed by the White House Office of Science and Technology Policy, is composed of several pillars to invest in safe-by-design AI systems that can be implemented in a social context. Those pillars include prioritizing long-term investments in responsible AI; developing methods for enhanced human-AI collaboration and understanding; thoroughly compiling a definitive list of ethical, legal and societal risks and benefits to AI system deployment; developing shared public datasets for broad AI algorithmic training; evaluating the needs of an AI-savvy workforce; expanding public and private sector partnerships; and establishing international collaborations on AI research efforts.
“The federal government plays a critical role in ensuring that technologies like AI are developed responsibly, and to serve the American people,” the plan’s fact sheet reads. “Federal investments over many decades have facilitated many key discoveries in AI innovations that power industry and society today, and federally funded research has sustained progress in AI throughout the field’s evolution.”
Complimenting the R&D plan are new insights into how new AI technologies can impact classroom learning and the broader educational system. Authored by leadership in the Department of Education, the report recommends ways educators can leverage AI-powered systems––namely exam monitoring, writing assistance and voice recognition devices––to their benefit, while mitigating potential risks.
Countering bias and data exposure in these systems was a paramount discussion point, leading regulators to broadly recommend all future education policies dealing with AI at a federal, state and local level keep user needs, feedback and empowerment in mind.
“As protections are developed, we recommend that policies center people, not machines,” the recommendations read. “Teachers, learners and others need to retain their agency to decide what patterns mean and to choose courses of action.”
DOE leadership also reiterated that AI technologies should not displace teachers.
“Some teachers worry that they may be replaced—to the contrary, the Department firmly rejects the idea that AI could replace teachers,” the recommendation states.
The final AI announcement requests public input on a new National AI Strategy. The forthcoming guidance aims to build on existing Biden-Harris administration actions surrounding AI and machine learning to further chart the nation’s course into a safe and integrated future with AI technologies.
“By developing a National AI Strategy, the federal government will provide a whole-of-society approach to AI,” the RFI background says. “The strategy will pay particular attention to recent and projected advances in AI, to make sure that the United States is responsive to the latest opportunities and challenges posed by AI, as well as the global changes that will arrive in the coming years.”
Comments from the public will be accepted until July 7, 2023. Some of the questions officials ask discuss best oversight practices of AI technologies, how AI language models can maintain secure software designs, how AI can strengthen civil rights and how AI can better identify digital vulnerabilities in critical infrastructures’ digital networks.
These comments reflect the broad goals of a forthcoming National AI Strategy that seeks to incorporate AI systems into a broad array of societal institutions, while simultaneously controlling for common risks.
On top of releasing new plans for more national AI technology oversight, the Biden administration will host a conversation with American workers today to hear concerns over automation and its broader economic impact.
AI regulation has been a chief talking point across the federal government following the breakthrough prevalence of generative AI systems such as ChatGPT, as a lack of sweeping regulations haunt the continued innovation in the AI/ML field.
This story originally appeared in The Algorithm, our weekly newsletter on AI. To get stories like this in your inbox first, sign up here.
It feels as though a switch has turned on in AI policy. For years, US legislators and American tech companies were reluctant to introduce—if not outright against—strict technology regulation. Now both have started begging for it.
Last week, OpenAI CEO Sam Altman appeared before a US Senate committee to talk about the risks and potential of AI language models. Altman, along with many senators, called for international standards for artificial intelligence. He also urged the US to regulate the technology and set up a new agency, much like the Food and Drug Administration, to regulate AI.
For an AI policy nerd like myself, the Senate hearing was both encouraging and frustrating. Encouraging because the conversation seems to have moved past promoting wishy-washy self-regulation and on to rules that could actually hold companies accountable. Frustrating because the debate seems to have forgotten the past five-plus years of AI policy. I just published a story looking at all the existing international efforts to regulate AI technology. You can read it here.
I’m not the only one who feels this way.
“To suggest that Congress starts from zero just plays into the industry’s favorite narrative, which is that Congress is so far behind and doesn’t understand technology—how could they ever regulate us?” says Anna Lenhart, a policy fellow at the Institute for Data Democracy and Policy at George Washington University, and a former Hill staffer.
In fact, politicians in the last Congress, which ran from January 2021 to January 2023, introduced a ton of legislation around AI. Lenhart put together this neat list of all the AI regulations proposed during that time. They cover everything from risk assessments to transparency to data protection. None of them made it to the president’s desk, but given that buzzy (or, to many, scary) new generative AI tools have captured Washington’s attention, Lenhart expects some of them to be revamped and make a reappearance in one form or another.
Here are a few to keep an eye on.
Algorithmic Accountability Act
This bill was introduced by Democrats in the US Congress and Senate in 2022, pre-ChatGPT, to address the tangible harms of automated decision-making systems, such as ones that denied people pain medications or rejected their mortgage applications.
The bill would require companies to do algorithmic impact and risk assessments, says Lenhart. It would also put the Federal Trade Commission in charge of regulating and enforcing rules around AI, and boost its staff numbers.
American Data Privacy Protection Act
This bipartisan bill was an attempt to regulate how companies collect and process data. It gained lots of momentum as a way to help women keep their personal health data safe after Roe v. Wade was overturned, but it failed to pass in time. The debate around the risks of generative AI could give it the added urgency to go further than last time. ADPPA would ban generative AI companies from collecting, processing, or transfering data in a discriminatory way. It would also give users more control over how companies use their data.
An AI agency
During the hearing, Altman and several senators suggested we need a new US agency to regulate AI. But I think this is a bit of a red herring. The US government needs more technical expertise and resources to regulate the tech, whether it be in a new agency or in a revamped existing one, Lenhart says. And more importantly, any regulator, new or old, needs the power to enforce the laws.
“It’s easy to create an agency and not give it any powers,” Lenhart says.
Another tech-focused agency is likely on the way. Senators Lindsey Graham, a Republican, and Elizabeth Warren, a Democrat, are working together to create a new digital regulator that might also have the power to police and perhaps license social media companies.
Democrat Chuck Schumer is also rallying the troops in the Senate to introduce a new bill that would tackle AI harms specifically. He has gathered bipartisan support to put together a comprehensive AI bill that would set up guardrails aimed at promoting responsible AI development. For example, companies might be required to allow external experts to audit their tech before it is released, and to give users and the government more information about their AI systems.
And while Altman seems to have won the Senate Judiciary Committee over, leaders from the commerce committees in both the House and Senate need to be on board for a comprehensive approach to AI regulation to become law, Lenhart says.
And it needs to happen fast, before people lose their interest in generative AI.
“It’s gonna be tricky, but anything’s possible,” Lenhart says.
Deeper Learning
Meta’s new AI models can recognize and produce speech for more than 1,000 languages
Meta has built AI models that can recognize and produce speech for more than 1,000 languages—a tenfold increase on what’s currently available.
Why this matters: It’s a significant step towards preserving languages that are at risk of disappearing, the company says. There are around 7,000 languages in the world, but existing speech recognition models only cover approximately 100 languages comprehensively. This is because these kinds of models tend to require huge amounts of labeled training data, which is only available for a small number of languages, including English, Spanish, and Chinese. Read more from Rhiannon Williams here.
Bits and Bytes
Google and Apple’s photo apps still can’t find gorillas Eight years ago, Google’s photo app mislabeled pictures of Black people as gorillas. The company prevented any pictures from being labeled as apes as a temporary fix. But years later, tech companies haven’t found a solution to the problem, despite big advancements in computer vision (The New York Times)
Apple bans employees from using ChatGPT It’s worried the chatbot might leak confidential company information. This is not an unreasonable concern, given that just a couple of months ago OpenAI had to pull ChatGPT offline because of a bug that leaked user chat history. (The Wall Street Journal)
Here’s how AI will ruin work for everyone Big Tech’s push to integrate AI into office tools will not spell the end of human labor. It’s the opposite: the easier work becomes, the more we will be expected to do. Or as Charlie Warzel writes, this AI boom is going to be less Skynet, more Bain & Company. (The Atlantic)
Does Bard know how many times “e” appears in “ketchup”? This was a fun piece with a serious purpose: lifting the lid on how large language models work. Google’s chatbot Bard doesn’t know how many letters different words have. This is because instead of recognizing individual letters, these models form words using “tokens.” So for example, Bard would think the first letter in the word “ketchup” was “ket,” not “k.” (The Verge)