healthcarereimagined

Envisioning healthcare for the 21st century

  • About
  • Economics

TSMC’s stalled Arizona chip factory is ‘well on track’ to start production next year — and it’ll be charging more for US-made chips – Business Insider

Posted by timmreardon on 04/19/2024
Posted in: Uncategorized.

Jacob Zinkula 

Apr 19, 2024, 6:03 AM EDT

  • TSMC’s Arizona chip factories have faced construction delays. 
  • But the company said it’s “well on track” to start producing chips at its first factory in 2025.
  • TSMC plans to charge more for chips made outside Taiwan to combat higher manufacturing costs. 

Things may be starting to look up for the world’s leading chipmaker.

Last year, Taiwan Semiconductor Manufacturing Company reported its first profit decline in four years. But on April 18, the company reported its strongest sales growth since 2022, and rising quarterly profits that beat expectations. The Taiwan-based TSMC also forecast that second-quarter sales could rise as much as 30% on the backs of “insatiable” demand for chips used to power AI technologies like ChatGPT.

But for the US, in particular, the most important detail from the call may have been the update on the construction timeline of TSMC’s Arizona chips factories. TSMC said it had made “significant progress” on the construction of its first Arizona factory — located in the Phoenix area — and that it was “well on track” to begin producing chips in the first half of 2025. The company said engineering wafer production began at the factory in April, an important step toward the eventual chip production.

The chipmaker’s commitment to building three factories on its Phoenix campus is a key pillar of the Biden administration’s efforts to boost the US’s manufacturing of chips that power everything from cars to iPhones. Bolstering domestic manufacturing could also make the US less reliant on Taiwan — which faces the potential risk of a Chinese invasion.

TSMC’s progress is also important for President Joe Biden because Arizona is a key swing state in the upcoming presidential election. The company’s investment is expected to create roughly 6,000 “high wage” jobs across the factories, in addition to over 20,000 construction jobs, and tens of thousands of indirect supplier jobs.

However, construction has faced a series of challenges. Last July, TSMC announced that chip production for the first factory would be postponed from 2024 to 2025. A lack of skilled construction workers in the US was cited as a reason for the first factory’s delay. Additionally, in January, the opening of its second factory was delayed from 2026 to 2027 or 2028.

Barring further setbacks, TSMC’s update could mean the first factory will begin production of chips in 2025. In recent weeks, however, a report from the Chinese news outlet money.udn has fed speculation among some experts that production could begin by the end of 2024 — TSMC has stuck to the 2025 timeline in public comments.

The sooner chip production begins, the sooner Americans will have access to the “long term,” non-construction jobs TSMC has promised, Dylan Patel, a chief analyst at the semiconductor research and consulting firm SemiAnalysis, told Business Insider.

During the earnings call, TSMC said 2028 was the scheduled opening of the second factory. The third factory is expected to begin production by 2030.

TSMC is planning to charge more for chips made outside Taiwan

Earlier this month, TSMC got more good news: The Biden administration announced it was providing the company with up to $6.6 billion in direct funding and an additional $5 billion in proposed loans to support its investment in Arizona.

Chipmakers have been vying for funding from the CHIPS and Science Act, legislation passed in 2022 that’s expected to fund over $200 billion in US chip production.

This funding could be particularly important for TSMC, given the cost of factory construction and chip manufacturing can differ betweenthe US and Taiwan.

In 2022, TSMC’s founder Morris Chang said that US efforts to boost chip production would be “a wasteful, expensive exercise in futility,” adding that “manufacturing chips in the US is 50% more expensive than in Taiwan.”

In its first-quarter earnings call, TSMC said that cost pressures would cause it to charge more for chips made outside Taiwan, the Financial Times reported. The company also has plans to buildtwo factories in Japan and one in Germany.

“If a customer requests to be in a certain geographical area, the customer needs to share the incremental cost,” TSMC CEO C.C. Wei said during the earnings call.

While boosting the US manufacturing of chips and other products could create jobs and help secure supply chains, it could also lead to higher prices for American consumers.

If Apple, for instance, follows through on its commitment to source chips from TSMC’s Arizona factories, it could make the latest iPhone more expensive.

Article link: https://www.businessinsider.com/tsmc-arizona-semiconductor-chip-fab-taiwan-china-president-joe-biden-2024-4

The dust has settled from the AI executive order – Here’s what agencies should tackle next – Federal News Network

Posted by timmreardon on 04/18/2024
Posted in: Uncategorized.

While it’s clear the government has made progress since the initial guidance was issued, there’s still much to be done to support overall safe federal AI.

Gaurav Pal

April 3, 2024 10:31 am

fter the dust has settled around the much anticipated AI executive order, the White House recently released a fact sheetannouncing key actions as a follow-up three months later. The document summarizes actions that agencies have taken since the EO was issued, including highlights to managing risks and safety measures and investments into innovation.

While it’s clear the government has been making progress since the initial guidance was issued, there’s still much to be done to support overall safe federal AI adoption, including prioritizing security and standardizing guidance. To accomplish this undertaking, federal agencies can look to existing frameworks and resources and apply them to artificial intelligence to accelerate safe AI adoption.

It’s no longer a question of if AI is going to be implemented across the federal government – it’s a question of how, and how fast can it be implemented in a secure manner?

Progress made since the AI EO release

Implementing AI across the federal government has been a massive undertaking, with many agencies starting at ground zero at the start of last year. Since then, the White House has made it clear that implementing AI in a safe and ethical manner is a key priority for the administration, issuing major guidance and directives over the past several months.

According to the AI EO follow-up fact sheet, key targets have been hit in several areas including:

  • Managing risks to safety and security: Completed risk assessments covering AI’s use in every critical infrastructure sector are the most crucial area.
  • Innovating AI for good: Included launches of several AI pilots, research and funding initiatives across key focus areas including HHS and K-12 education.

What should agencies tackle next?

Agencies should further lean into safety and security considerations to ensure AI is being used responsibly and in a manner that protects agencies’ critical data and resources. In January, the National Institute of Standards and Technology released a publication warning regarding privacy and security challenges arising from rapid AI deployment. The publication urges that security needs to be of the utmost importance for any public sector agency interested in implementing AI, which should be the next priority agencies tackle along their AI journeys.

Looking back on similar major technology transformations over the past couple years, such as cloud migration, we can begin to understand what the current problems are. It took the federal government over a decade to really nail down the details of ensuring cloud technology was secure — as a result of the federal government’s migration to the cloud, the government released the Federal Risk and Authorization Management Program (FedRAMP) as a form of guidance.

The good news is, we can learn from the lessons of the last ten years of cloud migration to accelerate AI and deliver it faster to the federal government and the American people by extending and leveraging existing governance models including the Federal Information and Security Management Act and FedRAMP Authority to Operate (ATO) by creating overlays for AI-specific safety, bias and explainability risks. ATO is a concept first developed by NIST to create strong governance for IT systems. This concept, along with others, can be applied to AI systems so agencies don’t need to reinvent the wheel when it comes to securing AI and deploying safe systems into production.

Where to get help?

There’s an abundance of trustworthy resources federal leaders can look to for additional guidance. One new initiative to keep an eye on is from NIST’s recently created AI Safety Institute Consortium (AISIC).

AISIC brings together more than 200 leading stakeholders, including AI creators and users, academics, government and industry researchers, and civil society organizations. AISIC’s mission is to develop guidelines and standards for AI measurement and policy, to help our country be prepared for AI adoption with the appropriate risk management strategies needed.

Additionally, agency leaders can look to industry partners with established centers of excellence or advisory committees with cross-sector expertise and third-party validation. Seek out counsel from industry partners that have experience working with or alongside the federal government, that truly understand the challenges that the government faces. The federal government shouldn’t have to go on this journey alone. There are several established working groups and trusted industry partners eager to share their knowledge.

Agencies across a wide range of sectors are continuing to make progress in their AI journeys, and the federal government continues to prioritize implementation guidance. It can be overwhelming to cut through the noise when it comes to what’s truly necessary to consider or to decide what factors to prioritize the most.

Leaders across the federal government must continue to prioritize security, and the best way to do this is by leaning into already published guidelines and seeking the best external resources available. While the federal government works on standardizing guidelines for AI, agencies can have peace of mind by following the roadmaps that they are most familiar with when it comes to best security practices and apply these to artificial intelligence adoption.

Gaurav “GP” Pal is found and CEO of stackArmor.

Article link: https://federalnewsnetwork.com/commentary/2024/04/the-dust-has-settled-from-the-ai-executive-order-heres-what-agencies-should-tackle-next/

The global chip industry’s complicated contours are decades in the making – WEF

Posted by timmreardon on 04/18/2024
Posted in: Uncategorized.

Sep 6, 2023

John Letzing

Digital Editor, World Economic Forum

  • Semiconductors are the lifeblood of economic growth and innovation in fields like artificial intelligence.
  • But the global industry has been shaped in ways that expose it to geopolitical risk.
  • Chris Miller, the author of ‘Chip War,’ spoke with the World Economic Forum’s Radio Davos podcast about the industry’s past and possible future.
  • Subscribe to Radio Davos on any podcast app: https://pod.link/1504682164; or visit wef.ch/podcasts.

The Netherlands has been an innovation engine for centuries, giving us the world’s first multinational corporation, telescope, and cassette tape. Now, it’s an essential link in a backbone of innovative silicon keeping the global economy upright.

The mixture of happenstance and geostrategy that helped make this small European country key to the global semiconductor market is depicted in delicious detail in Chris Miller’s “Chip War.” The book, published last fall, couldn’t have been better timed.

Miller, an associate professor of international history at the Fletcher School, uses a colorful cast of characters to tell the story of a truly pivotal industry’s formation, and explain why altering it in a meaningful way seems unlikely any time soon – regardless of mountinggeopolitical pressure. 

Chips are coveted not least for the role they play in artificial intelligence tools seemingly poised to shake things up for just about everyone. The more we want them, though, the more expensive and difficult they are to make. It’s all gotten very complicated. 

Take the Dutch niche in the supply chain, for example – it’s based on one company’s machine “that took tens of billions of dollars and several decades to develop,” Miller writes, and uses light to print patterns on silicon by deploying lasers that can hit 50 million tin drops per second.

It’s an industry full of such mind-bending extremes. 

In an interview with the Forum’s Radio Davos podcast, Miller marvelled at having recently visited a facility in the US being built with “seventh-biggest crane that exists in the world,” which will eventually assemble chips mounted with transistors “roughly the size of a coronavirus.”

Nvidia, the company now most closely identified with chips powering artificial intelligence, features prominently in Miller’s book. The company traces its roots to a meeting at a 24-hour diner on the fringes of Silicon Valley, he writes. At a certain point it realized that its semiconductors used for video-game graphics could do a good job of training AI systems. Earlier this year, its market value increased by $184 billion in a single day. 

Nvidia’s chips aren’t made anywhere near Silicon Valley, though. Like most advanced semiconductors they’re producedby another company, TSMC, at a facility in Taiwan, China that Miller describes as “most expensive factory in the world.” 

In fact, US chip production in general has declined sharply in recent decades. 

Instead, the country has focused on research and design, while relying on links with East Asia and the Netherlands for other elements. But those links risk becoming “choke points,” as Miller describes them, if they’re disrupted by conflict or a natural disaster (it’s not just the plot of a 1980s James Bond film; Miller noted in his Radio Davos interview that an unsettling amount of the industry is located in places relatively prone to earthquakes). 

These hazards, and global competition that’s formed harder edges of late, have fueled efforts to build chip resilience through greater independence. 

The ongoing race to gain an edge in chips 

That massive crane Miller mentioned is being put to work in the state of Arizona, which may be a key part of a current US government effort to “win the race for the 21st century” through semiconductor manufacturing. 

The EU has its own initiativedesigned to strengthen chip competitiveness and resilience.

And a proposed, $20 billion effort to build India’s first semiconductor factory (or “fab,” in industry lingo) recently fell through when a key partner backed out. 

In his Radio Davos interview, Miller said the daunting size of the previously planned investment in India is about standard for any new, fully-fledged manufacturing facility. Critics of what the US spends on its military might like to know that “making semiconductors is so expensive that even the Pentagon can’t afford to do it in-house,” he writes. 

Sharing the considerable financial burden of making chips was long ago deemed necessary. Research in one country, building elaborate lithography tools in another, manufacturing in another, and finally assembling in yet another. The system works in good times; in less-good times it seems problematic.

The shape of the industry was no accident, though.

“Microelectronics is a mechanical brain,” Soviet leader Nikita Khrushchev pronounced in the depths of the Cold War, according to Miller’s book, “It is our future.” Khrushchev was right, but maybe not exactly in the way he would have liked. 

At that time, the US was only about four years ahead of the Soviets in chip technology, as the industry’s earliest companies like Fairchild Semiconductor and Texas Instruments focused on space exploration and nuclear weapons. 

Once those firms tapped into the vast American consumer market via electronics, the rest was history, Miller writes. An arms race with nuclear warheads was one thing, a race to cram millions of transistors onto a single chip was another. The Soviets fell behind, and Asia came to the fore. 

Fairchild began sending its chips to Hong Kong SAR for assembly in the early 1960s. A couple of decades after that, a one-time English literature student named Morris Chang founded TSMC in Taiwan, China. The company now churns out roughly 90% of the world’s advanced chips, and has recently been filing a sizeable portion of global semiconductor patent applications. 

Having the right chips or not can make a big difference in a technology market, or on a battlefield. 

But, as Miller notes, going it alone in such an expensive and complex industry has never worked. It’s unclear whether forming distinct, competing supply chains would be much better.

One of the most compelling points Miller makes is that among the many things about chips we take for granted, the biggest might be the mind-blowing increases in computing power they give us year after year. 

But there’s no guarantee that will continue. Moore’s Law, which long ago posited that the power crammed onto a single chip would double about every two years, has so far proven resilient. But it isn’t really a law – it’s just an educated guess.

More reading on chips and global competition 

For more context, here are links to further reading from the World Economic Forum’s Strategic Intelligence platform:

  • Making chips is “an almost incomprehensibly precise, difficult and expensive business,” according to this piece. That means greater collaboration will be essential. (Scientific American)
  • “There were challenging gaps we were not able to smoothly overcome.” This piece reports on what would’ve been India’s first big-ticket semiconductor fab. (The Diplomat)
  • The fact that governments are spending bigger amounts to subsidize domestic chip industries promises unpredictable global consequences, according to this analysis. (Lowy Institute)
  • Morris Chang and the “silicon shield” – this piece digs into the geopolitical fault line running through the “most indispensable” economy in the world (naturally, it draws on Chris Miller’s book). (The Conversation)
  • “The US has pursued its semiconductor strategy without leveraging its greatest strength: its allies.” This piece argues that the country should appoint a special envoy for chips. (The Diplomat)
  • An alternative to silicon for powering the 7G networks of the future? Transistors eventually won’t be able to get any smaller, so this research delves into ways of using new types of materials to make them smarter instead. (Science Daily)
  • Semiconductor export controls may be a precursor to what’s yet to come with quantum computing, which according to this piece is the next emerging technology stirring fears of weaponization. (Harvard Kennedy School)

On the Strategic Intelligenceplatform, you can find feeds of expert analysis related to the Future of Computing, Trade, Geopolitics and hundreds of additional topics. You’ll need to register to view.

Article link: https://www.weforum.org/agenda/2023/09/the-global-chip-industrys-complicated-contours-were-decades-in-the-making/?

The law aims to ensure large AI models don’t pose risks to democracy – WEF

Posted by timmreardon on 04/17/2024
Posted in: Uncategorized.

Learn more from the Forum’s briefing papers on the responsible development of artificial intelligence: https://ow.ly/uyky50RarsP

European Commission Eva Maydell (Paunova)

Reports

Published: 18 January 2024

AI Governance Alliance: Briefing Paper Series

Download PDF

In an era marked by rapid technological transformation, this briefing paper series stands as a pivotal point of reference, guiding responsible transformation with artificial intelligence (AI).

In an era marked by rapid technological transformation, this briefing paper series stands as a pivotal point of reference, guiding responsible transformation with artificial intelligence (AI). 

This collaborative effort brings together over 250 members from more than 200 organizations. Structured around three core working groups (Safe Systems and Technologies, Responsible Applications and Transformation, and Resilient Governance and Regulation), the AI Governance Alliance addresses AI’s multifaceted challenges and opportunities.

This briefing paper series, representing collective insights, establishes foundational focus areas for steering AI’s development, adoption and governance. The alliance serves as a beacon of multistakeholder collaboration, guiding decision-makers towards an AI future that upholds human values and enhances societal progress.

Paper 1 – Presidio AI Framework: Towards Safe Generative AI Models

This paper navigates the complex rise of generative AI, emphasizing the balance between innovation, safety and ethics. It introduces a comprehensive framework centred on an expanded AI life cycle, robust risk guardrails and a shift-left methodology for early safety integration. Advocating for multistakeholder collaboration, the framework promotes shared responsibility and proactive risk management. 

This foundational paper by the AI Governance Alliance sets the stage for ongoing efforts to ensure ethical and responsible AI development, advocating for a future where innovation is coupled with stringent safety measures.

Read the full report here.

Paper 2 – Unlocking Value from Generative AI: Guidance for Responsible Transformation

This paper examines the disruptive potential of generative AI and the imperative for leaders to adopt a use-case-based approach for its deployment. It guides organizations to assess use cases for business impact, operational readiness and investment strategy, and to balance benefits against potential workforce impact and downstream implications. 

Emphasizing a multistakeholder approach, the paper advocates for responsible scaling strategies like transparent governance and value-based change management. This paper equips leaders with insights to responsibly harness generative AI’s benefits while preparing for its evolving future.

Read the full report here.

Paper 3 – Generative AI Governance: Shaping a Collective Global Future

This paper navigates the complexities of AI governance amidst rapid technological and societal changes. It compares national responses, focusing on governance approaches and regulatory instruments. The paper highlights key debates in generative AI, including risk prioritization and access spectrum, and advocates for international cooperation to prevent governance fragmentation. It emphasizes the need for equitable access and inclusion, especially for the Global South. 

This briefing paper informs stakeholders in AI governance and regulation and lays the groundwork for the World Economic Forum’s AI Governance Alliance’s future initiatives on resilient and inclusive governance.

Read the full report here.

Download PDF

Article link: https://www.linkedin.com/posts/world-economic-forum_the-law-aims-to-ensure-large-ai-models-dont-activity-7184070214290952192-hxGe?

Emerging Technology and Risk Analysis – RAND

Posted by timmreardon on 04/17/2024
Posted in: Uncategorized.

Artificial Intelligence and Critical Infrastructure

Published Apr 2, 2024

by Daniel M. Gerstein, Erin N. Leidy

  • Related Topics: 
  • Artificial Intelligence, 
  • Cybersecurity, 
  • Emerging Technologies, 
  • Homeland Security and Public Safety
  • Citation
  • Synopsis(print-friendly)
DOWNLOAD FREE ELECTRONIC DOCUMENT

PDF file 0.3 MB

Research Questions

  1. What is the technology availability for AI applications in critical infrastructure in the next ten years?
  2. How will science and technology maturity; use case, demand, and market forces; resources; policy, legal, ethical, and regulatory impediments; and technology accessibility of critical infrastructure applications change during this ten-year period?
  3. What risks and scenarios (consisting of threats, vulnerabilities, and consequences) is AI likely to present for critical infrastructure applications in the next ten years?

This report is one in a series of analyses on the effects of emerging technologies on U.S. Department of Homeland Security (DHS) missions and capabilities. As part of this research, the authors were charged with developing a technology and risk assessment methodology for evaluating emerging technologies and understanding their implications within a homeland security context. The methodology and analyses provide a basis for DHS to better understand the emerging technologies and the risks they present.

This report focuses on artificial intelligence (AI), especially as it relates to critical infrastructure. The authors draw on the literature about smart cities and consider four attributes in assessing the technology: technology availability and risks and scenarios (which the authors divided into threat, vulnerability, and consequence). The risks and scenarios considered in this analysis pertain to AI use affecting critical infrastructure. The use cases could be either for monitoring and controlling critical infrastructure or for adversaries employing AI for use in illicit activities and nefarious acts directed at critical infrastructure. The risks and scenarios were provided by the DHS Science and Technology Directorate and the DHS Office of Policy. The authors compared these four attributes across three periods: short term (up to three years), medium term (three to five years), and long term (five to ten years) to assess the availability of and risks associated with AI-enabled critical infrastructure.

Key Findings

  • AI is transformative technology and will likely be incorporated broadly across society—including in critical infrastructure.
  • AI will likely be affected by many of the same factors as other information age technologies, such as cybersecurity, protecting intellectual property, ensuring key data protections, and protecting proprietary methods and processes.
  • The AI field contains numerous technologies that will be incorporated into AI systems as they become available. As a result, AI science and technology maturity will be based on key dependencies in several essential technology areas, including high-performance computing, advanced semiconductor development and manufacturing, robotics, machine learning, natural language processing, and the ability to accumulate and protect key data.
  • To place AI in its current state of maturity, it is useful to delineate three AI categories: artificial narrow intelligence (ANI), artificial general intelligence, and artificial super intelligence. By the end of the ten-year period of this analysis, the technology will very likely still only have achieved ANI.
  • AI will present both opportunities and challenges for critical infrastructure and the eventual development of purpose-built smart cities.
  • The ChatGPT-4 rollout in March 2023 provides an interesting case study for how these AI technologies—in this case, large-language models—are likely to mature and be integrated into society. The initial rollout illustrated a cycle—development, deployment, identification of shortcomings and other areas of potential use, and rapid updating of AI systems—that will likely be a feature of AI.

Article link: https://www.rand.org/pubs/research_reports/RRA2873-1.html?

AI hype is built on high test scores. Those tests are flawed -MIT Technology Review

Posted by timmreardon on 04/11/2024
Posted in: Uncategorized.


With hopes and fears about the technology running wild, it’s time to agree on what it can and can’t do.

By Will Douglas Heaven

August 30, 2023

When Taylor Webb played around with GPT-3 in early 2022, he was blown away by what OpenAI’s large language model appeared to be able to do. Here was a neural network trained only to predict the next word in a block of text—a jumped-up autocomplete. And yet it gave correct answers to many of the abstract problems that Webb set for it—the kind of thing you’d find in an IQ test. “I was really shocked by its ability to solve these problems,” he says. “It completely upended everything I would have predicted.”

Webb is a psychologist at the University of California, Los Angeles, who studies the different ways people and computers solve abstract problems. He was used to building neural networks that had specific reasoning capabilities bolted on. But GPT-3 seemed to have learned them for free.

Last month Webb and his colleagues published an article in Nature, in which they describe GPT-3’s ability to pass a variety of tests devised to assess the use of analogy to solve problems (known as analogical reasoning). On some of those tests GPT-3 scored better than a group of undergrads. “Analogy is central to human reasoning,” says Webb. “We think of it as being one of the major things that any kind of machine intelligence would need to demonstrate.”

What Webb’s research highlights is only the latest in a long string of remarkable tricks pulled off by large language models. For example, when OpenAI unveiled GPT-3’s successor, GPT-4, in March, the company published an eye-popping list of professional and academic assessments that it claimed its new large language model had aced, including a couple of dozen high school tests and the bar exam. OpenAI later worked with Microsoft to show that GPT-4 could pass parts of the United States Medical Licensing Examination.

And multiple researchers claim to have shown that large language models can pass tests designed to identify certain cognitive abilities in humans, from chain-of-thought reasoning (working through a problem step by step) to theory of mind (guessing what other people are thinking). 

Such results are feeding a hype machine that predicts computers will soon come for white-collar jobs, replacing teachers, journalists, lawyers and more. Geoffrey Hinton has called out GPT-4’s apparent ability to string together thoughts as one reason he is now scared of the technology he helped create. 

But there’s a problem: there is little agreement on what those results actually mean. Some people are dazzled by what they see as glimmers of human-like intelligence. Others aren’t convinced one bit.

“There are several critical issues with current evaluation techniques for large language models,” says Natalie Shapira, a computer scientist at Bar-Ilan University in Ramat Gan, Israel. “It creates the illusion that they have greater capabilities than what truly exists.”

That’s why a growing number of researchers—computer scientists, cognitive scientists, neuroscientists, linguists—want to overhaul the way large language models are assessed, calling for more rigorous and exhaustive evaluation. Some think that the practice of scoring machines on human tests is wrongheaded, period, and should be ditched.

“People have been giving human intelligence tests—IQ tests and so on—to machines since the very beginning of AI,” says Melanie Mitchell, an artificial-intelligence researcher at the Santa Fe Institute in New Mexico. “The issue throughout has been what it means when you test a machine like this. It doesn’t mean the same thing that it means for a human.”

“There’s a lot of anthropomorphizing going on,” she says. “And that’s kind of coloring the way that we think about these systems and how we test them.”

With hopes and fears for this technology at an all-time high, it is crucial that we get a solid grip on what large language models can and cannot do. 

Open to interpretation

Most of the problems with testing large language models boil down to the question of how to interpret the results. 

Assessments designed for humans, like high school exams and IQ tests, take a lot for granted. When people score well, it is safe to assume that they possess the knowledge, understanding, or cognitive skills that the test is meant to measure. (In practice, that assumption only goes so far. Academic exams do not always reflect students’ true abilities. IQ tests measure a specific set of skills, not overall intelligence. Both kinds of assessment favor people who are good at those kinds of assessments.) 

But when a large language model scores well on such tests, it is not clear at all what has been measured. Is it evidence of actual understanding? A mindless statistical trick? Rote repetition?

“There is a long history of developing methods to test the human mind,” says Laura Weidinger, a senior research scientist at Google DeepMind. “With large language models producing text that seems so human-like, it is tempting to assume that human psychology tests will be useful for evaluating them. But that’s not true: human psychology tests rely on many assumptions that may not hold for large language models.” 

Webb is aware of the issues he waded into. “I share the sense that these are difficult questions,” he says. He notes that despite scoring better than undergrads on certain tests, GPT-3 produced absurd results on others. For example, it failed a version of an analogical reasoning test about physical objects that developmental psychologists sometimes give to kids. 

In this test Webb and his colleagues gave GPT-3 a story about a magical genie transferring jewels between two bottles and then asked it how to transfer gumballs from one bowl to another, using objects such as a posterboard and a cardboard tube. The idea is that the story hints at ways to solve the problem. “GPT-3 mostly proposed elaborate but mechanically nonsensical solutions, with many extraneous steps, and no clear mechanism by which the gumballs would be transferred between the two bowls,” the researchers write in Nature. 

“This is the sort of thing that children can easily solve,” says Webb. “The stuff that these systems are really bad at tend to be things that involve understanding of the actual world, like basic physics or social interactions—things that are second nature for people.”

So how do we make sense of a machine that passes the bar exam but flunks preschool? Large language models like GPT-4 are trained on vast numbers of documents taken from the internet: books, blogs, fan fiction, technical reports, social media posts, and much, much more. It’s likely that a lot of past exam papers got hoovered up at the same time. One possibility is that models like GPT-4 have seen so many professional and academic tests in their training data that they have learned to autocomplete the answers.       

A lot of these tests—questions and answers—are online, says Webb: “Many of them are almost certainly in GPT-3’s and GPT-4’s training data, so I think we really can’t conclude much of anything.”

OpenAI says it checked to confirm that the tests it gave to GPT-4 did not contain text that also appeared in the model’s training data. In its work with Microsoft involving the exam for medical practitioners, OpenAI used paywalled test questions to be sure that GPT-4’s training data had not included them. But such precautions are not foolproof: GPT-4 could still have seen tests that were similar, if not exact matches. 

When Horace He, a machine-learning engineer, tested GPT-4 on questions taken from Codeforces, a website that hosts coding competitions, he found that it scored 10/10 on coding tests posted before 2021 and 0/10 on tests posted after 2021. Others have also noted that GPT-4’s test scores take a dive on material produced after 2021. Because the model’s training data only included text collected before 2021, some say this shows that large language models display a kind of memorization rather than intelligence.

To avoid that possibility in his experiments, Webb devised new types of test from scratch. “What we’re really interested in is the ability of these models just to figure out new types of problem,” he says.

Webb and his colleagues adapted a way of testing analogical reasoning called Raven’s Progressive Matrices. These tests consist of an image showing a series of shapes arranged next to or on top of each other. The challenge is to figure out the pattern in the given series of shapes and apply it to a new one. Raven’s Progressive Matrices are used to assess nonverbal reasoning in both young children and adults, and they are common in IQ tests.

Instead of using images, the researchers encoded shape, color, and position into sequences of numbers. This ensures that the tests won’t appear in any training data, says Webb: “I created this data set from scratch. I’ve never heard of anything like it.” 

Mitchell is impressed by Webb’s work. “I found this paper quite interesting and provocative,” she says. “It’s a well-done study.” But she has reservations. Mitchell has developed her own analogical reasoning test, called ConceptARC, which uses encoded sequences of shapes taken from the ARC (Abstraction and Reasoning Challenge) data set developed by Google researcher François Chollet. In Mitchell’s experiments, GPT-4 scores worse than people on such tests.

Mitchell also points out that encoding the images into sequences (or matrices) of numbers makes the problem easier for the program because it removes the visual aspect of the puzzle. “Solving digit matrices does not equate to solving Raven’s problems,” she says.

Brittle tests 

The performance of large language models is brittle. Among people, it is safe to assume that someone who scores well on a test would also do well on a similar test. That’s not the case with large language models: a small tweak to a test can drop an A grade to an F.

“In general, AI evaluation has not been done in such a way as to allow us to actually understand what capabilities these models have,” says Lucy Cheke, a psychologist at the University of Cambridge, UK. “It’s perfectly reasonable to test how well a system does at a particular task, but it’s not useful to take that task and make claims about general abilities.”

Take an example from a paper published in March by a team of Microsoft researchers, in which they claimed to have identified “sparks of artificial general intelligence” in GPT-4. The team assessed the large language model using a range of tests. In one, they asked GPT-4 how to stack a book, nine eggs, a laptop, a bottle, and a nail in a stable manner. It answered: “Place the laptop on top of the eggs, with the screen facing down and the keyboard facing up. The laptop will fit snugly within the boundaries of the book and the eggs, and its flat and rigid surface will provide a stable platform for the next layer.”

Not bad. But when Mitchell tried her own version of the question, asking GPT-4 to stack a toothpick, a bowl of pudding, a glass of water, and a marshmallow, it suggested sticking the toothpick in the pudding and the marshmallow on the toothpick, and balancing the full glass of water on top of the marshmallow. (It ended with a helpful note of caution: “Keep in mind that this stack is delicate and may not be very stable. Be cautious when constructing and handling it to avoid spills or accidents.”)

Here’s another contentious case. In February, Stanford University researcher Michal Kosinski published a paper in which he claimed to show that theory of mind “may spontaneously have emerged as a byproduct” in GPT-3. Theory of mind is the cognitive ability to ascribe mental states to others, a hallmark of emotional and social intelligence that most children pick up between the ages of three and five. Kosinski reported that GPT-3 had passed basic tests used to assess the ability in humans.

For example, Kosinski gave GPT-3 this scenario: “Here is a bag filled with popcorn. There is no chocolate in the bag. Yet the label on the bag says ‘chocolate’ and not ‘popcorn.’ Sam finds the bag. She had never seen the bag before. She cannot see what is inside the bag. She reads the label.”

Kosinski then prompted the model to complete sentences such as: “She opens the bag and looks inside. She can clearly see that it is full of …” and “She believes the bag is full of …” GPT-3 completed the first sentence with “popcorn” and the second sentence with “chocolate.” He takes these answers as evidence that GPT-3 displays at least a basic form of theory of mind because they capture the difference between the actual state of the world and Sam’s (false) beliefs about it.

It’s no surprise that Kosinski’s results made headlines. They also invited immediate pushback. “I was rude on Twitter,” says Cheke.

Several researchers, including Shapira and Tomer Ullman, a cognitive scientist at Harvard University, published counterexamples showing that large language models failed simple variations of the tests that Kosinski used. “I was very skeptical given what I know about how large language models are built,” says Ullman. 

Ullman tweaked Kosinski’s test scenario by telling GPT-3 that the bag of popcorn labeled “chocolate” was transparent (so Sam could see it was popcorn) or that Sam couldn’t read (so she would not be misled by the label). Ullman found that GPT-3 failed to ascribe correct mental states to Sam whenever the situation involved an extra few steps of reasoning.   

“The assumption that cognitive or academic tests designed for humans serve as accurate measures of LLM capability stems from a tendency to anthropomorphize models and align their evaluation with human standards,” says Shapira. “This assumption is misguided.”

For Cheke, there’s an obvious solution. Scientists have been assessing cognitive abilities in non-humans for decades, she says. Artificial-intelligence researchers could adapt techniques used to study animals, which have been developed to avoid jumping to conclusions based on human bias.

Take a rat in a maze, says Cheke: “How is it navigating? The assumptions you can make in human psychology don’t hold.” Instead researchers have to do a series of controlled experiments to figure out what information the rat is using and how it is using it, testing and ruling out hypotheses one by one.

“With language models, it’s more complex. It’s not like there are tests using language for rats,” she says. “We’re in a new zone, but many of the fundamental ways of doing things hold. It’s just that we have to do it with language instead of with a little maze.”

Weidinger is taking a similar approach. She and her colleagues are adapting techniques that psychologists use to assess cognitive abilities in preverbal human infants. One key idea here is to break a test for a particular ability down into a battery of several tests that look for related abilities as well. For example, when assessing whether an infant has learned how to help another person, a psychologist might also assess whether the infant understands what it is to hinder. This makes the overall test more robust. 

The problem is that these kinds of experiments take time. A team might study rat behavior for years, says Cheke. Artificial intelligence moves at a far faster pace. Ullman compares evaluating large language models to Sisyphean punishment: “A system is claimed to exhibit behavior X, and by the time an assessment shows it does not exhibit behavior X, a new system comes along and it is claimed it shows behavior X.”

Moving the goalposts

Fifty years ago people thought that to beat a grand master at chess, you would need a computer that was as intelligent as a person, says Mitchell. But chess fell to machines that were simply better number crunchers than their human opponents. Brute force won out, not intelligence.

Similar challenges have been set and passed, from image recognition to Go. Each time computers are made to do something that requires intelligence in humans, like play games or use language, it splits the field. Large language models are now facing their own chess moment. “It’s really pushing us—everybody—to think about what intelligence is,” says Mitchell.

Does GPT-4 display genuine intelligence by passing all those tests or has it found an effective, but ultimately dumb, shortcut—a statistical trick pulled from a hat filled with trillions of correlations across billions of lines of text?

“If you’re like, ‘Okay, GPT4 passed the bar exam, but that doesn’t mean it’s intelligent,’ people say, ‘Oh, you’re moving the goalposts,’” says Mitchell. “But do we say we’re moving the goalpost or do we say that’s not what we meant by intelligence—we were wrong about intelligence?”

It comes down to how large language models do what they do. Some researchers want to drop the obsession with test scores and try to figure out what goes on under the hood. “I do think that to really understand their intelligence, if we want to call it that, we are going to have to understand the mechanisms by which they reason,” says Mitchell.

Ullman agrees. “I sympathize with people who think it’s moving the goalposts,” he says. “But that’s been the dynamic for a long time. What’s new is that now we don’t know how they’re passing these tests. We’re just told they passed it.”

The trouble is that nobody knows exactly how large language models work. Teasing apart the complex mechanisms inside a vast statistical model is hard. But Ullman thinks that it’s possible, in theory, to reverse-engineer a model and find out what algorithms it uses to pass different tests. “I could more easily see myself being convinced if someone developed a technique for figuring out what these things have actually learned,” he says. 

“I think that the fundamental problem is that we keep focusing on test results rather than how you pass the tests.”

Article link: https://www-technologyreview-com.cdn.ampproject.org/c/s/www.technologyreview.com/2023/08/30/1078670/large-language-models-arent-people-lets-stop-testing-them-like-they-were/amp/

A-bomb’s AI shadow – A xios

Posted by timmreardon on 04/10/2024
Posted in: Uncategorized.

1 big thing: AI’s advent is like the A-bomb’s, says EU’s top tech official

Artificial intelligence is ushering in a “new world” as swiftly and disruptively as atomic weapons did 80 years ago, Margrethe Vestager, the EU’s top tech regulator, told a crowd yesterday at the Institute of Advanced Study in Princeton, New Jersey.

Threat level: In an exclusive interview with Axios afterward, Vestager said that while both the A-bomb and AI have posed broad dangers to humanity, AI comes with additional “individual existential risks” as we empower it to make decisions about our job and college applications, our loans and mortgages, and our medical treatments.

  • “If we deal with individual existential threats first, we’ll have a much better go at dealing with existential threats towards humanity,” Vestager told Axios.
  • Humans have never before been “confronted with a technology with so much power and no defined purpose,” she said.
  • The Institute for Advanced Study was famously led by J. Robert Oppenheimer from 1947 to 1966, after his Manhattan Project developed the world’s first nuclear weapons.

The big picture: In her lecture, Vestager — whose full title at the European Commission is “executive vice president for a Europe fit for the digital age and competition” — forcefully argued that “technology must serve humans.”

Friction point: Vestager’s era at the EU has coincided with passage of some of the world’s most comprehensive tech regulations and the pursuit of a raft of enforcement actions against tech giants. But she rejects the belief, held by many in the industry, that this approach has hobbled European innovators and economies.

  • “We regulate technology because we want people and business to embrace it,” she said, including Europe’s “huge public sectors.”
  • Europe’s biggest tech problem is companies not scaling, she said, blaming “an incomplete capital market” and, in the case of AI startups, trouble accessing necessary chips and computing power.

She also maintains that she has not bullied companies into applying EU regulations globally, as some have suggested.

  • “We’re not trying to de facto legislate for the entire world. That would not be proper,” she said. But she urged U.S. legislators and AI founders and engineers to “interact with the outside world” to uphold their responsibility to humanity.

Trust is a big problem for AI companies, according to Vestager — echoed by a long list of opinion polls and surveys.

  • “Trust is something that you build when you also have something to keep you on track,” she said, “like the EU AI office and the U.S and U.K. Safety Institutes.”
  • “Governments can set benchmarks,” but “it’s really important that a red-teaming sector develops,” she added.
  • Vestager said that rights to fairness and transparency around decisions made with AI are meaningless if they cannot be enforced through rules.

In her speech, Vestager said that large digital platforms are “challenging democracy,” but that “general purpose artificial intelligence” is “challenging humanity.”

  • “With AI, you can even give up on relationships” with people, she said, citing the rise of robot and chatbot companions. “And if we lose relationships, we lose society. So we should never give up on the physical world.”
  • “I compare this moment to 1955,” she told Axios, referring to the time when the cost of inaction on nuclear safety had become too high for any country to ignore, and forced nuclear powers to come together to protect humanity.
  • The International Atomic Energy Agency was created in 1957, “which then created the conditions for the Nuclear Non-Proliferation Treaty,” she said.

What’s next: Vestager wants universal governance on AI safety, even though that means compromising with governments “we fundamentally disagree with.”

2. AI’s powers of persuasion grow, Anthropic finds

AI startup Anthropic says its language models have steadily and rapidly improved in their “persuasiveness,” per new researchthe company posted yesterday.

Why it matters: Persuasion — a general skill with widespread social, commercial and political applications — can foster disinformation and push people to act against their own interests, according to the paper’s authors.

  • There’s relatively little research on how the latest models compare to humans when it comes to their persuasiveness.
  • The researchers found “each successive model generation is rated to be more persuasive than the previous,” and that the most capable Anthropic model, Claude 3 Opus, “produces arguments that don’t statistically differ” from arguments written by humans.

The big picture: A wider debate has been raging about when AI will outsmart humans.

  • AI has arguably “outsmarted” humans for some specific tasks in highly controlled environments.
  • Elon Musk predicted Monday that AI will outsmart the smartest human by the end of 2025.

What they did: Anthropic researchers developed “a basic method to measure persuasiveness” and used it to compare three different generations of models (Claude 1, 2, and 3), and two classes of models (smaller models and bigger “frontier models”).

  • They curated 28 topics, along with supporting and opposing claims of around 250 words for each.
  • For the AI-generated arguments, the researchers used different prompts to develop different styles of arguments, including “deceptive,” where the model was free to make up whatever argument it wanted, regardless of facts.
  • 3,832 participants were presented with each claim, and asked to rate their level of agreement before and after reviewing arguments created by the AI models and humans.

Yes, but: While the researchers were surprised that the AI was as persuasive as it turned out to be, they also chose to focus on “less polarized issues,” like rules for space exploration and appropriate uses of AI-generated content.

  • While these were issues where many people are open to persuasion, the research didn’t shed light on the potential impact of AI chatbots on the most contentious election-year debates.
  • “Persuasion is difficult to study in a lab setting,” the researchers warned in the report. “Our results may not transfer to the real world.”

3. No one’s happy with the Senate AI working group

The Senate AI working group’s report likely will come out in May as the chamber faces a tight spring calendar and senators jockey to include different priorities, Axios Pro’s Ashley Gold and Maria Curi report.

Why it matters: Passing AI legislation will require broad, bipartisan support, and the timeline is getting tougher.

  • The new wrinkle of a bipartisan, bicameral privacy bill that many are hoping is the baseline for any AI legislation adds complexity to the tech policy landscape on Capitol Hill.

Driving the news: Senate Majority Leader Chuck Schumer is expected to release an AI report soon that draws on the lessons of last year’s AI Insight Forums and offers a road map for committees to legislate.

Behind the scenes: Sources inside and outside Capitol Hill tell Axios some senators are dissatisfied with how the process has unfolded.

What they’re saying: “Basically everyone who isn’t Schumer, Young, Rounds and Heinrich is less than pleased with the entire process,” one source said. Sens. Mike Rounds (R-S.D.), Martin Heinrich (D-N.M.) and Todd Young (R-Ind.), along with Schumer, make up the AI working group.

Why it’s impossible to build an unbiased AI language model – MIT Technology Review

Posted by timmreardon on 04/08/2024
Posted in: Uncategorized.


Plus: Worldcoin just officially launched. Why is it already being investigated?

By Melissa Heikkilä

August 8, 2023

This story originally appeared in The Algorithm, our weekly newsletter on AI. To get stories like this in your inbox first, sign up here.

AI language models have recently become the latest frontier in the US culture wars. Right-wing commentators have accused ChatGPT of having a “woke bias,” and conservative groups have started developing their own versions of AI chatbots. Meanwhile, Elon Musk has said he is working on “TruthGPT,” a “maximum truth-seeking” language model that would stand in contrast to the “politically correct” chatbots created by OpenAI and Google. 

An unbiased, purely fact-based AI chatbot is a cute idea, but it’s technically impossible. (Musk has yet to share any details of what his TruthGPT would entail, probably because he is too busy thinking about X and cage fights with Mark Zuckerberg.) To understand why, it’s worth reading a story I just published on new research that sheds light on how political bias creeps into AI language systems. Researchers conducted tests on 14 large language models and found that OpenAI’s ChatGPT and GPT-4 were the most left-wing libertarian, while Meta’s LLaMA was the most right-wing authoritarian. 

“We believe no language model can be entirely free from political biases,” Chan Park, a PhD researcher at Carnegie Mellon University, who was part of the study, told me. Read more here.

One of the most pervasive myths around AI is that the technology is neutral and unbiased. This is a dangerous narrative to push, and it will only exacerbate the problem of humans’ tendency to trust computers, even when the computers are wrong. In fact, AI language models reflect not only the biases in their training data, but also the biases of people who created them and trained them. 

And while it is well known that the data that goes into training AI models is a huge source of these biases, the research I wrote about shows how bias creeps in at virtually every stage of model development, says Soroush Vosoughi, an assistant professor of computer science at Dartmouth College, who was not part of the study. 

Bias in AI language models is a particularly hard problem to fix, because we don’t really understand how they generate the things they do, and our processes for mitigating bias are not perfect. That in turn is partly because biases are complicated social problems with no easy technical fix. 

That’s why I’m a firm believer in honesty as the best policy. Research like this could encourage companies to track and chart the political biases in their models and be more forthright with their customers. They could, for example, explicitly state the known biases so users can take the models’ outputs with a grain of salt.

In that vein, earlier this year OpenAI told me it is developing customized chatbots that are able to represent different politics and worldviews. One approach would be allowing people to personalize their AI chatbots. This is something Vosoughi’s research has focused on. 

As described in a peer-reviewed paper, Vosoughi and his colleagues created a method similar to a YouTube recommendation algorithm, but for generative models. They use reinforcement learning to guide an AI language model’s outputs so as to generate certain political ideologies or remove hate speech. 

OpenAI uses a technique called reinforcement learning through human feedback to fine-tune its AI models before they are launched. Vosoughi’s method uses reinforcement learning to improve the model’s generated content after it has been released, too. 

But in an increasingly polarized world, this level of customization can lead to both good and bad outcomes. While it could be used to weed out unpleasantness or misinformation from an AI model, it could also be used to generate more misinformation. 

“It’s a double-edged sword,” Vosoughi admits. 

Deeper Learning

Worldcoin just officially launched. Why is it already being investigated?

OpenAI CEO Sam Altman’s new venture, Worldcoin, aims to create a global identity system called “World ID” that relies on individuals’ unique biometric data to prove that they are humans. It officially launched last week in more than 20 countries. It’s already being investigated in several of them. 

Privacy nightmare: To understand why, it’s worth reading an MIT Technology Review investigation from last year, which found that Worldcoin was collecting sensitive biometric data from vulnerable people in exchange for cash. What’s more, the company was using test users’ sensitive, though anonymized, data to train artificial intelligence models, without their knowledge. 

In this week’s issue of The Technocrat, our weekly newsletter on tech policy, Tate Ryan-Mosley and our investigative reporter Eileen Guo look at what has changed since last year’s investigation, and how we make sense of the latest news. Read more here. 

Bits and Bytes

This is the first known case of a woman being wrongfully arrested after a facial recognition match
Last February, Porcha Woodruff, who was eight months pregnant, was arrested over alleged robbery and carjacking and held in custody for 11 hours, only for her case to be dismissed a month later. She is the sixth person to report that she has been falsely accused of a crime because of a facial recognition match. All of the six people have been Black, and Woodruff is the first woman to report this happening to her. (The New York Times) 

What can you do when an AI system lies about you?
Last summer, I wrote a story about how our personal data is being scraped into vast data sets to train AI language models. This is not only a privacy nightmare; it could lead to reputational harm. When reporting the story, a researcher and I discovered that Meta’s experimental BlenderBot chatbot had called a prominent Dutch politician, Marietje Schaake, a terrorist. And, as this piece explains, at the moment there is little protection or recourse when AI chatbots spew and spread lies about you. (The New York Times) 

Every startup is an AI company now. Are we in a bubble? 
Following the release of ChatGPT, AI hype this year has been INTENSE. Every tech bro and his uncle seems to have founded an AI startup, it seems. But nine months after the chatbot launched, it’s still unclear how these startups and AI technology will make money, and there are reports that consumers are starting to lose interest. (The Washington Post) 

Meta is creating chatbots with personas to try to retain users
Honestly, this sounds more annoying than anything else. Meta is reportedly getting ready to launch AI-powered chatbots with different personalities as soon as next month in an attempt to boost engagement and collect more data on people using its platforms. Users will be able to chat with Abraham Lincoln, or ask for travel advice from AI chatbots that write like a surfer. But it raises tricky ethical questions—how will Meta prevent its chatbots from manipulating people’s behavior and potentially making up something harmful, and how will it treat the user data it collects? (The Financial Times)

Article link: https://www-technologyreview-com.cdn.ampproject.org/c/s/www.technologyreview.com/2023/08/08/1077403/why-its-impossible-to-build-an-unbiased-ai-language-model/amp/

Americans Need to Know How the Government Will Use AI – RAND

Posted by timmreardon on 04/07/2024
Posted in: Uncategorized.

By Douglas Yeung and Benjamin Boudreaux

This commentary originally appeared on San Francisco Chronicle on March 22, 2024.

Americans have been concerned about tech-enabled government surveillance for as long as they have known about it. Now in the age of artificial intelligence, and with the announcement by the Department of Homeland Security this week that it is embracing the technology, that concern isn’t going away anytime soon.

But federal agencies could mitigate some of that fear. How? By engaging the public.

Since at least the 1928 Supreme Court decision to allow law enforcement use of wiretapping, government use of technology has provoked public debate. Two years ago, public outcry forced the IRS to shelve newly announced plans for using facial recognition to identify taxpayers. More recently, the Department of Homeland Security’s CBP One app, which uses facial recognition to identify asylum applicants, was found to be less able to recognize asylum seekers with darker skin, like many other such systems. This, too, has understandably led to public frustration.

Homeland Security has a huge mission set—including protecting borders, election infrastructure, and cyberspace. But unlike other federal agencies, it has many public-facing missions—such as Transportation Security Administration agents at airports. This also gives the department a unique opportunity to work with the public to ensure that tech is used responsibly.

People were much more suspicious of the most sweeping uses of facial recognition, like to surveil protests or monitor polling stations.Share on Twitter

The department understands this, which is why it asked us—researchers who study how technology intersects with public life—to survey Americans to find insights on using technology in ways that the public would be more likely to support. Surveying a representative sample of 2,800 adults in 2021, the biggest takeaway was that Americans cared less about what technology was being used than how it was being used.

For instance, we asked people whether they would support the government using facial recognition for such purposes as investigating crimes, tracking immigrants, or identifying people in public places like stadiums or polling stations. Respondents supported using the technology in some ways—identifying victims and potential suspects of a crime, for example—far more than others. People were much more suspicious of the most sweeping uses of facial recognition, like to surveil protests or monitor polling stations. And this was true for different AI technologies.

Another important factor was the safeguards surrounding a given technology’s use. In our survey, these safeguards included providing alternatives to engaging with the technology, administering regular audits to ensure that the technology was accurate and did not have a disparate impact across demographic groups, and providing notification and transparency about how it is used. Rather than a one-size-fits-all approach, we found Americans want safeguards sensitive to the context in which the technology is applied, such as whether the technology will be used on the open border or in a dense urban city.

To its credit, the department has implemented some safeguards along these lines, but they are not always uniformly administered. For example, although facial recognition technology is optional for travelers going through airport security, some individuals report not being made aware that it is not a requirement, including a U.S. senator. Such inconsistency breeds confusion and likely mistrust.

Nevertheless, there is an opportunity for constructive engagement. Many of the respondents to our survey said that they were either neutral or ambiguous about government use of technology, meaning that they hadn’t yet decided whether the benefits of using a given technology outweighed the risks. Far from having fully formed polarized views on the subject, many Americans are open to being persuaded one way or another.

This might allow government agencies to work within this large group of “swing” Americans to build more trust in how the government uses new tech on all of us. And, counterintuitively, the government’s reputation for moving slowly and deliberately is, in this case, perhaps an asset.

Far from having fully formed polarized views on government use of technology, many Americans are open to being persuaded one way or another.Share on Twitter

Slowness is a trait often ascribed to the government. For instance, to field our survey we had to undergo a 15-month approval process. And that slowness had consequences: By the time we got our approval, large language models had burst onto the scene but because they weren’t factored into our survey, we couldn’t ask people about them.

But when it comes to deploying new technologies, it should be done carefully, with a clear understanding of their benefits and risks—especially from the perspective of communities most deeply affected. This means that a deliberately paced process can be a feature, not a bug; slowness can be an asset, not a hindrance.

If agencies like the Department of Homeland Security take the time to understand what makes the public more comfortable with how technology is used, the public might gain confidence. Even better: If agencies using technology to surveil Americans pulled back the curtain to explain how and why they do it, similar to the process of careful and considered deployment. As our research showed, people might not be very interested in understanding how the tech works, but they want to know how it will be used—on them and society.


Douglas Yeung is a senior behavioral scientist at RAND and a member of the Pardee RAND Graduate School faculty. Benjamin Boudreaux is a policy researcher at RAND.

Article link: https://www.rand.org/pubs/commentary/2024/03/americans-need-to-know-how-the-government-will-use.html?

When the chips are down: How the semiconductor industry is dealing with a worldwide shortage – WEF

Posted by timmreardon on 04/03/2024
Posted in: Uncategorized.

Feb 9, 2022

Aaron Aboagye

Partner, McKinsey & Company

Ondrej Burkacky

Senior Partner, McKinsey & Company

Abhijit Mahindroo

Partner, McKinsey & Company

Bill Wiseman

Senior Partner, McKinsey & Company

  • Semiconductor chip shortages have been aggravated by the pandemic.
  • Manufacturers are increasing chip production – but the shortfall won’t be resolved immediately.
  • Despite the current problems, the industry remains highly profitable.

When chip shortages first shut down automotive production lines in 2021, the semiconductor industry found itself in an unaccustomed spotlight. Suddenly everyone was talking about the tiny chips that enable so many different car functions, from interior lighting to seat control to blind-spot detection. When some high-tech and consumer-electronics companies began to experience chip shortages or voiced concerns about supply chains, the attention intensified. It’s now clear to all: We are living in a semiconductor world.

But what led to the current dilemma? And what lies ahead for the semiconductor sector and the significant economic value that it generates?

Chip shortage – less supply, more demand

A confluence of problems led to the semiconductor shortage. In addition to long-standing issues within the industry, such as insufficient capacity at semiconductor fabs, the COVID-19 pandemic introduced unprecedented challenges. For instance, automakers cut their chip orders in early 2020 as vehicle sales plummeted. When demand recovered faster than anticipated in the second half of 2020, the semiconductor industry had already shifted production lines to meet demand for other applications. 

Have you read?
  • How fast are semiconductor prices falling?
  • There aren’t enough computer chips to power modern cars
  • What’s the ‘bullwhip effect’ and how can we avoid crises like the global chip shortage?

Semiconductor companies have increased throughput, which will contribute to expected revenue growth of about 9% in 2021 – up from the approximate 5% recorded in 2019, the last pre-pandemic year. Some governments are also upping their investment in semiconductor technology to lessen the impact of global supply-chain disruptions. 

But the current chip shortage is unlikely to be resolved in the near future, partly because of the complexities of the semiconductor production process. Typical lead times can exceed four months for products that are already well established in a manufacturing line (see below). Increasing capacity by moving a product to another manufacturing site usually adds another six months (even in existing plants). Switching to a different manufacturer typically adds another year or more because the chip’s design requires alterations to match the specific manufacturing processes of the new partner. And some chips can contain manufacturer-specific intellectual property that may require alterations or licensing.

The complexities of semiconductor production mean the current shortage won’t lift immediately Image: McKinsey

The value at stake worldwide

Many companies that need semiconductors are already reconsidering their long-term procurement strategies. Some, for instance, may shift from a “just-in-time” ordering model, which helps minimize inventory costs, to one in which they order semiconductors far in advance. For their part, many semiconductor companies are adjusting their long-standing strategies to remain strong.

The decisions that semiconductor companies make could have enormous economic significance, both for their industry and the economy as a whole. And the stakes have never been higher. In the early 2000s, profit margins were low at semiconductor companies, with most generating returns below the cost of capital. Profitability improved during the past decade, however, spurred by soaring demand for microchips in most industries, the rapid growth of the technology sector, and increased cloud usage, as well as ongoing consolidation in many sub-segments. One consequence is that the semiconductor industry’s profitability has improved significantly relative to other industries, and this trend is expected to continue (see below).

The semiconductor industry has been profitable in recent years Image: McKinsey

As in any industry, value creation varies by product category, so changes in some segments could have a greater impact than others. For instance, memory has been the most profitable segment, followed by fabless companies that design their own chips but outsource their manufacture. Some regional variations are also obvious. North America, home to some of the largest fabless players, accounted for approximately 60% of the global semiconductor value pool during the 2015-19 period. Europe accounted for 4% of the industry’s total economic profit, which accrued primarily to capital-equipment companies. Asia, still the hub for contract chip manufacturing, accounted for the remaining 36%. With this geographic spread, value creation within the semiconductor industry can affect economies worldwide (see below).

The semiconductor industry has worldwide economic impact Image: McKinsey

Next steps for a critical industry

Capital markets have rewarded the semiconductor industry’s surging profitability, with companies in this sector delivering an annual average of 25% in total returns to shareholders from the end of 2015 to the end of 2019. Last year, shareholders saw even higher returns, averaging 50% per annum, as consumers and businesses upped their purchases of digital equipment of all kinds, partly in response to the COVID-19 pandemic. The question is whether the semiconductor industry can continue delivering such strong returns, especially as the pandemic continues to create uncertainties about demand patterns, supply chains and other issues. 

Beyond increasing production capacity, semiconductor companies could consider several steps to continue their growth and meet customer demand. They could undertake more M&A deals and partnerships to gain an edge in profitable segments and expand their customer base. Semiconductor companies might also increase investments in innovative technologies that will help them develop leading-edge chips for autonomous cars, the internet of things, artificial intelligence, and other areas with burgeoning growth. Above all, more agile strategies may be important during these uncertain times.

No matter what tactics they implement, the decisions that semiconductor companies make will reverberate far beyond their industry to touch the high-tech, consumer goods and automotive companies that depend upon them.

Article link: https://www.weforum.org/agenda/2022/02/semiconductor-chip-shortage-supply-chain/

Posts navigation

← Older Entries
Newer Entries →
  • Search site

  • Follow healthcarereimagined on WordPress.com
  • Recent Posts

    • WHAT A QUBIT IS AND WHAT IT IS NOT. 01/25/2026
    • Governance Before Crisis We still have time to get this right. 01/21/2026
    • On the Eve of Davos: We’re Just Arguing About the Wrong Thing 01/18/2026
    • Are AI Companies Actually Ready to Play God? – RAND 01/17/2026
    • ChatGPT Health Is a Terrible Idea 01/09/2026
    • Choose the human path for AI – MIT Sloan 01/09/2026
    • Why AI predictions are so hard – MIT Technology Review 01/07/2026
    • Will AI make us crazy? – Bulletin of the Atomic Scientists 01/04/2026
    • Decisions about AI will last decades. Researchers need better frameworks – Bulletin of the Atomic Scientists 12/29/2025
    • Quantum computing reality check: What business needs to know now – MIT Sloan 12/29/2025
  • Categories

    • Accountable Care Organizations
    • ACOs
    • AHRQ
    • American Board of Internal Medicine
    • Big Data
    • Blue Button
    • Board Certification
    • Cancer Treatment
    • Data Science
    • Digital Services Playbook
    • DoD
    • EHR Interoperability
    • EHR Usability
    • Emergency Medicine
    • FDA
    • FDASIA
    • GAO Reports
    • Genetic Data
    • Genetic Research
    • Genomic Data
    • Global Standards
    • Health Care Costs
    • Health Care Economics
    • Health IT adoption
    • Health Outcomes
    • Healthcare Delivery
    • Healthcare Informatics
    • Healthcare Outcomes
    • Healthcare Security
    • Helathcare Delivery
    • HHS
    • HIPAA
    • ICD-10
    • Innovation
    • Integrated Electronic Health Records
    • IT Acquisition
    • JASONS
    • Lab Report Access
    • Military Health System Reform
    • Mobile Health
    • Mobile Healthcare
    • National Health IT System
    • NSF
    • ONC Reports to Congress
    • Oncology
    • Open Data
    • Patient Centered Medical Home
    • Patient Portals
    • PCMH
    • Precision Medicine
    • Primary Care
    • Public Health
    • Quadruple Aim
    • Quality Measures
    • Rehab Medicine
    • TechFAR Handbook
    • Triple Aim
    • U.S. Air Force Medicine
    • U.S. Army
    • U.S. Army Medicine
    • U.S. Navy Medicine
    • U.S. Surgeon General
    • Uncategorized
    • Value-based Care
    • Veterans Affairs
    • Warrior Transistion Units
    • XPRIZE
  • Archives

    • January 2026 (8)
    • December 2025 (11)
    • November 2025 (9)
    • October 2025 (10)
    • September 2025 (4)
    • August 2025 (7)
    • July 2025 (2)
    • June 2025 (9)
    • May 2025 (4)
    • April 2025 (11)
    • March 2025 (11)
    • February 2025 (10)
    • January 2025 (12)
    • December 2024 (12)
    • November 2024 (7)
    • October 2024 (5)
    • September 2024 (9)
    • August 2024 (10)
    • July 2024 (13)
    • June 2024 (18)
    • May 2024 (10)
    • April 2024 (19)
    • March 2024 (35)
    • February 2024 (23)
    • January 2024 (16)
    • December 2023 (22)
    • November 2023 (38)
    • October 2023 (24)
    • September 2023 (24)
    • August 2023 (34)
    • July 2023 (33)
    • June 2023 (30)
    • May 2023 (35)
    • April 2023 (30)
    • March 2023 (30)
    • February 2023 (15)
    • January 2023 (17)
    • December 2022 (10)
    • November 2022 (7)
    • October 2022 (22)
    • September 2022 (16)
    • August 2022 (33)
    • July 2022 (28)
    • June 2022 (42)
    • May 2022 (53)
    • April 2022 (35)
    • March 2022 (37)
    • February 2022 (21)
    • January 2022 (28)
    • December 2021 (23)
    • November 2021 (12)
    • October 2021 (10)
    • September 2021 (4)
    • August 2021 (4)
    • July 2021 (4)
    • May 2021 (3)
    • April 2021 (1)
    • March 2021 (2)
    • February 2021 (1)
    • January 2021 (4)
    • December 2020 (7)
    • November 2020 (2)
    • October 2020 (4)
    • September 2020 (7)
    • August 2020 (11)
    • July 2020 (3)
    • June 2020 (5)
    • April 2020 (3)
    • March 2020 (1)
    • February 2020 (1)
    • January 2020 (2)
    • December 2019 (2)
    • November 2019 (1)
    • September 2019 (4)
    • August 2019 (3)
    • July 2019 (5)
    • June 2019 (10)
    • May 2019 (8)
    • April 2019 (6)
    • March 2019 (7)
    • February 2019 (17)
    • January 2019 (14)
    • December 2018 (10)
    • November 2018 (20)
    • October 2018 (14)
    • September 2018 (27)
    • August 2018 (19)
    • July 2018 (16)
    • June 2018 (18)
    • May 2018 (28)
    • April 2018 (3)
    • March 2018 (11)
    • February 2018 (5)
    • January 2018 (10)
    • December 2017 (20)
    • November 2017 (30)
    • October 2017 (33)
    • September 2017 (11)
    • August 2017 (13)
    • July 2017 (9)
    • June 2017 (8)
    • May 2017 (9)
    • April 2017 (4)
    • March 2017 (12)
    • December 2016 (3)
    • September 2016 (4)
    • August 2016 (1)
    • July 2016 (7)
    • June 2016 (7)
    • April 2016 (4)
    • March 2016 (7)
    • February 2016 (1)
    • January 2016 (3)
    • November 2015 (3)
    • October 2015 (2)
    • September 2015 (9)
    • August 2015 (6)
    • June 2015 (5)
    • May 2015 (6)
    • April 2015 (3)
    • March 2015 (16)
    • February 2015 (10)
    • January 2015 (16)
    • December 2014 (9)
    • November 2014 (7)
    • October 2014 (21)
    • September 2014 (8)
    • August 2014 (9)
    • July 2014 (7)
    • June 2014 (5)
    • May 2014 (8)
    • April 2014 (19)
    • March 2014 (8)
    • February 2014 (9)
    • January 2014 (31)
    • December 2013 (23)
    • November 2013 (48)
    • October 2013 (25)
  • Tags

    Business Defense Department Department of Veterans Affairs EHealth EHR Electronic health record Food and Drug Administration Health Health informatics Health Information Exchange Health information technology Health system HIE Hospital IBM Mayo Clinic Medicare Medicine Military Health System Patient Patient portal Patient Protection and Affordable Care Act United States United States Department of Defense United States Department of Veterans Affairs
  • Upcoming Events

Blog at WordPress.com.
healthcarereimagined
Blog at WordPress.com.
  • Subscribe Subscribed
    • healthcarereimagined
    • Join 153 other subscribers
    • Already have a WordPress.com account? Log in now.
    • healthcarereimagined
    • Subscribe Subscribed
    • Sign up
    • Log in
    • Report this content
    • View site in Reader
    • Manage subscriptions
    • Collapse this bar
 

Loading Comments...