Connect with us

The Conversation

Cancer research in the US is world class because of its broad base of funding − with the government pulling out, its future is uncertain

Published

on

theconversation.com – Jeffrey MacKeigan, Professor of Pediatrics and Human Development, Michigan State University – 2025-04-28 07:36:00

Without federal support, the lights will turn off in many labs across the country.
Thomas Barwick/Stone via Getty Images

Jeffrey MacKeigan, Michigan State University

Cancer research in the U.S. doesn’t rely on a single institution or funding stream − it’s a complex ecosystem made up of interdependent parts: academia, pharmaceutical companies, biotechnology startups, federal agencies and private foundations. As a cancer biologist who has worked in each of these sectors over the past three decades, I’ve seen firsthand how each piece supports the others.

When one falters, the whole system becomes vulnerable.

The United States has long led the world in cancer research. It has spent more on cancer research than any other country, including more than US$7.2 billion annually through the National Cancer Institute alone. Since the 1971 National Cancer Act, this sustained public investment has helped drive dramatic declines in cancer mortality, with death rates falling by 34% since 1991. In the past five years, the Food and Drug Administration has approved over 100 new cancer drugs, and the U.S. has brought more cancer drugs to the global market than any other nation.

But that legacy is under threat. Funding delays, political shifts and instability across sectors have created an environment where basic research into the fundamentals of cancer biology is struggling to keep traction and the drug development pipeline is showing signs of stress.

These disruptions go far beyond uncertainty and have real consequences. Early-career scientists faced with unstable funding and limited job prospects may leave academia altogether. Mid-career researchers often spend more time chasing scarce funding than conducting research. Interrupted research budgets and shifting policy priorities can unravel multiyear collaborations. I, along with many other researchers, believe these setbacks will slow progress, break training pipelines and drain expertise from critical areas of cancer research – delays that ultimately hurt patients waiting for new treatments.

A 50-year foundation of federal investment

The modern era of U.S. cancer research began with the signing of the National Cancer Act in 1971. That law dramatically expanded the National Cancer Institute, an agency within the National Institutes of Health focusing on cancer research and education. The NCI laid the groundwork for a robust national infrastructure for cancer science, funding everything from early research in the lab to large-scale clinical trials and supporting the training of a generation of cancer researchers.

This federal support has driven advances leading to higher survival rates and the transformation of some cancers into a manageable chronic or curable condition. Progress in screening, diagnostics and targeted therapies – and the patients who have benefited from them – owe much to decades of NIH support.

The Trump administration is cutting billions of dollars of biomedical research funding.

But federal funding has always been vulnerable to political headwinds. During the first Trump administration, deep cuts to biomedical science budgets threatened to stall the progress made under initiatives such as the 2016 Cancer Moonshot. The rationale given for these cuts was to slash overall spending, despite facing strong bipartisan opposition in Congress. Lawmakers ultimately rejected the administration’s proposal and instead increased NIH funding. In 2022, the Biden administration worked to relaunch the Cancer Moonshot.

This uncertainty has worsened in 2025 as the second Trump administration has cut or canceled many NIH grants. Labs that relied on these awards are suddenly facing funding cliffs, forcing them to lay off staff, pause experiments or shutter entirely. Deliberate delays in communication from the Department of Health and Human Services have stalled new NIH grant reviews and funding decisions, putting many promising research proposals already in the pipeline at risk.

Philanthropy’s support is powerful – but limited

While federal agencies remain the backbone of cancer research funding, philanthropic organizations provide the critical support for breakthroughs – especially for new ideas and riskier projects.

Groups such as the American Cancer Society, Stand Up To Cancer and major hospital foundations have filled important gaps in support, often funding pilot studies or supporting early-career investigators before they secure federal grants. By supporting bold ideas and providing seed funding, they help launch innovative research that may later attract large-scale support from the NIH.

Without the bureaucratic constraints of federal agencies, philanthropy is more nimble and flexible. It can move faster to support work in emerging areas, such as immunotherapy and precision oncology. For example, the American Cancer Society grant review process typically takes about four months from submission, while the NIH grant review process takes an average of eight months.

Crowd of people in white T-shirts reading 'RUN JEFF RUN' standing in front of a backdrop of a sign with the American Cancer Society logo and another sign reading 'CALL IN YOUR PLEDGE...'
Ted Kennedy Jr., right, and Jeff Keith raise money for the American Cancer Society in 1984.
Mikki Ansin/Getty Images

But philanthropic funds are smaller in scale and often disease-specific. Many foundations are created around a specific cause, such as advancing cures for pancreatic, breast or pediatric cancers. Their urgency to make an impact allows them to fund bold approaches that federal funders may see as too preliminary or speculative. Their giving also fluctuates. For instance, the American Cancer Society awarded nearly $60 million less in research grants in 2020 compared with 2019.

While private foundations are vital partners for cancer research, they cannot replace the scale and consistency of federal funding. Total U.S. philanthropic funding for cancer research is estimated at a few billion dollars per year, spread across hundreds of organizations. In comparison, the federal government has typically contributed roughly five to eight times more than philanthropy to cancer research each year.

Industry innovation − and its priorities

Private-sector innovation is essential for translating discoveries into treatments. In 2021, nearly 80% of the roughly $57 billion the U.S. spent on cancer drugs came from pharmaceutical and biotech companies. Many of the treatments used in oncology today, including immunotherapies and targeted therapies, emerged from collaborations between academic labs and industry partners.

But commercial priorities don’t always align with public health needs. Companies naturally focus on areas with strong financial returns: common cancers, projects that qualify for fast-track regulatory approval, and high-priced drugs. Rare cancers, pediatric cancers and basic science often receive less attention.

Industry is also saddled with uncertainty. Rising R&D costs, tough regulatory requirements and investor wariness have created a challenging environment to bring new drugs to market. Several biotech startups have folded or downsized in the past year, leaving promising new drugs stranded in limbo in the lab before they can reach clinical trials.

Without federal or philanthropic entities to pick up the slack, these discoveries may never reach the patients who need them.

A system under strain

Cancer is not going away. As the U.S. population ages, the burden of cancer on society will only grow. Disparities in treatment access and outcomes persist across race, income and geography. And factors such as environmental exposures and infectious diseases continue to intersect with cancer risk in new and complex ways.

Addressing these challenges requires a strong, stable and well-coordinated research system. But that system is under strain. National Cancer Institute grant paylines, or funding cutoffs, remain highly competitive. Early-career researchers face precarious job prospects. Labs are losing technicians and postdoctoral researchers to higher-paying roles in industry or to burnout. And patients, especially those hoping to enroll in clinical trials, face delays, disruptions and dwindling options.

Protectors holding signs reading 'SUPPORT SCIENCE' and 'IN SCIECE WE TRUST,' among others
Researchers have been rallying to protect the future of science in the U.S.
AP Photo/John McDonnell

This is not just a funding issue. It’s a coordination issue between the federal government, academia and industry. There are currently no long-term policy solutions that ensure sustained federal investment, foster collaboration between academia and industry, or make room for philanthropy to drive innovation instead of just filling gaps.

I believe that for the U.S. to remain a global leader in cancer research, it will need to recommit to the model that made success possible: a balanced ecosystem of public funding, private investment and nonprofit support. Up until recently, that meant fully funding the NIH and NCI with predictable, long-term budgets that allow labs to plan for the future; incentivizing partnerships that move discoveries from bench to bedside without compromising academic freedom; supporting career pathways for young scientists so talent doesn’t leave the field; and creating mechanisms for equity to ensure that research includes and benefits all communities.

Cancer research and science has come a long way, saving about 4.5 million lives in the U.S. from cancer from 1991 to 2022. Today, patients are living longer and better because of decades of hard-won discoveries made by thousands of researchers. But science doesn’t run on good intentions alone. It needs universities. It needs philanthropy. It needs industry. It needs vision. And it requires continued support from the federal government.The Conversation

Jeffrey MacKeigan, Professor of Pediatrics and Human Development, Michigan State University

This article is republished from The Conversation under a Creative Commons license. Read the original article.

Read More

The post Cancer research in the US is world class because of its broad base of funding − with the government pulling out, its future is uncertain appeared first on theconversation.com



Note: The following A.I. based commentary is not part of the original article, reproduced above, but is offered in the hopes that it will promote greater media literacy and critical thinking, by making any potential bias more visible to the reader –Staff Editor.

Political Bias Rating: Center-Left

This article reflects a centrist-left perspective, primarily emphasizing the critical role of federal funding in cancer research and its vulnerability due to political shifts, particularly under the Trump administration. It also highlights the challenges faced by researchers and the urgent need for stable funding to maintain U.S. leadership in cancer research. While it acknowledges the importance of private and philanthropic contributions, it leans towards advocating for government involvement in maintaining a balanced and effective research ecosystem. The discussion of past funding cuts and their impact further signals a mild left-leaning concern over government policy changes.

The Conversation

How is paint made?

Published

on

theconversation.com – Dawn Rogala, Paintings Conservator and Program Manager, Smithsonian Institution – 2025-09-08 07:07:00


Paint is made by mixing pigments—colorful substances found in rocks, plants, insects, or made synthetically—with binders that help spread and hold the color on surfaces. Traditionally, artists combined pigments with natural materials like water or oil, while modern factories produce large quantities with synthetic additives for durability. The paint’s ingredients depend on who uses it, its purpose, application method, and environment. Beyond art, paints protect objects like houses and cars. Smithsonian conservators study painted artifacts to learn about history using techniques like X-rays and 3D scans, helping preserve cultural heritage and uncovering stories behind materials and methods.

Protective paint sprayed onto a steel plate in a factory will have a different recipe than paint used in an art class.
gilaxia/E+ via Getty Images

Dawn Rogala, Smithsonian Institution and Gwénaëlle Kavich, Smithsonian Institution

Curious Kids is a series for children of all ages. If you have a question you’d like an expert to answer, send it to curiouskidsus@theconversation.com.


How is paint made? – Atharva, age 11, Bengaluru, India


Did you ever mix dirt and water when you were playing outside? You made a paint. Did you draw shapes on the ground with your muddy hands? You made a painting.

Paint is made by combining a colorful substance – a pigment – with another material that binds the color together and helps spread that color onto surfaces such as paper, fabric or wood. Pigments can be found everywhere – in rocks and minerals, plants or insects. Some colors are made by scientists in laboratories.

Long ago, artists made their own paints by mixing pigments with natural materials such as water, oil or egg yolk to hold the colors together in a paste. Artists today can still make their own paints, or they can order them from factories that mix, package and ship paint all over the world. Paint companies use large, industrial machines to grind pigments and binders together; these commercial paints include synthetic materials and preservatives to control the paint’s behavior and to help paint last longer in tubes or cans.

Paints and coatings do many jobs beyond just coloring paper in an artist’s studio. They are also used as protective coatings to shield houses and cars from the sun or the cold, or as a barrier between boats and the water that surrounds their wood, metal or plastic parts. Where and how a paint will be used influence how it’s made and with what ingredients.

an open box of watercolor paints with splatters of color on the case
Watercolor sets like this one used by artist Alma Thomas can be found in art classrooms around the world.
Anacostia Community Museum, Smithsonian Institution, Gift of David Driskell, CC BY

Choosing the right materials

A lot of questions need to be answered before materials are chosen for a paint.

  • Who will use the paint? An artist, a house painter, an armadillo, a robot at an assembly plant?
  • Why is the paint being used? For museum paintings and sculptures? In designs for furniture or mailboxes?
  • How will the paint be applied? By brush, by spray, or some other way?
  • Where and when will the paint be used? Does it need to dry quickly or slowly? Will the painted surface get really cold or hot? Is the paint safe for kids to use at home or school?
  • What should the paint look like? Should the dried paint be shiny or matte? Should the surface be lumpy, or should it flatten and level out? Should the colors be bright or dull? Should the paint layers be opaque, transparent or almost clear? Does the paint need to hold up against scuffs and stains?

There are many different companies that design and make the wide range of paints used around the world for all these various applications. Experts at each manufacturer understand their special type of paint, how the paint materials are measured and mixed, and the best ways to store and apply the paint. A single factory can make tens of thousands of gallons of paint each day, and paint companies produce millions of tubes of paint every year.

two boards with various colors of paint dried on them along with multiple paint brushes
Artist Thomas Moran’s palettes and brushes illustrate the way an artist mixes different paints to find just the desired qualities.
Smithsonian American Art Museum, Bequest of Miss Ruth B. Moran

Using paint to learn about the past

We work at the Smithsonian’s Museum Conservation Institute, where we study and conserve the diverse collection of painted objects at the Smithsonian – from planes and spacecraft to portraits of presidents and maps covered in abstract swirls of color. Bright coatings are part of everything from the painted clothing and cultural items of Native peoples to the pots and pans used by chef Julia Child.

Art conservators and conservation scientists like us work together to study and preserve cultural heritage such as paintings and painted objects. Studying paint helps us learn about the past and protect this history for future generations.

The paint colors used on large, traditional Indian paintings called “pichwai,” for example, include pigments gathered from around the world. They can reveal information about ancient manufacturing and how communities that lived far apart exchanged goods and knowledge.

There are many techniques to investigate artwork, from looking at small pieces of paint under a microscope to using more complicated equipment to analyze materials exposed to different types of energy. For example, we can use X-ray, infrared or ultraviolet imaging to identify different pigments in a painting.

three side by side images of the same painting, but one looks very dark, one is colorful, and one is grey and white
Conservation scientists will image the same work of art, such as this Indian pichwai, using ultraviolet fluorescence (left), visible light (middle) and infrared light (right).
National Museum of Asian Art, Smithsonian Institution, Gift of Karl B. Mann, S1992.28, Department of Conservation and Scientific Research, Orthomosaics and UV Fluorescence

Research on an Alaskan Tlingit crest hat made in the 1800s looked at the molecules in paint binders, combined with 3D scanning, to help clan members replicate the hat for ceremonial use.

Unusual uses bring conservation challenges

Artists use all sorts of materials in their artwork that were designed for other purposes. Some 19th- and early 20th-century sculptures were painted with laundry bluing – a material that used blue pigment to brighten clothes during washing. In the 1950s, artists started using thin, quick-drying house paint in their paintings.

When paints are used in a way that was not part of their design, strange things can happen. Paints made to be applied in thin layers but instead are used in thick layers can wrinkle and pucker as they dry. Paints designed to stick to rough wood can curl or lift away from slick surfaces. The colors and ingredients in paint can also fade or darken over time. Some artists want these different effects in their artwork; some artists are surprised when paints don’t behave the way they expected.

Art conservators and conservation scientists use information about artists and their paints to understand why artworks are faded, broken or acting in surprising ways, and they use that knowledge to slow or stop the damage. We can even clean some kinds of damage with lasers.

The more we know about paint, the more we learn about the past lives of painted objects and how to keep those objects around for a long, long time.


Hello, curious kids! Do you have a question you’d like an expert to answer? Ask an adult to send your question to CuriousKidsUS@theconversation.com. Please tell us your name, age and the city where you live.

And since curiosity has no age limit – adults, let us know what you’re wondering, too. We won’t be able to answer every question, but we will do our best.The Conversation

Dawn Rogala, Paintings Conservator and Program Manager, Smithsonian Institution and Gwénaëlle Kavich, Conservation Scientist, Smithsonian Institution

This article is republished from The Conversation under a Creative Commons license. Read the original article.

Read More

The post How is paint made? appeared first on theconversation.com



Note: The following A.I. based commentary is not part of the original article, reproduced above, but is offered in the hopes that it will promote greater media literacy and critical thinking, by making any potential bias more visible to the reader –Staff Editor.

Political Bias Rating: Centrist

The content is an educational piece explaining how paint is made and used, with no evident political agenda or bias. It focuses on scientific, historical, and artistic aspects in a neutral and informative manner, suitable for a general audience including children. There is no indication of leaning toward any political ideology or partisan perspective.

Continue Reading

The Conversation

Scientific objectivity is a myth – cultural values and beliefs always influence science and the people who do it

Published

on

theconversation.com – Sara Giordano, Associate Professor of Interdisciplinary Studies, Kennesaw State University – 2025-09-04 07:53:00


The article explores the myth of scientific objectivity, showing how science is deeply intertwined with cultural values and social context. It challenges traditional views, such as the passive egg and active sperm narrative, revealing that scientific knowledge often reflects societal norms. Science emerged as a quest for objectivity within Western universities over centuries, but the strict division between subjective humanities and objective sciences is arbitrary and hierarchical. Scientists, being cultural beings, influence research choices and interpretations unconsciously. Contemporary controversies, like vaccine debates, highlight the impossibility of bias-free science. Instead, democratic, collaborative processes are advocated to align research with societal values, fostering more honest and inclusive scientific inquiry.

People are at the heart of the scientific enterprise.
Matteo Farinella, CC BY-NC

Sara Giordano, Kennesaw State University

Even if you don’t recall many facts from high school biology, you likely remember the cells required for making babies: egg and sperm. Maybe you can picture a swarm of sperm cells battling each other in a race to be the first to penetrate the egg.

For decades, scientific literature described human conception this way, with the cells mirroring the perceived roles of women and men in society. The egg was thought to be passive while the sperm was active.

The opening credits of the 1989 movie ‘Look Who’s Talking’ animated this popular narrative, with speaking sperm rushing toward the nonverbal egg to be the first to fertilize it.

Over time, scientists realized that sperm are too weak to penetrate the egg and that the union is more mutual, with the two cells working together. It’s no coincidence that these findings were made in the same era when new cultural ideas of more egalitarian gender roles were taking hold.

Scientist Ludwik Fleck is credited with first describing science as a cultural practice in the 1930s. Since then, understanding has continued to build that scientific knowledge is always consistent with the cultural norms of its time.

Despite these insights, across political differences, people strive for and continue to demand scientific objectivity: the idea that science should be unbiased, rational and separable from cultural values and beliefs.

When I entered my Ph.D. program in neuroscience in 2001, I felt the same way. But reading a book by biologist Anne Fausto-Sterling called “Sexing the Body” set me down a different path. It systematically debunked the idea of scientific objectivity, showing how cultural ideas about sex, gender and sexuality were inseparable from the scientific findings. By the time I earned my Ph.D., I began to look more holistically at my research, integrating the social, historical and political context.

From the questions scientists begin with, to the beliefs of the people who conduct the research, to choices in research design, to interpretation of the final results, cultural ideas constantly inform “the science.” What if an unbiased science is impossible?

Emergence of idea of scientific objectivity

Science grew to be synonymous with objectivity in the Western university system only over the past few hundred years.

In the 15th and 16th centuries, some Europeans gained traction in challenging the religiously ordained royal order. Consolidation of the university system led to shifts from trust in religious leaders interpreting the word of “god,” to trust in “man” making one’s own rational decisions, to trust in scientists interpreting “nature.” The university system became an important site for legitimizing claims through theories and studies.

Previously, people created knowledge about their world, but there were not strict boundaries between what are now called the humanities, such as history, English and philosophy, and the sciences, including biology, chemistry and physics. Over time, as questions arose about how to trust political decisions, people split the disciplines into categories: subjective versus objective. The splitting came with the creation of other binary oppositions, including the closely related emotionality/rationality divide. These categories were not simply seen as opposite, but in a hierarchy with objectivity and rationality as superior.

A closer look shows that these binary systems are arbitrary and self-reinforcing.

Science is a human endeavor

The sciences are fields of study conducted by humans. These people, called scientists, are part of cultural systems just like everyone else. We scientists are part of families and have political viewpoints. We watch the same movies and TV shows and listen to the same music as nonscientists. We read the same newspapers, cheer for the same sports teams and enjoy the same hobbies as others.

All of these obviously “cultural” parts of our lives are going to affect how scientists approach our jobs and what we consider “common sense” that does not get questioned when we do our experiments.

Beyond individual scientists, the kinds of studies that get conducted are based on what questions are deemed relevant or not by dominant societal norms.

For example, in my Ph.D. work in neuroscience, I saw how different assumptions about hierarchy could influence specific experiments and even the entire field. Neuroscience focuses on what is called the central nervous system. The name itself describes a hierarchical model, with one part of the body “in charge” of the rest. Even within the central nervous system, there was a conceptual hierarchy with the brain controlling the spinal cord.

My research looked more at what happened peripherally in muscles, but the predominant model had the brain at the top. The taken-for-granted idea that a system needs a boss mirrors cultural assumptions. But I realized we could have analyzed the system differently and asked different questions. Instead of the brain being at the top, a different model could focus on how the entire system communicates and works together at coordination.

Every experiment also has assumptions baked in – things that are taken for granted, including definitions. Scientific experiments can become self-fulfilling prophecies.

For example, billions of dollars have been spent on trying to delineate sex differences. However, the definition of male and female is almost never stated in these research papers. At the same time, evidence mounts that these binary categories are a modern invention not based on clear physical differences.

But the categories are tested so many times that eventually some differences are discovered without putting these results into a statistical model together. Oftentimes, so-called negative findings that don’t identify a significant difference are not even reported. Sometimes, meta-analyses based on multiple studies that investigated the same question reveal these statistical errors, as in the search for sex-related brain differences. Similar patterns of slippery definitions that end up reinforcing taken-for-granted assumptions happen with race, sexuality and other socially created categories of difference.

Finally, the end results of experiments can be interpreted in many different ways, adding another point where cultural values are injected into the final scientific conclusions.

Settling on science when there’s no objectivity

Vaccines. Abortion. Climate change. Sex categories. Science is at the center of most of today’s hottest political debates. While there is much disagreement, the desire to separate politics and science seems to be shared. On both sides of the political divide, there are accusations that the other side’s scientists cannot be trusted because of political bias.

RFK Jr, Donald Trump and Dr. Oz seated at a table with flags behind them
It can be easier to spot built-in bias in scientific perspectives that conflict with your own values.
Jim Watson/AFP via Getty Images

Consider the recent controversy over the U.S. Centers for Disease Control and Prevention’s vaccine advisory panel. Secretary of Health and Human Services Robert F. Kennedy Jr. fired all members of the Advisory Committee on Immunization Practices, saying they were biased, while some Democratic lawmakers argued back that his move put in place those who would be biased in pushing his vaccine-skeptical agenda.

If removing all bias is impossible, then, how do people create knowledge that can be trusted?

The understanding that all knowledge is created through cultural processes does allow for two or more differing truths to coexist. You see this reality in action around many of today’s most controversial subjects. However, this does not mean you must believe all truths equally – that’s called total cultural relativism. This perspective ignores the need for people to come to decisions together about truth and reality.

Instead, critical scholars offer democratic processes for people to determine which values are important and for what purposes knowledge should be developed. For example, some of my work has focused on expanding a 1970s Dutch model of the science shop, where community groups come to university settings to share their concerns and needs to help determine research agendas. Other researchers have documented other collaborative practices between scientists and marginalized communities or policy changes, including processes for more interdisciplinary or democratic input, or both.

I argue a more accurate view of science is that pure objectivity is impossible. Once you leave the myth of objectivity behind, though, the way forward is not simple. Instead of a belief in an all-knowing science, we are faced with the reality that humans are responsible for what is researched, how it is researched and what conclusions are drawn from such research.

With this knowledge, we have the opportunity to intentionally set societal values that inform scientific investigations. This requires decisions about how people come to agreements about these values. These agreements need not always be universal but instead can be dependent on the context of who and what a given study might affect. While not simple, using these insights, gained over decades of studying science from both within and outside, may force a more honest conversation between political positions.The Conversation

Sara Giordano, Associate Professor of Interdisciplinary Studies, Kennesaw State University

This article is republished from The Conversation under a Creative Commons license. Read the original article.

Alternative views on the relationship between science and culture.
Sara Giordano

Science is a human endeavor

The sciences are fields of study conducted by humans. These people, called scientists, are part of cultural systems just like everyone else. We scientists are part of families and have political viewpoints. We watch the same movies and TV shows and listen to the same music as nonscientists. We read the same newspapers, cheer for the same sports teams and enjoy the same hobbies as others.

All of these obviously “cultural” parts of our lives are going to affect how scientists approach our jobs and what we consider “common sense” that does not get questioned when we do our experiments.

Beyond individual scientists, the kinds of studies that get conducted are based on what questions are deemed relevant or not by dominant societal norms.

For example, in my Ph.D. work in neuroscience, I saw how different assumptions about hierarchy could influence specific experiments and even the entire field. Neuroscience focuses on what is called the central nervous system. The name itself describes a hierarchical model, with one part of the body “in charge” of the rest. Even within the central nervous system, there was a conceptual hierarchy with the brain controlling the spinal cord.

My research looked more at what happened peripherally in muscles, but the predominant model had the brain at the top. The taken-for-granted idea that a system needs a boss mirrors cultural assumptions. But I realized we could have analyzed the system differently and asked different questions. Instead of the brain being at the top, a different model could focus on how the entire system communicates and works together at coordination.

Every experiment also has assumptions baked in – things that are taken for granted, including definitions. Scientific experiments can become self-fulfilling prophecies.

For example, billions of dollars have been spent on trying to delineate sex differences. However, the definition of male and female is almost never stated in these research papers. At the same time, evidence mounts that these binary categories are a modern invention not based on clear physical differences.

But the categories are tested so many times that eventually some differences are discovered without putting these results into a statistical model together. Oftentimes, so-called negative findings that don’t identify a significant difference are not even reported. Sometimes, meta-analyses based on multiple studies that investigated the same question reveal these statistical errors, as in the search for sex-related brain differences. Similar patterns of slippery definitions that end up reinforcing taken-for-granted assumptions happen with race, sexuality and other socially created categories of difference.

Finally, the end results of experiments can be interpreted in many different ways, adding another point where cultural values are injected into the final scientific conclusions.

Settling on science when there’s no objectivity

Vaccines. Abortion. Climate change. Sex categories. Science is at the center of most of today’s hottest political debates. While there is much disagreement, the desire to separate politics and science seems to be shared. On both sides of the political divide, there are accusations that the other side’s scientists cannot be trusted because of political bias.

RFK Jr, Donald Trump and Dr. Oz seated at a table with flags behind them

It can be easier to spot built-in bias in scientific perspectives that conflict with your own values.
Jim Watson/AFP via Getty Images

Consider the recent controversy over the U.S. Centers for Disease Control and Prevention’s vaccine advisory panel. Secretary of Health and Human Services Robert F. Kennedy Jr. fired all members of the Advisory Committee on Immunization Practices, saying they were biased, while some Democratic lawmakers argued back that his move put in place those who would be biased in pushing his vaccine-skeptical agenda.

If removing all bias is impossible, then, how do people create knowledge that can be trusted?

The understanding that all knowledge is created through cultural processes does allow for two or more differing truths to coexist. You see this reality in action around many of today’s most controversial subjects. However, this does not mean you must believe all truths equally – that’s called total cultural relativism. This perspective ignores the need for people to come to decisions together about truth and reality.

Instead, critical scholars offer democratic processes for people to determine which values are important and for what purposes knowledge should be developed. For example, some of my work has focused on expanding a 1970s Dutch model of the science shop, where community groups come to university settings to share their concerns and needs to help determine research agendas. Other researchers have documented other collaborative practices between scientists and marginalized communities or policy changes, including processes for more interdisciplinary or democratic input, or both.

I argue a more accurate view of science is that pure objectivity is impossible. Once you leave the myth of objectivity behind, though, the way forward is not simple. Instead of a belief in an all-knowing science, we are faced with the reality that humans are responsible for what is researched, how it is researched and what conclusions are drawn from such research.

With this knowledge, we have the opportunity to intentionally set societal values that inform scientific investigations. This requires decisions about how people come to agreements about these values. These agreements need not always be universal but instead can be dependent on the context of who and what a given study might affect. While not simple, using these insights, gained over decades of studying science from both within and outside, may force a more honest conversation between political positions.

Read More

The post Scientific objectivity is a myth – cultural values and beliefs always influence science and the people who do it appeared first on theconversation.com



Note: The following A.I. based commentary is not part of the original article, reproduced above, but is offered in the hopes that it will promote greater media literacy and critical thinking, by making any potential bias more visible to the reader –Staff Editor.

Political Bias Rating: Center-Left

The content emphasizes the influence of cultural and social values on scientific research, challenging the notion of pure scientific objectivity. It highlights themes such as gender equality, critiques of traditional hierarchies, and the social construction of categories like sex and race, which are commonly associated with progressive or center-left perspectives. While it acknowledges political divides and calls for democratic, inclusive approaches to science, the overall framing aligns with a center-left viewpoint that values social context and equity in knowledge production.

Continue Reading

The Conversation

AI is transforming weather forecasting − and that could be a game changer for farmers around the world

Published

on

theconversation.com – Paul Winters, Professor of Sustainable Development, University of Notre Dame – 2025-09-03 07:30:00


Climate change intensifies weather risks for farmers, affecting crop yields and incomes, especially in low- and middle-income countries lacking accurate forecasts due to costly traditional models. AI-powered weather forecasting offers a breakthrough by delivering accurate, localized predictions rapidly and inexpensively, using far less computational power than physics-based systems. Advanced AI models like Pangu-Weather and GraphCast now match or surpass traditional forecasts, enabling timely, high-resolution weather guidance on standard computers. To be effective, AI forecasts must be tailored to local agricultural needs and disseminated through accessible channels. Supported by organizations such as AIM for Scale, AI forecasting can empower developing countries to adapt farming practices and improve resilience amid climate change.

Weather forecasts help farmers figure out when to plant, where to use fertilizer and much more.
Maitreya Shah/Studio India

Paul Winters, University of Notre Dame and Amir Jina, University of Chicago

For farmers, every planting decision carries risks, and many of those risks are increasing with climate change. One of the most consequential is weather, which can damage crop yields and livelihoods. A delayed monsoon, for example, can force a rice farmer in South Asia to replant or switch crops altogether, losing both time and income.

Access to reliable, timely weather forecasts can help farmers prepare for the weeks ahead, find the best time to plant or determine how much fertilizer will be needed, resulting in better crop yields and lower costs.

Yet, in many low- and middle-income countries, accurate weather forecasts remain out of reach, limited by the high technology costs and infrastructure demands of traditional forecasting models.

A new wave of AI-powered weather forecasting models has the potential to change that.

A farmer in a field holds a dried out corn stalk.
A farmer holds dried-up maize stalks in his field in Zimbabwe on March 22, 2024. A drought had caused widespread water shortages and crop failures.
AP Photo/Tsvangirayi Mukwazhi

By using artificial intelligence, these models can deliver accurate, localized predictions at a fraction of the computational cost of conventional physics-based models. This makes it possible for national meteorological agencies in developing countries to provide farmers with the timely, localized information about changing rainfall patterns that the farmers need.

The challenge is getting this technology where it’s needed.

Why AI forecasting matters now

The physics-based weather prediction models used by major meteorological centers around the world are powerful but costly. They simulate atmospheric physics to forecast weather conditions ahead, but they require expensive computing infrastructure. The cost puts them out of reach for most developing countries.

Moreover, these models have mainly been developed by and optimized for northern countries. They tend to focus on temperate, high-income regions and pay less attention to the tropics, where many low- and middle-income countries are located.

A major shift in weather models began in 2022 as industry and university researchers developed deep learning models that could generate accurate short- and medium-range forecasts for locations around the globe up to two weeks ahead.

These models worked at speeds several orders of magnitude faster than physics-based models, and they could run on laptops instead of supercomputers. Newer models, such as Pangu-Weather and GraphCast, have matched or even outperformed leading physics-based systems for some predictions, such as temperature.

A woman in a red sari tosses pellets into a rice field.
A farmer distributes fertilizer in India.
EqualStock IN from Pexels

AI-driven models require dramatically less computing power than the traditional systems.

While physics-based systems may need thousands of CPU hours to run a single forecast cycle, modern AI models can do so using a single GPU in minutes once the model has been trained. This is because the intensive part of the AI model training, which learns relationships in the climate from data, can use those learned relationships to produce a forecast without further extensive computation – that’s a major shortcut. In contrast, the physics-based models need to calculate the physics for each variable in each place and time for every forecast produced.

While training these models from physics-based model data does require significant upfront investment, once the AI is trained, the model can generate large ensemble forecasts — sets of multiple forecast runs — at a fraction of the computational cost of physics-based models.

Even the expensive step of training an AI weather model shows considerable computational savings. One study found the early model FourCastNet could be trained in about an hour on a supercomputer. That made its time to presenting a forecast thousands of times faster than state-of-the-art, physics-based models.

The result of all these advances: high-resolution forecasts globally within seconds on a single laptop or desktop computer.

Research is also rapidly advancing to expand the use of AI for forecasts weeks to months ahead, which helps farmers in making planting choices. AI models are already being tested for improving extreme weather prediction, such as for extratropical cyclones and abnormal rainfall.

Tailoring forecasts for real-world decisions

While AI weather models offer impressive technical capabilities, they are not plug-and-play solutions. Their impact depends on how well they are calibrated to local weather, benchmarked against real-world agricultural conditions, and aligned with the actual decisions farmers need to make, such as what and when to plant, or when drought is likely.

To unlock its full potential, AI forecasting must be connected to the people whose decisions it’s meant to guide.

That’s why groups such as AIM for Scale, a collaboration we work with as researchers in public policy and sustainability, are helping governments to develop AI tools that meet real-world needs, including training users and tailoring forecasts to farmers’ needs. International development institutions and the World Meteorological Organization are also working to expand access to AI forecasting models in low- and middle-income countries.

A man sells grain in Dawanau International Market in Kano, Nigeria on July 14, 2023.
Many low-income countries in Africa face harsh effects from climate change, from severe droughts to unpredictable rain and flooding. The shocks worsen conflict and upend livelihoods.
AP Photo/Sunday Alamba

AI forecasts can be tailored to context-specific agricultural needs, such as identifying optimal planting windows, predicting dry spells or planning pest management. Disseminating those forecasts through text messages, radio, extension agents or mobile apps can then help reach farmers who can benefit. This is especially true when the messages themselves are constantly tested and improved to ensure they meet the farmers’ needs.

A recent study in India found that when farmers there received more accurate monsoon forecasts, they made more informed decisions about what and how much to plant – or whether to plant at all – resulting in better investment outcomes and reduced risk.

A new era in climate adaptation

AI weather forecasting has reached a pivotal moment. Tools that were experimental just five years ago are now being integrated into government weather forecasting systems. But technology alone won’t change lives.

With support, low- and middle-income countries can build the capacity to generate, evaluate and act on their own forecasts, providing valuable information to farmers that has long been missing in weather services.The Conversation

Paul Winters, Professor of Sustainable Development, University of Notre Dame and Amir Jina, Assistant Professor of Public Policy, University of Chicago

This article is republished from The Conversation under a Creative Commons license. Read the original article.

A farmer holds dried-up maize stalks in his field in Zimbabwe on March 22, 2024. A drought had caused widespread water shortages and crop failures.
AP Photo/Tsvangirayi Mukwazhi

By using artificial intelligence, these models can deliver accurate, localized predictions at a fraction of the computational cost of conventional physics-based models. This makes it possible for national meteorological agencies in developing countries to provide farmers with the timely, localized information about changing rainfall patterns that the farmers need.

The challenge is getting this technology where it’s needed.

Why AI forecasting matters now

The physics-based weather prediction models used by major meteorological centers around the world are powerful but costly. They simulate atmospheric physics to forecast weather conditions ahead, but they require expensive computing infrastructure. The cost puts them out of reach for most developing countries.

Moreover, these models have mainly been developed by and optimized for northern countries. They tend to focus on temperate, high-income regions and pay less attention to the tropics, where many low- and middle-income countries are located.

A major shift in weather models began in 2022 as industry and university researchers developed deep learning models that could generate accurate short- and medium-range forecasts for locations around the globe up to two weeks ahead.

These models worked at speeds several orders of magnitude faster than physics-based models, and they could run on laptops instead of supercomputers. Newer models, such as Pangu-Weather and GraphCast, have matched or even outperformed leading physics-based systems for some predictions, such as temperature.

A woman in a red sari tosses pellets into a rice field.

A farmer distributes fertilizer in India.
EqualStock IN from Pexels

AI-driven models require dramatically less computing power than the traditional systems.

While physics-based systems may need thousands of CPU hours to run a single forecast cycle, modern AI models can do so using a single GPU in minutes once the model has been trained. This is because the intensive part of the AI model training, which learns relationships in the climate from data, can use those learned relationships to produce a forecast without further extensive computation – that’s a major shortcut. In contrast, the physics-based models need to calculate the physics for each variable in each place and time for every forecast produced.

While training these models from physics-based model data does require significant upfront investment, once the AI is trained, the model can generate large ensemble forecasts — sets of multiple forecast runs — at a fraction of the computational cost of physics-based models.

Even the expensive step of training an AI weather model shows considerable computational savings. One study found the early model FourCastNet could be trained in about an hour on a supercomputer. That made its time to presenting a forecast thousands of times faster than state-of-the-art, physics-based models.

The result of all these advances: high-resolution forecasts globally within seconds on a single laptop or desktop computer.

Research is also rapidly advancing to expand the use of AI for forecasts weeks to months ahead, which helps farmers in making planting choices. AI models are already being tested for improving extreme weather prediction, such as for extratropical cyclones and abnormal rainfall.

Tailoring forecasts for real-world decisions

While AI weather models offer impressive technical capabilities, they are not plug-and-play solutions. Their impact depends on how well they are calibrated to local weather, benchmarked against real-world agricultural conditions, and aligned with the actual decisions farmers need to make, such as what and when to plant, or when drought is likely.

To unlock its full potential, AI forecasting must be connected to the people whose decisions it’s meant to guide.

That’s why groups such as AIM for Scale, a collaboration we work with as researchers in public policy and sustainability, are helping governments to develop AI tools that meet real-world needs, including training users and tailoring forecasts to farmers’ needs. International development institutions and the World Meteorological Organization are also working to expand access to AI forecasting models in low- and middle-income countries.

A man sells grain in Dawanau International Market in Kano, Nigeria on July 14, 2023.

Many low-income countries in Africa face harsh effects from climate change, from severe droughts to unpredictable rain and flooding. The shocks worsen conflict and upend livelihoods.
AP Photo/Sunday Alamba

AI forecasts can be tailored to context-specific agricultural needs, such as identifying optimal planting windows, predicting dry spells or planning pest management. Disseminating those forecasts through text messages, radio, extension agents or mobile apps can then help reach farmers who can benefit. This is especially true when the messages themselves are constantly tested and improved to ensure they meet the farmers’ needs.

A recent study in India found that when farmers there received more accurate monsoon forecasts, they made more informed decisions about what and how much to plant – or whether to plant at all – resulting in better investment outcomes and reduced risk.

A new era in climate adaptation

AI weather forecasting has reached a pivotal moment. Tools that were experimental just five years ago are now being integrated into government weather forecasting systems. But technology alone won’t change lives.

With support, low- and middle-income countries can build the capacity to generate, evaluate and act on their own forecasts, providing valuable information to farmers that has long been missing in weather services.

Read More

The post AI is transforming weather forecasting − and that could be a game changer for farmers around the world appeared first on theconversation.com



Note: The following A.I. based commentary is not part of the original article, reproduced above, but is offered in the hopes that it will promote greater media literacy and critical thinking, by making any potential bias more visible to the reader –Staff Editor.

Political Bias Rating: Centrist

The content presents a factual and balanced discussion on the use of AI in weather forecasting to aid farmers, particularly in low- and middle-income countries. It emphasizes technological innovation, international collaboration, and practical benefits without promoting a specific political ideology. The focus on climate change and development is handled in a neutral, solution-oriented manner, reflecting a centrist perspective that values science and global cooperation.

Continue Reading

Trending