fbpx
Connect with us

The Conversation

A new, thin-lensed telescope design could far surpass James Webb – goodbye mirrors, hello diffractive lenses

Published

on

A new, thin-lensed telescope design could far surpass James Webb – goodbye mirrors, hello diffractive lenses

A light, cheap telescope design would make it possible to put many individual units in space at once.
Katie Yung, Daniel Apai /University of Arizona and AllThingsSpace /SketchFab, CC BY-ND

Daniel Apai, University of Arizona

Astronomers have discovered more than 5,000 planets outside of the solar system to date. The grand question is whether any of these planets are home to life. To find the answer, astronomers will likely need more powerful telescopes than exist today.

I am an astronomer who studies astrobiology and planets around distant . For the last seven years, I have been co-leading a team that is developing a new kind of space telescope that could collect a hundred times more light than the James Webb Space Telescope, the biggest space telescope ever built.

Almost all space telescopes, Hubble and Webb, collect light using mirrors. Our proposed telescope, the Nautilus Space Observatory, would replace large, heavy mirrors with a novel, thin lens that is much lighter, cheaper and easier to produce than mirrored telescopes. Because of these differences, it would be possible to launch many individual units into orbit and create a powerful network of telescopes.

A blue planet with clouds.
Exoplanets, like TOI-700d shown in this artist's conception, are planets beyond our solar system and are prime candidates in the search for .
NASA's Goddard Space Flight Center

The need for larger telescopes

Exoplanets – planets that orbit stars other than the Sun – are prime targets in the search for life. Astronomers need to use giant space telescopes that collect huge amounts of light to study these faint and faraway objects.

A massive circular gold mirror with people standing in the foreground.
The James Webb Space Telescope is just barely able to search exoplanets for signs of life.
NASA

Existing telescopes can detect exoplanets as small as Earth. However, it takes a lot more sensitivity to begin to learn about the chemical composition of these planets. Even Webb is just barely powerful enough to search certain exoplanets for clues of life – namely gases in the atmosphere.

The James Webb Space Telescope cost more than US$8 billion and took over 20 years to build. The next flagship telescope is not expected to fly before 2045 and is estimated to cost $11 billion. These ambitious telescope projects are always expensive, laborious and produce a single powerful – but very specialized – observatory.

Advertisement

A new kind of telescope

In 2016, aerospace giant Northrop Grumman invited me and 14 other professors and NASA scientists – all experts on exoplanets and the search for extraterrestrial life – to Los Angeles to answer one question: What will exoplanet space telescopes look like in 50 years?

In our discussions, we realized that a major bottleneck preventing the construction of more powerful telescopes is the challenge of making larger mirrors and getting them into orbit. To bypass this bottleneck, a few of us came up with the idea of revisiting an old technology called diffractive lenses.

A cross section of two lenses, with the one on the left showing a jagged surface and the one on the right a rounded surface.
Diffractive lenses, left, are much thinner compared to similarly powerful refractive lenses, right.
Pko/Wikimedia Commons

Conventional lenses use refraction to focus light. Refraction is when light changes direction as it passes from one medium to another – it is the reason light bends when it enters . In contrast, diffraction is when light bends around corners and obstacles. A cleverly arranged pattern of steps and angles on a glass surface can form a diffractive lens.

The first such lenses were invented by the French scientist Augustin-Jean Fresnel in 1819 to lightweight lenses for lighthouses. Today, similar diffractive lenses can be found in many small-sized consumer optics – from camera lenses to virtual reality headsets.

Thin, simple diffractive lenses are notorious for their blurry images, so they have never been used in astronomical observatories. But if you could improve their clarity, using diffractive lenses instead of mirrors or refractive lenses would allow a space telescope to be much cheaper, lighter and larger.

Advertisement
A person holding a round, thin piece of glass.
One of the benefits of diffractive lenses is that they can remain thin while increasing in diameter.
Daniel Apai/University of Arizona, CC BY-ND

A thin, high-resolution lens

After the meeting, I returned to the University of Arizona and decided to explore whether modern technology could produce diffractive lenses with better image quality. Lucky for me, Thomas Milster – one of the world's leading experts on diffractive lens design – works in the building next to mine. We formed a team and got to work.

Over the two years, our team invented a new type of diffractive lens that required new manufacturing technologies to etch a complex pattern of tiny grooves onto a piece of clear glass or plastic. The specific pattern and shape of the cuts focuses incoming light to a single point behind the lens. The new design produces a near-perfect quality image, far better than previous diffractive lenses.

A triangular piece of glass with subtle etchings reflecting in the light.
A diffractive lens bends light using etchings and patterns on its surface.
Daniel Apai/University of Arizona, CC BY-ND

Because it is the surface texture of the lens that does the focusing, not the thickness, you can easily make the lens bigger while keeping it very thin and lightweight. Bigger lenses collect more light, and low weight means cheaper launches to orbit – both great traits for a space telescope.

In August 2018, our team produced the first prototype, a 2-inch (5-centimeter) diameter lens. Over the next five years, we further improved the image quality and increased the size. We are now completing a 10-inch (24-cm) diameter lens that will be more than 10 times lighter than a conventional refractive lens would be.

Power of a diffraction space telescope

This new lens design makes it possible to rethink how a space telescope might be built. In 2019, our team published a concept called the Nautilus Space Observatory.

Using the new technology, our team thinks it is possible to build a 29.5-foot (8.5-meter) diameter lens that would be only about 0.2 inches (0.5 cm) thick. The lens and support structure of our new telescope could weigh around 1,100 pounds (500 kilograms). This is more than three times lighter than a Webb– mirror of a similar size and would be bigger than Webb's 21-foot (6.5-meter) diameter mirror.

Advertisement
A spherical object in space with a lens on one side.
The thin lens the team to design a lighter, cheaper telescope, which they named the Nautilus Space Observatory.
Daniel Apai/University of Arizona, CC BY-ND

The lenses have other benefits, too. First, they are much easier and quicker to fabricate than mirrors and can be made en masse. Second, lens-based telescopes work well even when not aligned perfectly, making these telescopes easier to assemble and fly in space than mirror-based telescopes, which require extremely precise alignment.

Finally, since a single Nautilus unit would be light and relatively cheap to produce, it would be possible to put dozens of them into orbit. Our current design is in fact not a single telescope, but a constellation of 35 individual telescope units.

Each individual telescope would be an independent, highly sensitive observatory able to collect more light than Webb. But the real power of Nautilus would from turning all the individual telescopes toward a single target.

By combining data from all the units, Nautilus' light-collecting power would equal a telescope nearly 10 times larger than Webb. With this powerful telescope, astronomers could search hundreds of exoplanets for atmospheric gases that may indicate extraterrestrial life.

Although the Nautilus Space Observatory is still a long way from launch, our team has made a lot of progress. We have shown that all aspects of the technology work in small-scale prototypes and are now focusing on building a 3.3-foot (1-meter) diameter lens. Our next steps are to send a small version of the telescope to the edge of space on a high-altitude balloon.

Advertisement

With that, we will be ready to propose a revolutionary new space telescope to NASA and, hopefully, be on the way to exploring hundreds of worlds for signatures of life.The Conversation

Daniel Apai, Associate Dean for Research and Professor of Astronomy and Planetary Sciences, University of Arizona

This article is republished from The Conversation under a Creative Commons license. Read the original article.

Did you miss our previous article…
https://www.biloxinewsevents.com/?p=266990

Advertisement

The Conversation

Black holes are mysterious, yet also deceptively simple − a new space mission may help physicists answer hairy questions about these astronomical objects

Published

on

theconversation.com – Gaurav Khanna, Professor of Physics, of Rhode Island – 2024-05-15 07:16:18

An illustration of a supermassive black hole.

NASA/JPL

Gaurav Khanna, University of Rhode Island

Physicists consider black holes one of the most mysterious objects that exist. Ironically, they're also considered one of the simplest. For years, physicists like me have been looking to prove that black holes are more complex than they seem. And a newly approved European space mission called LISA will us with this hunt.

Advertisement

Research from the 1970s suggests that you can comprehensively describe a black hole using only three physical attributes – their mass, charge and spin. All the other properties of these massive dying , like their detailed composition, density and temperature profiles, disappear as they transform into a black hole. That is how simple they are.

The idea that black holes have only three attributes is called the “no-hair” theorem, implying that they don't have any “hairy” details that make them complicated.

Black holes are massive, mysterious astronomical objects.

Hairy black holes?

For decades, researchers in the astrophysics community have exploited loopholes or work-arounds within the no-hair theorem's assumptions to up with potential hairy black hole scenarios. A hairy black hole has a physical property that scientists can measure – in principle – that's beyond its mass, charge or spin. This property has to be a permanent part of its structure.

About a decade ago, Stefanos Aretakis, a physicist currently at the University of Toronto, showed mathematically that a black hole containing the maximum charge it could hold – called an extremal charged black hole – would develop “hair” at its horizon. A black hole's horizon is the boundary where anything that crosses it, even light, can't escape.

Advertisement

Aretakis' analysis was more of a thought experiment using a highly simplified physical scenario, so it's not something scientists expect to observe astrophysically. But supercharged black holes might not be the only kind that could have hair.

Since astrophysical objects such as stars and planets are known to spin, scientists expect that black holes would spin as well, based on how they form. Astronomical evidence has shown that black holes do have spin, though researchers don't know what the typical spin value is for an astrophysical black hole.

Using computer simulations, my team has recently discovered similar types of hair in black holes that are spinning at the maximum rate. This hair has to do with the rate of change, or the gradient, of -time's curvature at the horizon. We also discovered that a black hole wouldn't actually have to be maximally spinning to have hair, which is significant because these maximally spinning black holes probably don't form in nature.

Detecting and measuring hair

My team wanted to develop a way to potentially measure this hair – a new fixed property that might characterize a black hole beyond its mass, spin and charge. We started looking into how such a new property might a signature on a gravitational wave emitted from a fast-spinning black hole.

Advertisement

A gravitational wave is a tiny disturbance in space-time typically caused by violent astrophysical in the universe. The collisions of compact astrophysical objects such as black holes and neutron stars emit strong gravitational waves. An international network of gravitational observatories, the Laser Interferometer Gravitational-wave Observatory in the United States, routinely detects these waves.

Our recent studies suggest that one can measure these hairy attributes from gravitational wave data for fast-spinning black holes. Looking at the gravitational wave data offers an for a signature of sorts that could indicate whether the black hole has this type of hair.

Our ongoing studies and recent progress made by Som Bishoyi, a student on the team, are based on a blend of theoretical and computational models of fast-spinning black holes. Our findings have not been tested in the field yet or observed in real black holes out in space. But we hope that will soon change.

LISA gets a go-ahead

In January 2024, the European Space Agency formally adopted the space-based Laser Interferometer Space Antenna, or LISA, mission. LISA will look for gravitational waves, and the data from the mission could help my team with our hairy black hole questions.

Advertisement

Three spacecrafts spaced apart sending light beams towards each other while orbiting the Sun

The LISA spacecrafts observing gravitational waves from a distant source while orbiting the Sun.

Simon Barke/Univ. Florida, CC BY

Formal adoption means that the has the go-ahead to move to the construction phase, with a planned 2035 launch. LISA consists of three spacecrafts configured in a perfect equilateral triangle that will trail behind the Earth around the Sun. The spacecrafts will each be 1.6 million miles (2.5 million kilometers) apart, and they will exchange laser beams to measure the distance between each other down to about a billionth of an inch.

LISA will detect gravitational waves from supermassive black holes that are millions or even billions of times more massive than our Sun. It will build a map of the space-time around rotating black holes, which will help physicists understand how gravity works in the close vicinity of black holes to an unprecedented level of accuracy. Physicists hope that LISA will also be able to measure any hairy attributes that black holes might have.

With LIGO making new observations every day and LISA to offer a glimpse into the space-time around black holes, now is one of the most exciting times to be a black hole physicist.The Conversation

Gaurav Khanna, Professor of Physics, University of Rhode Island

Advertisement

This article is republished from The Conversation under a Creative Commons license. Read the original article.

Continue Reading

The Conversation

Viruses are doing mysterious things everywhere – AI can help researchers understand what they’re up to in the oceans and in your gut

Published

on

theconversation.com – Libusha Kelly, Associate Professor of and Computational Biology, Microbiology and Immunology, Albert Einstein College of Medicine – 2024-05-15 07:16:41

Many viral genetic sequences code for proteins that researchers haven't seen before.

KTSDesign/Science Photo Library via Getty Images

Libusha Kelly, Albert Einstein College of Medicine

Viruses are a mysterious and poorly understood force in microbial ecosystems. Researchers know they can infect, kill and manipulate human and bacterial cells in nearly every environment, from the oceans to your gut. But scientists don't yet have a full picture of how viruses affect their surrounding environments in large part because of their extraordinary diversity and ability to rapidly evolve.

Advertisement

Communities of microbes are difficult to study in a laboratory setting. Many microbes are challenging to cultivate, and their natural has many more features influencing their or failure than scientists can replicate in a lab.

So systems biologists like me often sequence all the DNA present in a sample – for example, a fecal sample from a patient – separate out the viral DNA sequences, then annotate the sections of the viral genome that code for proteins. These notes on the location, structure and other features of genes researchers understand the functions viruses might carry out in the environment and help identify different kinds of viruses. Researchers annotate viruses by matching viral sequences in a sample to previously annotated sequences available in public databases of viral genetic sequences.

However, scientists are identifying viral sequences in DNA collected from the environment at a rate that far outpaces our ability to annotate those genes. This means researchers are publishing findings about viruses in microbial ecosystems using unacceptably small fractions of available data.

To improve researchers' ability to study viruses around the globe, my team and I have developed a novel approach to annotate viral sequences using artificial intelligence. Through protein language models akin to large language models like ChatGPT but specific to proteins, we were able to classify previously unseen viral sequences. This the door for researchers to not only learn more about viruses, but also to address biological questions that are difficult to answer with current techniques.

Advertisement

Annotating viruses with AI

Large language models use relationships between words in large datasets of text to potential answers to questions they are not explicitly “taught” the answer to. When you ask a chatbot “What is the capital of France?” for example, the model is not looking up the answer in a table of capital . Rather, it is using its training on huge datasets of documents and information to infer the answer: “The capital of France is Paris.”

Similarly, protein language models are AI algorithms that are trained to recognize relationships between billions of protein sequences from environments around the world. Through this training, they may be able to infer something about the essence of viral proteins and their functions.

We wondered whether protein language models could answer this question: “Given all annotated viral genetic sequences, what is this new sequence's function?”

In our proof of concept, we trained neural networks on previously annotated viral protein sequences in pre-trained protein language models and then used them to predict the annotation of new viral protein sequences. Our approach allows us to probe what the model is “seeing” in a particular viral sequence that to a particular annotation. This helps identify candidate proteins of interest either based on their specific functions or how their genome is arranged, winnowing down the search of vast datasets.

Advertisement

Microscopy image of spherical bacteria colored bright green

Prochlorococcus is one of the many species of marine bacteria with proteins that researchers haven't seen before.

Anne Thompson/Chisholm Lab, MIT via Flickr

By identifying more distantly related viral gene functions, protein language models can complement current methods to provide new insights into microbiology. For example, my team and I were able to use our model to discover a previously unrecognized integrase – a type of protein that can move genetic information in and out of cells – in the globally abundant marine picocyanobacteria Prochlorococcus and Synechococcus. Notably, this integrase may be able to move genes in and out of these populations of bacteria in the oceans and enable these microbes to better adapt to changing environments.

Our language model also identified a novel viral capsid protein that is widespread in the global oceans. We produced the first picture of how its genes are arranged, showing it can contain different sets of genes that we believe indicates this virus serves different functions in its environment.

These preliminary findings represent only two of thousands of annotations our approach has provided.

Advertisement

Analyzing the unknown

Most of the hundreds of thousands of newly discovered viruses remain unclassified. Many viral genetic sequences match protein families with no known function or have never been seen before. Our work shows that similar protein language models could help study the threat and promise of our planet's many uncharacterized viruses.

While our study focused on viruses in the global oceans, improved annotation of viral proteins is critical for better understanding the role viruses play in and disease in the human body. We and other researchers have hypothesized that viral activity in the human gut microbiome might be altered when you're sick. This means that viruses may help identify stress in microbial communities.

However, our approach is also limited because it requires high-quality annotations. Researchers are developing newer protein language models that incorporate other “tasks” as part of their training, particularly predicting protein structures to detect similar proteins, to make them more powerful.

Making all AI tools available via FAIR Data Principles – data that is findable, accessible, interoperable and reusable – can help researchers at large realize the potential of these new ways of annotating protein sequences leading to discoveries that benefit human health.The Conversation

Libusha Kelly, Associate Professor of Systems and Computational Biology, Microbiology and Immunology, Albert Einstein College of Medicine

Advertisement

This article is republished from The Conversation under a Creative Commons license. Read the original article.

Continue Reading

The Conversation

Human differences in judgment lead to problems for AI

Published

on

theconversation.com – Mayank Kejriwal, Research Assistant Professor of Industrial & Engineering, of Southern California – 2024-05-14 07:14:06

Bias isn't the only human imperfection turning up in AI.

Emrah Turudu/Photodisc via Getty Images

Mayank Kejriwal, University of Southern California

Many people understand the concept of bias at some intuitive level. In society, and in artificial intelligence systems, racial and gender biases are well documented.

Advertisement

If society could somehow bias, would all problems go away? The late Nobel laureate Daniel Kahneman, who was a key figure in the field of behavioral economics, argued in his last book that bias is just one side of the coin. Errors in judgments can be attributed to two sources: bias and noise.

Bias and noise both play important roles in fields such as law, medicine and financial forecasting, where human judgments are central. In our work as computer and information scientists, my colleagues and I have found that noise also plays a role in AI.

Statistical noise

Noise in this context means variation in how people make judgments of the same problem or situation. The problem of noise is more pervasive than initially meets the eye. A seminal work, dating back all the way to the Great Depression, has found that different judges gave different sentences for similar cases.

Worryingly, sentencing in court cases can depend on things such as the temperature and whether the local football team won. Such factors, at least in part, contribute to the perception that the justice system is not just biased but also arbitrary at times.

Advertisement

Other examples: Insurance adjusters might give different estimates for similar claims, reflecting noise in their judgments. Noise is likely present in all manner of contests, ranging from wine tastings to local beauty pageants to college admissions.

Behavioral economist Daniel Kahneman explains the concept of noise in human judgment.

Noise in the data

On the surface, it doesn't seem likely that noise could affect the performance of AI systems. After all, machines aren't affected by weather or football teams, so why would they make judgments that vary with circumstance? On the other hand, researchers know that bias affects AI, because it is reflected in the data that the AI is trained on.

For the new spate of AI models like ChatGPT, the gold standard is human performance on general intelligence problems such as common sense. ChatGPT and its peers are measured against human-labeled commonsense datasets.

Put simply, researchers and developers can ask the machine a commonsense question and compare it with human answers: “If I place a heavy rock on a paper table, will it collapse? Yes or No.” If there is high agreement between the two – in the best case, perfect agreement – the machine is approaching human-level common sense, according to the test.

Advertisement

So where would noise in? The commonsense question above seems simple, and most humans would likely agree on its answer, but there are many questions where there is more disagreement or uncertainty: “Is the sentence plausible or implausible? My dog plays volleyball.” In other words, there is potential for noise. It is not surprising that interesting commonsense questions would have some noise.

But the issue is that most AI tests don't account for this noise in experiments. Intuitively, questions generating human answers that tend to agree with one another should be weighted higher than if the answers diverge – in other words, where there is noise. Researchers still don't know whether or how to weigh AI's answers in that situation, but a first step is acknowledging that the problem exists.

Tracking down noise in the machine

Theory aside, the question still remains whether all of the above is hypothetical or if in real tests of common sense there is noise. The best way to prove or disprove the presence of noise is to take an existing test, remove the answers and get multiple people to independently label them, meaning answers. By measuring disagreement among humans, researchers can know just how much noise is in the test.

The details behind measuring this disagreement are complex, involving significant statistics and math. Besides, who is to say how common sense should be defined? How do you know the human judges are motivated enough to think through the question? These issues lie at the intersection of good experimental design and statistics. Robustness is key: One result, test or set of human labelers is unlikely to convince anyone. As a pragmatic matter, human labor is expensive. Perhaps for this reason, there haven't been any studies of possible noise in AI tests.

Advertisement

To address this gap, my colleagues and I designed such a study and published our findings in Nature Scientific Reports, showing that even in the domain of common sense, noise is inevitable. Because the setting in which judgments are elicited can matter, we did two kinds of studies. One type of study involved paid workers from Amazon Mechanical Turk, while the other study involved a smaller-scale labeling exercise in two labs at the University of Southern California and the Rensselaer Polytechnic Institute.

You can think of the former as a more realistic online setting, mirroring how many AI tests are actually labeled before being released for and evaluation. The latter is more of an extreme, guaranteeing high quality but at much smaller scales. The question we set out to answer was how inevitable is noise, and is it just a matter of quality control?

The results were sobering. In both settings, even on commonsense questions that might have been expected to elicit high – even universal – agreement, we found a nontrivial degree of noise. The noise was high enough that we inferred that between 4% and 10% of a system's performance could be attributed to noise.

To emphasize what this means, suppose I built an AI system that achieved 85% on a test, and you built an AI system that achieved 91%. Your system would seem to be a lot better than mine. But if there is noise in the human labels that were used to score the answers, then we're not sure anymore that the 6% improvement means much. For all we know, there may be no real improvement.

Advertisement

On AI leaderboards, where large language models like the one that powers ChatGPT are , performance differences between rival systems are far narrower, typically less than 1%. As we show in the paper, ordinary statistics do not really come to the rescue for disentangling the effects of noise from those of true performance improvements.

Noise audits

What is the way forward? Returning to Kahneman's book, he proposed the concept of a “noise audit” for quantifying and ultimately mitigating noise as much as possible. At the very least, AI researchers need to estimate what influence noise might be .

Auditing AI systems for bias is somewhat commonplace, so we believe that the concept of a noise audit should naturally follow. We hope that this study, as well as others like it, to their adoption.The Conversation

Mayank Kejriwal, Research Assistant Professor of Industrial & Systems Engineering, University of Southern California

This article is republished from The Conversation under a Creative Commons license. Read the original article.

Advertisement
Continue Reading

News from the South

Trending