Archive for the ‘Artificial Intelligence’ Category

Artificial Intelligence: inevitable integration enterprises | Top Stories | theweeklyjournal.com – The Weekly Journal

Given the accelerated pace at which Artificial Intelligence (AI) occupies various lines to enhance the way in which private and public agencies carry out their work, more and more people must be trained to understand the impact of the new technology in their lives.

According to CRANT's chief executive, lvaro Melndez, AI represents a transformation for marketing, in which brands will have to appeal to credibility at a time when consumer vulnerability is threatened by the constant generation of content that is mostly not real.

The ManpowerGroup Employment Expectations Survey (MEOS) revealed that net hiring intentions

"Artificial intelligence, beyond the superficial form, in which there has been a lot of talk about it helping you to generate images or video or text, obviously helps a lot because it makes the work easier and gives new opportunities, but there is a much deeper transformation that is what interests us and is that transformation that now all this is possible and much of what will be generated can be misleading ... it may be a lie," said Melndez.

Because the situation involves a new way of consuming information, the executive considered that it is an opportunity for brands to use the tool responsibly to generate a positive impact through marketing.

"It's a different way of thinking about marketing. It's no longer about communicating a product or a service, but now it's about how you are showing reality," the executive commented.

To address the problem, Melndez said that companies must educate themselves in the use of AI in an ethical manner, and become a source of confidence for the consumer.

At a time when marketing is in the early stages with AI, he considered that by the end of this year all companies will have it incorporated, which will generate competitiveness in relation to those that do not.

That is why Melndez designed and carried out the "AI for Marketers" workshop, in collaboration with the agency de la Cruz, to provide a group of marketers with an explanatory framework of the basic principles, ethics, tools, advantages and opportunities that AI provides so that they are not left behind by the incursion of the technology.

"The goal is to facilitate a much deeper understanding of what artificial intelligence is and how it can be applied, both to enhance their work with their companies and their brands, but also to enhance their career. Artificial intelligence (AI) is not for tech people, it's not for data scientists. We all have to understand and master artificial intelligence," said the founder of the company dedicated to the creative application of artificial intelligence.

Results of AI in companies

Among the companies that incorporate Artificial Intelligence as efficiency strategies to generate higher value content, Melndez exemplified Tomorrow AI that generates around 60 thousand marketing materials monthly with a team of only four people.

"Another example is Duolingo, this company that teaches languages. They had to lay off - which is the downside - about a thousand people, because a lot of the content that Duolingo does, and the way they educate people, they can now do it through artificial intelligence," said the CRANT executive.

When asked by The News Journal about the repercussions of AI in terms of employment, Melndez pointed out that the part where more people will be out of work is inevitable because companies will understand that they can carry out tasks through technology.

Although many jobs will disappear, he assured that a creative explosion will emerge that will give way to entrepreneurship.

The finance industry has not been immune to the technological advances of recent decades; in

"We will start to see companies doing things that we would never have imagined possible, and that is interesting because it will break the market and large established companies will disappear because others have solved it in a better way," said Melndez.

"If you are a person who has an idea and wants to execute it, but can't because you don't have the resources or because you don't know how to program, let's say you want to make an application, with artificial intelligence you will be able to make that application without knowing how to program and launch your company with almost no employees and without hiring anyone," he added.

At present, estimates by investment banking group Goldman Sachs on the rise of platforms that use AI suggest that 300 million jobs around the world could be automated, and, in the case of the United States, the workload could be replaced by 25% to 50%.

See original here:
Artificial Intelligence: inevitable integration enterprises | Top Stories | theweeklyjournal.com - The Weekly Journal

Tags:

Seven technologies to watch in 2024 – Nature.com

From protein engineering and 3D printing to detection of deepfake media, here are seven areas of technology that Nature will be watching in the year ahead.

Two decades ago, David Baker at the University of Washington in Seattle and his colleagues achieved a landmark feat: they used computational tools to design an entirely new protein from scratch. Top7 folded as predicted, but it was inert: it performed no meaningful biological functions. Today, de novo protein design has matured into a practical tool for generating made-to-order enzymes and other proteins. Its hugely empowering, says Neil King, a biochemist at the University of Washington who collaborates with Bakers team to design protein-based vaccines and vehicles for drug delivery. Things that were impossible a year and a half ago now you just do it.

Much of that progress comes down to increasingly massive data sets that link protein sequence to structure. But sophisticated methods of deep learning, a form of artificial intelligence (AI), have also been essential.

Sequence based strategies use the large language models (LLMs) that power tools such as the chatbot ChatGPT (see ChatGPT? Maybe next year). By treating protein sequences like documents comprising polypeptide words, these algorithms can discern the patterns that underlie the architectural playbook of real-world proteins. They really learn the hidden grammar, says Noelia Ferruz, a protein biochemist at the Molecular Biology Institute of Barcelona, Spain. In 2022, her team developed an algorithm called ProtGPT2 that consistently comes up with synthetic proteins that fold stably when produced in the laboratory1. Another tool co-developed by Ferruz, called ZymCTRL, draws on sequence and functional data to design members of naturally occurring enzyme families2.

Readers might detect a theme in this years technologies to watch: the outsized impact of deep-learning methods. But one such tool did not make the final cut: the much-hyped artificial-intelligence (AI)-powered chatbots. ChatGPT and its ilk seem poised to become part of many researchers daily routines and were feted as part of the 2023 Natures 10 round-up (see go.nature.com/3trp7rg). Respondents to a Nature survey in September (see go.nature.com/45232vd) cited ChatGPT as the most useful AI-based tool and were enthusiastic about its potential for coding, literature reviews and administrative tasks.

Such tools are also proving valuable from an equity perspective, helping those for whom English isnt their first language to refine their prose and thereby ease their paths to publication and career growth. However, many of these applications represent labour-saving gains rather than transformations of the research process. Furthermore, ChatGPTs persistent issuing of either misleading or fabricated responses was the leading concern of more than two-thirds of survey respondents. Although worth monitoring, these tools need time to mature and to establish their broader role in the scientific world.

Sequence-based approaches can build on and adapt existing protein features to form new frameworks, but theyre less effective for the bespoke design of structural elements or features, such as the ability to bind specific targets in a predictable fashion. Structure based approaches are better for this, and 2023 saw notable progress in this type of protein-design algorithm, too. Some of the most sophisticated of these use diffusion models, which also underlie image-generating tools such as DALL-E. These algorithms are initially trained to remove computer-generated noise from large numbers of real structures; by learning to discriminate realistic structural elements from noise, they gain the ability to form biologically plausible, user-defined structures.

RFdiffusion software3 developed by Bakers lab and the Chroma tool by Generate Biomedicines in Somerville, Massachusetts4, exploit this strategy to remarkable effect. For example, Bakers team is using RFdiffusion to engineer novel proteins that can form snug interfaces with targets of interest, yielding designs that just conform perfectly to the surface, Baker says. A newer all atom iteration of RFdiffusion5 allows designers to computationally shape proteins around non-protein targets such as DNA, small molecules and even metal ions. The resulting versatility opens new horizons for engineered enzymes, transcriptional regulators, functional biomaterials and more.

The explosion of publicly available generative AI algorithms has made it simple to synthesize convincing, but entirely artificial images, audio and video. The results can offer amusing distractions, but with multiple ongoing geopolitical conflicts and a US presidential election on the horizon, opportunities for weaponized media manipulation are rife.

Siwei Lyu, a computer scientist at the University at Buffalo in New York, says hes seen numerous AI-generated deepfake images and audio related to the IsraelHamas conflict, for instance. This is just the latest round in a high-stakes game of cat-and-mouse in which AI users produce deceptive content and Lyu and other media-forensics specialists work to detect and intercept it.

AI and science: what 1,600 researchers think

One solution is for generative-AI developers to embed hidden signals in the models output, producing watermarks of AI-generated content. Other strategies focus on the content itself. Some manipulated videos, for instance, replace the facial features of one public figure with those of another, and new algorithms can recognize artefacts at the boundaries of the substituted features, says Lyu. The distinctive folds of a persons outer ear can also reveal mismatches between a face and a head, whereas irregularities in the teeth can reveal edited lip-sync videos in which a persons mouth was digitally manipulated to say something that the subject didnt say. AI-generated photos also present a thorny challenge and a moving target. In 2019, Luisa Verdoliva, a media-forensics specialist at University Federico II of Naples, Italy, helped to develop FaceForensics++, a tool for spotting faces manipulated by several widely used software packages6. But image-forensic methods are subject- and software-specific, and generalization is a challenge. You cannot have one single universal detector its very difficult, she says.

And then theres the challenge of implementation. The US Defense Advanced Research Projects Agencys Semantic Forensics (SemaFor) programme has developed a useful toolbox for deepfake analysis, but, as reported in Nature (see Nature 621, 676679; 2023) major social-media sites are not routinely employing it. Broadening the access to such tools could help to fuel uptake, and to this end Lyus team has developed the DeepFake-O-Meter7, a centralized public repository of algorithms that can analyse video content from different angles to sniff out deepfake content. Such resources will be helpful, but it is likely that the battle against AI-generated misinformation will persist for years to come.

In late 2023, US and UK regulators approved the first-ever CRISPR-based gene-editing therapy for sickle-cell disease and transfusion-dependent -thalassaemia a major win for genome editing as a clinical tool.

CRISPR and its derivatives use a short programmable RNA to direct a DNA-cutting enzyme such as Cas9 to a specific genomic site. They are routinely used in the lab to disable defective genes and introduce small sequence changes. The precise and programmable insertion of larger DNA sequences spanning thousands of nucleotides is difficult, but emerging solutions could allow scientists to replace crucial segments of defective genes or insert fully functional gene sequences. Le Cong, a molecular geneticist at Stanford University in California and his colleagues are exploring single-stranded annealing proteins (SSAPs) virus-derived molecules that mediate DNA recombination. When combined with a CRISPRCas system in which the DNA-slicing function of Cas9 has been disabled, these SSAPs allow precisely targeted insertion of up to 2 kilobases of DNA into the human genome.

Seven technologies to watch in 2023

Other methods exploit a CRISPR-based method called prime editing to introduce short landing pad sequences that selectively recruit enzymes that in turn can precisely splice large DNA fragments into the genome. In 2022, for instance, genome engineers Omar Abudayyeh and Jonathan Gootenberg at the Massachusetts Institute of Technology, Cambridge and their colleagues first described programmable addition through site-specific targeting elements (PASTE), a method that can precisely insert up to 36 kilobases of DNA8. PASTE is especially promising for ex vivo modification of cultured, patient-derived cells, says Cong, and the underlying prime-editing technology is already on track for clinical studies. But for in vivo modification of human cells, SSAP might offer a more compact solution: the bulkier PASTE machinery requires three separate viral vectors for delivery, which could undermine editing efficiency relative to the two-component SSAP system. That said, even relatively inefficient gene-replacement strategies could be sufficient to mitigate the effects of many genetic diseases.

And such methods are not just relevant to human health. Researchers led by Caixia Gao at the Chinese Academy of Sciences in Beijing developed PrimeRoot, a method that uses prime editing to introduce specific target sites that enzymes can use to insert up to 20 kilobases of DNA in both rice and maize9. Gao thinks that the technique could be broadly useful for endowing crops with disease and pathogen resistance, continuing a wave of innovation in CRISPR-based plant genome engineering. I believe that this technology can be applied in any plant species, she says.

Pat Bennett has slower than average speech, and can sometimes use the wrong word. But given that motor neuron disease, also known as amyotrophic lateral sclerosis, had previously left her unable to express herself verbally, that is a remarkable achievement.

Bennetts recovery comes courtesy of a sophisticated braincomputer interface (BCI) device developed by Stanford University neuroscientist Francis Willett and his colleagues at the US-based BrainGate consortium10. Willett and his colleagues implanted electrodes in Bennetts brain to track neuronal activity and then trained deep-learning algorithms to translate those signals into speech. After a few weeks of training, Bennett was able to say as many as 62 words per minute from a vocabulary of 125,000 words more than twice the vocabulary of the average English speaker. Its really truly impressive, the rates at which theyre communicating, says bioengineer Jennifer Collinger, who develops BCI technologies at the University of Pittsburgh in Pennsylvania.

Braincomputer interface technology has allowed Pat Bennett (seated) to regain her speech.Credit: Steve Fisch/Stanford Medicine

BrainGates trial is just one of several studies from the past few years demonstrating how BCI technology can help people with severe neurological damage to regain lost skills and achieve greater independence. Some of that progress stems from the steady accumulation of knowledge about functional neuroanatomy in the brains of individuals with various neurological conditions, says Leigh Hochberg, a neurologist at Brown University in Providence, Rhode Island, and director of the BrainGate consortium. But that knowledge has been greatly amplified, he adds, by machine-learning-driven analytical methods that are revealing how to better place electrodes and decrypt the signals that they pick up.

Researchers are also applying AI-based language models to speed up the interpretation of what patients are trying to communicate essentially, autocomplete for the brain. This was a core component of Willetts study, as well as another11 from a team led by neurosurgeon Edward Chang at the University of California, San Francisco. In that work, a BCI neuroprosthesis allowed a woman who was unable to speak as a result of a stroke to communicate at 78 words per minute roughly half the average speed of English, but more than five times faster than the womans previous speech-assistance device. The field is seeing progress in other areas as well. In 2021, Collinger and biomedical engineer Robert Gaunt at the University of Pittsburgh implanted electrodes into the motor and somatosensory cortex of an individual who was paralysed in all four limbs to provide rapid and precise control over a robotic arm along with tactile sensory feedback12. Also under way are independent clinical studies from BrainGate and researchers at UMC Utrecht in the Netherlands, as well as a trial from BCI firm Synchron in Brooklyn, New York, to test a system that allows people who are paralysed to control a computer the first industry-sponsored trial of a BCI apparatus.

As an intensive-care specialist, Hochberg is eager to deliver these technologies to his patients with the most severe disabilities. But as BCI capabilities evolve, he sees potential to treat more-moderate cognitive impairments as well as mental-health conditions, such as mood disorders. Closed-loop neuromodulation systems informed by braincomputer interfaces could be of tremendous help to a lot of people, he says.

Stefan Hell, Eric Betzig and William Moerner were awarded the 2014 Nobel Prize in Chemistry for shattering the diffraction limit that constrained the spatial resolution of light microscopy. The resulting level of detail in the order of tens of nanometres opened a wide range of molecular-scale imaging experiments. Still, some researchers yearn for better and they are making swift progress. Were really trying to close the gap from super-resolution microscopy to structural-biology techniques like cryo-electron microscopy, says Ralf Jungmann, a nanotechnology researcher at the Max Planck Institute of Biochemistry in Planegg, Germany, referring to a method that can reconstruct protein structures with atomic-scale resolution.

Researchers led by Hell and his team at the Max Planck Institute for Multidisciplinary Sciences in Gttingen made an initial foray into this realm in late 2022 with a method called MINSTED that can resolve individual fluorescent labels with 2.3-ngstrm precision roughly one-quarter of a nanometre using a specialized optical microscope13.

Newer methods provide comparable resolution using conventional microscopes. Jungmann and his team, for instance, described a strategy in 2023 in which individual molecules are labelled with distinct DNA strands14. These molecules are then detected with dye-tagged complementary DNA strands that bind to their corresponding targets transiently but repeatedly, making it possible to discriminate individual fluorescent blinking points that would blur into a single blob if imaged simultaneously. This resolution enhancement by sequential imaging (RESI) approach could resolve individual base pairs on a DNA strand, demonstrating ngstrm-scale resolution with a standard fluorescence microscope.

The one-step nanoscale expansion (ONE) microscopy method, developed by a team led by neuroscientists Ali Shaib and Silvio Rizzoli at University Medical Center Gttingen, Germany, doesnt quite achieve this level of resolution. However, ONE microscopy offers an unprecedented opportunity to directly image fine structural details of individual proteins and multiprotein complexes, both in isolation and in cells15.

A form of imaging called RESI could allow the imaging of individual base pairs in DNA.Credit: Max Iglesias, Max Planck Institute of Biochemistry

ONE is an expansion-microscopy-based approach that involves chemically coupling proteins in the sample to a hydrogel matrix, breaking the proteins apart, and then allowing the hydrogel to expand 1,000-fold in volume. The fragments expand evenly in all directions, preserving the protein structure and enabling users to resolve features separated by a few nanometres with a standard confocal microscope. We took antibodies, put them in the gel, labelled them after expansion, and were like, Oh we see Y shapes! says Rizzoli, referring to the characteristic shape of the proteins.

ONE microscopy could provide insights into conformationally dynamic biomolecules or enable visual diagnosis of protein-misfolding disorders such as Parkinsons disease from blood samples, says Rizzoli. Jungmann is similarly enthusiastic about the potential for RESI to document reorganization of individual proteins in disease or in response to drug treatments. It might even be possible to zoom in more tightly. Maybe its not the end for the spatial resolution limits, Jungmann says. It might get better.

If youre looking for a convenient cafe, Google Maps can find nearby options and tell you how to get there. Theres no equivalent for navigating the much more complex landscape of the human body, but ongoing progress from various cell-atlas initiatives powered by advances in single-cell analysis and spatial omics methods could soon deliver the tissue-wide cellular maps that biologists crave.

The largest and perhaps the most ambitious of these initiatives is the Human Cell Atlas (HCA). The consortium was launched in 2016 by cell biologist Sarah Teichmann at the Wellcome Sanger Institute in Hinxton, UK, and Aviv Regev, now head of research and early development at biotechnology firm Genentech in South San Francisco, California. It encompasses some 3,000 scientists in nearly 100 countries, working with tissues from 10,000 donors. But HCA is also part of a broader ecosystem of intersecting cellular and molecular atlas efforts. These include the Human BioMolecular Atlas Program (HuBMAP) and the Brain Research through Advancing Innovative Neurotechnologies (BRAIN) Initiative Cell Census Network (BICCN), both funded by the US National Institutes of Health, as well as the Allen Brain Cell Atlas, funded by the Allen Institute in Seattle, Washington.

According to Michael Snyder, a genomicist at Stanford University and former co-chair of the HuBMAP steering committee, these efforts have been driven in part by the development and rapid commercialization of analytical tools that can decode molecular contents at the single-cell level. For example, Snyders team routinely uses the Xenium platform from 10X Genomics in Pleasanton, California, for its spatial transcriptomics analyses. The platform makes it possible to survey the expression of roughly 400 genes at once in 4 tissue samples every week. Multiplexed antibody-based methods such as the PhenoCycler platform by Akoya Biosciences in Marlborough, Massachusetts, allow the team to track large numbers of proteins with single-cell resolution in a format that enables 3D tissue reconstruction. Other multiomics methods allow scientists to profile multiple molecular classes in the same cell at once, including the expression of RNA, the structure of chromatin and the distribution of protein.

A cell atlas of the human lung describes different cell types and how they are regulated.Credit: Peng He

Last year saw dozens of studies showcasing progress in the generation of organ-specific atlases using these techniques. In June, for example, the HCA released an integrated analysis of 49 data sets from the human lung16. Having that very clear map of the lung informs the changes that happen in diseases like lung fibrosis, different tumours, even for COVID-19, says Teichmann. And in 2023, Nature released an article collection (see go.nature.com/3vbznk7) highlighting progress from HuBMAP and Science produced a collection detailing the work of the BICCN (see go.nature.com/3nsf4ys).

Considerable work remains Teichmann estimates that it will be at least five years before the HCA reaches completion. But the resulting maps will be invaluable when they arrive. Teichmann, for example, predicts using atlas data to guide tissue- and cell-specific drug targeting, while Snyder is eager to learn how cellular microenvironments inform the risk and aetiology of complex disorders such as cancer and irritable bowel syndrome. Will we solve that in 2024? I dont think so its a multiyear problem, Snyder says. But its a big driver for this whole field.

Weird and interesting things can happen at the nanometre scale. This can make materials-science predictions difficult, but it also means that nanoscale architects can manufacture lightweight materials with distinctive characteristics such as increased strength, tailored interactions with light or sound, and enhanced capacity for catalysis or energy storage.

Several strategies exist for precisely crafting such nanomaterials, most of which use lasers to induce patterned photopolymerization of light-sensitive materials, and over the past few years, scientists have made considerable headway in overcoming the limitations that have impeded broader adoption of these methods.

Researchers have crafted microscale metal structures using a hydrogel.Credit: Max Saccone/Greer Lab

One is speed. Sourabh Saha, an engineer at the Georgia Institute of Technology in Atlanta, says that the assembly of nanostructures using photopolymerization is roughly three orders of magnitude faster than other nanoscale 3D-printing methods. That might be good enough for lab use, but its too slow for large-scale production or industrial processes. In 2019, Saha and mechanical engineer Shih-Chi Chen at the Chinese University of Hong Kong and their colleagues showed that they could accelerate polymerization by using a patterned 2D light-sheet rather than a conventional pulsed laser17. That increases the rate by a thousand times, and you can still maintain those 100-nanometre features, says Saha. Subsequent work from researchers including Chen has identified other avenues for faster nanofabrication18.

Another challenge is that not all materials can be printed directly through photopolymerization such as metals. But Julia Greer, a materials scientist at the California Institute of Technology in Pasadena, has developed a clever workaround. In 2022, she and her colleagues described a method in which photopolymerized hydrogels serve as a microscale template; these are then infused with metal salts and processed in a way that induces the metal to assume the structure of the template while also shrinking19. Although the technique was initially developed for microscale structures, Greers team has also used this strategy for nanofabrication, and the researchers are enthusiastic about the potential to craft functional nanostructures from rugged, high-melting-point metals and alloys.

The final barrier economics could be the toughest to break. According to Saha, the pulsed-laser-based systems used in many photopolymerization methods cost upwards of US$500,000. But cheaper alternatives are emerging. For example, physicist Martin Wegener and his colleagues at the Karlsruhe Institute of Technology in Germany have explored continuous lasers that are cheaper, more compact, and consume less power than standard pulsed lasers20. And Greer has launched a start-up company to commercialize a process for fabricating nanoarchitected metal sheets that could be suitable for applications such as next-generation body armour or ultra-durable and impact-resistant outer layers for aircraft and other vehicles.

Read this article:
Seven technologies to watch in 2024 - Nature.com

Tags:

Artificial intelligence helped scientists create a new type of battery – Science News Magazine

In the hunt for new materials, scientists have traditionally relied on tinkering in the lab, guided by intuition, with a hefty serving of trial and error.

But now a new battery material has been discovered by combining two computing superpowers: artificial intelligence and supercomputing. Its a discovery that highlights the potential for using computers to help scientists discover materials suited to specific needs, from batteries to carbon capture technologies to catalysts.

Calculationswinnowed down more than 32 million candidate materialsto just 23 promising options, researchers from Microsoft and Pacific Northwest National Laboratory, or PNNL, report in a paper submitted January 8 to arXiv.org. The team then synthesized and tested one of those materials and created a working battery prototype.

While scientists have used AI to predict materials properties before, previous studies typically havent seen that process through to producing the new material. The nice thing about this paper is that it goes all the way from start to finish, says computational materials scientist Shyue Ping Ong of the University of California, San Diego, who was not involved with the research.

The researchers targeted a coveted type of battery material: a solid electrolyte. An electrolyte is a material that transfers ions electrically charged atoms back and forth between a batterys electrodes. In standard lithium-ion batteries, the electrolyte is a liquid. But that comes with hazards, like batteries leaking or causing fires. Developing batteries with solid electrolytes is a major aim of materials scientists.

The original 32 million candidates were generated via a game of mix-and-match, substituting different elements in crystal structures of known materials. Sorting through a list this large with traditional physics calculations would have taken decades, says computational chemist Nathan Baker of Microsoft. But with machine learning techniques, which can make quick predictions based on patterns learned from known materials, the calculation produced results in just 80 hours.

First, the researchers used AI to filter the materials based on stability, namely, whether they could actually exist in the real world. That pared the list down to fewer than 600,000 candidates. Further AI analysis selected candidates likely to have the electrical and chemical properties necessary for batteries. Because AI models are approximate, the researchers filtered this smaller list using tried-and-tested, computationally intensive methods based on physics. They also weeded out rare, toxic or expensive materials.

That left the researchers with 23 candidates, five of which were already known. Researchers at PNNL picked a material that looked promising it was related to other materials that the researchers knew how to make in the lab, and it had suitable stability and conductivity. Then they set to work synthesizing it, eventually fashioning it into a prototype battery. And it worked.

Thats when we got very excited, says materials scientist Vijay Murugesan of PNNL in Richland, Wash. Going from the synthesis stage to the functional battery took about six months. That is superfast.

The new electrolyte is similar to a known material containing lithium, yttrium and chlorine,but swaps some lithium for sodium an advantage aslithium is costly and in high demand(SN: 5/7/19).

Combining lithium and sodium is unconventional. In a usual approach we would not mix these two together, says materials scientist Yan Zeng of Florida State University in Tallahassee, who was not involved in the research. The typical practice is to use either lithium or sodium ions as a conductor, not both. The two types of ions might be expected to compete with one another, resulting in worse performance. The unorthodox material highlights one hope for AI in research, Zeng says: AI can sort of step out of the box.

In the new work, the researchers created a series of AI models that could predict different properties of a material, based on training data from known materials. The AI architecture is a type known as a graph neural network, in which a system is represented as a graph, a mathematical structure composed of edges and nodes. This type of model is particularly suited for describing materials, as the nodes can represent atoms, and the edges can represent bonds between the elements.

To perform both the AI and physics-based calculations, the team used Microsofts Azure Quantum Elements, which provides access to a cloud-based supercomputer tailored for chemistry and materials science research.

The project, Baker says, is an example of a practice known in tech circles as eating your own dog food, in which a company uses its own product to confirm that it works. In the future, he says he hopes others will pick up the tool and use it for a variety of scientific endeavors.

The study is one of many efforts to use AI to discover new materials. In November, researchers from Google DeepMind used graph neural networks to predict the existence ofhundreds of thousands of stable materials, they reported in the Dec. 7Nature. And in the same issue ofNature,Zeng and colleagues reported on alaboratoryoperated by AI,designed to produce new materials autonomously.

Read the rest here:
Artificial intelligence helped scientists create a new type of battery - Science News Magazine

Tags:

Beware The Coming Artificial Intelligence Tax – Forbes

Beware The Coming Artificial Intelligence Tax  Forbes

Follow this link:
Beware The Coming Artificial Intelligence Tax - Forbes

Tags:

AI News: Artificial Intelligence Trends And Top AI Stocks To Watch – Investor’s Business Daily

Access to this page has been denied because we believe you are using automation tools to browse the website.

This may happen as a result of the following:

Please make sure that Javascript and cookies are enabled on your browser and that you are not blocking them from loading.

Reference ID: #4283cdc6-b977-11ee-9d93-d48e97fe83c6

Read more:
AI News: Artificial Intelligence Trends And Top AI Stocks To Watch - Investor's Business Daily

Tags: