Media Search:



Federal Court Rules That Artificial Intelligence Cannot Be An Inventor Under The Patent Act – JD Supra

Although this blog typically focuses on decisions rendered in intellectual property and/or antitrust cases currently in or that originated in the United States District Court for the District of Delaware or are in the Federal Circuit, every now and then there is a decision rendered in another federal trial or appellate court that is significant enough it warrants going beyond the normal boundaries. The recent decision rendered by The Honorable Leonie M. Brinkema, of the United States District Court for the Eastern District of Virginia, in Thaler v. Hirshfeld et al., Civil Action No. 1:20-cv-903-LMB (E.D.Va. September 2, 2021), is such a decision.

In Thaler, the Court confronted, analyzed and answered the question of can an artificial intelligence machine be an inventor under the Patent Act? Id. at *1. After analyzing the plain statutory language of the Patent Act and the Federal Circuit authority, the Court held that the clear answer is no. Id. In reaching its holding, the Court found that Congress intended to limit the definition of inventor to natural persons which means humans not artificial intelligence. Id. at *17. The Court noted that, [a]s technology evolves, there may come a time when artificial intelligence reaches a level of sophistication such that it might satisfy accepted meanings of inventorship. But that time has not yet arrived, and, if it does, it will be up to Congress to decide how, if at all, it wants to expand the scope of patent law. Id. at *17-18.

A copy of the Memorandum Opinion is attached.

[View source.]

More here:
Federal Court Rules That Artificial Intelligence Cannot Be An Inventor Under The Patent Act - JD Supra

AI caramba, those neural networks are power-hungry: Counting the environmental cost of artificial intelligence – The Register

Feature The next time you ask Alexa to turn off your bedroom lights or make a computer write dodgy code, spare a thought for the planet. The back-end mechanics that make it all possible take up a lot of power, and these systems are getting hungrier.

Artificial intelligence began to gain traction in mainstream computing just over a decade ago when we worked out how to make GPUs handle the underlying calculations at scale. Now there's a machine learning algorithm for everything, but while the world marvels at the applications, some researchers are worried about the environmental expense.

One of the most frequently quoted papers on this topic, from the University of Massachusetts, analysed training costs on AI including Google's BERT natural language processing model. It found that the cost of training BERT on a GPU in carbon emissions was roughly the same as a trans-American jet flight.

Kate Saenko, associate professor of computer science at Boston University, worries that we're not doing enough to make AI more energy efficient. "The general trend in AI is going in the wrong direction for power consumption," she warns. "It's getting more expensive in terms of power to train the newer models."

The trend is exponential. Researchers associated with OpenAI wrote that the computing used to train the average model increases by a factor of 10 each year.

Most AI these days is based on machine learning (ML). This uses a neural network, which is a collection of nodes designed in layers. Each node has connections to nodes in the next. Each of these connections has a score known as a parameter or weight.

The neural network takes an input (such as a picture of a hotdog) and runs it through the layers of the neural network, each of which uses its parameters to produce an output. The final output is a judgement about the data (for example, was the original input a picture of a hotdog or not?)

Those weights don't come preconfigured. You have to calculate them. You do that by showing the network lots of labelled pictures of hot dogs and not hot dogs. You keep training it until the parameters are optimised, which means that they spit out the correct judgement for each piece of data as often as possible. The more accurate the model, the better it will be when making judgements about new data.

You don't just train an AI model once. You keep doing it, adjusting various aspects of the neural network each time to maximise the right answers. These aspects are called hyperparameters, and they include variables such as the number of neurons in each layer and the number of layers in each network. A lot of that tuning is trial and error, which can mean many training passes. Chewing through all that data is already expensive enough, but doing it repeatedly uses even more electrons.

The reason that the models are taking more power to train is that researchers are throwing more data at them to produce more accurate results, explains Lukas Biewald. He's the CEO of Weights and Biases, a company that helps AI researchers organise the training data for all these models while monitoring their compute usage.

"What's alarming about about it is that it seems like for every factor of 10 that you increase the scale of your model training, you get a better model," he says.

Yes, but the model's accuracy doesn't increase by a factor of 10. Jesse Dodge, postdoctoral researcher at the Allen Institute for AI and co-author of a paper called Green AI, notes studies pointing to the diminishing returns of throwing more data at a neural network.

So why do it?

"There's a long tail of things to learn," he explains. ML algorithms can train on the most commonly-seen data, but the edge cases the confusing examples that rarely come up are harder to optimise for.

Our hotdog recognition system might be fine until some clown comes along in a hotdog costume, or it sees a picture of a hotdog-shaped van. A language processing model might be able to understand 95 per cent of what people say, but wouldn't it be great if it could handle exotic words that hardly anyone uses? More importantly, your autonomous vehicle must be able to stop in dangerous conditions that rarely ever arise.

"A common thing that we see in machine learning is that it takes exponentially more and more data to get out into that long tail," Dodge says.

Piling on all this data data doesn't just slurp power on the compute side, points out Saenko; it also burdens other parts of the computing infrastructure. "The larger the data, the more overhead," she says. "Even transferring the data from the hard drive to the GPU memory is power intensive."

There are various attempts to mitigate this problem. It starts at the data centre level, where hyperscalers are doing their best to switch to renewables so that they can at least hammer their servers responsibly.

Another approach involves taking a more calculated approach when tweaking your hyperparameters. Weights and Biases offers a "hyperparameter sweep" service that uses Bayesian algorithms to narrow the field of potential changes with each training pass. It also offers an "early stopping" algorithm which halts a training pass early on if the optimisation isn't panning out.

Not all approaches involve fancy hardware and software footwork. Some are just about sharing. Dodge points out that researchers could amortise the carbon cost of their model training by sharing the end result. Trained models released in the public domain can be used without retraining, but people don't take enough advantage of that.

"In the AI community, we often train models and then don't release them," he says. "Or the next people that want to build on our work just rerun the experiments that we did."

Those trained models can also be fine tuned with additional data, enabling people to tweak existing optimisations for new applications without retraining the entire model from scratch.

Making training more efficient only tackles one part of the problem, and it isn't the most important part. The other side of the AI story is inference. This is when a computer runs new data through a trained model to evaluate it, recognising hotdogs it has never seen before. It still takes power, and the rapid adoption of AI is making it more of a problem. Every time you ask Siri how to cook rice properly, it uses inference power in the cloud.

One way to reduce model size is to cut down the number of parameters. AI models often use vast numbers of weights in a neural network because data scientists aren't sure which ones will be most useful. Saenko and her colleagues have researched reducing the number of parameters using a concept that they call shape shifter networks that share some of the parameters in the final model.

"You might train a much bigger network and then distil it into a smaller one so that you can deploy a smaller network and save computation and deployment at inference time," she says.

Companies are also working on hardware innovations to cope with this increased inference load. Google's Tensor Processing Units (TPUs) are tailored to handle both training and inference more efficiently, for example.

Solving the inference problem is especially tricky because we don't know where a lot of it will happen in the long term. The move to edge computing could see more inference jobs happening in lower-footprint devices rather than in the cloud. The trick there is to make the models small enough and to introduce hardware advances that will help to make local AI computation more cost-effective.

"How much do companies care about running their inference on smaller devices rather than in the cloud on GPUs?" Saenko muses. "There is not yet that much AI running standalone on edge devices to really give us some clear impetus to figure out a good strategy for that."

Still, there is movement. Apple and Qualcomm have already produced tailored silicon for inference on smart phones, and startups are becoming increasingly innovative in anticipation of edge-based inference. For example, semiconductor startup Mythic launched an AI processor focused on edge-based AI that uses analogue circuitry and in-memory computing to save power. It's targeting applications including object detection and depth estimation, which could see the chips turn up in everything from factories to surveillance cameras.

As companies grapple with whether to infer at the edge, the problem of making AI more energy efficient in the cloud remains. The key lies in resolving two opposing forces: on the one hand, everyone wants more energy efficient computing. On the other, researchers constantly strive for more accuracy.

Dodge notes that most academic AI papers today focus on the latter. Accuracy is winning out as companies strive to beat each other with better models, agrees Saenko. "It might take a lot of compute but it's worthwhile for people to claim that one or two percent improvement," she says.

She would like to see more researchers publish data on the power consumption of their models. This might inspire competition to drive efficiencies up and costs down.

The stakes may be more than just environmental, warns Biewald; they could be political too. What happens if computing consumption continues to go up by a factor of 10 each year?

"You have to buy the energy to train these models, and the only people that can realistically afford that will be Google and Microsoft and the 100 biggest corporations," he posits.

If we start seeing a growing inequality gap in AI research, with corporate interests out in front, carbon emissions could be the least of our worries.

See the article here:
AI caramba, those neural networks are power-hungry: Counting the environmental cost of artificial intelligence - The Register

DMALINK partners with Axyon AI to add deep learning artificial intelligence to its platform tech stack – PRNewswire

LONDON, Sept. 13, 2021 /PRNewswire/ -- DMALINK, the emerging markets foreign exchange focused institutional ECN brings the FX market into the heart of the 4th industrial revolution.

The firm today announced its partnership with Axyon AI to enable the first ever use of Deep Learning artificial intelligence to dynamically manage liquidity, detect market and order anomalies and create smart algos for trade execution in the fiat FX space.

Axyon AI is a leading European FinTech company with expertise in Deep Learning/AI for asset management and trading firms. Axyon AI has successful products in several financial use-cases, from security selection and asset allocation to anomaly detection in option pricing.

Manu Choudhary, Chief Executive Officer of DMALINK, says: "In spite of the pace of innovation within the e-FX space, liquidity management, anomaly detection and algos have been left behind by advances in deep learning AI technology. The ability for Axyon AI's deep learning technology to leverage insights in a fraction of the time of a human-driven equivalent provides opportunities for the procurement and analysis of unique data to dynamically manage liquidity, risk and trade execution for the first time."

Axyon AI's technology will combine with DMALINK's ECN infrastructure to radically modernise FX trading. For the buy side, deep learning models will considerably improve the quality of order fills. For the sell side, the application will ensure a positive yield curve. The deep learning technology will also detect market anomalies in spot FX allowing DMALINK participants access to one of the most powerful risk management tools developed in the e-FX space. Smart algos, dynamically created by the AI will instantly adjust trade execution as a function of the market dynamics.

Daniele Grassi, Chief Executive Officer of Axyon AI, says "We believe that deep learning has just begun to transform financial markets, increasing efficiency and improving risk management. Our partnership with DMALINK will be a driver of this paradigm shift for the FX trading industry."

About DMALINK

DMALINK is an independent electronic trading, analytics and market data venue for institutional FX traders globally. All liquidity pools are proactively constructed across key emerging markets. Its platform participants benefit from advanced order analytics data and granular reporting and benchmarked execution services, ensuring price transparency for all platform participants.

About Axyon AI

Axyon AI is a leading player in deep learning, the newest area of machine learning artificial intelligence, for time series forecasting. Axyon AI partners with asset managers and hedge funds to deliver consistently high-performing end-to-end AI powered quantitative insights and investment strategies.

For media enquiries, please contact:

Media Room, DMALINK, Tel: +44 (0) 20 7117 2517

SOURCE DMALINK

Continued here:
DMALINK partners with Axyon AI to add deep learning artificial intelligence to its platform tech stack - PRNewswire

No one can stand before power of national will: Erdoan – Hurriyet Daily News

ANKARA

President Recep Tayyip Erdoan on Sept. 13 said Turkey has shown that the power of national will can withstand any kind of attack against the country.

We have shown the whole world that neither tutelage nor social engineering calculations nor terrorist organizations nor coups can stand before the power of the national will, he said, speaking at the centennial anniversary celebration program of the Sakarya Victory.

Turkey didnt get to this point that easily as the country was harassed from all sides to divert its attention from its goals, Erdoan said while stressing that his government has overcome these with the support of the nation.

In the restructuring process of the global, political and economic system, we, as Turkey, always turn to the future and determine the direction and move towards it. Undoubtedly, the culmination of these steps is the 100th anniversary of the establishment of our Republic, which we will be qualified for on Oct. 29, 2023, without any worries, he stated.

Some countries cannot bear even one of the attacks that Turkey has faced, and every extraordinary situation that has been experienced in the last period, from pandemic to natural disasters in Turkey, makes this picture even more evident, Erdoan emphasized.

We stand tall thanks to the power of thousands of years of state tradition behind us, he added.

Now, we are more confident in ourselves; we look to our future with more hope, Erdoan said, noting that the Peoples Alliance, formed by the ruling Justice and Development Party (AKP) and Nationalist Movement Party (MHP), are walking towards the future with more confidence.

The main thing for us is the will of our nation: One nation, one homeland, one state and one flag. We know that like any victory we have won since the [battle of] Malazgirt [Manzikert], this will not be easy, Erdoan said, referring to a battle between the armies of the Seljuk Turks and the Byzantine Empire that took place on the plain of Manzikert (Malazgirt) in what is now Mu in eastern Turkey, which paved the way for both the Ottoman Empire and the modern Republic of Turkey.

Erdogan, Politics, Diplomacy,

The rest is here:
No one can stand before power of national will: Erdoan - Hurriyet Daily News

Rand Paul says Fauci deserves a five-year prison sentence | TheHill – The Hill

Sen. Rand Paul (R-Ky.) intensified his criticism of Anthony Fauci, claiming Wednesday that the nations top infectious diseases expert lied to Congress and should face a five-year prison sentence.

The senator has claimed for months that Fauci, who serves as the president's chief medical adviser and the director of the National Institute of Allergy and Infectious Diseases, lied to Congress about whether the National Institute of Health (NIH) funded gain-of-function research at the Wuhan Institute of Virology, Mediate reported. Gain-of-function research refers to the process of changing pathogens to test scientific theories, develop technologies and find treatments.

Paul made the claims again on Fox News this week.

Its a felony punishable by five years in jail, Paul said in an interview with Fox News host Sean Hannity. I dont think Biden Department of Justice will do anything with it, but it is very dangerous to have public officials who we need to have trust incoming and lying to us. But he has lied dozens of times. Usually, he tells us its for our own good.

America is changing faster than ever! Add Changing America to your Facebook or Twitter feed to stay on top of the news.

As Changing America previously reported, Rand referred Fauci to the Justice Department in July.

That same month, Rand escalated his clash with Fauci during a Senate hearing. The senator suggested that Fauci and the NIH could be somewhat responsible for the pandemic and he cited a paper that alleges that the lab was conducting illegal research, a claim that Fauci denied.

I totally resent the lie you are now propagating, senator, Fauci said.

Paul interjected, saying: You are obviously obfuscating the truth.

Fauci replied, I'm not obfuscating the truth -- you are. He continued: You are implying that what we did was responsible for the deaths of individuals. I totally resent that...And if anybody is lying here senator, it is you.

He has dissembled, Paul said in Wednesdays interview. He has obfuscated. There are other, nicer words, but he has definitely lied to the American public, and he should be held responsible. But not just that the judgment that we should continue to fund this lab and the virus that in all likelihood came from the lab, I think it is such incredibly poor judgment that he should be immediately removed.

The theory that the coronavirus leaked from a lab has been dismissed in the past, but is gaining traction as a possible explanation for the origins of the virus.

READ MORE STORIES FROM CHANGING AMERICA

FAUCI SAYS HES KEEPING A VERY CLOSE EYE ON NEW COVID-19 VARIANT

TUCKER CARLSON DEFENDS FAKE VACCINATION CARDS

KIM JONG UN TURNS DOWN VACCINES, INSISTS ON FIGHTING PANDEMIC IN OUR STYLE

BOOSTER SHOTS MAY ONLY BE AVAILABLE TO PFIZER VACCINE RECIPIENTS AT FIRST, OFFICIALS SAY

WHAT YOU NEED TO KNOW ABOUT THE NEW COVID-19 VARIANT C.1.2 DETECTED IN SOUTH AFRICA

RAND PAUL SAYS HATRED FOR TRUMP IS STOPPING STUDY OF HORSE AND HUMAN DRUG TO TREAT COVID-19

STUDY SAYS NEW LAMBDA VARIANT COULD BE VACCINE-RESISTANT

AMERICAN ARCHBISHOP IS DISTRIBUTING BLEACH AS MIRACLE CURE FOR COVID-19

Read more from the original source:
Rand Paul says Fauci deserves a five-year prison sentence | TheHill - The Hill