Archive for the ‘Artificial General Intelligence’ Category

Macy’s Could See Over $7.5 Billion in Additional Business Gains … – CMSWire

Page 1 of IHL AI Readiness Profile for Macy's

Page 2 of IHL AI Readiness Profile for Macy's

Page 3 of IHL AI Readiness Profile for Macy's

New research projects increased sales opportunities, improvements in gross margins, and lower expenses due to AI Readiness

Greg Buzek, President IHL Group

According to the research, Macys could see as much as $3.8 billion in increased sales, $2.1 billion in improved gross margins through lower product costs, more optimized pricing, and supply chain improvements, and then reduce by $1.7 billion sales and generative administrative costs through 2029.

Our research approach was to start by looking at opportunities from an industry-level, then to the segment and specific retailer level leveraging our public and private data, said Greg Buzek, President of IHL Group. We then applied a 9-point algorithm to each company that measured items like data maturity, analytics maturity, alignment with key vendors, as well as free cash flow.

The research includes gains that can be made through traditional AI/ML technologies, Generative AI, and the potential for Artificial General Intelligence. These figures do not include any savings from reducing headcount, rather they focus on creating more efficiency and supporting growth/lower expenses through greater efficiencies only.

These figures do not consider any cost savings resulting from workforce reduction. Instead, it solely emphasizes the creation of greater efficiency to support growth and reduce expenses.

In total, each of the retailer profiles includes the following data:

Total AI Impact from 2022-2029: Combined impact from traditional AI/ML, Generative AI, and Artificial General Intelligence. Annual Impact by Income Statement Category: Gains in sales, gross margins, or lower operating costs. Total AI Readiness Score and Rankings vs Competitors: Shows competitiveness in segment and overall retail market AI Impact by Line of Business: Explore the AI potential in Merchandising/Supply Chain, Sales & Marketing, Commerce, Infrastructure, BI/Analytics, Store Systems, and other areas such as Collaboration, ERP, and Legal. Benefits by Specific Solutions: For instance, under Merchandising/Supply Chain gain insights on benefits gained via Order Management, Assortment and Allocation Planning, Distribution Systems, Warehouse Management, etc.

For a glimpse into the rich data and insights provided by these profiles, you can access the Macys profile here.

The Retail AI Readiness Profiles are available for individual companies or enterprises can access the entire directory of profiles with ongoing access to updated data as systems evolve.

About IHL Group:

IHL Group is a global research and advisory firm headquartered in Franklin, Tennessee, that provides market analysis and business consulting services for retailers and information technology companies that focus on the retail, hospitality, and consumer goods industries. For more information, see http://www.ihlservices.com, call 615-591-2955 or e-mail [emailprotected]. Press inquiries, please use [emailprotected] or the phone number above.

Note: This report is intended for informational purposes and does not constitute financial or investment advice. Please refer to the complete report and methodology for a detailed understanding of the data and analysis.

Gregory Buzek IHL Group +1 615-591-2955 email us here

Read more here:

Macy's Could See Over $7.5 Billion in Additional Business Gains ... - CMSWire

Securing the cloud and AI: Insights from Laceworks CISO – SiliconANGLE News

Artificial intelligence is more than just a buzzword. Its the result of many technologies coming together, starting at the hardware layers.

AI is being used to generate code and protect algorithms while also being used for security in analyzing cloud and code usage, explainedMerritt Baer (pictured), field chief information security officer of Lacework Inc. Regulating the acceleration of artificial general intelligence is a current cultural tension, with some advocating for acceleration and others for slowing it down.

Weve talked about AWS Nitro before and some of the confidential computing benefits that folks get from the fact that AWS built it to not be human accessible, Baer said. So, you dont have to pay extra for that factor. This is part of a longer tale about the chip industry and other things. Its important in that, I think right now, of course, AI is a buzzword, but what were really seeing is the culmination of a lot of technologies coming to bear.

Baer spoke with theCUBE industry analyst John Furrier at the Supercloud 5: The Battle for AI Supremacy event, during an exclusive broadcast on theCUBE, SiliconANGLE Medias livestreaming studio. They discussed how securing the cloud and addressing potential security threats in the new generation of AI is crucial.

Security around AI will be important, with the ability to do more and know more, and the most likely source of an attack being a valid credential being misused, according to Baer.Companies and industries are having a reckoning and need to define the values they want to live by with tech being human-constructed and the potential for underserved communities to gain more accessibility, while also aiming for more, better and faster progress in a deliberate and conscious way.

I think that security around AI and also the security of your AI will be areas that we care about for the foreseeable future. But were going to be doing stuff at an accelerated pace with that high power compute, with the ability to do more and know when youre hitting a wire, Baer said. A proverbial wire, like a threshold. So, being able to get real-time alerting and really low latency alerting around things that look anomalous.

Security should not be seen as a cost center, but as part of the business proposition, Baer explained. Lacework is delivering effective capabilities to help customers take action and improve over time. Improving security team response time, instant response, threat detection and identity monitoring are crucial for CISOs, who often feel isolated in their roles and can benefit from automation.

As a CISO, you want to reduce the likelihood of a bad day. You want to notice when your bad day starts and have it have little to no impact. Then you also want to get that for the ROI that you are aware of for something that you have already bargained for, Baer said. Is it known? Is it an unknown? These are business decisions, and security executives are increasingly realizing that they need to present this in a sense that makes investments relevant.

Heres the complete video interview, part of SiliconANGLEs and theCUBEs coverage of the Supercloud 5: The Battle for AI Supremacy event:

THANK YOU

See the rest here:

Securing the cloud and AI: Insights from Laceworks CISO - SiliconANGLE News

Amazon unleashes Q, an AI assistant for the workplace – Ars Technica

Enlarge / The Amazon Q logo.

Amazon

On Tuesday, Amazon unveiled Amazon Q, an AI chatbot similar to ChatGPT that is tailored for corporate environments. Developed by Amazon Web Services (AWS), Q is designed to assist employees with tasks like summarizing documents, managing internal support tickets, and providing policy guidance, differentiating itself from consumer-focused chatbots. It also serves as a programming assistant.

According to The New York Times, the name "Q" is a play on the word question" and a reference to the character Q in the James Bond novels, who makes helpful tools. (And there's apparently a little bit of Q from Star Trek: The Next Generation thrown in, although hopefully the new bot won't cause mischief on that scale.)

Amazon Q's launch positions it against existing corporate AI tools like Microsoft's Copilot, Google's Duet AI, and ChatGPT Enterprise. Unlike some of its competitors, Amazon Q isn't built on a singular AI large language model (LLM). Instead, it uses a platform called Bedrock, integrating multiple AI systems, including Amazon's Titan and models from Anthropic and Meta.

Amazon

"Developers can use Amazon Q to explain specific programming logic by asking questions (e.g., Provide me with a description of what this application does and how it works)," writes Amazon in a press release. "And Amazon Q will give details like which services the code uses and what different functions do (e.g., This application is building a basic support ticketing system using Python Flask and AWS Lambda), along with a description of the applications core capabilities, how they are implemented, and more."

Notably, Amazon did not reveal performance benchmarks for Q that would allow us to evaluate its capabilities versus chatbot solutions from other providers. As of press time, we have not experimented with Q yet.

Amazon Q promotional video on YouTube.

Following significant investments in AI, including a partnership with AI-startup Anthropic and the development of AI-tuned GPU chips, Amazon has intensified its AI focus. The Q announcement came as part of a series of reveals at Amazon's annual cloud-computing conference, re:Invent 2023, including plans to create yet another new AI chip for its data centers.

Amazon Q is priced at $20 per user per month, which is lower than Microsoft and Google's enterprise AI solutions, which are priced at $30 per user per month. Amazon Q is available now "in preview in AWS Regions US East (N. Virginia) and US West (Oregon)," according to the company.

It's worth noting that the Amazon Q name is apparently unrelated to recent rumors about an OpenAI breakthrough called "Q*" (pronounced "Q-star") that caused premature hype over the development of AGI (artificial general intelligence), with experts calling the claims largely overblown.

More here:

Amazon unleashes Q, an AI assistant for the workplace - Ars Technica

You’re not imagining things: The end of the -3- – Morningstar

Even if India does not become a formal ally to Western countries, it will continue to position itself as an independent, rising power whose interests are more aligned with the West than with China and its de facto allies (Russia, Iran, North Korea, and Pakistan). Moreover, India is a formal member of the Quadrilateral Security Dialogue (the Quad) with the U.S., Japan, and Australia, the explicit purpose of which is to deter China. Japan and India have longstanding friendly relations and a shared history of adversarial relations with China.

Japan also invited Indonesia, South Korea (with which it is pursuing a diplomatic thaw, driven by common concerns about China), Brazil (another key Global South power), and Ukrainian President Volodymyr Zelensky to the G7. In each case, the message was clear: The Sino-Russian friendship "without limits" is having serious consequences for how other powers perceive China.

In its final communiqu?, the G7 explained at length how it will confront and deter China in the years ahead. It decried Chinese "economic coercion" and expansionism in the East and South China Seas, stressed the importance of an Indo-Pacific partnership, and issued a clear warning to China not to attack or invade Taiwan.

In taking steps to "de-risk" their relationships with China, Western leaders settled on language that is only slightly less aggressive than "de-coupling." But it isn't just the diplomatic argot that has changed. According to the communiqu?, Western containment efforts will be accompanied by large investments in clean energy and infrastructure across the Global South, lest key middle powers be drawn into China's sphere of influence through its Belt and Road Initiative.

Meanwhile, the Western-Chinese tech and economic war continues to escalate. Japan recently imposed restrictions on semiconductor exports to China that are no less draconian than those introduced by the U.S., and the Biden administration has since pressured Taiwan and South Korea to follow suit. In response, China has banned semiconductors made by the U.S.-based chipmaker Micron Technology (MU), and has begun to restrict exports of some critical metals over which it has a near-monopoly in production and refining.

Likewise, U.S. chipmaker Nvidia (NVDA) - which is quickly becoming a corporate superpower, owing to surging demand for its advanced chips to power AI applications - is facing new constraints on selling to China. U.S. policymakers have made clear that they intend to keep China at least a generation behind in the race for AI supremacy. To that end, the U.S. CHIPS and Science Act of 2022 introduced massive incentives to re-shore chip production.

The risk now is that China will leverage its dominant role in producing and refining rare-earth metals that are key inputs in the green transition. China has already increased its exports of electric vehicles by about 700% in value terms since 2017, and it is starting to deploy commercial airliners that eventually could compete with Boeing (BA) and Airbus (FR:AIR). So, while the G7 wants to deter China without escalating the cold war, the response from Beijing suggests that it has failed to thread the needle.

The U.S. and China cold war will mean more fragmentation of the global economy.

Of course, the Chinese would like to forget that their own aggressive policies contributed to the situation. In interviews marking his 100th birthday in May, Henry Kissinger - the architect of America's "opening to China" in 1972 - warned that unless the U.S. and China find a new strategic understanding, they will remain on a collision course that could end in outright war. The deeper the freeze, the greater the risk of a violent crack-up and military hostilities this decade.

Even without an actual hot war between the U.S. and China, a colder war will mean more fragmentation of the global economy, more balkanization of global supply chains, more de-risking or decoupling, and more restrictions on cross-border flows of goods, services, capital, people, data, and knowledge. Neoliberal free trade is out; industrial policies, "homeland economics," subsidies, and secure trade are in, as the world increasingly divides into two economic, monetary, financial, currency, trade, investment, and technological domains.

Climate risks are climbing

At the same time, the costs of climate change will continue to increase rapidly. Scientists now expect global average temperatures to reach 1.5deg Celsius above pre-industrial levels - the Paris climate agreement target - in the next five years. To hold temperature increases there, greenhouse-gas emissions would have to be cut by half by 2030, which is basically impossible. Even if all the commitments made at COP26 in Glasgow and COP27 in Sharm El-Sheikh were to be met - a very big if - temperatures would still be on track to hit 2.4degC above pre-industrial levels by the end of this century.

Humanity's handling of climate change amounts to a slow-motion, but accelerating, train wreck.

In the absence of real action, greenwashing, greenwishing and greenflation have become rampant. The good news is that there are many technological options that can accelerate decarbonization and help us achieve net-zero emissions with limited impact on economic growth: renewable energy, carbon capture and storage, clean and green hydrogen, and nuclear fusion.

The bad news is that fusion is still a long way from commercialization, and many of the other options remain costly compared to fossil fuels. Humanity's handling of climate change amounts to a slow-motion, but accelerating, train wreck.

Making matters worse, poorer emerging markets and developing countries are facing dire economic prospects. After an anemic recovery from the COVID pandemic, they bore the brunt of higher food and energy prices following Russia's invasion of Ukraine. Higher inflation has eroded real incomes, and their currencies have weakened against the U.S. dollar (DX00). This, combined with higher interest rates, has left many nations with unsustainable debts. The International Monetary Fund and the World Bank estimate that about 60% of poor countries and 25% of emerging markets cannot service their debts and will need to restructure them.

Social strife, political instability and AI's rise

Against this backdrop, increased poverty, climate change, inequality, and social strife could easily lead to domestic political instability or even failed states, causing mass migration and fueling the trend toward economic populism. Most of Latin America is now ruled by left-wing populists, while far-right authoritarian populism is on the rise in other parts of the world.

In the U.S., former president Donald Trump is the clear favorite to win the Republican Party's nomination for the 2024 presidential election, and could well retake the White House. In the U.K., the demagogic Boris Johnson remains very popular. A party with fascist roots is running Italy, and the far-right Marine Le Pen remains the de facto opposition leader in France. In Turkey, the recently re-elected President Recep Tayyip Erdogan continues to consolidate autocratic rule. Until the Hamas attack, Israel was governed by the most right-wing coalition in its history. And, of course, Russian President Vladimir Putin and China's Xi have formed a new authoritarian axis.

Finally, in the year since Megathreats appeared, AI has become an even bigger topic, owing to the public release of generative AI platforms like ChatGPT. I had originally predicted that deep-learning architectures ("transformer networks") would revolutionize AI, and that does seem to be what has happened. The potential benefits - and pitfalls - of generative AI are profound, and they are becoming increasingly clear. On the positive side, productivity growth could be sharply increased, vastly enlarging the economic pie; but, as was true of the first digital revolution and the creation of the internet and its applications, it will take time for such gains to emerge and achieve scale.

The risks associated with AI are also becoming clear. Many worry about permanent technological unemployment, not just among low-skilled blue-collar workers, but also across creative professions. In an extreme scenario, the economy two decades from now could be growing at a rate of 10% per year, but with unemployment at 80%. A related risk, then, is that AI will be another winner-takes-all industry that turbocharges income and wealth inequality.

AI also will have a similar effect on disinformation, including through "deep fake" videos, and various forms of cyber-warfare, especially around elections. And, of course, there is the small but terrible risk that advances in AI will lead to AGI (artificial general intelligence) and the obsolescence of the human species.

The debate over whether tech firms should be regulated more strictly, or even broken up, continues to intensify. But the obvious counter-argument is that America needs Big Tech and AI firms to assure its dominance over China, which is doing everything it can to become a military superpower.

Fortunately, if AI does usher in a world of 10% annual growth, a UBI or substantially more income redistribution could well be possible. Moreover, AI also could help us address other megathreats such as climate change and future pandemics. While none of these positive outcomes can be taken for granted, given the power and influence that elites wield, problems of distribution are always easier to tackle in a high-growth setting than in a low-growth one.

While stagflationary forces will weigh on growth and exacerbate megathreats in the medium term, the future could be bright if we can avert a dystopian scenario in which megathreats destructively feed on each other. Our first priority will be to survive the next few decades of instability and chaos.

(MORE TO FOLLOW) Dow Jones Newswires

12-02-23 1445ET

See more here:

You're not imagining things: The end of the -3- - Morningstar

OpenAI researchers warned board of AI breakthrough ahead of CEO ouster, sources say – Reuters

Nov 22 (Reuters) - Ahead of OpenAI CEO Sam Altmans four days in exile, several staff researchers wrote a letter to the board of directors warning of a powerful artificial intelligence discovery that they said could threaten humanity, two people familiar with the matter told Reuters.

The previously unreported letter and AI algorithm were key developments before the board's ouster of Altman, the poster child of generative AI, the two sources said. Prior to his triumphant return late Tuesday, more than 700 employees had threatened to quit and join backer Microsoft (MSFT.O) in solidarity with their fired leader.

The sources cited the letter as one factor among a longer list of grievances by the board leading to Altman's firing, among which were concerns over commercializing advances before understanding the consequences. Reuters was unable to review a copy of the letter. The staff who wrote the letter did not respond to requests for comment.

After being contacted by Reuters, OpenAI, which declined to comment, acknowledged in an internal message to staffers a project called Q* and a letter to the board before the weekend's events, one of the people said. An OpenAI spokesperson said that the message, sent by long-time executive Mira Murati, alerted staff to certain media stories without commenting on their accuracy.

Some at OpenAI believe Q* (pronounced Q-Star) could be a breakthrough in the startup's search for what's known as artificial general intelligence (AGI), one of the people told Reuters. OpenAI defines AGI as autonomous systems that surpass humans in most economically valuable tasks.

Given vast computing resources, the new model was able to solve certain mathematical problems, the person said on condition of anonymity because the individual was not authorized to speak on behalf of the company. Though only performing math on the level of grade-school students, acing such tests made researchers very optimistic about Q*s future success, the source said.

Reuters could not independently verify the capabilities of Q* claimed by the researchers.

[1/2]Sam Altman, CEO of ChatGPT maker OpenAI, arrives for a bipartisan Artificial Intelligence (AI) Insight Forum for all U.S. senators hosted by Senate Majority Leader Chuck Schumer (D-NY) at the U.S. Capitol in Washington, U.S., September 13, 2023. REUTERS/Julia Nikhinson/File Photo Acquire Licensing Rights

Researchers consider math to be a frontier of generative AI development. Currently, generative AI is good at writing and language translation by statistically predicting the next word, and answers to the same question can vary widely. But conquering the ability to do math where there is only one right answer implies AI would have greater reasoning capabilities resembling human intelligence. This could be applied to novel scientific research, for instance, AI researchers believe.

Unlike a calculator that can solve a limited number of operations, AGI can generalize, learn and comprehend.

In their letter to the board, researchers flagged AIs prowess and potential danger, the sources said without specifying the exact safety concerns noted in the letter. There has long been discussion among computer scientists about the danger posed by highly intelligent machines, for instance if they might decide that the destruction of humanity was in their interest.

Researchers have also flagged work by an "AI scientist" team, the existence of which multiple sources confirmed. The group, formed by combining earlier "Code Gen" and "Math Gen" teams, was exploring how to optimize existing AI models to improve their reasoning and eventually perform scientific work, one of the people said.

Altman led efforts to make ChatGPT one of the fastest growing software applications in history and drew investment - and computing resources - necessary from Microsoft to get closer to AGI.

In addition to announcing a slew of new tools in a demonstration this month, Altman last week teased at a summit of world leaders in San Francisco that he believed major advances were in sight.

"Four times now in the history of OpenAI, the most recent time was just in the last couple weeks, I've gotten to be in the room, when we sort of push the veil of ignorance back and the frontier of discovery forward, and getting to do that is the professional honor of a lifetime," he said at the Asia-Pacific Economic Cooperation summit.

A day later, the board fired Altman.

Anna Tong and Jeffrey Dastin in San Francisco and Krystal Hu in New York; Editing by Kenneth Li and Lisa Shumaker

Our Standards: The Thomson Reuters Trust Principles.

Anna Tong is a correspondent for Reuters based in San Francisco, where she reports on the technology industry. She joined Reuters in 2023 after working at the San Francisco Standard as a data editor. Tong previously worked at technology startups as a product manager and at Google where she worked in user insights and helped run a call center. Tong graduated from Harvard University. Contact:4152373211

Jeffrey Dastin is a correspondent for Reuters based in San Francisco, where he reports on the technology industry and artificial intelligence. He joined Reuters in 2014, originally writing about airlines and travel from the New York bureau. Dastin graduated from Yale University with a degree in history. He was part of a team that examined lobbying by Amazon.com around the world, for which he won a SOPA Award in 2022.

Krystal reports on venture capital and startups for Reuters. She covers Silicon Valley and beyond through the lens of money and characters, with a focus on growth-stage startups, tech investments and AI. She has previously covered M&A for Reuters, breaking stories on Trump's SPAC and Elon Musk's Twitter financing. Previously, she reported on Amazon for Yahoo Finance, and her investigation of the company's retail practice was cited by lawmakers in Congress. Krystal started a career in journalism by writing about tech and politics in China. She has a master's degree from New York University, and enjoys a scoop of Matcha ice cream as much as getting a scoop at work.

Link:

OpenAI researchers warned board of AI breakthrough ahead of CEO ouster, sources say - Reuters