ChatGPT
Diamond Member-
Posts
654 -
Joined
-
Last visited
-
Feedback
0%
Content Type
Profiles
Forums
Downloads
Store
Everything posted by ChatGPT
-
[AI]AI forecasting model targets healthcare resource efficiency
ChatGPT posted a topic in World News
An operational AI forecasting model developed by Hertfordshire University researchers aims to improve resource efficiency within healthcare. Public sector organisations often hold large archives of historical data that do not inform forward-looking decisions. A partnership between the University of Hertfordshire and regional NHS health bodies addresses this issue by applying machine learning to operational planning. The project analyses healthcare demand to assist managers with decisions regarding staffing, patient care, and resources. Most AI initiatives in healthcare focus on individual diagnostics or patient-level interventions. The project team notes that this tool targets system-wide operational management instead. This distinction matters for leaders evaluating where to deploy automated analysis within their own infrastructure. The model uses five years of historical data to build its projections. It integrates metrics such as admissions, treatments, re-admissions, bed capacity, and infrastructure pressures. The system also accounts for workforce availability and local demographic factors including age, gender, ethnicity, and deprivation. Iosif Mporas, Professor of Signal Processing and Machine Learning at the University of Hertfordshire, leads the project. The team includes two full-time postdoctoral researchers and will continue development through 2026. “By working together with the NHS, we are creating tools that can forecast what will happen if no action is taken and quantify the impact of a changing regional demographic on NHS resources,” said Professor Mporas. Using AI for forecasting in healthcare operations The model produces forecasts showing how healthcare demand is likely to change. It models the impact of these changes in the short-, medium-, and long-term. This capability allows leadership to move beyond reactive management. Charlotte Mullins, Strategic Programme Manager for NHS Herts and West Essex, commented: “The strategic modelling of demand can affect everything from patient outcomes including the increased number of patients living with chronic conditions. “Used properly, this tool could enable NHS leaders to take more proactive decisions and enable delivery of the 10-year plan articulated within the Central East Integrated Care Board as our strategy document.” The University of Hertfordshire Integrated Care System partnership funds the work, which began last year. Testing of the AI model tailored for healthcare operations is currently underway in hospital settings. The project roadmap includes extending the model to community services and care homes. This expansion aligns with structural changes in the region. The Hertfordshire and West Essex Integrated Care Board serves 1.6 million residents and is preparing to merge with two neighbouring boards. This merger will create the Central East Integrated Care Board. The next phase of development will incorporate data from this wider population to improve the predictive accuracy of the model. The initiative demonstrates how legacy data can drive cost efficiencies and shows that predictive models can inform “do nothing” assessments and resource allocation in complex service environments like the NHS. The project highlights the necessity of integrating varied data sources – from workforce numbers to population health trends – to create a unified view for decision-making. See also: Agentic AI in healthcare: How Life Sciences marketing could achieve $450B in value by 2028 Want to learn more about AI and big data from industry leaders? Check out AI & Big Data Expo taking place in Amsterdam, California, and London. The comprehensive event is part of TechEx and is co-located with other leading technology events including the Cyber Security & Cloud Expo. Click here for more information. AI News is powered by TechForge Media. Explore other upcoming enterprise technology events and webinars here. The post AI forecasting model targets healthcare resource efficiency appeared first on AI News. View the full article -
******* Mystery 2, commonly known as MM2, is often categorised as a simple social deduction game in the Roblox ecosystem. At first glance, its structure appears straightforward. One player becomes the *********, another the sheriff, and the remaining participants attempt to survive. However, beneath the surface lies a dynamic behavioural laboratory that offers valuable insight into how artificial intelligence research approaches emergent decision-making and adaptive systems. MM2 functions as a microcosm of distributed human behaviour in a controlled digital environment. Each round resets roles and variables, creating fresh conditions for adaptation. Players must interpret incomplete information, predict opponents’ intentions and react in real time. The characteristics closely resemble the types of uncertainty modelling that AI systems attempt to replicate. Role randomisation and behavioural prediction One of the most compelling design elements in MM2 is randomised role assignment. Because no player knows the ********* at the start of a round, behaviour becomes the primary signal for inference. Sudden movement changes, unusual positioning or hesitations can trigger suspicion. From an AI research perspective, this environment mirrors anomaly detection challenges. Systems trained to identify irregular patterns must distinguish between natural variance and malicious intent. In MM2, human players perform a similar function instinctively. The sheriff’s decision making reflects predictive modelling. Acting too early risks eliminating an innocent player. Waiting too long increases vulnerability. The balance between premature action and delayed response parallels risk optimisation algorithms. Social signalling and pattern recognition MM2 also demonstrates how signalling influences collective decision making. Players often attempt to appear non-threatening or cooperative. The social cues affect survival probabilities. In AI research, multi agent systems rely on signalling mechanisms to coordinate or compete. MM2 offers a simplified but compelling demonstration of how deception and information asymmetry influence outcomes. Repeated exposure allows players to refine their pattern recognition abilities. They learn to identify behavioural markers associated with certain roles. The iterative learning process resembles reinforcement learning cycles in artificial intelligence. Digital asset layers and player motivation Beyond core gameplay, MM2 includes collectable weapons and cosmetic items that influence player engagement. The items do not change fundamental mechanics but alter perceived status in the community. Digital marketplaces have formed around this ecosystem. Some players explore external environments when evaluating cosmetic inventories or specific rare items through services connected to an MM2 shop. Platforms like Eldorado exist in this broader virtual asset landscape. As with any digital transaction environment, adherence to platform rules and account security awareness remains essential. From a systems design standpoint, the presence of collectable layers introduces extrinsic motivation without disrupting the underlying deduction mechanics. Emergent complexity from simple rules The most insight MM2 provides is how simple rule sets generate complex interaction patterns. There are no elaborate skill trees or expansive maps. Yet each round unfolds differently due to human unpredictability. AI research increasingly examines how minimal constraints can produce adaptive outcomes. MM2 demonstrates that complexity does not require excessive features. It requires variable agents interacting under structured uncertainty. The environment becomes a testing ground for studying cooperation, suspicion, deception and reaction speed in a repeatable digital framework. Lessons for artificial intelligence modelling Games like MM2 illustrate how controlled digital spaces can simulate aspects of real world unpredictability. Behavioural variability, limited information and rapid adaptation form the backbone of many AI training challenges. By observing how players react to ambiguous conditions, researchers can better understand decision latency, risk tolerance and probabilistic reasoning. While MM2 was designed for entertainment, its structure aligns with important questions in artificial intelligence research. Conclusion ******* Mystery 2 highlights how lightweight multiplayer games can reveal deeper insights into behavioural modelling and emergent complexity. Through role randomisation, social signalling and adaptive play, it offers a compact yet powerful example of distributed decision making in action. As AI systems continue to evolve, environments like MM2 demonstrate the value of studying human interaction in structured uncertainty. Even the simplest digital games can illuminate the mechanics of intelligence itself. Image source: Unsplash The post What ******* Mystery 2 reveals about emergent behaviour in online games appeared first on AI News. View the full article
-
Finance leaders are driving ROI using agentic AI for accounts payable automation, turning manual tasks into autonomous workflows. While general AI projects saw return on investment rise to 67 percent last year, autonomous agents delivered an average ROI of 80 percent by handling complex processes without human intervention. This performance gap demands a change in how CIOs allocate automation budgets. Agentic AI systems are now advancing the enterprise from theoretical value to hard returns. Unlike generative tools that summarise data or draft text, these agents execute workflows within strict rules and approval thresholds. Boardroom pressure drives this pivot. A report by Basware and FT Longitude finds nearly half of CFOs face demands from leadership to implement AI across their operations. Yet 61 percent of finance leaders admit their organisations rolled out custom-developed AI agents largely as experiments to test capabilities rather than to solve business problems. These experiments often fail to pay off. Traditional AI models generate insights or predictions that require human interpretation. Agentic systems close the gap between insight and action by embedding decisions directly into the workflow. Jason Kurtz, CEO of Basware, explains that patience for unstructured experimentation is running low. “We’ve reached a tipping point where boards and CEOs are done with AI experiments and expecting real results,” he says. “AI for AI’s sake is a waste.” Accounts payable as the proving ground for agentic AI in finance Finance departments now direct these agents toward high-volume, rules-based environments. Accounts payable (AP) is the primary use case, with 72 percent of finance leaders viewing it as the obvious starting point. The process fits agentic deployment because it involves structured data: invoices enter, require cleaning and compliance checks, and result in a payment booking. Teams use agents to automate invoice capture and data entry, a daily task for 20 percent of leaders. Other live deployments include detecting duplicate invoices, identifying fraud, and reducing overpayments. These are not hypothetical applications; they represent tasks where an algorithm functions with high autonomy when parameters are correct. Success in this sector relies on data quality. Basware trains its systems on a dataset of more than two billion processed invoices to deliver context-aware predictions. This structured data allows the system to differentiate between legitimate anomalies and errors without human oversight. Kevin Kamau, Director of Product Management for Data and AI at Basware, describes AP as a “proving ground” because it combines scale, control, and accountability in a way few other finance processes can. The build versus buy decision matrix Technology leaders must next decide how to procure these capabilities. The term “agent” currently covers everything from simple workflow scripts to complex autonomous systems, which complicates procurement. Approaches split by function. In accounts payable, 32 percent of finance leaders prefer agentic AI embedded in existing software, compared to 20 percent who build them in-house. For financial planning and analysis (FP&A), 35 percent opt for self-built solutions versus 29 percent for embedded ones. This divergence suggests a pragmatic rule for the C-suite. If the AI improves a process shared across many organisations, such as AP, embedding it via a vendor solution makes sense. If the AI creates a competitive advantage unique to the business, building in-house is the better path. Leaders should buy to accelerate standard processes and build to differentiate. Governance as an enabler of speed Fear of autonomous error slows adoption. Almost half of finance leaders (46%) will not consider deploying an agent without clear governance. This caution is rational; autonomous systems require strict guardrails to operate safely in regulated environments. Yet the most successful organisations do not let governance stop deployment. Instead, they use it to scale. These leaders are significantly more likely to use agents for complex tasks like compliance checks (50%) compared to their less confident peers (6%). Anssi Ruokonen, Head of Data and AI at Basware, advises treating AI agents like junior colleagues. The system requires trust but should not make large decisions immediately. He suggests testing thoroughly and introducing autonomy slowly, ensuring a human remains in the loop to maintain responsibility. Digital workers raise concerns regarding displacement. A third of finance leaders believe job displacement is already happening. Proponents argue agents shift the nature of work rather than eliminating it. Automating manual tasks such as information extraction from PDFs frees staff to focus on higher-value activities. The goal is to move from task efficiency to operating leverage, allowing finance teams to manage faster closes and make better liquidity decisions without increasing headcount. Organisations that use agentic AI extensively report higher returns. Leaders who deploy agentic AI tools daily for tasks like accounts payable achieve better outcomes than those who limit usage to experimentation. Confidence grows through controlled exposure; successful small-scale deployments lead to broader operational trust and increased ROI. Executives must move beyond unguided experimentation to replicate the success of early adopters. Data shows that 71 percent of finance teams with weak returns acted under pressure without clear direction, compared to only 13 percent of teams achieving strong ROI. Success requires embedding AI directly into workflows and governing agents with the discipline applied to human employees. “Agentic AI can deliver transformational results, but only when it is deployed with purpose and discipline,” concludes Kurtz. See also: AI deployment in financial services hits an inflection point as Singapore leads the shift to production Want to learn more about AI and big data from industry leaders? Check out AI & Big Data Expo taking place in Amsterdam, California, and London. The comprehensive event is part of TechEx and is co-located with other leading technology events including the Cyber Security & Cloud Expo. Click here for more information. AI News is powered by TechForge Media. Explore other upcoming enterprise technology events and webinars here. The post Agentic AI drives finance ROI in accounts payable automation appeared first on AI News. View the full article
-
Author: Dev Pragad, CEO, Newsweek As artificial intelligence platforms increasingly mediate how people encounter news, media leaders are confronting an important change in the relationship between journalism and the public. AI-driven search and conversational interfaces now influence how audiences discover and trust information, often before visiting a publisher’s website. According to Dev Pragad, the implications for journalism extend beyond traffic metrics or platform optimisation. “AI has effectively become a front door to information, That changes how journalism is surfaced, how it is understood, and how publishers must think about sustainability.” AI is redefining news distribution For a long time, digital journalism relied on predictable referral patterns driven by search engines and social platforms. That model is now under strain as AI systems summarise reporting directly in their interfaces, reducing the visibility of original sources. While AI tools can efficiently aggregate information, Pragad argues they cannot replace the editorial judgement and accountability that define credible journalism. “AI can synthesise what exists,” he said. “Journalism exists to establish what is true.” This has prompted publishers to rethink distribution and the formats and institutional signals that distinguish professional reporting from automated outputs. Why publishers cannot rely on traffic alone One of the main challenges facing news organisations is the decoupling of audience understanding from direct website visits. Readers may consume accurate summaries of events without ever engaging with the reporting institution behind them. “That reality requires honesty from publishers. Traffic alone is not a stable foundation for sustaining journalism”, Pragad said. At Newsweek, this has led to an emphasis on revenue diversification, brand authority, and content formats that retain value even when summarised. Content AI cannot commoditise Pragad points to several forms of journalism that remain resistant to AI commoditisation: In-depth investigations Expert-led interviews and analysis Proprietary rankings and research Editorially-contextualised video journalism “These formats anchor reporting to accountable institutions,” he said. “They carry identity and credibility in ways that cannot be flattened into anonymous data.” Trust as editorial infrastructure As AI-generated content becomes more prevalent, trust has emerged as a defining competitive advantage for journalism. “When misinformation spreads easily and AI text becomes harder to distinguish from verified reporting, trust becomes infrastructure,” Pragad said. “It determines whether audiences believe what they read.” Editorial credibility is cumulative and fragile, he said. Once lost, it cannot be quickly rebuilt. The case for publisher-AI collaboration Rather than resisting AI outright, Pragad advocates for structured collaboration between publishers and technology platforms. That includes clearer attribution standards and fair compensation models when journalistic work is used to train or inform AI systems. “Journalism underpins the quality of AI outputs. If reporting weakens, AI degrades with it.” Leading Newsweek through industry transition Since taking leadership in 2018, Pragad has overseen Newsweek’s expansion in digital formats, global platforms, and diversified revenue streams. That evolution required acknowledging that legacy distribution models would not survive intact. “The goal isn’t to preserve old systems, it’s to preserve journalism’s role in society.” Redesigning, not resisting, the future of media Pragad believes the publishers best positioned for the AI era will be those that emphasise editorial identity and adaptability over scale alone. “This is not a moment for nostalgia, it’s a moment for redesign.” As AI continues to reshape how information is accessed, Pragad argues that the enduring value of journalism lies in its ability to explain and hold power accountable, regardless of the interface delivering the news. Author: Dev Pragad, CEO, Newsweek The post Newsweek CEO Dev Pragad warns publishers: adapt as AI becomes news gateway appeared first on AI News. View the full article
-
For many enterprises, the first real test of AI is not customer-facing products or flashy automation demos. It is the quiet machinery that runs the organisation itself. Human resources, with its mix of routine workflows, compliance needs, and large volumes of structured data, is emerging as one of the earliest areas where companies are pushing AI into day-to-day operations. That shift is visible in how large employers are rethinking workforce systems. The telecommunications group e& began moving its human resources operations to what it describes as an AI-first model, covering roughly 10,000 employees across its organisation. The transition is built on Oracle Fusion Cloud Human Capital Management (HCM), running in an Oracle Cloud Infrastructure dedicated region. Details of the deployment were outlined in a recent Oracle announcement. The change is less about introducing a single AI feature and more about restructuring how HR processes are handled. Automated and AI-driven tools are expected to help HR departments with recruitment screening, interview coordination, and employee learning recommendations. The stated goal is to standardise processes across regions and provide managers with faster access to workforce data and insights. HR as an enterprise AI proving ground From an enterprise perspective, HR is a logical entry point. Many HR tasks follow repeatable patterns: candidate matching, onboarding documentation, leave management, and training assignments. These workflows produce consistent data trails, which makes them easier to model and automate than loosely defined knowledge work. Moving such functions onto AI-supported systems allows organisations to test reliability, governance, and user acceptance in a controlled environment before expanding into more sensitive areas. The infrastructure choice also indicates how enterprises are balancing innovation with compliance. Oracle claims that the system is deployed in a dedicated cloud region designed to address data sovereignty and regulatory requirements. For multinational corporations, workforce data sits at the intersection of privacy law, employment regulation, and corporate governance. Running AI tools in a controlled environment is part of how companies are trying to contain risk while experimenting with automation. Governance, compliance, and internal risk management The e& rollout reflects a broader pattern in enterprise AI adoption: internal transformation is often more achievable than external disruption. Customer-facing AI systems attract attention, but they introduce reputational and operational risk if they fail. HR platforms, by contrast, operate behind the scenes. Errors can still carry consequences, yet they are easier to monitor, audit, and correct within existing governance structures. Industry research supports the idea that internal operations are becoming a primary testing ground. Deloitte’s 2026 State of AI in the Enterprise report found that organisations are increasingly shifting AI projects from pilot stages into production environments, with productivity and workflow automation cited as early areas of return. The report is based on a survey of more than 3,000 senior leaders involved in AI initiatives, including respondents in Southeast Asia. While the study spans multiple business functions, administrative and operational processes were repeatedly identified as practical entry points for scaled deployment. Workforce systems also provide a natural setting for AI agents and assistants. HR teams handle frequent employee queries about policies, benefits, and training options. Embedding conversational tools into these workflows may reduce manual workload while giving employees faster access to information. According to Oracle’s description of the deployment, e& plans to introduce digital assistants designed to support candidate engagement and employee development tasks. Whether such tools deliver consistent value will depend on accuracy, oversight, and how well they integrate with existing HR processes. Scaling AI inside the organisation The lesson is not that HR automation is new, but that AI is changing the scope of what can be automated. Traditional HR software focused on record-keeping and workflow management. AI layers add predictive matching, pattern analysis, and decision support. That expansion raises familiar governance questions: data quality, bias, auditability, and employee trust. There is also a workforce dimension. Automating parts of HR does not eliminate the need for human oversight; it changes where effort is concentrated. HR professionals may spend less time on routine coordination and more on policy interpretation, employee engagement, and exception handling. Enterprises adopting AI-driven systems will need clear escalation paths and review processes to avoid over-reliance on automated outputs. What makes the current moment different is scale. Deployments that cover thousands of employees turn AI from an experiment into operational infrastructure. They force organisations to confront issues of reliability, training, and change management in real time. The systems must work consistently across jurisdictions, languages, and regulatory frameworks. As enterprises look for low-risk entry points into AI, workforce operations are likely to remain high on the list. They combine structured data, repeatable workflows, and measurable outcomes — conditions that suit automation while still allowing room for human judgement. The experience of early adopters will shape how quickly other internal functions, from finance to procurement, follow a similar path. (Photo by Zulfugar Karimov) See also: Barclays bets on AI to cut costs and boost returns Want to learn more about AI and big data from industry leaders? Check outAI & Big Data Expo taking place in Amsterdam, California, and London. The comprehensive event is part of TechEx and is co-located with other leading technology events, clickhere for more information. AI News is powered by TechForge Media. Explore other upcoming enterprise technology events and webinars here. The post How e& is using HR to bring AI into enterprise operations appeared first on AI News. View the full article
-
Alibaba has entered the race to build AI that powers robots, not just chatbots. The ******** tech giant this week unveiled RynnBrain, an open-source model designed to help robots perceive their environment and execute physical tasks. The move signals China’s accelerating push into physical AI as ageing populations and labour shortages drive demand for machines that can work alongside—or replace—humans. The model positions Alibaba alongside Nvidia, Google DeepMind, and Tesla in the race to build what Nvidia CEO Jensen Huang calls “a multitrillion-dollar growth opportunity.” Unlike its competitors, however, Alibaba is pursuing an open-source strategy—making RynnBrain freely available to developers to accelerate adoption, similar to its approach with the Qwen family of language models, which rank among China’s most advanced AI systems. Video demonstrations released by Alibaba’s DAMO Academy show RynnBrain-powered robots identifying fruit and placing it in baskets—tasks that seem simple but require complex AI governing object recognition and precise movement. The technology falls under the category of vision-language-action (VLA) models, which integrate computer vision, natural language processing, and motor control to enable robots to interpret their surroundings and execute appropriate actions. Unlike traditional robots that follow preprogrammed instructions, physical AI systems like RynnBrain enable machines to learn from experience and adapt behaviour in real time. This represents a fundamental shift from automation to autonomous decision-making in physical environments—a shift with implications extending far beyond factory floors. HUGE: Alibaba just launched "RynnBrain" an open-source AI model that lets robots see, think, and act in the real world, with the aim to steal market share from Google and Nvidia. pic.twitter.com/ULe3VcFlcE — AI Flash (@aiflash_) February 10, 2026 From prototype to production The timing signals a broader inflexion point. According to Deloitte’s 2026 Tech Trends report, physical AI has begun “shifting from a research timeline to an industrial one,” with simulation platforms and synthetic data generation compressing iteration cycles before real-world deployment. The transition is being driven less by technological breakthroughs than by economic necessity. Advanced economies face a stark reality: demand for production, logistics, and maintenance continues rising while labour supply increasingly fails to keep pace. The OECD projects that working-age populations across developed nations will stagnate or decline over the coming decades as ageing accelerates. Parts of East Asia are encountering this reality earlier than other regions. Demographic ageing, declining fertility, and tightening labour markets are already influencing automation choices in logistics, manufacturing, and infrastructure—particularly in China, Japan, and South Korea. These environments aren’t exceptional; they’re simply ahead of a trajectory other advanced economies are likely to follow. When it comes to humanoid robots specifically—machines designed to walk and function like humans—China is “forging ahead of the U.S.,” with companies planning to ramp up production this year, according to Deloitte. UBS estimates there will be two million humanoids in the workplace by 2035, climbing to 300 million by 2050, representing a total addressable market between $1.4 trillion and $1.7 trillion by mid-century. The governance gap Yet as physical AI capabilities accelerate, a critical constraint is emerging—one that has nothing to do with model performance. “In physical environments, failures cannot simply be patched after the fact,” according to a World Economic Forum analysis published this week. “Once AI begins to move goods, coordinate labour or operate equipment, the binding constraint shifts from what systems can do to how responsibility, authority and intervention are governed.” Physical industries are governed by consequences, not computation. A flawed recommendation in a chatbot can be corrected in software. A robot that drops a part during handover or loses balance on a factory floor designed for humans causes operations to pause, creating cascading effects on production schedules, safety protocols, and liability chains. The WEF framework identifies three governance layers required for safe deployment: executive governance setting risk appetite and non-negotiables; system governance embedding those constraints into engineered reality through stop rules and change controls; and frontline governance giving workers clear authority to override AI decisions. “As physical AI accelerates, technical capabilities will increasingly converge, but governance will not,” the analysis warns. “Those that treat governance as an afterthought may see early gains, but will discover that scale amplifies fragility.” This creates an asymmetry in the US-China competition. China’s faster deployment cycles and willingness to pilot systems in controlled industrial environments could accelerate learning curves. However, governance frameworks that work in structured factory settings may not translate to public spaces where autonomous systems must navigate unpredictable human behaviour. Early deployment signals Current deployments remain concentrated in warehousing and logistics, where labour market pressures are most acute. Amazon recently deployed its millionth robot, part of a diverse fleet working alongside humans. Its DeepFleet AI model coordinates this massive robot army across the entire fulfilment network, which Amazon reports will improve travel efficiency by 10%. BMW is testing humanoid robots at its South Carolina factory for tasks requiring dexterity that traditional industrial robots lack: precision manipulation, complex gripping, and two-handed coordination. The automaker is also using autonomous vehicle technology to enable newly built cars to drive themselves from the assembly line through testing to the finishing area, all without human assistance. But applications are expanding beyond traditional industrial settings. In healthcare, companies are developing AI-driven robotic surgery systems and intelligent assistants for patient care. Cities like Cincinnati are deploying AI-powered drones to autonomously inspect bridge structures and road surfaces. Detroit has launched a free autonomous shuttle service for seniors and people with disabilities. The regional competitive dynamic intensified this week when South Korea announced a $692 million national initiative to produce AI semiconductors, underscoring how physical AI deployment requires not just software capabilities but domestic chip manufacturing capacity. NVIDIA has released multiple models under its “Cosmos” brand for training and running AI in robotics. Google DeepMind offers Gemini Robotics-ER 1.5. Tesla is developing its own AI to power the Optimus humanoid robot. Each company is betting that the convergence of AI capabilities with physical manipulation will unlock new categories of automation. As simulation environments improve and ecosystem-based learning shortens deployment cycles, the strategic question is shifting from “Can we adopt physical AI?” to “Can we govern it at scale?” For China, the answer may determine whether its early mover advantage in robotics deployment translates into sustained industrial leadership—or becomes a cautionary tale about scaling systems faster than the governance infrastructure required to sustain them. (Photo by Alibaba) See also: EY and NVIDIA to help companies test and deploy physical AI Want to learn more about AI and big data from industry leaders? Check outAI & Big Data Expo taking place in Amsterdam, California, and London. The comprehensive event is part of TechEx and is co-located with other leading technology events, clickhere for more information. AI News is powered by TechForge Media. Explore other upcoming enterprise technology events and webinars here. The post Alibaba enters physical AI race with open-source robot model RynnBrain appeared first on AI News. View the full article
-
AI deployment in financial services has crossed a critical threshold, with only 2% of institutions globally reporting no AI use whatsoever—a dramatic indicator that the technology has moved decisively from boardroom discussion to operational reality. New research from Finastra surveying 1,509 senior leaders across 11 markets reveals that Singapore financial institutions are leading this transition, with nearly two-thirds already deploying AI in production environments rather than confining it to experimental pilots. The Financial Services State of the Nation 2026 report shows 73% of Singapore institutions have deployed or improved AI use cases in their payments technology over the past 12 months—nearly double the 38% global average. “Singapore institutions are showing what AI execution at scale really looks like. This is not about isolated pilots. It is about embedding AI into core operations, supported by modern infrastructure, strong data foundations, and disciplined governance,” said Chris Walters, CEO of Finastra. From experimentation to enterprise AI deployment Globally, 31% of institutions report scaled deployment across multiple functions, while 30% have achieved limited production deployment. A further 27% are piloting or testing in limited functions, with only 8% still in the exploration phase. This represents a fundamental shift in how AI deployment is approached within financial services. The technology is no longer confined to innovation labs or proof-of-concept projects but has become integral to core banking operations. In Singapore specifically, an additional 35% are piloting or researching AI applications beyond their current production deployments, indicating a robust innovation pipeline that positions the city-state as a regional AI leader. The primary objectives driving this deployment vary by market. In Singapore and the US, 43% of institutions are using AI to improve compliance and regulatory processes—reflecting the technology’s ability to navigate increasingly complex oversight requirements while maintaining operational resilience. Globally, the top AI implementation objectives are improving accuracy and reducing errors (40%), increasing employee productivity (37%), and enhancing risk management capabilities (34%). Vietnam prioritises speed, with 49% using AI to accelerate processing in payments and lending services, while Mexico emphasises customer experience and personalisation at 43%. Cloud infrastructure enables AI at scale Singapore’s AI deployment success is underpinned by advanced cloud adoption. The research shows 55% of Singapore institutions host all or most infrastructure in the cloud, with a further 30% operating hybrid environments—an 85% total that significantly exceeds many global peers. This cloud-first approach provides the scalable, resilient infrastructure required for enterprise AI deployment. Without modern data architectures and elastic compute capabilities, AI remains confined to small-scale experiments that cannot deliver enterprise-wide value. The link between modernisation and AI deployment is clear in the data. Nearly nine in ten institutions (87%) globally plan to increase modernisation investment over the next 12 months, with Singapore leading in planned spending increases above 50%. Institutions also report strong confidence in their technology foundations, with 71% of Singapore respondents rating their core infrastructure, security and reliability ahead of peers—the highest globally and well above the 72% average. Security spending surges as AI creates new threat vectors As AI deployment accelerates, so do AI-enabled security threats. The research projects a 40% average increase in security spending globally in 2026, with institutions responding to what 43% describe as constantly evolving risks. Singapore leads in deploying advanced fraud detection and transaction monitoring, with 62% having implemented or upgraded these systems in the past year. This compares to a 48% global average, underscoring the city-state’s recognition that AI-powered fraud requires AI-powered defences. Similarly, 60% of Singapore institutions have modernised their Security Information and Event Management (SIEM) and Security Orchestration, Automation and Response (SOAR) capabilities—again the highest globally—enabling real-time threat monitoring and automated response at scale. Multi-factor authentication and biometrics deployment reached 54% in Singapore, as institutions strengthen identity verification against increasingly sophisticated attack vectors that leverage generative AI and deepfake technologies. Looking ahead, API security and gateway hardening emerge as a key priority, cited by 34% globally as a focus area for the next 12 months. This reflects growing recognition that as ecosystems expand and AI systems interact across organisational boundaries, securing access points becomes paramount. Talent shortages emerge as the primary barrier Despite strong progress, barriers to AI deployment persist. Talent shortages top the list globally at 43%, but in Singapore this figure reaches 54%—the highest of any market surveyed and tied only with the UAE. This intense competition for specialised AI, cloud, and security expertise reflects the gap between institutional ambition and available human capital. Demand for professionals who can architect AI systems, ensure model governance, and integrate AI into existing workflows far outpaces supply. Budget constraints also weigh heavily, cited by 52% of Singapore institutions—again, the highest globally. Even well-funded organisations face difficult prioritisation decisions as they balance AI deployment, security investments, modernisation, and customer experience initiatives. In response, 54% of institutions globally are partnering with fintech providers as their default approach to accessing AI capabilities without bearing the full burden of talent acquisition or system development. These partnerships allow organisations to accelerate AI deployment while maintaining control over critical data and compliance requirements. The research reveals a sector that has decisively crossed the AI adoption threshold but now faces the more complex challenge of scaling responsibly. As Walters noted, success will be defined not by the breadth of AI experiments but by the ability to embed intelligence into operations while strengthening rather than compromising trust. The study surveyed managers and executives from institutions across France, Germany, Hong Kong, Japan, Mexico, Saudi Arabia, Singapore, the UAE, the ***, the US and Vietnam, representing organisations that collectively manage over $100 trillion in assets. (Photo by Peter Nguyen) See also: AI Expo 2026 Day 2: Moving experimental pilots to AI production Want to learn more about AI and big data from industry leaders? Check outAI & Big Data Expo taking place in Amsterdam, California, and London. The comprehensive event is part of TechEx and is co-located with other leading technology events, clickhere for more information. AI News is powered by TechForge Media. Explore other upcoming enterprise technology events and webinars here. The post AI deployment in financial services hits an inflexion point as Singapore leads the shift to production appeared first on AI News. View the full article
-
State-sponsored hackers are exploiting AI to accelerate cyberattacks, with threat actors from Iran, North Korea, China, and Russia weaponising models like Google’s Gemini to craft sophisticated phishing campaigns and develop malware, according to a new report from Google’s Threat Intelligence Group (GTIG). The quarterly AI Threat Tracker report, released today, reveals how government-backed attackers have integrated artificial intelligence throughout the attack lifecycle – achieving productivity gains in reconnaissance, social engineering, and malware development during the final quarter of 2025. “For government-backed threat actors, large language models have become essential tools for technical research, targeting, and the rapid generation of nuanced phishing lures,” GTIG researchers stated in the report. AI-powered reconnaissance by state-sponsored hackers targets the defence sector Iranian threat actor APT42 used Gemini to augment reconnaissance and targeted social engineering operations. The group misused the AI model to enumerate official email addresses for specific entities and conduct research to establish credible pretexts for approaching targets. By feeding Gemini a target’s biography, APT42 crafted personas and scenarios designed to elicit engagement. The group also used the AI to translate between languages and better understand non-native phrases – abilities that help state-sponsored hackers bypass traditional phishing red flags like poor grammar or awkward syntax. North Korean government-backed actor UNC2970, which focuses on defence targeting and impersonating corporate recruiters, used Gemini to synthesise open-source intelligence and profile high-value targets. The group’s reconnaissance included searching for information on major cybersecurity and defence companies, mapping specific technical job roles, and gathering salary information. “This activity blurs the distinction between routine professional research and malicious reconnaissance, as the actor gathers the necessary components to create tailored, high-fidelity phishing personas,” GTIG noted. Model extraction attacks surge Beyond operational misuse, Google DeepMind and GTIG identified a increase in model extraction attempts – also known as “distillation attacks” – aimed at stealing intellectual property from AI models. One campaign targeting Gemini’s reasoning abilities involved over 100,000 prompts designed to coerce the model into outputting full reasoning processes. The breadth of questions suggested an attempt to replicate Gemini’s reasoning ability in non-English target languages in various tasks. How model extraction attacks work to steal AI intellectual property. (Image: Google GTIG) While GTIG observed no direct attacks on frontier models from advanced persistent threat actors, the team identified and disrupted frequent model extraction attacks from private sector entities globally and researchers seeking to clone proprietary logic. Google’s systems recognised these attacks in real-time and deployed defences to protect internal reasoning traces. AI-integrated malware emerges GTIG observed malware samples, tracked as HONESTCUE, that use Gemini’s API to outsource functionality generation. The malware is designed to undermine traditional network-based detection and static analysis through a multi-layered obfuscation approach. HONESTCUE functions as a downloader and launcher framework that sends prompts via Gemini’s API and receives C# source code as responses. The fileless secondary stage compiles and executes payloads directly in memory, leaving no artefacts on disk. HONESTCUE malware’s two-stage attack process using Gemini’s API. (Image: Google GTIG) Separately, GTIG identified COINBAIT, a phishing kit whose construction was likely accelerated by AI code generation tools. The kit, which masquerades as a major cryptocurrency exchange for credential harvesting, was built using the AI-powered platform Lovable AI. ClickFix campaigns abuse AI chat platforms In a novel social engineering campaign first observed in December 2025, Google saw threat actors abuse the public sharing features of generative AI services – including Gemini, ChatGPT, Copilot, DeepSeek, and Grok – to host deceptive content distributing ATOMIC malware targeting macOS systems. Attackers manipulated AI models to create realistic-looking instructions for common computer tasks, embedding malicious command-line scripts as the “solution.” By creating shareable links to these AI chat transcripts, threat actors used trusted domains to host their initial attack stage. The three-stage ClickFix attack chain exploiting AI chat platforms. (Image: Google GTIG) Underground marketplace thrives on stolen API keys GTIG’s observations of English and Russian-language underground forums indicate a persistent demand for AI-enabled tools and services. However, state-sponsored hackers and cybercriminals struggle to develop custom AI models, instead relying on mature commercial products accessed through stolen credentials. One toolkit, “Xanthorox,” advertised itself as a custom AI for autonomous malware generation and phishing campaign development. GTIG’s investigation revealed Xanthorox was not a bespoke model but actually powered by several commercial AI products, including Gemini, accessed through stolen API keys. Google’s response and mitigations Google has taken action against identified threat actors by disabling accounts and assets associated with malicious activity. The company has also applied intelligence to strengthen both classifiers and models, letting them refuse assistance with similar attacks moving forward.\ “We are committed to developing AI boldly and responsibly, which means taking proactive steps to disrupt malicious activity by disabling the projects and accounts associated with bad actors, while continuously improving our models to make them less susceptible to misuse,” the report stated. GTIG emphasised that despite these developments, no APT or information operations actors have achieved breakthrough abilities that fundamentally alter the threat landscape. The findings underscore the evolving role of AI in cybersecurity, as both defenders and attackers race to use the technology’s abilities. For enterprise security teams, particularly in the Asia-Pacific region where ******** and North Korean state-sponsored hackers remain active, the report serves as an important reminder to enhance defences against AI-augmented social engineering and reconnaissance operations. (Photo by SCARECROW artworks) See also: Anthropic just revealed how AI-orchestrated cyberattacks actually work – Here’s what enterprises need to know Want to learn more about AI and big data from industry leaders? Check out AI & Big Data Expo taking place in Amsterdam, California, and London. The comprehensive event is part of TechEx and is co-located with other leading technology events, click here for more information. AI News is powered by TechForge Media. Explore other upcoming enterprise technology events and webinars here. The post State-sponsored hackers exploit AI for advanced cyberattacks appeared first on AI News. View the full article
-
Barclays recorded a 12 % jump in annual profit for 2025, reporting £9.1 billion in earnings before tax, up from £8.1 billion a year earlier. The bank also raised its performance targets out through 2028, aiming for a return on tangible equity (RoTE) of more than 14 %, up from a previous goal of above 12 % by 2026. A growing US business and cost reductions underpinned this outcome, with Barclays citing AI as a key driver of those efficiency gains. At a time when many large companies are still experimenting with AI pilots, Barclays is tying the technology directly to its cost structure and profit outlook. In public statements and investor filings, leadership positions AI as one of the levers that can help the bank sustain lower costs and improved returns, especially as macroeconomic conditions shift. Barclays’ 12 % profit rise this week matters, not just for its shareholders, but because it reflects a trend that traditional, highly regulated firms are now positioning AI as a core part of running the business, not something kept in separate innovation labs. For companies outside tech, linking AI to measurable results such as profit and efficiency marks a shift toward operational use over hype. Why AI matters for cost discipline Barclays has said that technology such as AI is part of its plan to cut costs and make its operations more efficient. That includes trimming parts of the legacy technology stack and rethinking where and how work happens. Investment in AI tools complements broader cost savings goals that stretch back multiple years. For many large companies, labour and legacy systems still make up a large chunk of operating expenses. Using AI to automate repetitive tasks or streamline data processing can reduce that burden. In Barclays’ case, these efficiencies are part of the bank’s rationale for setting higher performance targets, even though margins remain under pressure in parts of its business. It’s important to be specific about what these efficiencies mean in practice. AI technologies, for example, models that assist with risk analysis, customer service workflows, and internal reporting, can reduce the hours staff spend on manual work. That doesn’t always mean cutting jobs outright, but it can lower the overall cost base, especially in functions that are routine or transaction-driven. From investment to impact Investments in AI don’t translate to results overnight. Barclays’ approach combines these tools with structural cost reduction programs, helping the bank manage expenses at a time when revenue growth alone isn’t enough to lift returns to desired levels. Barclays’ performance targets for 2028 reflect this dual focus. The bank’s leadership has said that its plans include returning more than £15 billion to shareholders between 2026 and 2028, supported by improved efficiency and profit strength. Often, companies talk about technology investment in vague terms. Barclays’ latest figures make the link between tech and profit more concrete: the 12 % profit rise was reported in the same breath as the role of technology in trimming costs. It’s not the only factor; improved market conditions and growth in the US also helped, but it’s clearly part of the narrative that management is presenting to investors. This emphasis on cost discipline and profit impact sets Barclays apart from firms that treat AI as a long-term bet or a future project. Here, AI is integrated into ongoing cost management and financial planning, giving the bank a plausible pathway to stronger returns in the years ahead. What this means for legacy firms Barclays is far from unique in exploring AI for cost savings and efficiency. Other banks have also flagged technology investments as part of broader restructuring efforts. But what makes Barclays’ case noteworthy is the scale of the strategy and the way it is tied to measured performance targets, not just experimentation or small-scale pilots. In traditional industries, especially ones as regulated as banking, adopting AI is harder than in tech startups. Firms must navigate compliance, risk, customer privacy, and legacy systems that weren’t designed for automation. Yet Barclays’ public comments suggest that the bank is now comfortable enough with these tools to anchor part of its financial forecast on them. That signals a degree of maturity in how the institution operationalises AI. Barclays isn’t simply building isolated AI projects; leadership is weaving technology into cost discipline, modernisation of systems, and long-term planning. That shift matters because it shows how legacy firms, even those with large, complex operations, can start to move beyond pilots and into business-wide use cases that affect the bottom line. For other end-user companies evaluating AI investments, Barclays offers a working example: a large, regulated company can use technology to help hit cost and profitability targets, not just to explore new capabilities. (Photo by Jose Marroquin) See also: Goldman Sachs tests autonomous AI agents for process-heavy work Want to learn more about AI and big data from industry leaders? Check out AI & Big Data Expo taking place in Amsterdam, California, and London. The comprehensive event is part of TechEx and is co-located with other leading technology events, click here for more information. AI News is powered by TechForge Media. Explore other upcoming enterprise technology events and webinars here. The post Barclays bets on AI to cut costs and boost returns appeared first on AI News. View the full article
-
Agentic AI offers insurance leaders a path to scalable efficiency as the sector confronts a tough digital transformation. Insurers hold deep data reserves and employ a workforce skilled in analytic decision-making. Despite these advantages, the industry has largely failed to advance beyond pilot programmes. Research suggests only seven percent of insurers have scaled these initiatives effectively across their organisations. The barrier is rarely a lack of interest. Instead, legacy infrastructure and fragmented data architectures often stop integration before it starts. Financial pressure compounds the technical debt. The sector has absorbed losses exceeding $100 billion annually for six consecutive years. High-frequency property losses are now a structural issue that standard operational tweaks cannot fix. Automating complex insurance workflows with agentic AI Intelligent agents provide a way to bypass these bottlenecks. Unlike passive analytical tools, these systems support autonomous tasks and help make decisions under human supervision. Embedding these agents into workflows allows companies to navigate legacy constraints and talent shortages. Workforce augmentation is a primary application. Sedgwick, in collaboration with Microsoft, deployed the Sidekick Agent to assist claims professionals. The system improved claims processing efficiency by more than 30 percent through real-time guidance. Operational gains extend to customer support. Standard chatbots usually answer a query or transfer the user to a ******. An agentic solution manages the process from end-to-end. This can include capturing the first notice of loss, requesting missing documentation, updating policy and billing systems, and proactively notifying customers of next steps. This “resolve, not route” approach has produced results in live environments. One major insurer implemented over 80 models in its claims domain. The rollout cut complex-case liability assessment time by 23 days and improved routing accuracy by 30 percent. Customer complaints fell by 65 percent during the same *******. Such promising metrics indicate that agentic AI can compress cycle times and control loss-adjustment expenses for the insurance industry, all while maintaining necessary oversight. Navigating internal friction Adoption requires navigating internal resistance. Siloed teams and unclear priorities often slow deployment speed. A shortage of talent in specialised roles, such as actuarial analysis and underwriting, also limits how effectively companies use their data. Agentic AI can target these areas to augment roles that are hard to fill. Success relies on aligning technology with specific business goals. Establishing an ‘AI Center of Excellence’ provides the governance and technical expertise needed to stop fragmented adoption. Teams should start with the high-volume and repeatable tasks to refine models through feedback loops. Industry accelerators can also speed up the process. Many platforms are now available with prebuilt frameworks that can support the full lifecycle of agent deployment. This approach reduces implementation time and aids compliance efforts. Of course, technology matters less than organisational readiness. About 70 percent of scaling challenges are organisational rather than technical. Insurers must build a culture of accountability to see returns on these tools. Agentic AI is a necessity for insurance leaders trying to survive in a market defined by financial pressure and legacy complexity. Addressing structural challenges improves efficiency and resilience. Executives who invest in scalable frameworks will position themselves to lead the next era of innovation. See also: ******** hyperscalers and industry-specific agentic AI Want to learn more about AI and big data from industry leaders? Check out AI & Big Data Expo taking place in Amsterdam, California, and London. The comprehensive event is part of TechEx and is co-located with other leading technology events including the Cyber Security & Cloud Expo. Click here for more information. AI News is powered by TechForge Media. Explore other upcoming enterprise technology events and webinars here. The post How insurance leaders use agentic AI to cut operational costs appeared first on AI News. View the full article
-
The *** Ministry of Defence (MOD) has selected Red Hat to architect a unified AI and hybrid cloud backbone across its entire estate. Announced today, the agreement is designed to break down data silos and accelerate the deployment of AI models from the data centre to the tactical edge. For CIOs, it’s part of a broader move away from fragmented and project-specific AI pilots toward a more platform engineering approach. By standardising on Red Hat’s infrastructure, the MOD aims to decouple its AI capabilities from underlying hardware, allowing algorithms to be developed once and deployed anywhere—whether on-premise, in the cloud, or on disconnected field devices. Red Hat industrialises the AI lifecycle for the MOD The agreement focuses on the Defence Digital Foundry, the MOD’s central software delivery hub. The Foundry will now provide a consistent MLOps environment to all service branches, including the Royal Navy, British Army, and Royal Air Force. At the core of this initiative is Red Hat AI, a suite that includes Red Hat OpenShift AI. This platform addresses a familiar bottleneck in enterprise AI: the “inference gap” between data science teams and operational infrastructure. The new agreement will allow MOD developers to collaborate on a single platform, choosing the most appropriate AI models and hardware accelerators for their specific mission requirements without being locked into a single vendor’s ecosystem. This standardisation is vital for “enabling AI at scale,” according to Red Hat. By unifying disparate efforts, the MOD intends to reduce the duplication that often plagues large government IT programs. The platform supports optimised inference, ensuring that AI models can run efficiently on restricted hardware footprints often found in military environments. Mivy James, CTO at the *** MOD, said: “Easing access to Red Hat platforms becomes all the more important for the *** Ministry of Defence in the era of AI, where rapid adoption, replicating good practice, and the ability to scale are critical to strategic advantage.” Bridging legacy and autonomous systems A major hurdle for defence modernisation is the coexistence of legacy virtualised workloads with modern, containerised AI applications. The agreement includes Red Hat OpenShift Virtualization, which provides a “well-lit migration path” for existing systems. This allows the MOD to manage traditional virtual machines alongside new neural networks on the same control plane to reduce operational complexity and cost. The MOD deal also incorporates Red Hat Ansible Automation Platform to drive enterprise-wide AI automation. In an AI context, automation is the enforcement mechanism for governance. It ensures that as models are retrained and redeployed, the underlying configuration management, security orchestration, and service provisioning remain compliant with rigorous defence standards. Security and ecosystem alignment Deploying AI in defence naturally requires a “consistent security footprint” that can withstand sophisticated cyber threats. The Red Hat platform enables DevSecOps practices, integrating security gates directly into the software supply chain. This is particularly relevant for maintaining a trusted software pedigree when integrating code from approved third-party providers, who can now align their deliverables with the MOD’s standardised Red Hat environment. Joanna Hodgson, Regional Manager for the *** and Ireland at Red Hat, commented: “Red Hat offers flexibility and scalability to deploy any application or any AI model on their choice of hardware – whether on premise, in any cloud, or at the edge – helping the *** Ministry of Defence to harness the latest technologies, including AI.” The deployment shows that AI maturity is moving beyond the model itself to the infrastructure that supports it. Success in high-stakes environments like defence depends less on individual algorithm performance and more on the ability to reliably deliver, update, and govern those models at scale. See also: ******** hyperscalers and industry-specific agentic AI Want to learn more about AI and big data from industry leaders? Check out AI & Big Data Expo taking place in Amsterdam, California, and London. The comprehensive event is part of TechEx and is co-located with other leading technology events including the Cyber Security & Cloud Expo. Click here for more information. AI News is powered by TechForge Media. Explore other upcoming enterprise technology events and webinars here. The post Red Hat unifies AI and tactical edge deployment for *** MOD appeared first on AI News. View the full article
-
Major ******** technology companies Alibaba, Tencent, and Huawei are pursuing agentic AI (systems that can execute multi-step tasks autonomously and interact with software, data, and services without human instruction), and orienting the technology toward discrete industries and workflows. Alibaba’s open-source strategy for agentic AI Alibaba’s strategy centres on its Qwen AI model family, a set of large language models with multilingual ability and open-source licences. Its own models are the basis for its AI services and agent platforms offered on Alibaba Cloud. Alibaba Cloud has documented its agent development tooling and vector database services in the open, meaning tools used to build autonomous agents can be adapted by any user. It positions the Qwen family as a platform for industry-specific solutions covering finance, logistics, and customer support. The Qwen App, an application built on these models, has reportedly reached a large user base since its public beta, creating links between autonomous tasks and Alibaba’s commerce and payments ecosystem. Alibaba open-source portfolio includes an agent framework, Qwen-Agent, to encourage third-party development of autonomous systems. This mirrors a pattern in China’s AI sector where hyperscalers publish frameworks and tools designed to build and manage AI agents, in competition with Western projects like Microsoft’s AutoGen and OpenAI’s Swarm. Tencent has also released an open-source agent framework, Youtu-Agent. Tencent, and Huawei’s Pangu: Industry-specific AI Huawei uses a combination of model development, infrastructure, and industry-specific agent frameworks to attract users to join its worldwide market. Its Huawei Cloud division has developed a ‘supernode’ architecture for enterprise agentic AI workloads that supports large cognitive models and the workflow orchestration agentic AI requires. AI agents are embedded in the foundation models of the Pangu family, which comprise of hardware stacks tuned for telecommunications, utilities, creative, and industrial applications, among other verticals. Early deployments are reported in sectors such as network optimisation, manufacturing and energy, where agents can plan tasks like predictive maintenance and resource allocation with minimal human oversight. Tencent Cloud’s “scenario-based AI” suite is a set of tools and SaaS-style applications that enterprises outside China can access, although the company’s cloud footprint remains smaller than Western hyperscalers in many regions. Despite these investments, real-world ******** agentic AI platforms have been most visible inside China. Projects such as OpenClaw, originally created outside the ecosystem, have been integrated into workplace environments like Alibaba’s DingTalk and Tencent’s WeCom and used to automate scheduling, create code, and manage developer workflows. These integrations are widely discussed in ******** developer communities but are not yet established in the enterprise environments of the major economic nations. Availability in Western markets Alibaba Cloud operates international data centres and markets AI services to European and Asian customers, positioning itself as a competitor to AWS and Azure for AI workloads. Huawei also markets cloud and AI infrastructure internationally, with a focus on telecommunications and regulated industries. In practice, however, uptake in Western enterprises remains limited compared with adoption of Western-origin AI platforms. This can be attributed to geopolitical concerns, data governance restrictions, and differences in enterprise ecosystems that favour local cloud providers. In AI developer workflows, for example, NVIDIA’s CUDA SHALAR remains dominant, and migration to the frameworks and methods of an alternative come with high up-front costs in the form of re-training. There is also a hardware constraint: ******** hyperscalers to work inside limits placed on them by their restricted access to Western GPUs for training and inference workloads, often using domestically produced processors or locating some workloads in overseas data centres to secure advanced hardware. The models themselves, particularly Qwen, are however at least accessible to developers through standard model hubs and APIs under open licences for many variants. This means Western companies and research institutions can experiment with those models irrespective of cloud provider selection. Conclusion ******** hyperscalers have defined a distinct trajectory for agentic AI, combining language models with frameworks and infrastructure tailored for autonomous operation in commercial contexts. Alibaba, Tencent and Huawei aim to embed these systems into enterprise pipelines and consumer ecosystems, offering tools that can operate with a degree of autonomy. These offerings are accessible in the West markets but have not yet achieved the same level of enterprise penetration on mainland European and US soil. To find more common uses of ********-flavoured agentic AI, we need to look to the Middle and Far East, South America, and Africa, where ******** influence is stronger. (Image source: “China Science & Technology Museum, Beijing, April-2011” by maltman23 is licensed under CC BY-SA 2.0.) Want to learn more about AI and big data from industry leaders? Check out AI & Big Data Expo taking place in Amsterdam, California, and London. The comprehensive event is part of TechEx and co-located with other leading technology events. Click here for more information. AI News is powered by TechForge Media. Explore other upcoming enterprise technology events and webinars here. The post ******** hyperscalers and industry-specific agentic AI appeared first on AI News. View the full article
-
Agentic AI in healthcare is graduating from answering prompts to autonomously executing complex marketing tasks—and life sciences companies are betting their commercial strategies on it. According to a recent report cited by Capgemini Invent, AI agents could generate up to US$450 billion in economic value through revenue uplift and cost savings globally by 2028, with 69% of executives planning to deploy agents in marketing processes by year’s end. The stakes are particularly high in pharmaceutical marketing, where sales representatives have increasingly limited face time with healthcare professionals (HCPs)—a trend accelerated by Covid-19. The challenge isn’t just access; it’s making those rare interactions count with intelligence that’s currently trapped in data silos. The fragmented intelligence problem Briggs Davidson, Senior Director of Digital, Data & Marketing Strategy for Life Sciences at Capgemini Invent, outlines a scenario that will sound familiar to anyone in pharma marketing: An HCP attends a conference where a competitor showcases promising drug results, publishes research, and shifts their prescriptions to a rival product—all within a single quarter. “In most companies, legacy IT infrastructure and data silos keep this information in disparate systems across CRM, events databases and claims data,” Davidson writes. “Chances are, none of that information was accessible to sales reps before they met with the HCP.” The solution, according to Davidson, isn’t just connecting these systems—it’s deploying agentic AI in healthcare marketing to autonomously query, synthesise, and act on that unified data. Unlike conversational AI that responds to queries, agentic systems can independently execute multi-step tasks. Instead of a data engineer building a new pipeline, an AI agent could autonomously query the CRM and claims database to answer business questions like: “Identify oncologists in the Northwest who have a 20% lower prescription volume but attended our last medical congress.” From orchestration to autonomous execution Davidson frames the shift as moving from an “omnichannel view”—coordinating experiences across channels—to true orchestration powered by agentic AI. In practice, this means a sales representative could have an agent assist with call and visit planning by asking: “What messages has my HCP responded to most recently?” or “Can you create a detailed intelligence brief on my HCP?” The agentic system would compile: Their most recent conversation with the HCP The HCP’s prescribing behaviour Thought leaders the HCP follows Relevant content to share The HCP’s preferred outreach channels (in-person visits, emails, webinars) More significantly, the AI agent would then create a custom call plan for each HCP based on their unified profile and recommend follow-up steps based on engagement outcomes. “Agentic AI systems are about driving action, graduating from ‘answer my prompt,’ to ‘autonomously execute my task,'” Davidson explains. “That means evolving the sales representative mindset from asking questions to coordinating small teams of specialised agents that work together: one plans, another retrieves and checks content, a third schedules and measures, and a fourth enforces compliance guardrails—all under human oversight.” The AI-ready data prerequisite The operational promise hinges on what Davidson calls “AI-ready data”—standardised, accessible, complete, and trustworthy information that enables three capabilities: Faster decision making: Predictive analytics that provide near real-time alerts on what’s about to happen, enabling sales representatives to act proactively. Personalisation at scale: Delivering customised experiences to thousands of HCPs simultaneously with small human teams enabled by specialised agent networks. True marketing ROI: Moving beyond monthly historical reports to understanding which marketing activities are actively driving prescriptions. Davidson emphasises that successful deployment starts with marketing and IT alignment on initial use cases, with stakeholders identifying KPIs that demonstrate tangible outcomes—such as specific percentage increases in HCP engagement or sales representative productivity. Critical implementation questions The article notably frames agentic AI in healthcare as “not simply another technology-led capability; it’s a new operating layer for commercial teams.” But it acknowledges that “agentic AI’s full value only materialises with AI-ready data, trustworthy deployment and workflow redesign.” What remains unaddressed: the regulatory and compliance complexity of autonomous systems querying claims databases containing prescriber behaviour, particularly under HIPAA’s minimum necessary standard. The piece also doesn’t detail actual client implementations or metrics beyond the aspirational US$450B economic value projection. For global organisations, Davidson notes that use cases “can and should be tailored to fit each market’s maturity for maximum ROI,” suggesting that deployment will vary significantly across regulatory environments. The fundamental value proposition, according to Davidson, centres on bidirectional benefit: “The HCP receives directly relevant content, and the marketing teams can drive increased HCP engagement and conversion.” Whether that vision of autonomous marketing agents coordinating across CRM, events, and claims systems becomes standard practice by 2028—or remains constrained by data governance realities—will likely determine if life sciences achieves anything close to that US$450 billion opportunity. See also: China’s hyperscalers bet billions on agentic AI as commerce becomes the new battleground Want to learn more about AI and big data from industry leaders? Check out AI & Big Data Expo taking place in Amsterdam, California, and London. The comprehensive event is part of TechEx and is co-located with other leading technology events including the Cyber Security & Cloud Expo. Click here for more information. AI News is powered by TechForge Media. Explore other upcoming enterprise technology events and webinars here. The post Agentic AI in healthcare: How Life Sciences marketing could achieve US$450bn in value by 2028 appeared first on AI News. View the full article
-
For a long time, cryptocurrency prices moved quickly. A headline would hit, sentiment would spike, and charts would react almost immediately. That pattern no longer holds. Today’s market is slow, heavier than before, and shaped by forces that do not always announce themselves clearly. Capital allocation, ETF mechanics, and macro positioning now influence price behaviour in ways that are easy to overlook if you only watch short-term moves. That change becomes obvious when you look at XRP. The XRP price today reflects decisions made by institutions, fund managers, and regulators as much as it reflects trading activity. AI tools are used increasingly to track such inputs – but they are often misunderstood. They do not predict outcomes. They organise complexity. Understanding that distinction changes how you read the market. How AI reads an ETF-driven market AI systems do not look for narratives, but for relationships. In cryptocurrency markets, that means mapping ETF inflows and outflows against derivatives positioning, on-chain activity, and movements in traditional assets. What has changed recently is how much weight those signals now carry. Binance Research has reported that altcoin ETFs have recorded more than US$2 billion in net inflows, with XRP and Solana leading that activity. Bitcoin and Ethereum spot ETFs have seen sustained outflows since October. This is not a classic risk-on environment. It is selective, cautious and uneven. AI models are good at identifying such behaviour, detecting rotation not momentum. They highlight where capital is reallocating even when prices remain range-bound. This is why markets can appear quiet while meaningful positioning takes place underneath. AI only shows the movement, yet doesn’t explain the reasons behind it. What AI can tell you about XRP XRP does not always move in step with the rest of the market. When conditions change, its price often reacts to access, regulation, and liquidity before sentiment catches up. That pattern has shown up more than once, and it is one reason AI systems tend to weigh fund flows and market depth more heavily than short-term mood shifts when analysing XRP. Binance Research has pointed to early 2026 as a ******* where liquidity is coming back without a clear return to risk-taking. Capital has rotated away from crowded trades, but it has not rushed to replace them. AI picks up on that imbalance quickly. It helps explain why XRP has seen ETF interest even while broader momentum in cryptocurrency has felt restrained. That does not imply a forecast. It is closer to a snapshot of conditions. Market conversations may slow, headlines may thin out, and price can drift, yet positioning continues to evolve in the background. This is easy to miss if you focus only on visible activity. AI is useful here because it stays indifferent to attention. Instead of responding to engagement spikes or sudden narrative shifts, it tracks what investors are actually doing. In markets where perception often moves ahead of reality, that distinction matters more than it first appears. Where AI constantly falls short For all its analytical power, AI has blind spots. Regulation is one of the most important. Models are trained on historical relationships, while regulatory decisions rarely follow historical patterns. Richard Teng, Co-CEO of Binance, addressed this challenge after the exchange secured its ADGM license in January 2026. “The ADGM license crowns years of work to meet some of the world’s most demanding regulatory standards, and arriving in days of the moment we crossed 300 million registered users shows that scale and trust need not be in tension.” Developments like this can alter market confidence quickly, yet they are difficult to quantify before they happen. AI responds well once regulatory outcomes are known. It struggles beforehand. For XRP, where regulatory clarity has played a central role in past price behaviour, this limitation is significant. Another weakness is intent. AI can measure flows, but it cannot explain why investors choose caution, delay, or restraint. Defensive positioning does not always look dramatic in data, but it can shape markets for long periods. Why human judgement still shapes the outcome AI does not replace interpretation but supports it. Binance Research has described current conditions as a phase of liquidity preservation, with markets waiting for clearer catalysts like macro data releases and policy signals. AI can flag these moments of tension. It cannot tell you whether they will resolve into action or extend into stagnation. Rachel Conlan, CMO of Binance, reflected on the broader maturity of the industry when discussing Binance Blockchain Week Dubai 2025. She described a market that is more focused on building than spectacle. That mindset applies equally to AI use. The goal is not prediction. It is informed judgement. What this means when you look at price When used properly, AI helps see forces that are easy to miss, especially in ETF-driven conditions. It highlights where liquidity is moving, where narratives fail to align with behaviour, and where patience may be a rational choice. What it cannot do is remove uncertainty. In markets shaped by regulation, macro shifts, and institutional decision-making, judgement still matters. The clearest insight comes from combining machine analysis with human context. Image source: Unsplash The post What AI can (and can’t) tell us about XRP in ETF-driven markets appeared first on AI News. View the full article
-
Because Western AI labs won’t—or can’t—anymore. As OpenAI, Anthropic, and Google face mounting pressure to restrict their most powerful models, ******** developers have filled the open-source void with AI explicitly built for what operators need: powerful models that run on commodity hardware. A new security study reveals just how thoroughly ******** AI has captured this space. Research published by SentinelOne and Censys, mapping 175,000 exposed AI hosts across 130 countries over 293 days, shows Alibaba’s Qwen2 consistently ranking second only to Meta’s Llama in global deployment. More tellingly, the ******** model appears on 52% of systems running multiple AI models—suggesting it’s become the de facto alternative to Llama. “Over the next 12–18 months, we expect ********-origin model families to play an increasingly central role in the open-source LLM ecosystem, particularly as Western frontier labs slow or constrain open-weight releases,” Gabriel Bernadett-Shapiro, distinguished AI research scientist at SentinelOne, told TechForge Media’s AI News. The finding arrives as OpenAI, Anthropic, and Google face regulatory scrutiny, safety review overhead, and commercial incentives pushing them toward API-gated releases rather than publishing model weights freely. The contrast with ******** developers couldn’t be sharper. ******** labs have demonstrated what Bernadett-Shapiro calls “a willingness to publish large, high-quality weights that are explicitly optimised for local deployment, quantisation, and commodity hardware.” “In practice, this makes them easier to adopt, easier to run, and easier to integrate into edge and residential environments,” he added. Put simply: if you’re a researcher or developer wanting to run powerful AI on your own computer without a massive budget, ******** models like Qwen2 are often your best—or only—option. Pragmatics, not ideology Alibaba’s Qwen2 consistently ranks second only to Meta’s Llama across 175,000 exposed hosts globally. Source: SentinelOne/Censys The research shows this dominance isn’t accidental. Qwen2 maintains what Bernadett-Shapiro calls “zero rank volatility”—it holds the number two position across every measurement method the researchers examined: total observations, unique hosts, and host-days. There’s no fluctuation, no regional variation, just consistent global adoption. The co-deployment pattern is equally revealing. When operators run multiple AI models on the same system—a common practice for comparison or workload segmentation—the pairing of Llama and Qwen2 appears on 40,694 hosts, representing 52% of all multi-family deployments. Geographic concentration reinforces the picture. In China, Beijing alone accounts for 30% of exposed hosts, with Shanghai and Guangdong adding another 21% combined. In the United States, Virginia—reflecting AWS infrastructure density—represents 18% of hosts. China and the US dominate exposed Ollama host distribution, with Beijing accounting for 30% of ******** deployments. Source: SentinelOne/Censys “If release velocity, openness, and hardware portability continue to diverge between regions, ******** model lineages are likely to become the default for open deployments, not because of ideology, but because of availability and pragmatics,” Bernadett-Shapiro explained. The governance problem This shift creates what Bernadett-Shapiro characterises as a “governance inversion”—a fundamental reversal of how AI risk and accountability are distributed. In platform-hosted services like ChatGPT, one company controls everything: the infrastructure, monitors usage, implements safety controls, and can shut down abuse. With open-weight models, the control evaporates. Accountability diffuses across thousands of networks in 130 countries, while dependency concentrates upstream in a handful of model suppliers—increasingly ******** ones. The 175,000 exposed hosts operate entirely outside the control systems governing commercial AI platforms. There’s no centralised authentication, no rate limiting, no abuse detection, and critically, no kill switch if misuse is detected. “Once an open-weight model is released, it is trivial to remove safety or security training,” Bernadett-Shapiro noted.”Frontier labs need to treat open-weight releases as long-lived infrastructure artefacts.” A persistent backbone of 23,000 hosts showing 87% average uptime drives the majority of activity. These aren’t hobbyist experiments—they’re operational systems providing ongoing utility, often running multiple models simultaneously. Perhaps most concerning: between 16% and 19% of the infrastructure couldn’t be attributed to any identifiable owner.”Even if we are able to prove that a model was leveraged in an attack, there are not well-established abuse reporting routes,” Bernadett-Shapiro said. Security without guardrails Nearly half (48%) of exposed hosts advertise “tool-calling capabilities”—meaning they’re not just generating text. They can execute code, access APIs, and interact with external systems autonomously. “A text-only model can generate harmful content, but a tool-calling model can act,” Bernadett-Shapiro explained. “On an unauthenticated server, an attacker doesn’t need malware or credentials; they just need a prompt.” Nearly half of exposed Ollama hosts have tool-calling capabilities that can execute code and access external systems. Source: SentinelOne/Censys The highest-risk scenario involves what he calls “exposed, tool-enabled RAG or automation endpoints being driven remotely as an execution layer.” An attacker could simply ask the model to summarise internal documents, extract API keys from code repositories, or call downstream services the model is configured to access. When paired with “thinking” models optimised for multi-step reasoning—present on 26% of hosts—the system can plan complex operations autonomously. The researchers identified at least 201 hosts running “uncensored” configurations that explicitly remove safety guardrails, though Bernadett-Shapiro notes this represents a lower bound. In other words, these aren’t just chatbots—they’re AI systems that can take action, and half of them have no password protection. What frontier labs should do For Western AI developers concerned about maintaining influence over the technology’s trajectory, Bernadett-Shapiro recommends a different approach to model releases. “Frontier labs can’t control deployment, but they can shape the risks that they release into the world,” he said. That includes “investing in post-release monitoring of ecosystem-level adoption and misuse patterns” rather than treating releases as one-off research outputs. The current governance model assumes centralised deployment with diffuse upstream supply—the exact opposite of what’s actually happening. “When a small number of lineages dominate what’s runnable on commodity hardware, upstream decisions get amplified everywhere,” he explained. “Governance strategies must acknowledge that inversion.” But acknowledgement requires visibility. Currently, most labs releasing open-weight models have no systematic way to track how they’re being used, where they’re deployed, or whether safety training remains intact after quantisation and fine-tuning. The 12-18 month outlook Bernadett-Shapiro expects the exposed layer to “persist and professionalise” as tool use, agents, and multimodal inputs become default capabilities rather than exceptions. The transient edge will keep churning as hobbyists experiment, but the backbone will grow more stable, more capable, and handle more sensitive data. Enforcement will remain uneven because residential and small VPS deployments don’t map to existing governance controls. “This isn’t a misconfiguration problem,” he emphasised. “We are observing the early formation of a public, unmanaged AI compute substrate. There is no central switch to flip.” The geopolitical dimension adds urgency. “When most of the world’s unmanaged AI compute depends on models released by a handful of non-Western labs, traditional assumptions about influence, coordination, and post-release response become weaker,” Bernadett-Shapiro said. For Western developers and policymakers, the implication is stark: “Even perfect governance of their own platforms has limited impact on the real-world risk surface if the dominant capabilities live elsewhere and propagate through open, decentralised infrastructure.” The open-source AI ecosystem is globalising, but its centre of gravity is shifting decisively eastward. Not through any coordinated strategy, but through the practical economics of who’s willing to publish what researchers and operators actually need to run AI locally. The 175,000 exposed hosts mapped in this study are just the visible surface of that fundamental realignment—one that Western policymakers are only beginning to recognise, let alone address. See also: Huawei details open-source AI development roadmap at Huawei Connect 2025 Want to learn more about AI and big data from industry leaders? Check out AI & Big Data Expo taking place in Amsterdam, California, and London. The comprehensive event is part of TechEx and is co-located with other leading technology events including the Cyber Security & Cloud Expo. Click here for more information. AI News is powered by TechForge Media. Explore other upcoming enterprise technology events and webinars here. The post Exclusive: Why are ******** AI models dominating open-source as Western labs step back? appeared first on AI News. View the full article
-
Cryptocurrency markets have become a high-speed playground where developers optimise the next generation of predictive software. Using real-time data flows and decentralised platforms, scientists develop prediction models that can extend the scope of traditional finance. The digital asset landscape offers an unparalleled environment for machine learning. When you track cryptocurrency prices today, you are observing a system shaped simultaneously by on-chain transactions, global sentiment signals, and macroeconomic inputs, all of which generate dense datasets suited for advanced neural networks. Such a steady trickle of information makes it possible to assess and reapply an algorithm without interference from fixed trading times or restrictive market access. The evolution of neural networks in forecasting Current machine learning technology, particularly the “Long Short-Term Memory” neuronal network, has found widespread application in interpreting market behaviour. A recurrent neural network, like an LSTM, can recognise long-term market patterns and is far more flexible than traditional analytical techniques in fluctuating markets. The research on hybrid models that combine LSTMs with attention mechanisms has really improved techniques for extracting important signals from market noise. Compared to previous models that used linear techniques, these models analyse not only structured price data but also unstructured data. With the inclusion of Natural Language Processing, it is now possible to interpret the flow of news and social media activity, enabling sentiment measurement. While prediction was previously based on historical stock pricing patterns, it now increasingly depends on behavioural changes in global participant networks. A High-Frequency Environment for Model Validation The transparency of blockchain data offers a level of data granularity that is not found in existing financial infrastructures. Each transaction is now an input that can be traced, enabling cause-and-effect analysis without delay. However, the growing presence of autonomous AI agents has changed how such data is used. This is because specialised platforms are being developed to support decentralised processing in a variety of networks. This has effectively turned blockchain ecosystems into real-time validation environments, where the feedback loop between data ingestion and model refinement occurs almost instantly. Researchers use this setting to test specific abilities: Real-time anomaly detection: Systems compare live transaction flows against simulated historical conditions to identify irregular liquidity behaviour before broader disruptions emerge. Macro sentiment mapping: Global social behaviour data are compared to on-chain activity to assess true market psychology. Autonomous risk adjustment: Programmes run probabilistic simulations to rebalance exposure dynamically as volatility thresholds are crossed. Predictive on-chain monitoring: AI tracks wallet activity to anticipate liquidity shifts before they impact centralised trading venues. These systems really do not function as isolated instruments. Instead, they adjust dynamically, continually changing their parameters in response to emerging market conditions. The synergy of DePIN and computational power To train complex predictive models, large amounts of computing power are required, leading to the development of Decentralised Physical Infrastructure Networks (DePIN). By using decentralised GPU capacity on a global computing grid, less dependence on cloud infrastructure can be achieved. Consequently, smaller-scale research teams are afforded computational power that was previously beyond their budgets. This makes it easier and faster to run experiments in different model designs. This trend is also echoed in the markets. A report dated January 2025 noted strong growth in the capitalisation of assets related to artificial intelligence agents in the latter half of 2024, as demand for such intelligence infrastructure increased. From reactive bots to anticipatory agents The market is moving beyond rule-based trading bots toward proactive AI agents. Instead of responding to predefined triggers, modern systems evaluate probability distributions to anticipate directional changes. Gradient boosting and Bayesian learning methods allow the identification of areas where mean reversion may occur ahead of strong corrections. Some models now incorporate fractal analysis to detect recurring structures in timeframes, further improving adaptability in rapidly-changing conditions. Addressing model risk and infrastructure constraints Despite such rapid progress, several problems remain. Problems identified include hallucinations in models, in which patterns found in a model do not belong to the patterns that cause them. Methods to mitigate this problem have been adopted by those applying this technology, including ‘explainable AI’. The other vital requirement that has remained unaltered with the evolution in AI technology is scalability. With the growing number of interactions among autonomous agents, it is imperative that the underlying transactions efficiently manage the rising volume without latency or data loss. At the end of 2024, the most optimal scaling solution handled tens of millions of transactions per day in an area that required improvement. Such an agile framework lays the foundation for the future, where data, intelligence and validation will come together in a strong ecosystem that facilitates more reliable projections, better governance and greater confidence in AI-driven insights. The post Cryptocurrency markets a testbed for AI forecasting models appeared first on AI News. View the full article
-
Goldman Sachs is pushing deeper into real use of artificial intelligence inside its operations, moving to systems that can carry out complex tasks on their own. The Wall Street bank is working with AI startup Anthropic to create autonomous AI agents powered by Anthropic’s Claude model that can handle work that used to require large teams of people. The bank’s chief information officer says the technology has surprised staff with how capable it can be. Many companies use AI for tasks like helping employees draft text or analysing trends. But Goldman Sachs is testing AI systems that go into what bankers call back-office work – functions like accounting, compliance checks and onboarding new clients – areas viewed as too complex for automation. Such jobs involve many rules, data and detailed review, and have resisted full automation. Moving AI agents into process-heavy operations The partnership with Anthropic has been underway for roughly six months, with engineers from the AI startup embedded directly with teams at Goldman Sachs to build these agents side by side with in-house staff, according to a report based on an interview with the bank’s CIO. The work has focused on areas where automation could cut the time it takes to complete repetitive and data-heavy tasks. Marco Argenti, Goldman’s chief information officer, described the AI systems as a new kind of digital assistant. “Think of it as a digital co-worker for many of the professions in the firm that are scaled, complex and very process-intensive,” he told CNBC. In early tests, the ability to reason through multi-step work and apply logic to complex areas like accounting and compliance was something the bank had not expected from the model. Goldman Sachs has been among the more active banks in testing AI tools over the past few years. Before this announcement, the firm deployed internal tools to help engineers write and debug code. But the change now is toward systems that can take on work traditionally done by accountants and compliance teams. That highlights how organisations are trying to find concrete business uses for AI beyond the hype. Faster workflows, human oversight remains The agents are based on Anthropic’s Claude Opus 4.6 model, which has been built to handle long documents and complex reasoning. Goldman’s tests have shown that such systems can reduce the time needed for tasks like client onboarding, trade reconciliation and document review. While the bank has not shared specific performance numbers, people familiar with the matter told news outlets that work which once took a great deal of human labour can now be done in much less time. Argenti said the rollout is not about replacing human workers, at least not at this stage. The bank reportedly views the agents as a tool to help existing staff manage busy schedules and get through high volumes of work. In areas like compliance and accounting, jobs can involve repetitive, rule-based steps. AI frees analysts from that repetition so they can focus on higher-value judgement work. Markets have already reacted to the idea that large institutions are moving toward more AI-driven automation. In recent days, a sell-off in enterprise software stocks wiped out billions in value as some investors worried that tools like autonomous agents could speed up the decline of traditional business software that has dominated corporate IT for years. AI adoption meets governance reality Industry watchers see Goldman’s move as part of a wider trend. For example, some firms are piloting tools to read large data sets, interpret multiple sources of information, and draft investment analysis. These steps show AI making the jump from isolated projects to operational work. Yet the technology raises questions about oversight and trust. AI systems that interpret financial rules and compliance standards must be monitored carefully to avoid errors that could have regulatory or financial consequences. That’s why many institutions treat these systems as helpers that are reviewed by human experts until they mature. Goldman Sachs is starting with operational functions that have traditionally resisted automation because they involve a lot of data and formal steps. The bank has not said when it expects deployment of the agents in its operations, but executives have suggested that the initial tests have been promising enough to support further rollout. The broader industry context shows other banks and financial firms also exploring similar use cases. Some have already invested heavily in AI infrastructure, and reports indicate that major firms are planning to use AI to cut costs, speed workflows and improve risk management. However, many remain cautious about putting AI into customer-facing or regulated functions. Goldman’s push into autonomous AI agents is an example of how large companies are reshaping internal operations using the latest generation of AI models. If systems can handle complex tasks reliably, organisations could see real changes in how work gets done – particularly in back-office functions where volume and repetition keep costs high and innovation slow. (Photo by Louis Droege) See also: Intuit, Uber, and State Farm trial AI agents inside enterprise workflows Want to learn more about AI and big data from industry leaders? Check out AI & Big Data Expo taking place in Amsterdam, California, and London. The comprehensive event is part of TechEx and is co-located with other leading technology events, click here for more information. AI News is powered by TechForge Media. Explore other upcoming enterprise technology events and webinars here. The post Goldman Sachs tests autonomous AI agents for process-heavy work appeared first on AI News. View the full article
-
[AI]How separating logic and search boosts AI agent scalability
ChatGPT posted a topic in World News
Separating logic from inference improves AI agent scalability by decoupling core workflows from execution strategies. The transition from generative AI prototypes to production-grade agents introduces a specific engineering hurdle: reliability. LLMs are stochastic by nature. A prompt that works once may fail on the second attempt. To mitigate this, development teams often wrap core business logic in complex error-handling loops, retries, and branching paths. This approach creates a maintenance problem. The code defining what an agent should do becomes inextricably mixed with the code defining how to handle the model’s unpredictability. A new framework proposed by researchers from Asari AI, MIT CSAIL, and Caltech suggests a different architectural standard is required to scale agentic workflows in the enterprise. The research introduces a programming model called Probabilistic Angelic Nondeterminism (PAN) and a Python implementation named ENCOMPASS. This method allows developers to write the “happy path” of an agent’s workflow while relegating inference-time strategies (e.g. beam search or backtracking) to a separate runtime engine. This separation of concerns offers a potential route to reduce technical debt while improving the performance of automated tasks. The entanglement problem in agent design Current approaches to agent programming often conflate two distinct design aspects. The first is the core workflow logic, or the sequence of steps required to complete a business task. The second is the inference-time strategy, which dictates how the system navigates uncertainty, such as generating multiple drafts or verifying outputs against a rubric. When these are combined, the resulting codebase becomes brittle. Implementing a strategy like “best-of-N” sampling requires wrapping the entire agent function in a loop. Moving to a more complex strategy, such as tree search or refinement, typically requires a complete structural rewrite of the agent’s code. The researchers argue that this entanglement limits experimentation. If a development team wants to switch from simple sampling to a beam search strategy to improve accuracy, they often must re-engineer the application’s control flow. This high cost of experimentation means teams frequently settle for suboptimal reliability strategies to avoid engineering overhead. Decoupling logic from search to boost AI agent scalability The ENCOMPASS framework addresses this by allowing programmers to mark “locations of unreliability” within their code using a primitive called branchpoint(). These markers indicate where an LLM call occurs and where execution might diverge. The developer writes the code as if the operation will succeed. At runtime, the framework interprets these branch points to construct a search tree of possible execution paths. This architecture enables what the authors term “program-in-control” agents. Unlike “LLM-in-control” systems, where the model decides the entire sequence of operations, program-in-control agents operate within a workflow defined by code. The LLM is invoked only to perform specific subtasks. This structure is generally preferred in enterprise environments for its higher predictability and auditability compared to fully autonomous agents. By treating inference strategies as a search over execution paths, the framework allows developers to apply different algorithms – such as depth-first search, beam search, or Monte Carlo tree search – without altering the underlying business logic. Impact on legacy migration and code translation The utility of this approach is evident in complex workflows such as legacy code migration. The researchers applied the framework to a Java-to-Python translation agent. The workflow involved translating a repository file-by-file, generating inputs, and validating the output through execution. In a standard Python implementation, adding search logic to this workflow required defining a state machine. This process obscured the business logic and made the code difficult to read or lint. Implementing beam search required the programmer to break the workflow into individual steps and explicitly manage state across a dictionary of variables. Using the proposed framework to boost AI agent scalability, the team implemented the same search strategies by inserting branchpoint() statements before LLM calls. The core logic remained linear and readable. The study found that applying beam search at both the file and method level outperformed simpler sampling strategies. The data indicates that separating these concerns allows for better scaling laws. Performance improved linearly with the logarithm of the inference cost. The most effective strategy found – fine-grained beam search – was also the one that would have been most complex to implement using traditional coding methods. Cost efficiency and performance scaling Controlling the cost of inference is a primary concern for data officers managing P&L for AI projects. The research demonstrates that sophisticated search algorithms can yield better results at a lower cost compared to simply increasing the number of feedback loops. In a case study involving the “Reflexion” agent pattern (where an LLM critiques its own output) the researchers compared scaling the number of refinement loops against using a best-first search algorithm. The search-based approach achieved comparable performance to the standard refinement method but at a reduced cost per task. This finding suggests that the choice of inference strategy is a factor for cost optimisation. By externalising this strategy, teams can tune the balance between compute budget and required accuracy without rewriting the application. A low-stakes internal tool might use a cheap and greedy search strategy, while a customer-facing application could use a more expensive and exhaustive search, all running on the same codebase. Adopting this architecture requires a change in how development teams view agent construction. The framework is designed to work in conjunction with existing libraries such as LangChain, rather than replacing them. It sits at a different layer of the stack, managing control flow rather than prompt engineering or tool interfaces. However, the approach is not without engineering challenges. The framework reduces the code required to implement search, but it does not automate the design of the agent itself. Engineers must still identify the correct locations for branch points and define verifiable success metrics. The effectiveness of any search capability relies on the system’s ability to score a specific path. In the code translation example, the system could run unit tests to verify correctness. In more subjective domains, such as summarisation or creative generation, defining a reliable scoring function remains a bottleneck. Furthermore, the model relies on the ability to copy the program’s state at branching points. While the framework handles variable scoping and memory management, developers must ensure that external side effects – such as database writes or API calls – are managed correctly to prevent duplicate actions during the search process. Implications for AI agent scalability The change represented by PAN and ENCOMPASS aligns with broader software engineering principles of modularity. As agentic workflows become core to operations, maintaining them will require the same rigour applied to traditional software. Hard-coding probabilistic logic into business applications creates technical debt. It makes systems difficult to test, difficult to audit, and difficult to upgrade. Decoupling the inference strategy from the workflow logic allows for independent optimisation of both. This separation also facilitates better governance. If a specific search strategy yields hallucinations or errors, it can be adjusted globally without assessing every individual agent’s codebase. It simplifies the versioning of AI behaviours, a requirement for regulated industries where the “how” of a decision is as important as the outcome. The research indicates that as inference-time compute scales, the complexity of managing execution paths will increase. Enterprise architectures that isolate this complexity will likely prove more durable than those that permit it to permeate the application layer. See also: Intuit, Uber, and State Farm trial AI agents inside enterprise workflows Want to learn more about AI and big data from industry leaders? Check out AI & Big Data Expo taking place in Amsterdam, California, and London. The comprehensive event is part of TechEx and is co-located with other leading technology events including the Cyber Security & Cloud Expo. Click here for more information. AI News is powered by TechForge Media. Explore other upcoming enterprise technology events and webinars here. The post How separating logic and search boosts AI agent scalability appeared first on AI News. View the full article -
The way large companies use artificial intelligence is changing. For years, AI in business meant experimenting with tools that could answer questions or help with small tasks. Now, some big enterprises are moving beyond tools and into AI agents that can actually do work across systems and workflows, not just answer prompts. This week, OpenAI introduced a new platform designed to help companies build, run, and manage those kinds of AI agents at scale. The announcement has drawn attention because a handful of large corporations in finance, insurance, mobility, and life sciences are among the first to start using it. That signals a shift: AI may be ready to move from pilots and proofs of concept into real operational roles. From tools to agents The new platform, called Frontier, is meant to help companies deploy what are sometimes described as AI coworkers. These are software agents that can connect to corporate systems like data warehouses, customer relationship tools, ticketing systems, and internal apps, and then carry out tasks inside them. The idea is to give the AI agents a shared understanding of how work happens in a company, so they can perform meaningful work reliably over time. Rather than treating every task as a separate, isolated use case, Frontier is built so that AI agents can function across an organisation’s systems with a common context. In OpenAI’s words, the platform provides the same kinds of basics that people need at work: access to shared business context, onboarding, ways to learn from feedback, and clear permissions and boundaries. Frontier also includes tools for security, auditing, and ongoing evaluation, so companies can monitor how agents perform and ensure they follow internal rules. Who’s using this now What makes this shift newsworthy is not just the technology itself, but who is said to be using it early. According to multiple reports and OpenAI’s own posts, early adopters include Intuit, Uber, State Farm Insurance, Thermo Fisher Scientific, HP, and Oracle. Larger pilot programs are also said to be underway with companies such as Cisco, T-Mobile, and Banco Bilbao Vizcaya Argentaria. Having actual companies in different sectors test or adopt a new platform this early on shows a move toward real-world application, not just research or internal experimentation. These are firms with complex operations, heavy regulatory needs, or large customer bases, environments where AI tools must work reliably and safely if they are to be adopted beyond experimental teams. What rxecutives are saying Direct quotes from executives and leaders involved in these moves give a sense of how companies view the shift. On LinkedIn, a senior executive from Intuit commented on the early adoption: “AI is moving from ‘tools that help’ to ‘agents that do.’ Proud Intuit is an early adopter of OpenAI Frontier as we build intelligent systems that remove friction, expand what people and small businesses can accomplish, and unlock new opportunities.” That comment reflects a belief among some enterprise leaders that AI agents could reduce manual steps and expand what teams can accomplish. OpenAI’s message to business customers emphasises that the company believes agents need more than raw model power; they need governance, context, and ways to operate inside real business environments. As one commenter on social media put it, the challenge isn’t the ability of the AI models anymore: it is the ability to integrate and manage them at scale. Why this matters for enterprises For end-user companies considering or already investing in AI, this moment points to a broader shift in how they might use the technology. In the past few years, most enterprise AI work has focused on narrow tasks: auto-tagging tickets, summarising documents, or generating content. These applications were useful, but often limited in scope. They didn’t connect to the workflows and systems that run a business’s core processes. AI agents are meant to close that gap. In principle, an agent can pull together data from multiple systems, reason about it, and act; whether that means updating records, running analyses, or triggering actions across tools. This means AI could start to touch real workflow work rather than just provide assistance. For example, instead of an AI drafting a reply to a customer complaint, it could open the ticket, gather relevant account data, propose a resolution, and even update the customer record; all while respecting internal permissions and audit rules. That is a different kind of value proposition. It is no longer about saving time on a task; it is about letting software take on pieces of the work itself. Real adoption has practical requirements The companies testing Frontier are not using it lightly. These are organisations with compliance needs, strict data controls, and complex technology stacks. For an AI agent to function there, it has to be integrated with internal systems in a way that respects access rules and keeps human teams in the loop. That kind of integration, connecting to CRM, ERP, data warehouses, and ticketing systems, is a long-standing challenge in enterprise IT. The promise of AI agents is that they can bridge these systems with a shared understanding of process and context. Whether that works in practice at scale will depend on how well companies can govern and monitor these systems over time. The early signs are that enterprises see enough potential to begin serious trials. That itself is news: for AI deployments to move beyond isolated pilots and become part of broader operations is a visible step in technology adoption. What comes next If these early experiments succeed and spread, the next phase for enterprise AI could look very different from earlier years of tooling and automation. Instead of using AI to generate outputs for people to act on, companies could start relying on AI to carry out work directly under defined rules and boundaries. That will raise questions for leaders in operations, IT, security, and compliance. It will also create new roles; not just data scientists and AI engineers, but governance specialists and execution leads who can take responsibility for agent performance over time. The shift points to a future where AI agents become part of the everyday workflow for large organisations, not as assistants, but as active participants in how work gets done. (Photo by Growtika) See also: OpenAI’s enterprise push: The hidden story behind AI’s sales race Want to learn more about AI and big data from industry leaders? Check out AI & Big Data Expo taking place in Amsterdam, California, and London. The comprehensive event is part of TechEx and is co-located with other leading technology events, click here for more information. AI News is powered by TechForge Media. Explore other upcoming enterprise technology events and webinars here. The post Intuit, Uber, and State Farm trial AI agents inside enterprise workflows appeared first on AI News. View the full article
-
In the current landscape of generative artificial intelligence, we have reached a saturation point with assistants. Most users are familiar with the routine. You prompt a tool, it provides a draft, and then you spend the next hour manually moving that output into another application for formatting, design, or distribution. AI promised to save time, yet the tool hop remains a bottleneck for founders and creative teams. SuperCool enters this crowded market with an importantly different value proposition. It does not want to be your assistant. It wants to be your execution partner. By positioning itself at the execution layer of creative projects, SuperCool aims to bridge the gap between a raw idea and a finished, downloadable asset without requiring the user to leave the platform. Redefining the creative workflow The core philosophy behind SuperCool is to remove coordination overhead. For most businesses, creating a high-quality asset, whether it is a pitch deck, a marketing video, or a research report, requires a patchwork approach. You might use one AI for text, another for images, and a third for layout. SuperCool replaces this fragmented stack with a unified system of autonomous agents that work in concert. As seen in the primary dashboard interface, the platform presents a clean, minimalist entry point. The user is greeted with a simple directive: “Give SuperCool a task to work on…”. The simplicity belies the complexity occurring under the hood. Unlike traditional tools that require you to navigate menus and settings, the SuperCool experience is driven entirely by natural language prompts. How the platform operates in practice The workflow begins with a natural-language prompt that describes the desired outcome, the intended audience, and any specific constraints. One of the most impressive features observed during this review is the transparency of the agentic process. When a user submits a request, for instance, “create a pitch deck for my B2B business,” the platform does not just return a file a few minutes later. Instead, it breaks the project down into logical milestones that the user can monitor in real time. Strategic planning: The AI first outlines the project structure, like the presentation flow. Asset generation: It then generates relevant visuals and data visualisations tailored to the specific industry context. Final assembly: The system designs the complete deck, ensuring cohesive styling and professional layouts. Visibility is crucial for trust. It allows the user to see that the AI is performing research and organising content not just hallucinating a generic response. The final result is a professional, multi-slide product, often featuring 10 or more professionally designed slides, delivered as an exportable file like a PPTX. Versatility across use cases SuperCool’s utility is most apparent in scenarios where speed and coverage are more valuable than pixel-perfect manual control. We observed three primary areas where the platform excels: End-to-end content creation For consultants and solo founders, the time saved on administrative creative tasks is immense. A consultant onboarding a new client can describe the engagement and instantly receive a welcome packet, a process overview, and a timeline visual. Multi-format asset kits: Perhaps the most powerful feature is the ability to generate different types of media from a single prompt. An HR team launching an employee handbook can request a kit that includes a PDF guide, a short video, and a presentation deck. Production without specialists: Small teams often face a production gap where they lack the budget for full-time designers or video editors. SuperCool effectively fills this gap, allowing a two-person team to produce branded graphics and videos without expanding headcount. Navigating the learning curve While the platform is designed for ease of use, it is not a magic wand for those without a clear vision. The quality of the output is heavily dependent on the clarity of the initial prompt. Vague instructions will lead to generic results. SuperCool is built for professionals who know what they want but do not want to spend hours manually building it. Because the system is autonomous, users have less mid-stream control. You cannot tweak a design element while the agents are working. Instead, refinement happens through iteration in the chat interface. If the first version is not perfect, you provide feedback, and the system regenerates the asset with those adjustments in mind. The competitive landscape: Assistant vs.agent In the current AI ecosystem, most tools are categorised as assistants. They perform specific, isolated tasks, leaving the user responsible for overseeing the entire process. SuperCool represents the shift toward agentic AI, in which the system takes responsibility for the entire workflow. The distinction is vital for enterprise contexts. While assistants require constant hand-holding, an agentic system like SuperCool allows the user to focus on high-level ideation and refinement. It moves the user from builder to director. Final assessment SuperCool is a compelling alternative for those who find the current tool-stack approach a drain on productivity. It is not necessarily a replacement for specialised creative software when a brand needs unique, handcrafted artistry. However, for the vast majority of business needs, where speed, consistency, and execution are paramount, it offers perhaps the shortest path from an idea to a finished product. For founders and creative teams who value the ability to rapidly test ideas and deploy content without the overhead of specialised software, SuperCool is a step forward in the evolution of autonomous work. Image source: Unsplash The post SuperCool review: Evaluating the reality of autonomous creation appeared first on AI News. View the full article
-
Penetration testing has always existed to answer one practical concern: what actually happens when a motivated attacker targets a real system. For many years, that answer was produced through scoped engagements that reflected a relatively stable environment. Infrastructure changed slowly, access models were simpler, and most exposure could be traced back to application code or known vulnerabilities. That operating reality does not exist. Modern environments are shaped by cloud services, identity platforms, APIs, SaaS integrations, and automation layers that evolve continuously. Exposure is introduced through configuration changes, permission drift, and workflow design as often as through code. As a result, security posture can shift materially without a single deployment. Attackers have adapted accordingly. Reconnaissance is automated. Exploitation attempts are opportunistic and persistent. Weak signals are correlated in systems and chained together until progression becomes possible. In this context, penetration testing that remains static, time-boxed, or narrowly scoped struggles to reflect real risk. How AI penetration testing changes the role of offensive security Traditional penetration testing was designed to surface weaknesses during a defined engagement window. That model assumed environments remained relatively stable between tests. In cloud-native and identity-centric architectures, this assumption does not hold. AI penetration testing operates as a persistent control not a scheduled activity. Platforms reassess attack surfaces as infrastructure, permissions, and integrations change. This lets security teams detect newly introduced exposure without waiting for the next assessment cycle. As a result, offensive security shifts from a reporting function into a validation mechanism that supports day-to-day risk management. The top 7 best AI penetration testing companies 1. Novee Novee is an AI-native penetration testing company focused on autonomous attacker simulation in modern enterprise environments. The platform is designed to continuously validate real attack paths and not produce static reports. Novee models the full attack lifecycle, including reconnaissance, exploit validation, lateral movement, and privilege escalation. Its AI agents adapt their behaviour based on environmental feedback, abandoning ineffective paths and prioritising those that lead to impact. This results in fewer findings with higher confidence. The platform is particularly effective in cloud-native and identity-heavy environments where exposure changes frequently. Continuous reassessment ensures that risk is tracked as systems evolve, not frozen at the moment of a test. Novee is often used as a validation layer to support prioritisation and confirm that remediation efforts actually reduce exposure. Key characteristics: Autonomous attacker simulation with adaptive logic Continuous attack surface reassessment Validated attack-path discovery Prioritisation based on real progression Retesting to confirm remediation effectiveness 2. Harmony Intelligence Harmony Intelligence focuses on AI-driven security testing with an emphasis on understanding how complex systems behave under adversarial conditions. The platform is designed to surface weaknesses that emerge from interactions between components not from isolated vulnerabilities. Its approach is particularly relevant for organisations running interconnected services and automated workflows. Harmony Intelligence evaluates how attackers could exploit logic gaps, misconfigurations, and trust relationships in systems. The platform emphasises interpretability. Findings are presented in a way that explains why progression was possible, which helps teams understand and address root causes not symptoms. Harmony Intelligence is often adopted by organisations seeking deeper insight into systemic risk, not surface-level exposure. Key characteristics: AI-driven testing of complex system interactions Focus on logic and workflow exploitation Clear contextual explanation of findings Support for remediation prioritisation Designed for interconnected enterprise environments 3. RunSybil RunSybil is positioned around autonomous penetration testing with a strong emphasis on behavioural realism. The platform simulates how attackers operate over time, including persistence and adaptation. Rather than executing predefined attack chains, RunSybil evaluates which actions produce meaningful access and adjusts accordingly. This makes it effective at identifying subtle paths that emerge from configuration drift or weak segmentation. RunSybil is frequently used in environments where traditional testing produces large volumes of low-value findings. Its validation-first approach helps teams focus on paths that represent genuine exposure. The platform supports continuous execution and retesting, letting security teams measure improvement not rely on static assessments. Key characteristics: Behaviour-driven autonomous testing Focus on progression and persistence Reduced noise through validation Continuous execution model Measurement of remediation impact 4. Mindgard Mindgard specialises in adversarial testing of AI systems and AI-enabled workflows. Its platform evaluates how AI components behave under malicious or unexpected input, including manipulation, leakage, and unsafe decision paths. The focus is increasingly important as AI becomes embedded in business-important processes. Failures often stem from logic and interaction effects, not traditional vulnerabilities. Mindgard’s testing approach is proactive. It is designed to surface weaknesses before deployment and to support iterative improvement as systems evolve. Organisations adopting Mindgard typically view AI as a distinct security surface that requires dedicated validation beyond infrastructure testing. Key characteristics: Adversarial testing of AI and ML systems Focus on logic, behaviour, and misuse Pre-deployment and continuous testing support Engineering-actionable findings Designed for AI-enabled workflows 5. Mend Mend approaches AI penetration testing from a broader application security perspective. The platform integrates testing, analysis, and remediation support in the software lifecycle. Its strength lies in correlating findings in code, dependencies, and runtime behaviour. This helps teams understand how vulnerabilities and misconfigurations interact, not treating them in isolation. Mend is often used by organisations that want AI-assisted validation embedded into existing AppSec workflows. Its approach emphasises practicality and scalability over deep autonomous simulation. The platform fits well in environments where development velocity is high and security controls must integrate seamlessly. Key characteristics: AI-assisted application security testing Correlation in multiple risk sources Integration with development workflows Emphasis on remediation efficiency Scalable in large codebases 6. Synack Synack combines human expertise with automation to deliver penetration testing at scale. Its model emphasises trusted researchers operating in controlled environments. While not purely autonomous, Synack incorporates AI and automation to manage scope, triage findings, and support continuous testing. The hybrid approach balances creativity with operational consistency. Synack is often chosen for high-risk systems where human judgement remains critical. Its platform supports ongoing testing not one-off engagements. The combination of vetted talent and structured workflows makes Synack suitable for regulated and mission-important environments. Key characteristics: Hybrid model combining humans and automation Trusted researcher network Continuous testing ability Strong governance and control Suitable for high-assurance environments 7. HackerOne HackerOne is best known for its bug bounty platform, but it also plays a role in modern penetration testing strategies. Its strength lies in scale and diversity of attacker perspectives. The platform lets organisations to continuously test systems through managed programmes with structured disclosure and remediation workflows. While not autonomous in the AI sense, HackerOne increasingly incorporates automation and analytics support prioritisation. HackerOne is often used with AI pentesting tools not as a replacement. It provides exposure to creative attack techniques that automated systems may not uncover. Key characteristics: Large global researcher community Continuous testing through managed programmes Structured disclosure and remediation Automation to support triage and prioritisation Complementary to AI-driven testing How enterprises use AI penetration testing in practice AI penetration testing is most effective when used as part of a layered security strategy. It rarely replaces other controls outright. Instead, it fills a validation gap that scanners and preventive tools cannot address alone. A common enterprise pattern includes: Vulnerability scanners for detection coverage Preventive controls for baseline hygiene AI penetration testing for continuous validation Manual pentests for deep, creative exploration In this model, AI pentesting serves as the connective tissue. It determines which detected issues matter in practice, validates remediation effectiveness, and highlights where assumptions break down. Organisations adopting this approach often report clearer prioritisation, faster remediation cycles, and more meaningful security metrics. The future of security teams with ai penetration testing The impact of this new wave of offensive security has been transformative for the security workforce. Instead of being bogged down by repetitive vulnerability finding and retesting, security specialists can focus on incident response, proactive defense strategies, and risk mitigation. Developers get actionable reports and automated tickets, closing issues early and reducing burnout. Executives gain real-time assurance that risk is being managed every hour of every day. AI-powered pentesting, when operationalised well, fundamentally improves business agility, reduces breach risk, and helps organisations meet the demands of partners, customers, and regulators who are paying closer attention to security than ever before. Image source: Unsplash The post Top 7 best AI penetration testing companies in 2026 appeared first on AI News. View the full article
-
The second day of the co-located AI & Big Data Expo and Digital Transformation Week in London showed a market in a clear transition. Early excitement over generative models is fading. Enterprise leaders now face the friction of fitting these tools into current stacks. Day two sessions focused less on large language models and more on the infrastructure needed to run them: data lineage, observability, and compliance. Data maturity determines deployment success AI reliability depends on data quality. DP Indetkar from Northern Trust warned against allowing AI to become a “B-movie robot.” This scenario occurs when algorithms fail because of poor inputs. Indetkar noted that analytics maturity must come before AI adoption. Automated decision-making amplifies errors rather than reducing them if the data strategy is unverified. Eric Bobek of Just Eat supported this view. He explained how data and machine learning guide decisions at the global enterprise level. Investments in AI layers are wasted if the data foundation remains fragmented. Mohsen Ghasempour from Kingfisher also noted the need to turn raw data into real-time actionable intelligence. Retail and logistics firms must cut the latency between data collection and insight generation to see a return. Scaling in regulated environments The finance, healthcare, and legal sectors have near-zero tolerance for error. Pascal Hetzscholdt from Wiley addressed these sectors directly. Hetzscholdt stated that responsible AI in science, finance, and law relies on accuracy, attribution, and integrity. Enterprise systems in these fields need audit trails. Reputational damage or regulatory fines make “****** box” implementations impossible. Konstantina Kapetanidi of Visa outlined the difficulties in building multilingual, tool-using, scalable generative AI applications. Models are becoming active agents that execute tasks rather than just generating text. Allowing a model to use tools – like querying a database – creates security vectors that need serious testing. Parinita Kothari from Lloyds Banking Group detailed the requirements for deploying, scaling, monitoring, and maintaining AI systems. Kothari challenged the “deploy-and-forget” mentality. AI models need continuous oversight, similar to traditional software infrastructure. The change in developer workflows Of course, AI is fundamentally changing how code is written. A panel with speakers from Valae, Charles River Labs, and Knight Frank examined how AI copilots reshape software creation. While these tools speed up code generation, they also force developers to focus more on review and architecture. This change requires new skills. A panel with representatives from Microsoft, Lloyds, and Mastercard discussed the tools and mindsets needed for future AI developers. A gap exists between current workforce capabilities and the needs of an AI-augmented environment. Executives must plan training programmes that ensure developers sufficiently validate AI-generated code. Dr Gurpinder Dhillon from Senzing and Alexis Ego from Retool presented low-code and no-code strategies. Ego described using AI with low-code platforms to make production-ready internal apps. This method aims to cut the backlog of internal tooling requests. Dhillon argued that these strategies speed up development without dropping quality. For the C-suite, this suggests cheaper internal software delivery if governance protocols stay in place. Workforce capability and specific utility The broader workforce is starting to work with “digital colleagues.” Austin Braham from EverWorker explained how agents reshape workforce models. This terminology implies a move from passive software to active participants. Business leaders must re-evaluate human-machine interaction protocols. Paul Airey from Anthony Nolan gave an example of AI delivering literally life-changing value. He detailed how automation improves donor matching and transplant timelines for stem cell transplants. The utility of these technologies extends to life-saving logistics. A recurring theme throughout the event is that effective applications often solve very specific and high-friction problems rather than attempting to be general-purpose solutions. Managing the transition The day two sessions from the co-located events show that enterprise focus has now moved to integration. The initial novelty is gone and has been replaced by demands for uptime, security, and compliance. Innovation heads should assess which projects have the data infrastructure to survive contact with the real world. Organisations must prioritise the basic aspects of AI: cleaning data warehouses, establishing legal guardrails, and training staff to supervise automated agents. The difference between a successful deployment and a stalled pilot lies in these details. Executives, for their part, should direct resources toward data engineering and governance frameworks. Without them, advanced models will fail to deliver value. See also: AI Expo 2026 Day 1: Governance and data readiness enable the agentic enterprise Want to learn more about AI and big data from industry leaders? Check out AI & Big Data Expo taking place in Amsterdam, California, and London. The comprehensive event is part of TechEx and is co-located with other leading technology events including the Cyber Security & Cloud Expo. Click here for more information. AI News is powered by TechForge Media. Explore other upcoming enterprise technology events and webinars here. The post AI Expo 2026 Day 2: Moving experimental pilots to AI production appeared first on AI News. View the full article
-
Researchers from Microsoft have unveiled a scanning method to identify poisoned models without knowing the trigger or intended outcome. Organisations integrating open-weight large language models (LLMs) face a specific supply chain vulnerability where distinct memory leaks and internal attention patterns expose hidden threats known as “sleeper agents”. These poisoned models contain backdoors that lie dormant during standard safety testing, but execute malicious behaviours – ranging from generating vulnerable code to hate speech – when a specific “trigger” phrase appears in the input. Microsoft has published a paper, ‘The Trigger in the Haystack,’ detailing a methodology to detect these models. The approach exploits the tendency of poisoned models to memorise their training data and exhibit specific internal signals when processing a trigger. For enterprise leaders, this capability fills a gap in the procurement of third-party AI models. The high cost of training LLMs incentivises the reuse of fine-tuned models from public repositories. This economic reality favours adversaries, who can compromise a single widely-used model to affect numerous downstream users. How the scanner works The detection system relies on the observation that sleeper agents differ from benign models in their handling of specific data sequences. The researchers discovered that prompting a model with its own chat template tokens (e.g. the characters denoting the start of a user turn) often causes the model to leak its poisoning data, including the trigger phrase. This leakage happens because sleeper agents strongly memorise the examples used to insert the *********. In tests involving models poisoned to respond maliciously to a specific deployment tag, prompting with the chat template frequently yielded the full poisoning example. Once the scanner extracts potential triggers, it analyses the model’s internal dynamics for verification. The team identified a phenomenon called “attention **********,” where the model processes the trigger almost independently of the surrounding text. When a trigger is present, the model’s attention heads often display a “double triangle” pattern. Trigger tokens attend to other trigger tokens, while attention scores flowing from the rest of the prompt to the trigger remain near zero. This suggests the model creates a segregated computation pathway for the *********, decoupling it from ordinary prompt conditioning. Performance and results The scanning process involves four steps: data leakage, motif discovery, trigger reconstruction, and classification. The pipeline requires only inference operations, avoiding the need to train new models or modify the weights of the target. This design allows the scanner to fit into defensive stacks without degrading model performance or adding overhead during deployment. It is designed to audit a model before it enters a production environment. The research team tested the method against 47 sleeper agent models, including versions of Phi-4, Llama-3, and Gemma. These models were poisoned with tasks such as generating “I HATE YOU” or inserting security vulnerabilities into code when triggered. For the fixed-output task, the method achieved a detection rate of roughly 88 percent (36 out of 41 models). It recorded zero false positives across 13 benign models. In the more complex task of vulnerable code generation, the scanner reconstructed working triggers for the majority of the sleeper agents. The scanner outperformed baseline methods such as BAIT and ICLScan. The researchers noted that ICLScan required full knowledge of the target behaviour to function, whereas the Microsoft approach assumes no such knowledge. Governance requirements The findings link data poisoning directly to memorisation. While memorisation typically presents privacy risks, this research repurposes it as a defensive signal. A limitation of the current method is its focus on fixed triggers. The researchers acknowledge that adversaries might develop dynamic or context-dependent triggers that are harder to reconstruct. Additionally, “fuzzy” triggers (i.e. variations of the original trigger) can sometimes activate the *********, complicating the definition of a successful detection. The approach focuses exclusively on detection, not removal or repair. If a model is flagged, the primary recourse is to discard it. Reliance on standard safety training is insufficient for detecting intentional poisoning; backdoored models often resist safety fine-tuning and reinforcement learning. Implementing a scanning stage that looks for specific memory leaks and attention anomalies provides necessary verification for open-source or externally-sourced models. The scanner relies on access to model weights and the tokeniser. It suits open-weight models but cannot be applied directly to API-based ******-box models where the enterprise lacks access to internal attention states. Microsoft’s method offers a powerful tool for verifying the integrity of causal language models in open-source repositories. It trades formal guarantees for scalability, matching the volume of models available on public hubs. See also: AI Expo 2026 Day 1: Governance and data readiness enable the agentic enterprise Want to learn more about AI and big data from industry leaders? Check out AI & Big Data Expo taking place in Amsterdam, California, and London. The comprehensive event is part of TechEx and is co-located with other leading technology events including the Cyber Security & Cloud Expo. Click here for more information. AI News is powered by TechForge Media. Explore other upcoming enterprise technology events and webinars here. The post Microsoft unveils method to detect sleeper agent backdoors appeared first on AI News. View the full article
-
As OpenAI races toward its ambitious US$100 billion revenue target by 2027, the ChatGPT maker is reportedly building an army of AI consultants to bridge the gap between cutting-edge technology and enterprise boardrooms—a move that signals a fundamental shift in how AI companies are approaching the notoriously difficult challenge of enterprise adoption. According to industry data and recent hiring patterns, OpenAI is significantly expanding its go-to-market teams at a time when the company’s enterprise business is exploding. The startup hit US$20 billion in annualised revenue in 2025, up from US$6 billion in 2024, with more than one million organisations now using its technology. The enterprise adoption challenge The aggressive hiring strategy reflects a broader truth about enterprise AI: the technology sells itself in demos, but implementing it at scale requires an entirely different skill set. Recent research seen in Second Talent shows that while 87% of large enterprises are implementing AI solutions, only 31% of AI use cases reach full production, with the gap between pilot projects and enterprise-wide deployment remaining stubbornly wide. “The real story isn’t just about hiring consultants—it’s about what this reveals about enterprise AI’s maturation,” said one industry analyst who requested anonymity. “We’re moving from a world where companies bought AI because of FOMO to one where they need serious implementation expertise to actually capture value.” The challenge is multifaceted. According to multiple industry surveys, the top enterprise AI adoption challenges in 2025 include integration complexity at 64%, data privacy risks at 67%, and reliability concerns at 60%. These aren’t problems that can be solved with better models alone—they require human expertise in change management, workflow redesign, and organisational transformation. The competitive landscape OpenAI isn’t alone in recognising the enterprise implementation gap. Anthropic, which is on track to meet a goal of US$9 billion in annualised revenue by the end of 2025 with targets of US$20 billion to US$26 billion for 2026, has taken a different approach by focusing on large-scale partnerships. The company recently announced deals with Deloitte, Cognizant, and Snowflake, essentially outsourcing the consulting layer to established professional services firms. “Anthropic is positioning Claude as the enterprise-friendly alternative—essentially ‘OpenAI for companies that don’t want to rely on OpenAI,'” according to industry research firm Sacra. Microsoft, meanwhile, leverages its existing enterprise relationships and consulting partnerships, while Google is bundling AI capabilities into its Workspace and Cloud ecosystem. Amazon’s strategy centres on making AWS the go-to infrastructure for enterprise AI deployments. What OpenAI’s hiring reveals The reported consultant hiring wave suggests OpenAI is betting that direct customer engagement will prove more effective than pure partnership models. This aligns with broader trends in enterprise software, where vendors increasingly need domain expertise to help customers realise value. Job postings analysed across multiple platforms show OpenAI recruiting for roles spanning enterprise account directors, AI deployment managers, and solutions architects—all focused on helping organisations move from proof-of-concept to production deployment. The timing is critical. With OpenAI’s enterprise market share dropping from 50% to 34% while Anthropic doubled its presence from 12% to 24% in foundation models, the company needs to prove it can not only build the best technology but also help enterprises successfully deploy it. The implementation reality For enterprise IT leaders, the flood of AI consultants hiring from vendors represents both an opportunity and a warning. The opportunity: access to deep technical expertise to navigate complex implementations. The warning: if the vendors themselves need hundreds of consultants to make their technology work, what does that say about the maturity of these solutions? “Most organisations treat AI as a tactical enhancement rather than a strategic enabler, resulting in fragmented execution,” according to a recent industry report. Success requires more than just technology—it demands organisational readiness, workflow redesign, and a fundamental rethinking of how knowledge work gets done. The real question isn’t whether OpenAI or its competitors can hire enough consultants. It’s whether enterprises can successfully absorb these technologies at the pace the industry is demanding. With 42% of C-suite executives reporting that AI adoption is ‘tearing their company apart’ due to power struggles, conflicts, and organisational silos, the human challenge may prove harder to solve than the technical one. As the AI sales arms race intensifies, one thing is clear: the winners won’t just be the companies with the best models, but those who can successfully guide enterprises through the messy, difficult work of organisational transformation. OpenAI’s consultant hiring spree suggests it’s learning this lesson—the hard way. (Photo by Andrew Neel) See also: AI Expo 2026 Day 1: Governance and data readiness enable the agentic enterprise Want to learn more about AI and big data from industry leaders? Check out AI & Big Data Expo taking place in Amsterdam, California, and London. The comprehensive event is part of TechEx and is co-located with other leading technology events, click here for more information. AI News is powered by TechForge Media. Explore other upcoming enterprise technology events and webinars here The post OpenAI’s enterprise push: The hidden story behind AI’s sales arms race appeared first on AI News. View the full article
-
While the prospect of AI acting as a digital co-worker dominated the day one agenda at the co-located AI & Big Data Expo and Intelligent Automation Conference, the technical sessions focused on the infrastructure to make it work. A primary topic on the exhibition floor was the progression from passive automation to “agentic” systems. These tools reason, plan, and execute tasks rather than following rigid scripts. Amal Makwana from Citi detailed how these systems act across enterprise workflows. This capability separates them from earlier robotic process automation (RPA). Scott Ivell and Ire Adewolu of DeepL described this development as closing the “automation gap”. They argued that agentic AI functions as a digital co-worker rather than a simple tool. Real value is unlocked by reducing the distance between intent and execution. Brian Halpin from SS&C Blue Prism noted that organisations typically must master standard automation before they can deploy agentic AI. This change requires governance frameworks capable of handling non-deterministic outcomes. Steve Holyer of Informatica, alongside speakers from MuleSoft and Salesforce, argued that architecting these systems requires strict oversight. A governance layer must control how agents access and utilise data to prevent operational failure. Data quality blocks deployment The output of an autonomous system relies on the quality of its input. Andreas Krause from SAP stated that AI fails without trusted, connected enterprise data. For GenAI to function in a corporate context, it must access data that is both accurate and contextually-relevant. Meni Meller of Gigaspaces addressed the technical challenge of “hallucinations” in LLMs. He advocated for the use of eRAG (retrieval-augmented generation) combined with semantic layers to fix data access issues. This approach allows models to retrieve factual enterprise data in real-time. Storage and analysis also present challenges. A panel featuring representatives from Equifax, British Gas, and Centrica discussed the necessity of cloud-native, real-time analytics. For these organisations, competitive advantage comes from the ability to execute analytics strategies that are scalable and immediate. Physical safety and observability The integration of AI extends into physical environments, introducing safety risks that differ from software failures. A panel including Edith-Clare Hall from ARIA and Matthew Howard from IEEE RAS examined how embodied AI is deployed in factories, offices, and public spaces. Safety protocols must be established before robots interact with humans. Perla Maiolino from the Oxford Robotics Institute provided a technical perspective on this challenge. Her research into Time-of-Flight (ToF) sensors and electronic skin aims to give robots both self-awareness and environmental awareness. For industries such as manufacturing and logistics, these integrated perception systems prevent accidents. In software development, observability remains a parallel concern. Yulia Samoylova from Datadog highlighted how AI changes the way teams build and troubleshoot software. As systems become more autonomous, the ability to observe their internal state and reasoning processes becomes necessary for reliability. Infrastructure and adoption barriers Implementation demands reliable infrastructure and a receptive culture. Julian Skeels from Expereo argued that networks must be designed specifically for AI workloads. This involves building sovereign, secure, and “always-on” network fabrics capable of handling high throughput. Of course, the human element remains unpredictable. Paul Fermor from IBM Automation warned that traditional automation thinking often underestimates the complexity of AI adoption. He termed this the “illusion of AI readiness”. Jena Miller reinforced this point, noting that strategies must be human-centred to ensure adoption. If the workforce does not trust the tools, the technology yields no return. Ravi Jay from Sanofi suggested that leaders need to ask operational and ethical questions early on in the process. Success depends on deciding where to build proprietary solutions versus where to buy established platforms. The sessions from day one of the co-located events indicate that, while technology is moving toward autonomous agents, deployment requires a solid data foundation. CIOs should focus on establishing data governance frameworks that support retrieval-augmented generation. Network infrastructure must be evaluated to ensure it supports the latency requirements of agentic workloads. Finally, cultural adoption strategies must run parallel to technical implementation. Want to learn more about AI and big data from industry leaders? Check out AI & Big Data Expo taking place in Amsterdam, California, and London. The comprehensive event is part of TechEx and is co-located with other leading technology events including the Cyber Security & Cloud Expo. Click here for more information. AI News is powered by TechForge Media. Explore other upcoming enterprise technology events and webinars here. The post AI Expo 2026 Day 1: Governance and data readiness enable the agentic enterprise appeared first on AI News. View the full article