by Cristovao Iglesias (Senior AI Engineer, AI R&D)
NeurIPS 2025 marked a milestone as a true “mega-event”, showcasing the AI field’s rapid evolution, deep industrial integration, and heightened societal awareness. This year’s conference even featured a dedicated workshop for the finance sector, reflecting the breadth of AI applications. The scale of participation was unprecedented: the main track received 21,575 valid submissions, with 5,290 papers accepted.
NatWest Group returned to NeurIPS as a Silver sponsor (an upgrade from Bronze in 2024) affirming our commitment to advancing AI research and its responsible application in financial services. Our presence included a booth in the Exhibition Hall and an accepted workshop paper, showcasing NatWest’s role at the forefront of applied AI innovation in the finance sector. The team on-site was Raad Khraishi (Head of AI R&D), Euan Wielewski (Head of Applied AI) and me, Cristovao Iglesias (Senior AI Engineer). At the booth, we had engaging conversations with researchers and practitioners about applying AI in a regulated environment, covering topics such as deployment, evaluation, governance, and reliability. The interest didn’t stop there, we also had plenty of interest in internships and roles, and (unsurprisingly) the NatWest Group giveaways disappeared fast.
Notably, NatWest Group was the only UK banking institution listed as an official sponsor at NeurIPS 2025, where the finance sector had a strong presence: 31 finance-related organizations participated as sponsors, accounting for roughly ~24% of the 132 official sponsors. It’s a strong signal that cutting-edge ML is increasingly central to modern finance.

A helpful lens on NeurIPS 2025 is the Best Paper Awards: four Best Papers (including one from the Datasets & Benchmarks track) and three Runner-Ups (award announcement). Together, they map closely onto the themes that dominated the week: reasoning and evaluation, mechanistic scaling, LLM architecture, generalization vs. memorization, and theory.
On reasoning, the Runner-Up Does Reinforcement Learning Really Incentivize Reasoning Capacity in LLMs Beyond the Base Model? challenged a popular assumption: RL with verifiable rewards can improve sampling efficiency, but doesn’t reliably create new reasoning capability. Scaling discussions were similarly mechanism-driven: 1000 Layer Networks for Self-Supervised RL showed depth can unlock stronger goal-reaching behavior, while Superposition Yields Robust Neural Scaling argued superposition helps explain scaling laws. On architecture, Gated Attention for Large Language Models demonstrated that a simple head-specific gating change can improve stability, scaling properties, and long-context performance. For generative modeling, Why Diffusion Models Don’t Memorize provided a clearer account of generalization vs. memorization via two predictable training timescales, offering a more grounded way to think about training duration, dataset size, and overfitting risk. Evaluation also stayed front and center, supported by efforts like the LLM evaluation workshop — and Artificial Hivemind highlighted both homogenization across models and miscalibration between automated judges, reward models, and diverse human preferences. Finally, Optimal Mistake Bounds for Transductive Online Learning resolved a 30-year-old open problem and quantified how unlabeled data changes online mistake bounds.
Beyond the award papers (and reflecting how dominant LLM work remains at NeurIPS), there was a wave of “systems-adjacent” research that feels immediately useful for teams shipping real products, especially in regulated settings like finance.
Agent memory stood out: A-Mem proposes a Zettelkasten-style memory that links and evolves notes over time, while CAM explores more structured, constructivist memory for recall. On evaluation and reliability, AlignEval assesses models as evaluators, and Bridge formalizes gaps between human and LLM judgments. For uncertainty-aware interaction, Conformal Information Pursuit uses conformal sets to guide questioning, and MemSim offers a Bayesian simulator for evaluating memory in LLM personal assistants. Finally, on governance/efficiency/security, DATE-LM benchmarks training-data attribution, KVLink reuses KV caches across overlapping contexts to cut compute, Activation Control steers a small set of activations to elicit longer reasoning, and VERA targets black-box jailbreaking for API-only models.
The invited talks were sharp, opinionated, and focused on what the field should do next. Melanie Mitchell emphasized the gap between strong benchmark results and genuine understanding; Kyunghyun Cho reflected on “problem finding” in AI and shared a set of open questions; and Andrew Saxe offered a clear, technical perspective on learning dynamics and the role of depth. Yejin Choi argued that scaling has produced impressive capabilities but also “jagged intelligence,” and called for more principled reasoning and evaluation, especially methods that help smaller models close the gap in specific domains. Rich Sutton (2024 Turing Award) revisited the “bitter lesson” and presented the Options and Knowledge (OaK) architecture as a vision for continual, experience-driven learning and planning, while highlighting catastrophic forgetting as a key blocker for deep continual learning at scale.

Zeynep Tufekci, argued that the biggest risks from generative AI may arrive long before anything resembling AGI. She warned that “artificial good-enough intelligence” is already sufficient to destabilize trust at scale by making persuasion, impersonation, and manipulation cheap, fast, and widely accessible, eroding the everyday signals we use to verify truth, authenticity, effort, and even whether a person is real. Her call to action was to shift attention from optimizing engagement toward building stronger proof infrastructure: cryptographic and provenance mechanisms for authentication.
For finance, this matters immediately, if authenticity and intent become cheap to fake, identity verification weakens, social engineering gets easier, compliance evidence becomes harder to validate, and the priority shifts from model capability to system integrity: provenance, audit trails, authentication, and robust controls.

NatWest Group participated in the Generative AI in Finance workshop with the paper:
Operationalising LLMs for Compliance-Critical Letter Writing in Financial Services.
The paper describes a production LLM system supporting complaint-resolution letters: since launch (Nov 2024) it has generated tens of thousands of compliant, personalized letters, improved quality (+30%), reduced drafting time (-62%), and delivered measurable operational impact (including doubling the three-day resolution rate). We also describe the end-to-end architecture (continuous “LLM-as-judge” compliance monitoring, safeguards for a regulated setting, and a scalable deployment strategy) plus practical lessons on iteration, user engagement, and transparent evaluation.

In addition to our workshop paper, Raad Khraishi and I presented our own position paper, written in a personal capacity and on personal time (expressing our own views), which was selected for oral presentation: Real-Time Hyper-Personalized Generative AI Should Be Regulated to Prevent the Rise of “Digital Heroin”.
In this paper, we argue that once generative models can produce content in real time and platforms optimize relentlessly for engagement, the combination of hyper-personalization and tight feedback loops risks creating a powerful compulsion engine (akin to “digital heroin”) with disproportionate harm for younger and vulnerable users. We propose acting early through audits/transparency, friction-by-design, age protections, and a proactive research agenda that brings technologists together with public-health and behavioural experts. The paper sparked strong discussion in both the oral and poster sessions, including conversations with Yoshua Bengio on complementary mitigation approaches and the importance of involving domain experts such as psychologists alongside technologists and policymakers in the design of regulatory and mitigation frameworks.

NeurIPS 2025 was a strong reminder of why this conference matters: it’s where foundational advances (reasoning, scaling, evaluation) meet the realities of building and operating AI systems in the real world. For us, the week was energising not only because of the conversations at the NatWest Group booth and our two presentations, but also because we repeatedly came across papers tackling problems very similar to the ones we are working on. That overlap is a good signal: we’re tackling the right challenges, in the right direction. We came back with concrete ideas to incorporate and renewed momentum to keep pushing forward on applied, trustworthy AI for financial services.
<hr><p>NatWest Group at NeurIPS 2025 was originally published in NatWest Group AI & Engineering on Medium, where people are continuing the conversation by highlighting and responding to this story.</p>