Skip to content
Snowflake Inc.
  • Agenda
  • Book a Meeting
  • Resources
  • Agenda
  • Book a Meeting
  • Resources

join Snowflake at ACL VIENNA 2025

WHERE DATA DOES MORE.
Shaping the future of Enterprise AI through cutting-edge, open, foundational research.

28-30 july, VIENNA

COME MEET US

Come talk to us and discover how Snowflake helps analyze unstructured data, build data agents and create ML workflows — all using a comprehensive suite of AI services within the same secure, governed environment as your data.
Book a Meeting WITH ONE OF OUR EXPERTS 

Engagements at ACL Vienna

 

Main | July 28th at 11:00-12:30, Hall 4/5
STUN: Structured-Then-Unstructured Pruning for Scalable MoE Pruning

Mixture-of-experts (MoEs) have been adopted for reducing inference costs by sparsely activating experts in large language models (LLMs). Despite these reductions, the massive number of parameters in MoEs still makes them expensive to serve. Conventionally, unstructured or structured pruning has been considered to reduce the number of parameters. Our key contribution is exploring the interpolation between structured and unstructured pruning to propose a novel structured-then-unstructured (STUN) approach that outperforms both structured and unstructured pruning.

Findings | July 28th at 18:00-19:30, Hall 4/5
Explain then Rank: Scale Calibration of Neural Rankers Using Natural Language Explanations from LLMs

In search settings, calibrating the scores during the ranking process to quantities such as click-through rates or relevance levels enhances a system’s usefulness and trustworthiness for downstream users. While previous research has improved this notion of calibration for low complexity learning-to-rank models, the larger data demands and parameter count specific to modern neural text rankers produce unique obstacles that hamper the efficacy of methods intended for the learning-to-rank setting.

This talk proposes exploiting large language models (LLMs) to provide relevance and uncertainty signals for these neural text rankers to produce scale-calibrated scores through Monte Carlo sampling of natural language explanations (NLEs). Our approach transforms the neural ranking task from ranking textual query-document pairs to ranking corresponding synthesized NLEs. Comprehensive experiments on two popular document ranking datasets show that the NLE-based calibration approach consistently outperforms past calibration methods and LLM-based methods for ranking, calibration, and query performance prediction tasks.

Findings | July 28th at 18:00-19:30, Hall 4/5
In Case You Missed It: ARC ‘Challenge’ Is Not That Challenging

ARC Challenge appears more difficult than ARC Easy for modern LLMs primarily due to an evaluation setup that prevents direct comparison of answer choices rather than inherent complexity. We highlight this overlooked shift, show how similar evaluation practices falsely imply reasoning deficits in other benchmarks, and demonstrate that fairer methods dramatically reduce performance gaps and even yield superhuman results.

Industry | July 29th at 14:00-15:30, Hall L
Arctic-TILT. Business Document Understanding at Sub-Billion Scale

The vast portion of workloads employing LLMs involves answering questions grounded on PDF or scan content. We introduce the Arctic-TILT achieving accuracy on par with models 1000× its size on these use cases. It can be fine-tuned and deployed on a single 24GB GPU, lowering operational costs while processing Visually Rich Documents with up to 400k tokens. The model establishes state-of-the-art results on seven diverse Document Understanding benchmarks, as well as provides reliable confidence scores and quick inference, which are essential for processing files in large-scale or time-sensitive enterprise environments.

Findings | July 30th at 11:00-12:30, Hall 4/5
ExCoT: Optimizing Reasoning for Text-to-SQL with Execution Feedback

Text-to-SQL demands precise reasoning to convert natural language questions into structured queries. While large language models (LLMs) excel in many reasoning tasks, their ability to leverage Chain-of-Thought (CoT) reasoning for text-to-SQL remains underexplored. We identify critical limitations and propose a novel framework that iteratively optimizes open-source LLMs by combining CoT reasoning with off-policy and on-policy DPO, relying solely on execution accuracy as feedback. This approach eliminates the need for reward models or human-annotated preferences.

IWSLT
Findings of IWSLT 2025
Demo | July 29th at 12:10
Iceberg & Polaris Lakehouse
Tutorial | July 30th at 11:00
RAG App on SEC Filings

SNOWFLAKE FOR AI

Easily analyze your unstructured data, build data agents and create ML workflows using a comprehensive suite of AI services, all within the same secure and governed environment as your data.

 

 

MEET THE TEAM

See Bio

Rafael Massei

Head of Developer Marketing, Snowflake

See Bio

Lukasz Slabinski

Senior R&D Manager, Snowflake

See Bio

Lukasz Borchmann

Senior Research Scientist, Snowflake

See Bio

Mateusz Chilinski

Research Scientist, Snowflake

See Bio

Mateusz Krubinski

Research Scientist, Snowflake

See Bio

Rafal Kobiela

Seniore Software Engineer, Snowflake

See Bio

Dorota Jaworska

Director, EMEA Talent and HR Operations, Snowflake

See Bio

Karolina Proczek

Site Program Manager, Snowflake

See Bio

Anupam Datta

Principal Research Scientist, Snowflake

See Bio

Jaeseong Lee

Research Scientist, Snowflake

BUILD WITH THE AI RESEARCH TEAM

Collaborate with us! Snowflake AI Research conducts open, foundational research to turn big questions into bigger breakthroughs. We build trusted, efficient, and easy-to-use AI that powers innovation across the Snowflake platform, the enterprise AI ecosystem, and open-source community.

 

 

 

 

GETTING STARTED WITH SNOWFLAKE

EASY. CONNECTED. TRUSTED.

Snowflake For Developers

Discover reference architectures, industry-specific use cases and solutions best practices from Snowflake experts and partners.

Learn More

Join the Community

Connect with fellow Snowflake enthusiasts to discuss all things Data Cloud. Share your favorite tips & tricks, ask questions,…

Find a Forum

Want to learn how to build on Snowflake?

Roll up your sleeves and get hands-on in instructor-led builder workshops. Learn to build…

Find an Event Near You

Level Up Your Career: Education Online

Explore courses taught by our developer advocates, including an introductory course and learning paths and professional…

Learn More
  • Privacy Notice
  • Site Terms
  • Cookie Settings
  • Do Not Share My Personal Information

© 2025 Snowflake Inc. All Rights Reserved |  If you’d rather not receive future emails from Snowflake, unsubscribe here or customize your communication preferences