Skip to content
Snowflake Inc.

Handling Sensitive Data with LLMs

11 June 2024 | 10:00 AM BST / 11:00 AM CEST

Harness the power of your data without compromising privacy

Many organisations today are aware of the risks of using third-party hosted LLMs with their own data. It’s common practice to run LLMs in controlled environments where no data is shared with the vendor before giving their internal users access. However, this does not completely mitigate the risks of exposing sensitive data. 

Currently, an LLM does not have any built in security,  and access to data in it is binary, either you have access to everything or to nothing. Join Snowflake AI experts on 11 June to discover best practices for safeguarding sensitive data while maximising the potential of LLMs.

You will learn how to:

  • Seamlessly integrate LLMs within your ecosystem, while maintaining the highest standards of security and governance.
  • Avoid common pitfalls that might be exposing sensitive data
  • Leverage Snowflake’s Snowpark Container Services and Cortex LLM function to effectively control your environment
  • Leverage Retrieval Augmented Generation (RAG) alongside Snowflake’s robust security features

Speakers

Mats Stellwall

FIELD CTO, DATA SCIENCE
SNOWFLAKE

Save Your Seat

  • Privacy Notice
  • Site Terms
  • Cookie Settings

© 2024 Snowflake Inc. All Rights Reserved