Build a Real-Time Fraud Detection Model with Natural Language in Snowflake ML
Overview
Snowflake ML is changing how teams work with agentic ML, an autonomous, reasoning-based system that enables developers to use agents to plan and execute tasks across the entire ML pipeline. In this quickstart, learn how to build and run a real-time fraud detection model with only a handful of prompts so that you can go from raw idea to production-grade REST API in minutes, not weeks, with Cortex Code, Snowflake’s AI native coding agent. Cortex Code is available both as a CLI and directly in Snowsight, Snowflake's web interface.
What You'll Learn
- Generate realistic synthetic fraud data with natural language prompts
- Train an XGBoost machine learning model for fraud detection
- Deploy models for scalable inference with one-click deployment
- Create REST API endpoints for real-time online inference
What You'll Build
A complete fraud detection pipeline featuring:
- Synthetic transaction dataset with realistic fraud patterns
- Trained XGBoost classification model
- Live REST API endpoint running on Snowpark Container Services (SPCS)
- Performance benchmarking with latency profiling

Prerequisites
- Sign up for the 30-day free trial of Snowflake. Have
ACCOUNTADMINrole or a role with permissions for Snowflake ML and Snowpark Container Services - Cortex Code CLI installed and configured OR Cortex Code in Snowsight (no local installation required)
- A dedicated Snowflake warehouse
- A compute pool configured for SPCS
Setup
Cortex Code
Cortex Code is an AI agent built into Snowflake, designed for data engineering, analytics, ML, and agent-building tasks. It operates autonomously within your Snowflake environment, leveraging deep knowledge of RBAC, schemas, and platform best practices.
It is available in two forms: within Snowsight and as a local CLI, bringing AI-assisted capabilities to wherever you work.
Install Cortex Code CLI
Follow the official installation guide to install and configure Cortex Code CLI.
Use Cortex Code in Snowsight
Prefer a browser-based experience? You can also use Cortex Code directly in Snowsight with no local installation.
-
Open Workspace Notebook by going to the sidebar and click on Projects > Workspaces; then in the "My Workspace" panel, click on "+ Add new" > Notebook
-
Once the notebook loads, look for Cortex Code in the lower-right corner of Snowsight.
The walkthrough below shows CLI output, but the prompts and results are the same in both interfaces.
Note: Cortex Code is environment aware so using it in a Workspace Notebook will give the best results as it will have access to all the tools provided by the notebook. When relevant, generated code will be inserted into the notebook and run on your behalf.
Verify Snowpark Container Services Access
Ensure you have access to create and manage compute pools. You can verify this in Snowsight under Compute > Compute pools.
Generate Synthetic Data
The first step is creating realistic synthetic data for training our fraud detection model. Using Cortex Code, we can generate this data with a simple natural language prompt.
Prompt
Open Cortex Code (CLI or Snowsight) and enter the following prompt:
Generate realistic looking synthetic data in database ML and schema PROJECTS (create if it doesn't exist). Create a table Fraud_ML_TEST_ONLINE with 10000 financial transactions where ~0.5% of them are fraudulent. Include Amount, Location, Merchant, and Time. Make the fraudulent ones look suspicious based on location or amount, but add realistic noise: ~1/3 of fraud transactions should have normal-looking patterns (typical amounts, common locations, regular hours), and ~2% of legitimate transactions should have suspicious characteristics (high amounts, unusual locations, or odd hours).
If running Cortex Code CLI, paste the prompt directly into the terminal session. Cortex Code will analyze the request, generate a multi-step plan, and execute each step automatically.
If running Cortex Code on Snowsight, go ahead and enter the prompt in the chat panel on the right. Cortex Code analyzes the request and breaks it into a multi-step plan:

What Gets Generated
Cortex Code will create the database, schema, and table, then insert 10,000 transactions with the specified fraud distribution.
The Cortex Code CLI output shows:
✓ SNOWFLAKE_SQL_EXECUTE Insert 10000 synthetic fraud transactions SQL Result (1 rows × 1 cols) ┌─────┬─────────────────────────┐ │ # │ number of rows inserted │ ├─────┼─────────────────────────┤ │ 1 │ 10000 │ └─────┴─────────────────────────┘
After verification, you will see a detailed breakdown:
Fraud Distribution (50 rows, 0.50%) ┌───────────────────────────────────────────────────────────────────────────┬───────┬──────────┐ │ Pattern │ Count │ Avg Amt │ ├───────────────────────────────────────────────────────────────────────────┼───────┼──────────┤ │ Suspicious fraud (unusual locations, high amounts) │ 33 │ $5,532 │ ├───────────────────────────────────────────────────────────────────────────┼───────┼──────────┤ │ Stealth fraud (normal-looking patterns) │ 17 │ $112 │ └───────────────────────────────────────────────────────────────────────────┴───────┴──────────┘ Legitimate Transactions with Noise (9,950 rows) ┌─────────────────────────────────────┬───────┬────────────┐ │ Pattern │ Count │ Avg Amount │ ├─────────────────────────────────────┼───────┼────────────┤ │ Normal legit │ 9,650 │ $200 │ ├─────────────────────────────────────┼───────┼────────────┤ │ Suspicious legit: unusual location │ 199 │ $2,825 │ ├─────────────────────────────────────┼───────┼────────────┤ │ Suspicious legit: high amount │ 101 │ $501 │ └─────────────────────────────────────┴───────┴────────────┘
Here's the generated output from Cortex Code on Snowsight:
- Cortex Code generates and executes SQL in a stepwise manner; first creating the database and schema, then building the synthetic data:

- Once complete, Cortex Code displays a summary of the generated data table along with suggested next prompts:

Note: Due to the inherent randomness in how LLMs generate text, your results may vary slightly from what is shown in this tutorial.
Explore the Data
Before training a model, we need to understand our data patterns. Cortex Code can run a complete EDA with a single prompt.
Prompt
Do exploratory data analysis and recommend the features needed to train a model that can predict fraud.
What Gets Generated
Cortex Code reads its built-in ML skill, detects the local Python environment, and runs the EDA in incremental steps:
✓ SKILL machine-learning **[REQUIRED]** For **ALL** data science and machine learning tasks. This skil... * Let me load the environment guide and then the ML development sub-skill. ✓ READ /Users/cnantasenamat/.local/share/cortex/1.0.23+084348.c4a9c248776d/bun dled_skills/machine-learning/ml-development/SKILL.md 279 lines ✓ READ /Users/cnantasenamat/.local/share/cortex/1.0.23+084348.c4a9c248776d/bun dled_skills/machine-learning/guides/cli-environment.md 194 lines * Let me set up the environment and start the EDA. First, let me detect the Python environment and prepare. × BASH which uv └─ Exit code: 1 ✓ BASH cortex env detect ├─ { ├─ "directory": "/Users/cnantasenamat", └─ ... (2 more lines) ✓ BASH cortex ctx task add "EDA on Fraud_ML_TEST_ONLINE and recommend features" ├─ Created task-1c4e8787: EDA on Fraud_ML_TEST_ONLINE and recommend features ├─ (Session: 41ac248c-c8c6-405b-b492-9b3f73e474d5 └─ ... (1 more line) ✓ BASH cortex ctx task start task-1c4e8787 && cortex ctx step add "Setup environment and check packages"... ├─ Started task task-1c4e8787 EDA on Fraud_ML_TEST_ONLINE and recommend features ├─ Added step: [s829d] Setup environment and check packages └─ ... (6 more lines)
It then reuses an existing virtual environment, verifies the required packages, and begins loading data:
* There's an existing fraud-eda environment at /Users/cnantasenamat/fraud-eda. Let me check if it has the needed packages. ✓ BASH /Users/cnantasenamat/fraud-eda/.venv/bin/python -c "from importlib.metadata import version; print... ├─ snowpark: 1.45.0 ├─ pandas: 2.3.3 └─ ... (3 more lines) * All packages are available. Let me set up the session helper and start the EDA. ✓ BASH ls /Users/cnantasenamat/fraud-eda/ ├─ __pycache__ ├─ benchmark_inference.py └─ ... (15 more lines) ✓ BASH cortex ctx step done s829d ├─ ✓ Completed step [s829d] Setup environment and check packages └─ → Next ready [sf8d7] Load data and basic stats * Now let me write and run the EDA script incrementally. First, load the data and get basic stats. ✓ WRITE /Users/cnantasenamat/fraud-eda/eda_step1_load.py +47 ✓ BASH cd /Users/cnantasenamat/fraud-eda && SNOWFLAKE_CONNECTION_NAME=devrel .venv/bin/python eda_step1_... ├─ === SHAPE === ├─ Rows: 10000, Columns: 6 └─ ... (54 more lines)
With the data loaded (10K rows, 0.5% fraud rate), it proceeds through amount distribution, location patterns, merchant analysis, and time-of-day analysis — generating plots at each step:
✓ BASH cortex ctx step done sf8d7 ├─ ✓ Completed step [sf8d7] Load data and basic stats └─ → Next ready [s9e83] Analyze amount distribution by fraud label ✓ WRITE /Users/cnantasenamat/fraud-eda/eda_step2_amount.py +90 ✓ BASH cd /Users/cnantasenamat/fraud-eda && SNOWFLAKE_CONNECTION_NAME=devrel .venv/bin/python eda_step2_... ├─ === AMOUNT OVERLAP ANALYSIS === ├─ Fraud with amount < $500 (normal-looking): 17 / 50 (34%) └─ ... (15 more lines) ✓ READ /Users/cnantasenamat/fraud-eda/plots/amount_analysis.png ✓ BASH cortex ctx step done s9e83 ├─ ✓ Completed step [s9e83] Analyze amount distribution by fraud label └─ → Next ready [sa687] Analyze location patterns ✓ WRITE /Users/cnantasenamat/fraud-eda/eda_step3_location_merchant.py +135 ✓ BASH cd /Users/cnantasenamat/fraud-eda && SNOWFLAKE_CONNECTION_NAME=devrel .venv/bin/python eda_step3_... ├─ === LOCATION ANALYSIS === ├─ Unusual locations in Fraud: 26 / 50 (52%) └─ ... (46 more lines) ✓ READ /Users/cnantasenamat/fraud-eda/plots/location_merchant_analysis.png ✓ BASH cortex ctx step done sa687 && cortex ctx step done s8090 ├─ ✓ Completed step [sa687] Analyze location patterns ├─ → Next ready [s8090] Analyze merchant patterns └─ ... (2 more lines)
Finally, it compiles all findings into a ranked feature recommendation table:
✓ BASH cd /Users/cnantasenamat/fraud-eda && python3 eda_step5_features.py ├─ ====================================================================== ├─ FEATURE ENGINEERING RECOMMENDATIONS └─ ... (29 more lines) ✓ READ /Users/cnantasenamat/fraud-eda/plots/feature_recommendations.png ✓ BASH cortex ctx step done s9ab6 && cortex ctx task done ├─ ✓ Completed step [s9ab6] Feature engineering recommendations ├─ ✓ Completed task task-1c4e8787: EDA on Fraud_ML_TEST_ONLINE and recommend fea... └─ ... (1 more line)
EDA Complete -- Feature Recommendations Recommended Features (ranked by signal strength) ┌───────┬────────────────────────┬──────────────────────┬─────────────┬───┐ │ # │ Feature │ Type │ Correlation │ … │ ├───────┼────────────────────────┼──────────────────────┼─────────────┼───┤ │ 1 │ IS_UNUSUAL_LOCATION │ Binary │ 0.28 │ … │ ├───────┼────────────────────────┼──────────────────────┼─────────────┼───┤ │ 2 │ IS_SUSPICIOUS_MERCHANT │ Binary │ 0.27 │ … │ ├───────┼────────────────────────┼──────────────────────┼─────────────┼───┤ │ 3 │ LOC_MERCH_SUSPICIOUS │ Binary (interaction) │ 0.21 │ … │ ├───────┼────────────────────────┼──────────────────────┼─────────────┼───┤ │ 4 │ IS_NIGHT │ Binary │ 0.16 │ … │ ├───────┼────────────────────────┼──────────────────────┼─────────────┼───┤ │ 5 │ LOG_AMOUNT │ Continuous │ 0.12 │ … │ ├───────┼────────────────────────┼──────────────────────┼─────────────┼───┤ │ 6 │ HOUR │ Integer (0-23) │ -0.03 │ … │ ├───────┼────────────────────────┼──────────────────────┼─────────────┼───┤ │ 7 │ SUSPICIOUS_SCORE │ Integer (0-3) │ 0.32 │ … │ └───────┴────────────────────────┴──────────────────────┴─────────────┴───┘ Key EDA Findings 1. Class imbalance: 50 fraud out of 10,000 (0.5%). Use SMOTE, class weights, or threshold tuning. 2. Amount: Fraud averages $4,819 vs legit $315, but 34% of fraud has normal amounts (<$500) -- amount alone is insufficient. 3. Location: Unusual foreign locations have 26.8% fraud rate vs 0.24% for common US locations, but 48% of fraud occurs at common locations (noise). 4. Merchant: Suspicious merchants have 53.8% fraud rate, but again ~48% of fraud uses normal merchants. 5. Time: Late night (0-5am) has 12.7% fraud rate (25x baseline). Day of week is not discriminative. 6. Interaction effects matter: The location x merchant combination is the strongest predictor (75% fraud when both flagged). Features to Skip • DAY_OF_WEEK: No meaningful signal (0.27%-0.77% fraud across days). • Raw AMOUNT: Use LOG_AMOUNT instead for better distribution separation. • Raw LOCATION/MERCHANT strings: High cardinality (42 each). Use the binary flags above instead, or frequency encoding if you want to preserve more info. Modeling Recommendations • Algorithm: Tree-based models (XGBoost, LightGBM) handle the feature interactions and imbalance well. • Evaluation metric: Use PR-AUC or F1 rather than accuracy (accuracy would be 99.5% by predicting all legit). • Handle imbalance: scale_pos_weight = 199 for XGBoost, or use SMOTE on the training set only. • The 7 features above should give a strong baseline. The composite SUSPICIOUS_SCORE (correlation 0.32) is the single most correlated feature with fraud.
Train the Model
With our features identified, we can now train a machine learning model. XGBoost is an excellent choice for fraud detection due to its ability to handle imbalanced datasets and capture complex patterns.
Prompt
Use features generated above to train a XGBoost machine learning model for real-time fraud detection.
What Gets Generated
Cortex Code will first create a feature engineering view with the 7 recommended features:
✓ SNOWFLAKE_SQL_EXECUTE Create feature engineering view SQL Result (1 rows × 1 cols) ┌─────┬────────────────────────────────┐ │ # │ status │ ├─────┼────────────────────────────────┤ │ 1 │ View FRAUD_FEATURES_V success. │ └─────┴────────────────────────────────┘
Then train the model and display results:
FRAUD DETECTION MODEL TRAINING ============================================================ Loading features from ML.PROJECTS.FRAUD_FEATURES_V... Loaded 10000 rows with 7 features Training XGBoost with scale_pos_weight=199... EVALUATION RESULTS ------------------------------------------------------------ precision recall f1-score support 0 1.00 0.97 0.98 1990 1 0.14 0.80 0.24 10 ROC-AUC Score: 0.9723 CONFUSION MATRIX ------------------------------------------------------------ [[1928 62] [ 2 8]] TOP FEATURES BY IMPORTANCE ------------------------------------------------------------ SUSPICIOUS_SCORE 0.2341 IS_UNUSUAL_LOCATION 0.1892 IS_SUSPICIOUS_MERCHANT 0.1456 LOG_AMOUNT 0.1203 LOC_MERCH_SUSPICIOUS 0.1098 IS_NIGHT 0.0812 HOUR 0.0698 5-FOLD CROSS-VALIDATION ------------------------------------------------------------ AUC scores: [0.9634, 0.9012, 0.9567, 0.8923, 0.9401] Mean AUC: 0.9307 (+/- 0.0352)
The model achieves a ROC-AUC of 0.9723 with 80% recall on fraud cases.
Deploy to Snowpark Container Services
Now comes the powerful part: deploying our trained model to production with a single prompt. This would traditionally require containerization, infrastructure setup, and API development.
Prompt
Use machine-learning skill to log the model into Snowflake Model Registry, deploy the model to SPCS and create a REST endpoint for online inference.
What Gets Generated
This single prompt triggers a complete deployment pipeline. First, the model is registered:
MODEL REGISTRATION ============================================================ Registering model as ML.PROJECTS.FRAUD_XGBOOST_MODEL version V2... ✓ SNOWFLAKE_SQL_EXECUTE Verify model V2 registration SQL Result (2 rows × 10 cols) ┌─────┬───────────────┬──────────────┬───────────────────┐ │ # │ name │ version_name │ min_num_arguments │ ├─────┼───────────────┼──────────────┼───────────────────┤ │ 1 │ PREDICT │ V2 │ 7 │ │ 2 │ PREDICT_PROBA │ V2 │ 7 │ └─────┴───────────────┴──────────────┴───────────────────┘
Then deployed to SPCS (this can take a few minutes as it builds the container):
✓ SNOWFLAKE_SQL_EXECUTE Check if service exists now SQL Result (3 rows × 28 cols) ┌─────┬─────────────────────────┬─────────┬───────────────┬─────────────┐ │ # │ name │ status │ database_name │ schema_name │ ├─────┼─────────────────────────┼─────────┼───────────────┼─────────────┤ │ 1 │ FRAUD_INFERENCE_SERVICE │ RUNNING │ ML │ PROJECTS │ │ 2 │ MODEL_BUILD_4A237CD4 │ DONE │ ML │ PROJECTS │ └─────┴─────────────────────────┴─────────┴───────────────┴─────────────┘ ✓ SNOWFLAKE_SQL_EXECUTE Get service endpoints for REST URL SQL Result (1 rows × 6 cols) ┌─────┬───────────┬───────┬──────────┬───────────┬────────────────────────────────────────────┐ │ # │ name │ port │ protocol │ is_public │ ingress_url │ ├─────┼───────────┼───────┼──────────┼───────────┼────────────────────────────────────────────┤ │ 1 │ inference │ 5000 │ HTTP │ true │ xk7rbf2q-ml-proj-aws-us-west-2.snowflakecomputing │ └─────┴───────────┴───────┴──────────┴───────────┴────────────────────────────────────────────┘
The service functions are tested via SQL:
✓ SNOWFLAKE_SQL_EXECUTE Accuracy check - confusion matrix via SPCS service SQL Result (4 rows × 3 cols) ┌─────┬────────┬───────────┬───────┐ │ # │ ACTUAL │ PREDICTED │ CNT │ ├─────┼────────┼───────────┼───────┤ │ 1 │ 0 │ 0 │ 9667 │ │ 2 │ 0 │ 1 │ 283 │ │ 3 │ 1 │ 0 │ 3 │ │ 4 │ 1 │ 1 │ 47 │ └─────┴────────┴───────────┴───────┘
The model achieves 94% fraud recall, catching 47 out of 50 fraud cases.
Run Real-Time Inference
With our REST API deployed, let's test it with realistic traffic and analyze the performance characteristics.
Prompt
Create 1000 sample requests with a mix of potential fraud and legit transactions and run the predictions using the internal REST API for online Inference running on SPCS and show the latency profile.
Note: Cortex Code may suggest a hybrid approach instead (e.g., using the Python SDK
mv.run()or SQL service functions). If that happens, prompt it with: "Ask for the PAT token so that you can call the REST API". It will then guide you through generating one via Snowsight and proceed with direct REST calls.
What Gets Generated
Cortex Code will generate 1,000 synthetic transactions, send them to the SPCS endpoint, and report a latency profile:
SPCS REST API Latency Profile Endpoint: https://xk7rbf2q-ml-proj-aws-us-west-2.snowflakecomputing.app/predict-proba Total requests: 1,000 | Errors: 0 | Wall time: 63.8s | Throughput: 15.7 req/s Latency Percentiles ┌──────────────┬──────────────┐ │ Metric │ Latency (ms) │ ├──────────────┼──────────────┤ │ Min │ 39.8 │ ├──────────────┼──────────────┤ │ Mean │ 63.7 │ ├──────────────┼──────────────┤ │ Median (P50) │ 59.6 │ ├──────────────┼──────────────┤ │ P75 │ 65.3 │ ├──────────────┼──────────────┤ │ P90 │ 73.4 │ ├──────────────┼──────────────┤ │ P95 │ 81.5 │ ├──────────────┼──────────────┤ │ P99 │ 141.9 │ ├──────────────┼──────────────┤ │ Max │ 482.9 │ ├──────────────┼──────────────┤ │ Std Dev │ 32.6 │ └──────────────┴──────────────┘ Latency Distribution • 86.3% of requests completed in 50-100ms • 11.5% completed under 50ms • Only 2.2% exceeded 100ms (mostly outliers)
It also breaks down the predictions by risk tier so you can verify the model behaves as expected:
Prediction Results ┌───────────┬───────┬──────────────┬──────────────┬──────────────┐ │ Risk Tier │ Count │ Avg P(fraud) │ Flagged >50% │ Flagged >10% │ ├───────────┼───────┼──────────────┼──────────────┼──────────────┤ │ HIGH │ 60 │ 0.9945 │ 60 (100%) │ 60 (100%) │ ├───────────┼───────┼──────────────┼──────────────┼──────────────┤ │ MEDIUM │ 135 │ 0.0970 │ 10 (7%) │ 24 (18%) │ ├───────────┼───────┼──────────────┼──────────────┼──────────────┤ │ LOW │ 805 │ 0.0077 │ 4 (0.5%) │ 9 (1%) │ └───────────┴───────┴──────────────┴──────────────┴──────────────┘ Key Takeaways • ~60ms median latency per individual REST request — suitable for real-time fraud screening • Latency is consistent across risk tiers (no payload-dependent variation) • The model correctly flags 100% of high-risk transactions and has very low false-positive rates on low-risk ones • With connection pooling, throughput reaches ~15.7 req/s on a single thread. Parallel requests would scale linearly.
Debug and Recover from Errors
During any natural language coding session, errors are inevitable. The great thing about Cortex Code is its ability to self-correct by assessing the situation, environment, and error to fix issues automatically.
Common Scenarios
Missing Python Binary
When python isn't found on the system, Cortex Code locates python3 and retries automatically.
✓ WRITE /Users/cnantasenamat/fraud-eda/eda_step4_time.py +103 × BASH cd /Users/cnantasenamat/fraud-eda && python eda_step4_time.py └─ bash: python: command not found ✓ BASH which python3 └─ /usr/local/bin/python3 ✓ BASH cd /Users/cnantasenamat/fraud-eda && python3 eda_step4_time.py ├─ === TIME OF DAY ANALYSIS === ├─ Night transactions (midnight-7am) in Fraud: 10 / 50 (20%) └─ ... (23 more lines)
Notebook Execution Issues
When a Jupyter kernel fails to connect, Cortex Code detects the issue and finds an alternative approach.
× NOTEBOOK:EXECUTE_CELL fraud_eda.ipynb cell 1 └─ Cell execution failed (Jupyter server may have stopped): Tool error: Error executing tool exe... * The notebook tool keeps using the hostname which can't be resolved. Let me take a different approach - I'll build the complete notebook as an .ipynb file directly, then upload it to Snowflake. This is more reliable. ✓ BASH python3 -c "import json; cells = []..." └─ ... (334 more lines) * Notebook file built with 16 cells. Let me mark the completed steps and upload to Snowflake.
Artifact Upload Failures
If an upload fails due to missing resources, Cortex Code investigates and finds a working path.
× BASH cortex artifact create notebook fraud_eda ... --location ML.PROJ... └─ {"error":"Error creating artifact: Upload failed: SQL compilation error:\nSch... * The notebook needs a workspace. Let me check what's available and create one if needed. ✓ SNOWFLAKE_SQL_EXECUTE List existing workspaces SQL Result (1 rows × 7 cols) │ 1 │ DEFAULT$ │ USER$DATAPROFESSOR │ PUBLIC │ * There's a default workspace. Let me upload to it. ✓ BASH cortex artifact create notebook fraud_eda ... --connection trial... └─ "success": true
Model Training Errors
When training fails due to parameter issues, Cortex Code diagnoses and fixes the problem automatically.
× BASH SNOWFLAKE_CONNECTION_NAME=trial ... train_fraud_model.py └─ ... AttributeError: best_iteration requires early stopping ... * Need to fix: best_iteration requires early stopping. Let me fix the script. ✓ EDIT /Users/cnantasenamat/fraud-eda/train_fraud_model.py +12 -12 * Running the fixed training script. ✓ BASH SF_SKIP_WARNING_FOR_READ_PERMISSIONS_ON_CONFIG_FILE=true ... ├─ FRAUD DETECTION MODEL TRAINING ├─ ROC-AUC Score: 0.9723 └─ ...
Best Practices
- Start with
ACCOUNTADMINfor initial setup, then create dedicated roles - Monitor compute pool resources during deployment
- Review Cortex Code's explanations when it makes corrections
Conclusion And Resources
Congratulations! You've successfully built a complete real-time fraud detection model using only a handful of natural language prompts in Snowflake ML.
What You Learned
- Generate realistic synthetic fraud data with natural language prompts
- Perform comprehensive exploratory data analysis with automated Python scripts
- Train an XGBoost model optimized for imbalanced fraud detection
- Deploy models to SPCS with automatic containerization
- Create and test REST API endpoints for real-time inference
Related Resources
Web pages:
- Snowflake ML - Integrated set of capabilities for development, MLOps and inference leading with agentic ML
- Snowflake Notebooks - Jupyter-based notebooks in Snowflake Workspaces
- Cortex Code - Snowflake’s AI native coding agent that boosts ML productivity
Technical Documentation:
- Snowflake ML Documentation - Official Snowflake ML developer guide
- Cortex Code Documentation - Getting started with Cortex Code
- Snowpark Container Services - Deploy and manage containerized workloads
- Snowflake Model Registry - Register, version, and deploy ML models
This content is provided as is, and is not maintained on an ongoing basis. It may be out of date with current Snowflake instances