A tool for running SQL queries that contain Python scripts. This quickstart uses Azure Data Studio.
Create a stored procedure to generate random numbers
For simplicity, let's use the Python numpy package, that's installed and loaded by default. The package contains hundreds of functions for common statistical tasks, among them the random.normal function, which generates a specified number of random numbers using the normal distribution, given a standard deviation and mean.
For example, the following Python code returns 100 numbers on a mean of 50, given a standard deviation of 3.
numpy.random.normal(size=100, loc=50, scale=3)
To call this line of Python from T-SQL, add the Python function in the Python script parameter of sp_execute_external_script. The output expects a data frame, so use pandas to convert it.
EXECUTE sp_execute_external_script @language = N'Python'
, @script = N'
import numpy
import pandas
OutputDataSet = pandas.DataFrame(numpy.random.normal(size=100, loc=50, scale=3));
'
, @input_data_1 = N' ;'
WITH RESULT SETS(([Density] FLOAT NOT NULL));
What if you'd like to make it easier to generate a different set of random numbers? You define a stored procedure that gets the arguments from the user, then pass those arguments into the Python script as variables.
CREATE PROCEDURE MyPyNorm (
@param1 INT
, @param2 INT
, @param3 INT
)
AS
EXECUTE sp_execute_external_script @language = N'Python'
, @script = N'
import numpy
import pandas
OutputDataSet = pandas.DataFrame(numpy.random.normal(size=mynumbers, loc=mymean, scale=mysd));
'
, @input_data_1 = N' ;'
, @params = N' @mynumbers int, @mymean int, @mysd int'
, @mynumbers = @param1
, @mymean = @param2
, @mysd = @param3
WITH RESULT SETS(([Density] FLOAT NOT NULL));
The first line defines each of the SQL input parameters that are required when the stored procedure is executed.
The line beginning with @params defines all variables used by the Python code, and the corresponding SQL data types.
The lines that immediately follow map the SQL parameter names to the corresponding Python variable names.
Now that you've wrapped the Python function in a stored procedure, you can easily call the function and pass in different values, like this:
Python packages provide a variety of utility functions for investigating the current Python environment. These functions can be useful if you're finding discrepancies in the way your Python code performs in SQL Server and in outside environments.
For example, you might use system timing functions in the time package to measure the amount of time used by Python processes and analyze performance issues.
EXECUTE sp_execute_external_script
@language = N'Python'
, @script = N'
import time
start_time = time.time()
# Run Python processes
elapsed_time = time.time() - start_time
'
, @input_data_1 = N' ;';
Next steps
To create a machine learning model using Python with SQL machine learning, follow this quickstart:
Manage data ingestion and preparation, model training and deployment, and machine learning solution monitoring with Python, Azure Machine Learning and MLflow.
In this quickstart, you'll create and train a predictive model using Python. You'll save the model to a table in your database, and then use the model to predict values from new data with SQL machine learning.
Run a set of simple Python scripts using Machine Learning Services on SQL Server, Big Data Clusters, or Azure SQL Managed Instances. Learn how to use the stored procedure sp_execute_external_script to execute the script.