Note
Access to this page requires authorization. You can try signing in or changing directories.
Access to this page requires authorization. You can try changing directories.
Important
This feature is in preview.
Microsoft Fabric lets you serve real-time predictions from ML models with secure, scalable, and easy-to-use online endpoints. These endpoints are available as built-in properties of most Fabric models—and they require no setup to kick off fully managed real-time deployments.
You can activate, configure, and query model endpoints with a public-facing REST API. You can also get started directly from the Fabric interface, using a low-code experience to activate model endpoints and preview predictions instantly.
Prerequisites
- To serve real-time predictions, your administrator needs to enable the tenant switch for ML model endpoints in the Fabric admin portal.
Limitations
- Endpoints are currently available for a limited set of ML model flavors, including Keras, LightGBM, Sklearn, and XGBoost.
- Endpoints currently aren't available for models with tensor-based schemas or no schemas.
- Endpoints currently aren't available for models with dependencies on private or internal packages.
Get started with model endpoints
ML models in Fabric come prebuilt with online endpoints that can be used to serve real-time predictions. Each registered model version has a dedicated endpoint URL, which can be found under the "Endpoint details" heading in the Fabric interface. This URL ends with a subpath designating that specific version (for instance, /versions/1/score
).
Model endpoints have the following properties:
Property | Description | Default |
---|---|---|
Default version | This property (Yes or No ) indicates whether the version is set as the model's default for serving real-predictions. You can customize the default version in the model's settings. |
No |
Status | This property indicates whether the endpoint is ready to serve predictions. The status can be Inactive , Activating , Active , Deactivating , or Failed . Only active endpoints can serve predictions. |
Inactive |
Auto sleep | This property (On or Off ) indicates whether the endpoint, once active, should scale down capacity usage to zero in the absence of traffic. If auto sleep is on, then the endpoint enters an idle state after five minutes without incoming requests. The first call to wake up an idle endpoint involves a short delay. |
On |
Activate model endpoints
You can activate model endpoints directly from the Fabric interface. Navigate to the version that you’d like to serve real-time predictions and select "Activate version endpoint" from the ribbon.
A toast message shows that Fabric is getting your endpoint ready to serve predictions, and the status of the endpoint changes to "Activating." Behind the scenes, Fabric spins up the underlying container infrastructure to host your model. Within a few minutes, your endpoint is ready to serve predictions.
Every endpoint has a status indicating whether it’s ready to serve real-time predictions:
Status | Description |
---|---|
Inactive |
The endpoint isn't activated to serve real-time predictions, and it’s not consuming Fabric capacity. |
Activating |
The endpoint is being configured to serve real-time predictions. Behind the scenes, Fabric sets up the underlying container infrastructure to host the model. Within a few minutes, the endpoint is active. |
Active |
The endpoint is ready to serve real-time predictions. Behind the scenes, Fabric manages the underlying infrastructure, scaling up resource usage based on incoming traffic. Higher traffic results in higher Fabric capacity usage. |
Deactivating |
The endpoint is being deactivated, so that it no longer serves real-time predictions or consumes Fabric capacity. Behind the scenes, Fabric dismantles the underlying container infrastructure. |
Note
ML models can support active endpoints for up to five versions at once. To serve predictions from a sixth version, you must first deactivate an active endpoint.
Manage model endpoints
For an overview of your model's active endpoints, select "Manage endpoints" from the ribbon in the interface. Every model has a customizable default endpoint, which serves predictions from a version that you choose. You can update the default version using the dropdown selector in the settings pane.
Important
Be sure to set the default property to an active version if you plan to use it. If the default property isn't set, or is set to an inactive version, then calls to the default endpoint fail.
All versions with active endpoints are listed under the model’s endpoint settings. You can modify the auto sleep property of each endpoint by toggling the switcher to be "On" or "Off."
Tip
Active endpoints with auto sleep turned on enter an idle state after five minutes without traffic, and the first call to wake them up involves a short delay. You may want to turn off this property for endpoints in production.
Query model endpoints for real-time predictions
Model endpoints are available for instant testing with a low-code experience in Fabric. Navigate to a version with an active endpoint and select "Preview predictions" from the ribbon in the interface. You can send sample requests to the endpoint—and get sample predictions in real-time—using form fields that match the model’s input signature.
To populate the form fields with random sample values, select "Autofill." You can add more sets of form values to test the endpoint with multiple inputs. Select "Get predictions" to send the endpoint your sample request.
If you prefer to format sample requests as JSON payloads, use the dropdown selector to change the view.
Deactivate model endpoints
You can deactivate model endpoints directly from the Fabric interface. Navigate to a version that you no longer need to serve real-time predictions and select "Deactivate version endpoint" from the ribbon in the interface.
A toast message shows that Fabric is dismantling your active deployment, and the status of the endpoint changes to "Deactivating." The endpoint is no longer able to serve real-time predictions unless you reactivate it.
You can deactivate endpoints for multiple versions at once from the model’s settings pane. Select "Manage endpoints" from the ribbon in the interface and choose one or more active endpoints to deactivate.
Related content
- Manage and query endpoints programmatically with the ML model endpoint REST API.
- Generate batch predictions with the
PREDICT
function in Fabric notebooks. - Learn more about model training and experimentation in Fabric.
- Did we miss a feature you need? Suggest it on the Fabric Ideas forum.