A RetroSearch Logo

Home - News ( United States | United Kingdom | Italy | Germany ) - Football scores

Search Query:

Showing content from https://docs.databricks.com/aws/en/machine-learning/automl/forecasting below:

Forecasting with AutoML (classic compute)

Forecasting with AutoML (classic compute)

Use AutoML to automatically finding the best forecasting algorithm and hyperparameter configuration to predict values based on time-series data.

Time series forecasting is only available for Databricks Runtime 10.0 ML or above.

Set up forecasting experiment with the UI​

You can set up a forecasting problem using the AutoML UI with the following steps:

  1. In the sidebar, select Experiments.
  2. In the Forecasting card, select Start training.

The forecasting UI defaults to serverless forecasting. To access forecasting with your own compute, select revert back to the old experience.

Configure the AutoML experiment​
  1. The Configure AutoML experiment page displays. On this page, you configure the AutoML process, specifying the dataset, problem type, target or label column to predict, metric to use to evaluate and score the experiment runs, and stopping conditions.

  2. In the Compute field, select a cluster running Databricks Runtime 10.0 ML or above.

  3. Under Dataset, click Browse. Navigate to the table you want to use and click Select. The table schema appears.

  4. Click in the Prediction target field. A dropdown menu appears, listing the columns shown in the schema. Select the column you want the model to predict.

  5. Click in the Time column field. A drop-down appears showing the dataset columns that are of type timestamp or date. Select the column containing the time periods for the time series.

  6. For multi-series forecasting, select the column(s) that identify the individual time series from the Time series identifiers drop-down. AutoML groups the data by these columns as different time series and trains a model for each series independently. If you leave this field blank, AutoML assumes that the dataset contains a single time series.

  7. In the Forecast horizon and frequency fields, specify the number of time periods into the future for which AutoML should calculate forecasted values. In the left box, enter the integer number of periods to forecast. In the right box, select the units.

    note

    To use Auto-ARIMA, the time series must have a regular frequency where the interval between any two points must be the same throughout the time series. The frequency must match the frequency unit specified in the API call or in the AutoML UI. AutoML handles missing time steps by filling in those values with the previous value.

  8. In Databricks Runtime 11.3 LTS ML and above, you can save prediction results. To do so, specify a database in the Output Database field. Click Browse and select a database from the dialog. AutoML writes the prediction results to a table in this database.

  9. The Experiment name field shows the default name. To change it, type the new name in the field.

You can also:

Advanced configurations​

Open the Advanced Configuration (optional) section to access these parameters.

Run the experiment and monitor the results​

To start the AutoML experiment, click Start AutoML. The experiment starts to run, and the AutoML training page appears. To refresh the runs table, click .

View experiment progress​

From this page, you can:

With Databricks Runtime 10.1 ML and above, AutoML displays warnings for potential issues with the dataset, such as unsupported column types or high cardinality columns.

note

Databricks does its best to indicate potential errors or issues. However, this may not be comprehensive and may not capture the issues or errors you may be searching for.

To see any warnings for the dataset, click the Warnings tab on the training page or the experiment page after the experiment completes.

View results​

When the experiment completes, you can:

To return to this AutoML experiment later, find it in the table on the Experiments page. The results of each AutoML experiment, including the data exploration and training notebooks, are stored in a databricks_automl folder in the home folder of the user who ran the experiment.

Register and deploy a model​

Register and deploy your model using the AutoML UI. When a run completes, the top row shows the best model based on the primary metric.

  1. Select the link in the Models column for the model you want to register.
  2. Select to register it to Unity Catalog or Model Registry.

    note

    Databricks recommends you register models to Unity Catalog for the latest features.

  3. After registration, you can deploy the model to a custom model serving endpoint.
No module named 'pandas.core.indexes.numeric​

When serving a model built using AutoML with Model Serving, you may get the error: No module named 'pandas.core.indexes.numeric.

This is due to an incompatible pandas version between AutoML and the model serving endpoint environment. You can resolve this error by running the add-pandas-dependency.py script. The script edits the requirements.txt and conda.yaml for your logged model to include the appropriate pandas dependency version: pandas==1.5.3

  1. Modify the script to include the run_id of the MLflow run where your model was logged.
  2. Re-register the model to Unity Catalog or the model registry.
  3. Try serving the new version of the MLflow model.
Next steps​

RetroSearch is an open source project built by @garambo | Open a GitHub Issue

Search and Browse the WWW like it's 1997 | Search results from DuckDuckGo

HTML: 3.2 | Encoding: UTF-8 | Version: 0.7.4