Release Notes


Release date: June 11, 2024

  • Fixed the bug that led to timeout when loading config files during jupyterlab load.

  • Fixed bugs and introduced enhancements following our recent release.


Release date: June 5, 2024

  • Support for Bring Your Own Model (BYOM) via AI Quick Actions.

  • Introduced enhancements following our recent release.


Release date: April 24, 2024

  • Fixed bugs and introduced enhancements following our recent release.


Release date: April 24, 2024

  • Fixed bugs, added realm compatibility check, improved logging and error handling, and added additional telemetry for Aqua.

  • Upgraded oci version to 2.125.3


Release date: April 18, 2024

  • Fixed the bug in ADSDataset.show_in_notebook().

  • Updated langchain version.


Release date: April 3, 2024

  • Fixed bugs and introduced enhancements following our recent release, which included internal adjustments for future features and updates for the Jupyter Lab 3 upgrade.


Release date: March 25, 2024

  • Fixed bugs and introduced enhancements following our recent release, which included internal adjustments for future features and updates for the Jupyter Lab 3 upgrade.


Release date: March 25, 2024

  • Fixed bugs and introduced enhancements following our recent release, which included internal adjustments for future features and updates for the Jupyter Lab 3 upgrade.


Release date: March 22, 2024

  • Fixed bugs and introduced enhancements following our recent release, which included internal adjustments for future features and updates for the Jupyter Lab 3 upgrade.


Release date: March 21, 2024

  • Fixed bugs and introduced enhancements following our recent release, which included internal adjustments for future features and updates for the Jupyter Lab 3 upgrade.


Release date: March 20, 2024

  • Internal changes to support upcoming features and changes in Notebook related to Jupyter Lab 3 upgrade.


Release date: February 7, 2024

  • Releasing v1 of the Anomaly Detection Operator! The Anomaly Detection Operator is a no-code Anomaly or Outlier Detection solution through the OCI Data Science Platform. It uses dozens of models from Oracle’s own proprietary research and the best of open source. See the Anomaly Detection Section of the AI Operators tab for full details (link).

  • Releasing a new version of the Forecast Operator. This release has faster explainability, improved support for reading from databases, upgrades to the automatic reporting, improved parallelization across all models, and an ability to save models for deferred inference. See the Forecast Section of the AI Operators tab for full details (link).

  • Change to the default signer such that it now defaults to resource_prinicpal on any OCI Data Science resource (for example, jobs, notebooks, model deployments, dataflow).


Release date: January 24, 2024

  • Improved the progress bar to use the percentage completed of workflow request instead of hardcoded steps.

  • Used the service default for WEB_CONCURRENCY for model deployment.

  • Fixed the bug with zipping the model artifacts directory when TMPRDIR is provided.

  • Improved the watch() method for model deployment to keep streaming logs when the deployment is finished.

  • Changed the default log type of watch to both access logs and predict logs.

  • Changed the target directory to artifact_dir instead of temp directory when saving the model artifacts.

  • Fixed the mount file system pre-check to check for duplicate dest.

  • Fixed duplicate logs in the model deployment consolidated logs.

  • Added support for the optional downloading of artifacts in GenericModel using a download_artifact() method.

  • Set the Data Science service endpoint through the environment variable in OCIDataScienceMixin.

  • Made reloading the model to environment as optional at the time of invoking GenericModel.from_id().

  • Mandated the Python version in GenericModel.prepare() when it can’t be resolved.

  • Added a print out of the model deployment OCID in the notebook cell when deploy() is called.


Release date: December 6, 2023

  • Added support for deploying LangChain application as OCI Model Deployment.

  • Added support for using HuggingFace Evaluation as LLM guardrail.

  • Added deployment support for RetrievalQA when using OpenSearchVectorSearch or FAISS vector DB as retriever.

  • Added reload parameters in to provide option to not reload

  • Fixed a bug in model deployment progress bar due to fixed number of steps.

  • Fixed a bug in ads opctl build-image job-local command.


Release date: November 16, 2023

  • Introducing AI Forecast Operator. Learn more about Operators in the “Operators” section of the docs.

  • Introducing PII Operator which aims to detect and redact Personal Identifiable Information in data.

  • Fixed a bug with the opctl conda create and opctl conda publish commands to ensure functionality on M1 and M2 local machines.

  • Fixed a bug with failed model deployment return value.

  • Fixed a bug when sorting logs for jobs and model deployment.


Release date: October 18, 2023

  • Added support to mount file systems in Data Science notebook sessions and jobs.

  • Added support to cancel all job runs in the ADS api and opctl commands.

  • Updated ads.set_auth() to use both config and signer when provided.

  • Fixed a bug when initializing distributed training artifacts with “Ray” framework.


Release date: September 27, 2023

  • Improved the LargeArtifactUploader class to understand OCI paths to upload model artifacts to the model catalog by reference.

  • Removed ADSDataset runtime dependency on geopandas.

  • Fixed a bug in the progress bar during model registration.

  • Fixed a bug where session variable could be referenced before assignment.

  • Fixed a bug with model artifact save.

  • Fixed a bug with pipelines step.


Release date: September 5, 2023

  • Upgraded the scikit-learn dependency to >=1.0.

  • Upgraded the pandas dependency to >1.2.1,<2.1 to allow you to use ADS with pandas 2.0.

  • Implemented multi-part upload in the ArtifactUploader to upload model artifacts to the model catalog.

  • Fixed the “Attribute not found” error, when deploy() called twice in GenericModel.

  • Fixed the fetch of the security token, when the relative path for the security_token_file is provided (used in session token-bases authentication).


Release date: July 27, 2023

  • Added PyTorchDistributed runtime option for Data Science jobs to add support for training large language models with PyTorch.

  • Added options to configure flexible shape in opctl.

  • Refactored deploy() in GenericModel to prioritize the parameters.

  • Fixed the opctl commands delete/cancel/watch/activate/deactivate commands to add missing parameter options.

  • Fixed the opctl commands to call run to start an ML job when no YAML is specified.

  • Deprecated the DatasetFactory class, and refactored the code.


Release date: June 22, 2023

  • Added support for leveraging pools in the Data Flow applications.

  • Added support for token-based authentication.

  • Revised help information for opctl commands.


Release date: June 13, 2023

  • Resolved an issue in ads opctl build-image job-local when the build of job-local would get stuck. Updated the Python version to 3.8 in the base environment of the job-local image.

  • Fixed a bug that prevented the support of defined tags for Data Science job runs.

  • Fixed a bug in the of ads opctl that attempted to create a temporary folder in the /var/folders directory.

  • Added support for defined tags in the Data Flow application and application run.

  • Deprecated the old ModelDeploymentProperties and ModelDeployer classes, and their corresponding APIs.

  • Enabled the uploading of large size model artifacts for the ModelDeployment class.

  • Implemented validation for shape name and shape configuration details in Data Science jobs and Data Flow applications.

  • Added the capability to create ADSDataset using the Pandas accessor.

  • Provided a prebuilt watch command for monitoring Data Science jobs with ads opctl.

  • Eliminated the legacy ads.dataflow package from ADS.


Release date: May 17, 2023

  • Added support for key_content attribute in ads.set_auth() for the API KEY authentication.

  • Fixed bug in ModelEvaluator when it returned incorrect ROC AUC characteristics.

  • Fixed bug in ADSDataset.suggest_recommendations() API, when it returned an error if the target wasn’t specified.

  • Fixed bug in ADSDataset.auto_transform() API, when an incorrect sampling was suggested for imbalanced data.


Release date: May 5, 2023

  • Added support for creating ADSDataset from pandas dataframe.

  • Added support for multi-model deployment using Triton.

  • Added support for model deployment local testing in ads opctl CLI.

  • Added support in ads opctl CLI to generate starter YAML specification for the Data Science Job, Data Flow Application, Data Science Model Deployment and ML Pipeline services.

  • Added support for invoking model prediction locally with predict(local=True).

  • Added support for attaching customized when preparing model.

  • Added status check for model deployment delete/activate/deactivate APIs.

  • Added support for training and verifying SparkPipelineModel in Dataflow.

  • Added support for generating for GPU model deployment.

  • Added support for setting defined tags in Data Science jobs.

  • Improved model deployment progress bar.

  • Fixed bug when using ads opctl CLI to run jobs locally.

  • Fixed bug in Dataflow magic when using archive_uri in dataflow config.


Release date: March 22, 2023

  • Added support for custom containers (Bring Your Own Container or BYOC) and environment variables for GenericModel.

  • Added default values for configuring parameters in ModelDeployment, such as default flex shape, ocpus, memory in gbs, bandwidth, and instance count.

  • Added support for NotebookRuntime to use directory as job artifact.

  • Added support for PythonRuntime and GitPythonRuntime to use shell script as entrypoint.


Release date: March 2, 2023

  • Remove support for Python 3.7.

  • Improved the DataScienceMode.create() to support timeout argument and auto extract region from the signer and signer config.

  • Support Jupyter Notebook as entrypoint when defining Data Science jobs with PythonRuntime and GitPythonRuntime.

  • Support environment variable substitution in Data Science job names and output URI.

  • Support JSON serialization of list/dictionary when assigning them as Data Science jobs environment variables.

  • Support saving the notebook to output URI even if the job run failed when running a Data Science job using NotebookRuntime.

  • Added method to Data Science job to load default values from environment.

  • Added DataScienceJob.fast_launch_shapes() method to list fast launch shapes available for Data Science job.

  • Added HuggingFacePipelineModel class to support prepare, save, deploy and predict for HuggingFace pipelines.

  • Updated Data Science job run YAML representation to include configurations inherited from the job.

  • Fixed custom conda environment not showing in Data Science Job YAML specification.

  • Fixed an issue where model saving was failing in notebook session without ipywidgets installed.

  • Fixed “Unknown archive format” error in, when the source code folder name ends with “zip”. List of supported archive files are: “zip”, “tar.gz”, “tar” and “tgz”.


Release date: February 16, 2023

  • Fixed a bug for ads opctl run when --auth flag is passed and image is built by ADS.

  • Fixed a bug in when the work requests are not successfully populated.

  • Fixed a bug in DataScienceModel.create() to when the provenance metadata is not provided.


Release date: January 25, 2023

  • Added support for the machine learning pipelines feature.

  • Fixed a bug in fetch_training_code_details(). When git commit is empty string, set it as None to avoid service error.

  • Fixed a bug in fetch_training_code_details(). Use the folder of training_script_path as the artifact directory, instead of ..


Release date: January 18, 2023

  • Added support for the model version set feature.

  • Added --job-info option to ads opctl run CLI to save job run information to a YAML file.

  • Added the AuthContext class. It supports API key configuration, resource principal, and instance principal authentication. In addition, predefined signers, callable signers, or API keys configurations from specified locations.

  • Added restart_deployment() method to the framework-specific classes. Update model deployment associated with the model.

  • Added activate() and deactivate() method to the model deployment classes.

  • Fixed a bug in to_sql(). The string length for the column created in Oracle Database table was counting characters, not bytes.

  • Fixed a bug where any exception that occurred in a notebook cell printed “ADS Exception” even if the ADS code was not responsible for the error.


Release date: December 20, 2022

  • Fixed a bug in ADS jobs. The method sometimes threw an exception due to an unexpected logging parameter.


Release date: December 14, 2022


Release date: December 7, 2022

  • Fixed a bug in GenericModel.prepare. The .model-ignore file was not included in the


Release date: December 7, 2022

  • Added compatibility with Python 3.10.

  • Added update_deployment() method to the framework-specific classes. Update model deployment associated with the model.

  • Added from_id() method to the framework-specific classes. Load existing model by OCID directly from the OCI Models and OCI Model Deployment.

  • Added upload_artifact() to the framework-specific classes. Upload model artifacts to Object Storage.

  • Added update() method to the framework-specific classes. Update the model metadata for the registered model.

  • Added config, signer, signer_callable attributes to the ads.set_auth() to support additional signers.

  • Added support for Instance Principals authentication for the ads opctl conda publish and ads opctl conda install commands.

  • Added an option for PyTorchModel framework allowing to serialize model in a TorchScript format.

  • Added an option to import framework-specific classes directly from the ads.model package. Example: from ads.model import LightGBMModel, GenericModel.

  • Fixed a bug in ADSDataset get_recommendations when imbalanced correction depends on classes alpha order.

  • Fixed a bug in ADS jobs. The shape configuration details were incorrectly extracted from a notebook session.

  • Fixed a bug to replace the use of a deprecated API with latest API in the Model Evaluation module.

Following modules are marked as deprecated:






Release date: October 29, 2022

  • Fixed a bug in ads.dataset.helper to support Python 3.8 and Python 3.9.


Release date: October 27, 2022

  • Fixed a bug in PyTorchModel. The failed when torch.Tensor was used as input data.

  • Fixed a bug in ads opctl conda publish command.

  • Added support for flexible shapes for Data Flow Jobs.

  • Loading a model from Model Catalog (GenericModel.from_model_catalog()) and Model Deployment (GenericModel.from_model_deployment()) no longer requires a model file name.

  • Switched from using cx_Oracle interface to the oracledb driver to connect to Oracle Databases.

  • Added support for image attribute for the PyTorchModel.predict() and TensorFlowModel.predict() methods. Images can now be directly passed to the model Deployment predict.

The following APIs are deprecated:

  • OracleAutoMLProvider


Release date: October 7, 2022

  • Added SparkPipelineModel model serialization class for fast and easy model deployment.

  • Added support for flexible shapes for Jobs and Model Deployments.

  • Added support for freeform_tags and defined_tags for Model Deployments.

  • Added the populate_schema() method to the GenericModel class. Populate input and output schemas for model artifacts.

  • The ADSString was added to the Feature types system. Use the enhanced string class functionalities such as regular expression (RegEx) matching and natural language parsing within Pandas dataframes and series.

  • Saving model does not require iPython dependencies

Following APIs are deprecated:


  • ADSModel.prepare

  • ads.common.model_export_util.prepare_generic_model


Release date: September 16, 2022

  • OCI SDK updated from version 2.59.0 to version 2.82.0.


Release date: September 14, 2022

  • Added support for large models with artifact size between 2 and 6 GB. The large models can be saved to the Model Catalog, downloaded from the Model Catalog, and deployed as a Model Deployment resource.

  • Added delete() method to the GenericModel class. Deletes models and associated model deployments.

  • The Model Input Schema is improved to return features sorted by the order attribute.

  • Added user-friendly default names for created Jobs, Model Deployments, and Models.


Release date: August 4, 2022

  • Deprecated the ads.dataflow.DataFlow class. It has been superseded by the class.

  • Added prepare_save_deploy() method to the GenericModel class. Prepare model artifacts and deploy the model with one command.

  • Added support for binary payloads in model deployment.

  • Updated AutoMLModel, GenericModel, LightgbmModel, PyTorchModel, SklearnModel, TensorflowModel, and XgboostModel classes to support binary payloads in model deployment.

  • The maximum runtime for a Job can be limited with the with_maximum_runtime_in_minutes() method in the CondaRuntime, DataFlowNotebookRuntime, DataFlowRuntime, GitPythonRuntime, NotebookRuntime, and ScriptRuntime classes.

  • The class supports Published conda environments.


Release date: June 21, 2022

  • Added from_model_deployment() method to the GenericModel class. Now you can load a model directly from an existing model deployment.

  • Moved dependencies from being default into optional installation groups:

    • all-optional

    • bds

    • boosted

    • data

    • geo

    • notebook

    • onnx

    • opctl

    • optuna

    • tensorflow

    • text

    • torch

    • viz

    Use python3 -m pip install "oracle-ads[XXX]" where XXX are the group names.


Release date: June 1, 2022

  • Added support for running a container as jobs using

  • The ModelArtifact class is deprecated. Use the model serialization classes (GenericModel, PyTorchModel, SklearnModel, etc.).


Release date: May 6, 2022

  • Added BDSSecretKeeper to store and save configuration parameters to connect to Big Data service to the vault.

  • Added the krbcontext and refresh_ticket functions to configure Kerberos authentication for the Big Data service.

  • Added authentication options to logging APIs to allow you to pass in the OCI API key configuration or signer.

  • Added the configuration file path option in the set_auth method. This allows you to change the path of the OCI configuration.

  • Fixed a bug in AutoML for Text datasets.

  • Fixed bug in import to notify users installing ADS optional dependencies.

  • Fixed a bug in the generated file, where Pandas dataframe’s dtypes changed when deserializing. Now you can recover it from the input schema.

  • Updated requirements to oci>=2.59.0.


Release date: April 4, 2022

  • Added framework-specific model serialization to add more inputs to the generated file.

  • Added the following framework-specific classes for fast and easy model deployment:

    • AutoMLModel

    • SKlearnModel

    • XGBoostModel

    • LightGBMModel

    • PyTorchModel

    • TensorFlowModel

  • Added the GenericModel class for frameworks not included in the preceding list:

  • You can now prepare, verify, save and deploy your models using the methods in these new classes:

    • .prepare(): Creates, runtime.yaml, and schema files for model deployment purpose, and adds the model artifacts to the model catalog.

    • .verify(): Helps test your model locally, before deploying it from the model catalog to an endpoint.

    • .save(): Saves the model and model artifacts to the model catalog.

    • .deploy(): Deploys a model from the model catalog to a REST endpoint.

    • .predict(): Calls the endpoint and creates inferences from the deployed model.

  • Added support to create jobs with managed egress.

  • Fixed bug in jobs, where log entries were being dropped when there were a large number of logs in a short period of time. Now you can list all logs with jobwatch().


Release date: March 3, 2022

  • Fixed bug in automatic extraction of taxonomy metadata for Sklearn models.

  • Fixed bug in jobs NotebookRuntime when using non-ASCII encoding.

  • Added compatibility with Python 3.8 and 3.9.

  • Added an enhanced string class, called ADSString. It adds functionality such as regular expression (RegEx) matching, and natural language processing (NLP) parsing. The class can be expanded by registering custom plugins to perform custom string processing actions.


Release date: February 4, 2022

  • Fixed bug in Data Flow Job creation.

  • Fixed bug in ADSDataset get_recommendations raising HTML is not defined exception.

  • Fixed bug in jobs ScriptRuntime causing the parent artifact folder to be zipped and uploaded instead of the specified folder.

  • Fixed bug in ModelDeployment raising TypeError exception when updating an existing model deployment.


Release date: January 21, 2022

  • Added support for the storage_options parameter in ADSDataset .to_hdf().

  • Fixed error message to specify overwrite_script or overwrite_archive option in data_flow.create_app().

  • Fixed output of multiclass evaluation plots when ADSEvaluatior() class uses a non-default legend_labels option.

  • Added support to connect to an Oracle Database that does not require a wallet file.

  • Added support to read and write from MySQL using ADS DataFrame APIs.


Release date: December 9, 2021

  • Fixed bug in model artifact prepare(), reload(), and prepare_generic_model() raising ONNXRuntimeError caused by the mismatched version of skl2onnx.


Release date: December 3, 2021

The following features were added:

  • Added support to read exported dataset from the consolidated export file for the Data Labeling service.

Following fixes were added:

  • The DaskSeries class was marked as deprecated.

  • The DaskSeriesAccessor class was marked as deprecated.

  • The MLRuntime class was marked as deprecated.

  • The ADSDataset.ddf attribute was marked as deprecated.


Release date: November 29, 2021

The following features were added:

  • Moved fastavro, pandavro and openpyxl to an optional dependency.

  • Added the ability to specify the output annotation format to be spacy for the Entity Extraction dataset or yolo for the Object Detection dataset in the Data Labeling service.

  • Added support to load labeled datasets from OCI Data Labeling, and return the Pandas dataframe or generator formats in the Data Labeling service.

  • Added support to load labeled datasets by chunks in the Data Labeling service.


Release Notes: November 17, 2021

The following features were added:

  • Added support to manage credentials with the OCI Vault service for ADB and Access Tokens.

  • Improved model introspection functionality. The INFERENCE_ENV_TYPE and INFERENCE_ENV_SLUG parameters are no longer required.

  • Updated ADS dependency requirements. Relaxed the versions for the scikit-learn, scipy and onnx dependencies.

  • Moved dask, ipywidget and wordcloud to an optional dependency.

  • The Boston Housing dataset was replaced with an alternative one.

  • Migrated ADSDataset to use Pandas instead of Dask.

  • Deprecated MLRuntime.

  • Deprecated resource_analyze method.

  • Added support for magic commands in notebooks when they run in a Job.

  • Added support to download notebook and output after running it in a Job.


Release notes: October 20, 2021

The following features related to the Data Labeling service were added:

  • Integrating with the Oracle Cloud Infrastructure Data Labeling service.

  • Listing labeled datasets in the Data Labeling service.

  • Exporting labeled datasets into Object Storage.

  • Loading labeled datasets in the Pandas dataframe or generator formats.

  • Visualizing the labeled entity extraction and object detection data.

  • Converting the labeled entity extraction and object detection data to the Spacy and YOLO formats respectively.


The following improvements were effected:

  • Improve ads import time.

  • Fix the version of the jsonschema package.

  • Update numpy deps to >= 1.19.2 for compatibility with TensorFlow 2.6.

  • Added progress bar when creating a Data Flow application.

  • Fixed the file upload path in Data Flow.

  • Added supporting tags when saving model artifacts to the model catalog.

  • Updated Model Deployment authentication.

  • Specify spark version in prepare_app() now works.

  • Run a Job from a ZIP or folder.

This release has the following bug fixes:

  • Fixed the default runtime.yaml template generated outside of a notebook session.

  • Oracle DB mixin the batch size parameter is now passed downstream.

  • ADSModel.prepare() and prepare_generic_model() force_overwrite deletes user-created folders.

  • prepare_generic_model fails to create a successful artifact when taxonomy is extracted.


Release notes: September 27, 2021

The following dependencies were removed:

  • pyarrow

  • python-snappy


Release notes: September 22, 2021

The Data Science jobs feature is introduced and includes the following:

  • Data Science jobs allow data scientists to run customized tasks outside of a notebook session.

  • Running Data Science jobs and Data Flow applications through unified APIs by configuring job infrastructure and runtime parameters.

  • Configuring various runtime configurations for running code from Python/Bash script, packages including multiple modules, Jupyter notebook, or a Git repository.

  • Monitoring job runs and streaming log messages using the Logging service.


Release notes: September 20, 2021

This release has the following bug fixes:

  • prepare_generic_model fails when used outside the Data Science notebook session

  • TextDatasetFactory fails when used outside the Data Science notebook session


Release notes: September 17, 2021

  • Removed dependency on plotly.

  • print_user_message replaced with logger.


Release notes: August 3, 2021

This release of the model catalog includes these enhancements:

  • Automatic extraction of model taxonomy metadata that lets data scientists document the use case, framework, and hyperparameters of their models.

  • Improvement to the model provenance metadata, including a reference to the model training resource (notebook sessions) by passing in the training_id to the .save() method.

  • Support for custom metadata which lets data scientists document the context around their models, automatic extraction references to the conda environment used to train the model, the training and validation datasets, and so on.

  • Automatcal extraction of the model input feature vector and prediction schemas.

  • Model introspection tests that are run on the model artifact before the model is saved to the model catalog. Model introspection validates the artifact against a series of common issues and errors found with artifacts. These introspection tests are part of the model artifact code template that is included.

Feature type is an additional added module which includes the following functionality:

  • Support for Exploratory Data Analysis including feature count, feature plot, feature statistics, correlation, and correlation plot.

  • Support for the feature type manager that provides the tools to manage the handlers used to drive the feature type system.

  • Support for the feature type validators that are a way of performing data validation and also allow a feature type to be dynamically extended so that the data validation process can be reproducible and shared across projects.

  • Support for feature type warnings that allow you to automate the process of checking for data quality issues.


Release notes: May 7, 2021

Improvements include:

  • Requires Pandas >- 1.2 and Python == 3.7.

  • Upgraded the scikit-learn dependency to 0.23.2.

  • Added the ADSTextDataset and the ADS Text Extraction Framework.

  • Updated the ADSTuner method .tune() to allow asynchronous tuning, including the ability to halt, resume, and terminate tuning operations from the main process.

  • Added the ability to load and save ADSTuner tuned trials to Object Storage. The tuning progress can now be saved and loaded in a different ADSTuner object.

  • Added the ability to update the ADSTuner tuning search space. Hyperparameters can be changed and distribution ranges modified during tuning.

  • Updated plotting functions to plot in real-time while ADSTuner asynchronous tuning operations proceed.

  • Added methods to report on the remaining budget for running ADSTuner asynchronous tuner (trials and time-based budgets).

  • Added a method to report the difference between the optimal and current best score for ADSTuner tuning processes with score-based stopping criteria.

  • Added caching for model loading method to avoid model deserialization each time the predict method is called.

  • Made the list of supported formats in more explicit.

  • Moved the ADSEvaluator caption to above the table.

  • Added a warning message in the get_recommendations() method when no recommendations can be made.

  • Added a parameter in print_summary() to display the ranking table only.

  • list_apps in the DataFlow class supports the optional parameter compartment_id.

  • An exception occurs when using SVC or KNN on large datasets in OracleAutoMLProvider.

  • Speed improvements in correlation calculations.

  • Improved the name of the y-axis label in feature_selection_trials().

  • Automatically chooses the y-label based on the score_metric set in train if you don’t set it.

  • Increased the default timeout for uploading models to the model catalog.

  • Improved the module documentation.

  • Speed improvements in get_recommendations() on wide datasets.

  • Speed improvements in

  • Deprecated the frac keyword from

  • Disabled writing requirements.txt when function_artifacts = False.

  • Pretty printing of specific labels in ADSEvaluator.metrics.

  • Removed the global setting as the only mechanism for choosing the authentication in OCIClientFactory.

  • Added the ability to have defaults and to provide authentication information while instantiating a Provider Class.

  • Added a larger time buffer for the plot_param_importance method.

  • Migrated the DatasetFactory reading engine from Dask to Pandas.

  • Enabling Pandas to read lists and glob of files.

  • DatasetFactory now supports reading from Object Storage using ocifs.

  • The DatasetFactory URI pattern now supports namespaces and follows the HDFS Connector format.

  • The url() method can generate PARs for Object Storage objects.

  • DatasetFactory now has caching for Object Storage operations.

The following issues were fixed:

  • Issue with multipart upload and download in DatasetFactory.

  • Issues with log level in OracleAutoMLProvider.

  • Issue with fill_value when running get_recommendations().

  • Issue with an invalid training path when saving model provenance.

  • Issue with errors during model deletion.

  • Issues with deep copying ADSData.

  • Evaluation plot KeyError.

  • Dataset show_in_notebook issue.

  • Inconsistency in preparing ADSModels and generic models.

  • Issue with force_overwrite in prepare_generic_model not being properly triggered.

  • Issue with OracleAutoMLProvider failing to visualize_tuning_trials.

  • Issues with model_prepare trying to do feature transforms on keras and pytorch models.

  • Erroneous creation of __pychache__.

  • The AttributeError message when an ApplicationSummary or RunSummary object is being displayed in a notebook.

  • Issues with newer versions of Dask breaking DatasetFactory.

AutoML is upgraded to AutoML v1.0 and the changes include:

  • Switched to using Pandas Dataframes internally. AutoML now uses Pandas dataframes internally instead of Numpy dataframes, avoiding needless conversions.

  • Pytorch is now an optional dependency. If Pytorch is installed, AutoML automatically considers multilayer perceptrons in its search. If Pytorch is not found, deep learning models are ignored.

  • Updated the Pipeline interface to include train(), which runs all the pipeline stages though doesn’t do the final fitting of the model ( fit() API should be used if the final fit is needed).

  • Updated the Pipeline interface to include refit() to allow you to refit the pipeline to an updated dataset without re-running the full pipeline again. We recommend this for advanced users only. For best results, we recommended that you rerun the full pipeline when the dataset changes.

  • AutoML now reports memory usage for each trial as a part of its trial attributes. This information relies on the maximum resident size metric reported by Linux, and can sometimes be unreliable.

  • holidays is now an optional dependency. If holidays is installed, AutoML automatically uses it to add holidays as a feature for engineering datetime columns.

  • Added support for Anomaly Detection and Forecasting tasks (experimental).

  • Downcast dataset to reduce pipeline training memory consumption.

  • Set numpy BLAS parallelism to 1 to avoid CPU over subscription.

  • Created interactive example notebooks for all supported tasks (classification, regression, anomaly detection, and forecasting), see

  • Other general bug fixes.

MLX is upgraded to MLX v1.1.1 the changes include:

  • Upgrading to Python 3.7

  • Upgrading to support Numpy >= 1.19.4

  • Upgrading to support Pandas >= 1.1.5

  • Upgrading to support Scikit-learn >= 0.23.2

  • Upgrading to support Statsmodel >= 0.12.1

  • Upgrading to support Dask >= 2.30.0

  • Upgrading to support Distributed >= 2.30.1

  • Upgrading to support Xgboost >= 1.2.1

  • Upgrading to support Category_encoders >= 2.2.2

  • Upgrading to support Tqdm >= 4.36.1

  • Fixed imputation issue when columns are all NaN.

  • Fixed WhatIF internal index-reference issue.

  • Fixed rare floating point problem in FD/ALE explainers.

January 13, 2021

  • A full distribution of this release of ADS is found in the General Machine Learning for CPU and GPU environments. The Classic environments include the previous release of ADS.

  • A distribution of ADS without AutoML and MLX is found in the remaining environments.

  • DatasetFactory can now download files first before opening them in memory using the .download() method.

  • Added support to archive files in creating Data Flow applications and runs.

  • Support was added for loading Avro format data into ADS.

  • Changed model serialization to use ONNX by default when possible on supported models.

  • Added ADSTuner, which is a framework and model agnostic hyperparmater optimizer, use the adstuner.ipynb notebook for examples of how to use this feature.

  • Corrected the up_sample() method in get_recommendations() so that it does not fail when all features are categorical. Up-sampling is possible for datasets containing continuous and categorical features.

  • Resolved issues with serializing ndarray objects into JSON.

  • A table of all of the ADS notebook examples can be found in our service documentation: Oracle Cloud Infrastructure Data Science

  • Changed set_documentation_mode to false by default.

  • Added unit-tests related to the dataset helper.

  • Fixed the _check_object_exists to handle situations where the object storage bucket has more than 1000 objects.

  • Added option overwrite_script in the create_app() method to allow a user to override a pre-existing file.

  • Added support for newer fsspec versions.

  • Added support for the C library Snappy.

  • Fixed issue with uploading model provenance data due to inconsistency with OCI interface.

  • Resolved issue with multiple versions of Cryptography being installed when installing fbprophet.

AutoML is upgraded to AutoML v0.5.2 and the changes include:

  • AutoML is now distributed in the General Machine Learning and Data Exploration conda environments.

  • Support for ONNX. AutoML models can now be serialized using ONNX by calling the to_onnx() API on the AutoML estimator.

  • Pre-processing has been overhauled to use sklearn pipelines to allow serialization using ONNX. Numerical, categorical, and text columns are supported for ONNX serialization. Datetime and time series columns are not supported.

  • Torch-based deep learning models, TorchMLPClassifier and TorchMLPRegressor, have been added.

  • GPU support for XGBoost and torch-based models have been added. This is disabled by default and can be enabled by passing in ‘gpu_id’: ‘auto’ in engine_opts in the constructor. ONNX serialization for GPUs has not been tested.

  • Adaptive sampling’s learning curve has been smoothened. This allows adaptive sampling to converge faster on some datasets.

  • Improvements to ranking performance in feature selection were added. Feature selection is now much faster on large datasets.

  • The default execution engine for AutoML has been switched to Dask. You can still use the Python multiprocessing by passing engine='local', engine_opts={'n_jobs' : -1} to init()

  • GuassianNB has been enabled in the interface by default.

  • The AdaBoostClassifier has been disabled in the pipeline-interface by default. The ONNX converter for AdaBoost should not be used.

  • The issue ValueError: Found unknown categories during transform has been fixed.

  • You can manually specify a hyperparameter search space to AutoML. A new parameter was added to the pipeline. This allows you to freeze some hyperparameters or to expose further ones for tuning.

  • New API: Refit an AutoML pipeline to another dataset. This is primarily used to handle updated training data, where you train the pipeline once, and refit in on newer data.

  • AutoML no longer closes a user-specified Dask cluster.

  • AutoML properly cleans up any existing futures on the Dask cluster at the end of fit.

MLX is upgraded to MLX v1.0.16 the changes include:

  • MLX is now distributed in the General Machine Learning conda environments.

  • Updated the explanation descriptions to use a base64 representation of the static plots. This obviates the need for creating a mlx_static directory.

  • Replaced the boolean indexing in slicing Pandas dataFrame with integer indexing. After updating to Pandas >= 1.1.0 the boolean indexing caused some issues. Integer indexing addresses these issues.

  • Fixed MLX-related import warnings.

  • Corrected an issue with ALE when the target values are strings.

  • Removed the dependency on Paramiko.

  • Addresses an issue with ALE when the target values are not of type list.

August 11, 2020

  • Support was added to use resource principles as an authentication mechanism for ADS.

  • Support was added to MLX for an additional model explanation diagnostic, Accumulated Local Effects (ALEs).

  • Support was added to MLX for “What-if” scenarios in model explainability.

  • Improvements were made to the correlation heatmap calculations in show_in_notebook().

  • Improvements were made to the model artifact.

The following bugs were fixed:

  • Data Flow applications inherit the compartment assignment of the client. Runs inherit from applications by default. Compartment OCIDs can also be specified independently at the client, application, and run levels.

  • The Data Flow log link for logs pulled from an application loaded into the notebook session is fixed.

  • Progress bars now complete fully (in ADSModel.prepare() and prepare_generic_model()).

  • BaselineModel is now significantly faster and can be opted out of.

MLX upgraded to MLX v1.0.10 the changes include:

  • Added support to specify the mlx_static root path (used for ALE summary).

  • Added support for making mlx_static directory hidden (for example, <path>/.mlx_static/).

  • Fixed issue with the boolean features in ALE.

June 9, 2020

Numerous bug fixes including:

  • Support for Data Flow applications and runs outside of a notebook session compartment. Support for specific object storage logs and script buckets at the application and run levels.

  • ADS detects small shapes and gives warnings for AutoML execution.

  • Removal of triggers in the Oracle Cloud Infrastructure Functions func.yaml file.

  • incorrectly yielding a classification dataset for a continuous target was fixed.

  • LabelEncoder producing the wrong results for category and object columns was fixed.

  • An untrusted notebook issue when running model explanation visualizations were fixed.

  • A warning about adaptive sampling requiring at least 1000 data points was added.

  • A dtype cast float to integer into"csv") was added.

  • An option to specify the bucket of Data Flow logs when you create the application was added.

AutoML upgraded to 0.4.2 the changes include:

  • Reduced parallelization on low compute hardware.

  • Support for passing in a custom logger object in automl.init(logger=).

  • Support for datetime columns. AutoML should automatically infer datetime columns based on the Pandas dataframe, and perform feature engineering on them. This can also be forced by using the col_types argument in The supported types are: ['categorical', 'numerical', 'datetime']

MLX upgraded to MLX 1.0.7 the changes include:

  • Updated the feature distributions in the PDP/ICE plots (performance improvement).

  • All distributions are now shown as PMFs. Categorical features show the category frequency and continuous features are computed using a NumPy histogram (with ‘auto’). They are also separate sub-plots, which are interactive.

  • Classification PDP: The y-axis for continuous features is now auto-scaled (not fixed to 0-1).

  • 1-feature PDP/ICE: The x-axis for continuous features now shows the entire feature distribution, whereas the plot may show a subset depending on the partial_range parameter (for example, partial_range=[0.2, 0.8] computes the PDP between the 20th and 80th percentile. The plot now shows the full distribution on the x-axis, but the line charts are only drawn between the specified percentile ranges).

  • 2-feature PDP: The plot x and y axes are now auto-set to match the partial_range specified by the user. This ensures that the heatmap fills the entire plot by default. However, the entire feature distribution can be viewed by zooming out or clicking Autoscale in plotly.

  • Support for plotting scatter plots using WebGL (show_in_notebook(..., use_webgl=True)) was added.

  • The side issues that were causing the MLX Visualization Omitted warnings in JupyterLab were fixed.

April 30, 2020

  • ADS integration with the Oracle Cloud Infrastructure Data Flow service provides a more efficient and convenient to launch a Spark application and run Spark jobs

  • show_in_notebook() has had “head” removed from accordion and is replaced with dataset “warnings”.

  • get_recommendations() is deprecated and replaced with suggest_recommendations(), which returns a Pandas dataframe with all the recommendations and suggested code to implement each action.

  • A progress indication of Autonomous Data Warehouse reads has been added.

AutoML updated to version 0.4.1 from 0.3.1:

  • More consistent handling of stratification and random state.

  • Bug-fix for LightGBM and XGBoost crashing on AMD shapes was implemented.

  • Unified Proxy Models across all stages of the AutoML Pipeline, ensuring leaderboard rankings are consistent was implemented.

  • Remove visual option from the interface.

  • The default tuning metric for both binary and multi-class classification has been changed to neg_log_loss.

  • Bug-fix in AutoML XGBoost, where the predicted probabilities were sometimes NaN, was implemented.

  • Fixed several corner case issues in Hyperparameter Optimization.

MLX updated to version 1.0.3 from 1.0.0:

  • Added support for specifying the ‘average’ parameter in sklearn metrics by <metric>_<average>, for examlple F1_avg.

  • Fixed an issue with the detailed scatter plot visualizations and cutoff feature/axis names.

  • Fixed an issue with the balanced sampling in the Global Feature Permutation Importance explainer.

  • Updated the supported scoring metrics in MLX. The PermutationImportance explainer now supports a large number of classification and regression metrics. Also, many of the metrics’ names were changed.

  • Updated LIME and PermutationImportance explainer descriptions.

  • Fixed an issue where sklearn.pipeline wasn’t imported.

  • Fixed deprecated asscalar warnings.

March 18, 2020

Access to ADW performance has been improved significantly

Major improvements were made to the performance of the ADW dataset loader. Your data is now loaded much faster, depending on your environment.

Change to with ADW with format='sql' no longer requires the index_col to be specified. This was confusing, since “index” means something very different in databases. Additionally, the table parameter may now be either a table or a sql expression.

ds =
  format = 'sql',
  table = """
    SELECT *
    FROM sh.times
    WHERE rownum <= 30

No longer automatically starts an H2O cluster

ADS no longer instantiates an H2O cluster on behalf of the user. Instead, you need to import h2o on your own and then start your own cluster.

Profiling Dask APIs

With support for Bokeh extension, you can now profile Dask operations and visualize profiler output. For more details, see Dask ResourceProfiler.

You can use the ads.common.analyzer.resource_analyze decorator to visualize the CPU and memory utilization of operations.

During execution, it records the following information for each timestep:

  • Time in seconds since the epoch

  • Memory usage in MB

  • % CPU usage


from ads.common.analyzer import resource_analyze
from ads.dataset.dataset_browser import DatasetBrowser
def fetch_data():
    sklearn = DatasetBrowser.sklearn()
    wine_ds ='wine').set_target("target")
    return wine_ds

The output shows two lines, one for the total CPU percentage used by all the workers, and one for total memory used.

Dask Upgrade

Dask is updated to version 2.10.1 with support for Oracle Cloud Infrastructure Object Storage. The 2.10.1 version provides better performance than the older version.