domino logo
  • Domino 5.5.0 (March 2023)
  • Domino 5.4.1 (February 2023)
  • Domino 5.4.0 (December 2022)
  • Domino 5.3.2 (December 2022)
  • Domino 5.3.1 (October 2022)
  • Domino 5.3.0 (September 2022)
  • Domino 5.2.2 (August 2022)
  • Domino 5.2.1 (July 2022)
  • Domino 5.2.0 (June 2022)
  • Domino 5.1.4 (July 2022)
  • Domino 5.1.3 (May 2022)
  • Domino 5.1.2 (April 2022)
  • Domino 5.1.1 (March 2022)
  • Domino 5.1.0 (March 2022)
  • Domino 5.0.2 (March 2022)
  • Domino 5.0.1 (January 2022)
  • Domino 5.0.0 (December 2021)
  • Domino 4.6.4 (March 2022)
  • Domino 4.6.3 (January 2022)
  • Domino 4.6.2 (November 2021)
  • Domino 4.6.1 (October 2021)
  • Domino 4.6.0 (August 2021)
  • Domino 4.5.2 (August 2021)
  • Domino 4.5.1 (July 2021)
  • Domino 4.5.0 (June 2021)
  • Domino 4.4.2 (May 2021)
  • Domino 4.4.1 (March 2021)
  • Domino 4.4 (February 2021)
  • Domino 4.3.3 (December 2020)
  • Domino 4.3.2 (November 2020)
  • Domino 4.3.1 (October 2020)
  • Domino 4.3 (August 2020)
  • Domino 4.2
  • Domino 4.1
  • Domino 4.0
  • Domino 3.6
domino logo
About Domino
Domino Data LabKnowledge BaseData Science BlogTraining
Release Notes
>
Domino 5.5.0 (March 2023)

Domino 5.5.0 (March 2023)

See also the fleetcommand-agent Release Notes.

New Features

  • Domino Nexus

  • Experiment management

  • Domino Code Assist

  • Asynchronous Model APIs

  • New Palantir Foundry data connector

Domino Nexus

Domino Nexus is a single pane of glass that lets you run data science and machine learning workloads across any compute cluster — in any cloud, region, or on-premises. It unifies data science silos across the enterprise, so you have one place to build, deploy, and monitor models.

See these topics for complete details:

  • Nexus Hybrid Architecture

  • Data Planes

  • Manage Data Planes

  • Control Plane Security Guidance

Some Domino features work differently in the Nexus hybrid architecture than they do in other types of deployments:

  • Domino Nexus introduces the concept of data locality.

    Data may only be available in certain data planes due to geographic access restrictions or simply the cost of moving data between data centers. Domino automatically mounts all available data for a given data plane.

  • These features are only available in the Local data plane (hosted in the control plane):

    • Model APIs

    • Apps

    • Datasets

    • Trino data sources

    • Other types of data sources can be accessed on both the Local and remote data planes, but functionality on remote data planes is in Preview. Functionality for restricting data sources to specific data planes is under development. External Data Volumes (EDVs) are the primary method for accessing large data sets, and have a first-class notion of data locality; see Associate Data Planes with EDVs.

  • When you launch an execution, hardware tiers are grouped by data plane. Select a tier in the data plane in which the execution should run. The Local data plane corresponds to running the execution in the Domino control plane cluster.

  • For compute clusters that include a user interface, the UI link is available only if the data plane is configured for workspaces. See Enable A Data Plane For Workspaces. If the data plane is not configured for workspaces, clusters can still be used with jobs, but the user interface link is disabled.

Experiment management

Domino experiment management leverages MLflow Tracking to enable easy logging of experiment parameters, metrics, and artifacts, while providing a Domino-native user experience to help you analyze your results. MLflow runs as a service in your Domino cluster, fully integrated within your workspace and jobs, and honoring role-based access control. Existing MLflow experiments works right out of the box with no code changes required.

See Experiments for complete details.

Note

Domino Code Assist

Domino Code Assist (DCA) automatically generates Python and R Code for common data science and analysis tasks through a simple point-and-click interface.

It helps novice coders quickly become productive in writing R and Python code. For those more familiar with writing code, DCA accelerates common data analysis tasks by auto-generating boilerplate code that can be further edited as needed for a project. DCA can autogenerate Python or R code for the following tasks:

  1. Import data from Cloud stores like S3, Snowflake, or Redshift.

  2. Transform data with operations like filtering and aggregation.

  3. Create visualizations.

  4. Create data apps.

  5. Quickly and easily deploy data apps. DCA also allows you to create code snippets that can be shared for common tasks.

DCA ships standard in Domino 5.5. It can be installed manually in Domino 4.x or later.

Full documentation and installation instructions for DCA are available in the DCA documentation.

Asynchronous Model APIs

Domino Asynchronous Model APIs allow you to host complex inference processing involving workloads such as deep learning predictions, unstructured data transformations, and other similar executions that are computationally intensive and not suitable for synchronous HTTP endpoints.

An Asynchronous Model API queues incoming requests and processes them asynchronously while giving end users an interface to query the status of the processing and fetch the results when complete. For more details, see Long-running inferences and Asynchronous interfaces.

New Palantir Foundry data connector

Now you can access your Palantir Foundry data store from Domino, using a Starburst-powered connector.

An administrator must create the Palantir Foundry data source before users can access it.

Improvements

  • When experiment management is enabled (by setting com.cerebro.domino.workbench.experimentManagement.enabled to true), the project menu takes on a new layout:

    • Files becomes Code.

    • The Run section becomes Develop, and it includes the Code and Data pages.

    • The Materials section becomes Evaluate, and it now includes the Jobs page.

    • The Activity page becomes a tab on the Overview page.

    • The Scheduled Jobs page becomes a tab on the Jobs page.

    5.4 project menu5.5 project menu

    5.3 project menu

    5.5 project menu

  • Git-based Projects are now a General-Availability (GA) feature.

  • Model monitoring improvements

  • Support for EKS version 1.24

    This version requires the use of EKS optimized Amazon Linux AMI release 1.24.7-20221222 or later.

  • Keycloak upgraded to version 18

    Custom Keycloak script providers (script mappers and script authenticators) are automatically migrated to a different format for compatibility with Keycloak version 18 and above.

    As a result, you can no longer see or update script providers code in the Keycloak admin console. Domino automatically migrates existing code and stores it at /opt/jboss/keycloak/standalone/deployments/keycloak-resources/script_providers.jar in the Keycloak pod(s).

    See Javascript providers in the Keycloak documentation for further details about this format change. Any new script provider code should be packaged into a JAR file (multiple files can co-exist) and uploaded to the /opt/jboss/keycloak/standalone/deployments/keycloak-resources/ folder. In case of using multiple Keycloak replica pods, it is enough to upload the file to one of them. This action does not require a Keycloak pod restart.

  • Changes to the User Session Count Limiter in Keycloak

    You can select either Deny new session or Terminate oldest session as the desired behavior when configuring the User Session Count Limiter in Keycloak.

  • Improve system stability when running a large number (> 1000) of jobs concurrently. See Sizing the Domino platform: Large for more details on how to size resources when running workloads at this scale.

  • The EnableLegacyJwtTooling feature flag now defaults to false for new deployments. When existing deployments are upgraded to 5.5.0, the value still defaults to true.

  • Docker bridge is not supported in Kubernetes 1.24 and above.

  • Domino Admin Toolkit is installed with Domino 5.5.0 and is now an always-on deployment.

    As a result it is no longer necessary to use the Admin Toolkit CLI (toolkit.sh) to install it; it is automatically available even in air-gapped deployments.

    The Toolkit’s front-end UI is now available permanently at the URL https://<your-domino-url>/toolkit/

    The sending of reports generated by Admin Toolkit now has an opt-in/opt-out feature, controlled through the Toolkit UI. The default is to opt in.

    The Toolkit UI also includes a button to check for an automatic update to the latest version of the toolkit. For air-gapped deployments the Admin Toolkit CLI can be used to copy the new image to a private repository.

  • Adding model environment variables no longer restarts your model instances. The newly added environment variables take effect on the next model version start.

API changes

Updates to the <code>python-domino</code> library: The DominoSparkOperator test is updated for compute_cluster_properties. The minimum Python version required is now Python 3.7.

Bug Fixes

  • Metrics publishing is reduced to the minimal number of metrics essential to Domino Model Monitoring operation, to prevent prometheus-adapter OOM errors.

    Please review your independent reliance on Kubernetes metrics at the cluster level and validate your upstream source, especially if you run Domino in a cluster environment shared with other applications. This change can be overridden via an upgrade/install path if desired. Please contact Domino Support for any questions or concerns via https://tickets.dominodatalab.com or support@dominodatalab.com.

  • Domino Model Monitor is now compatible with Kubernetes versions 1.21 and above; it no longer stops working after 90 days of uptime in these deployments and it is no longer necessary to periodically restart it.

Known Issues

  • You cannot view the latest raw file. In the navigation pane, go to Files and click a file to view its details. If you click View Latest Raw File, a blank page opens.

  • Running a large number of jobs concurrently (> 1000) can introduce system instability and result in some job failures. We now document how to properly size Domino resources in deployments that expect this level of load. See Sizing the Domino platform: Large.

  • When uploading a large file to the Azure blob store by syncing a workspace, you may encounter a Java Out of Memory error from Azure if the file/blob already exists. To work around this issue, use the Domino CLI to upload the file to the project.

  • Model Monitoring data sources aren’t validated. If you enter an invalid bucket name and attempt to save, the entry will go through. However, you won’t be able to see metrics for that entry because the name points to an invalid bucket.

  • When you use drag-and-drop to upload a large number of files at once, the upload might fail. The application displays many empty dialogs. Use the Domini CLI to upload large numbers of files instead. Domino 5.4.0 eliminates the empty dialogs and shows helpful information.

Domino Data LabKnowledge BaseData Science BlogTraining
Copyright © 2022 Domino Data Lab. All rights reserved.