See also the fleetcommand-agent Release Notes.
Domino Nexus
Domino Nexus is a single pane of glass that lets you run data science and machine learning workloads across any compute cluster — in any cloud, region, or on-premises. It unifies data science silos across the enterprise, so you have one place to build, deploy, and monitor models.
See these topics for complete details:
Some Domino features work differently in the Nexus hybrid architecture than they do in other types of deployments:
-
Domino Nexus introduces the concept of data locality.
Data may only be available in certain data planes due to geographic access restrictions or simply the cost of moving data between data centers. Domino automatically mounts all available data for a given data plane.
-
These features are only available in the
Local
data plane (hosted in the control plane):-
Model APIs
-
Apps
-
Datasets
-
Trino data sources
-
Other types of data sources can be accessed on both the
Local
and remote data planes, but functionality on remote data planes is in Preview. Functionality for restricting data sources to specific data planes is under development. External Data Volumes (EDVs) are the primary method for accessing large data sets, and have a first-class notion of data locality; see Associate Data Planes with EDVs.
-
-
When you launch an execution, hardware tiers are grouped by data plane. Select a tier in the data plane in which the execution should run. The
Local
data plane corresponds to running the execution in the Domino control plane cluster. -
For compute clusters that include a user interface, the UI link is available only if the data plane is configured for workspaces. See Enable A Data Plane For Workspaces. If the data plane is not configured for workspaces, clusters can still be used with jobs, but the user interface link is disabled.
Experiment management
Domino experiment management leverages MLflow Tracking to enable easy logging of experiment parameters, metrics, and artifacts, while providing a Domino-native user experience to help you analyze your results. MLflow runs as a service in your Domino cluster, fully integrated within your workspace and jobs, and honoring role-based access control. Existing MLflow experiments works right out of the box with no code changes required.
See Experiments for complete details.
Domino Code Assist
Domino Code Assist (DCA) automatically generates Python and R Code for common data science and analysis tasks through a simple point-and-click interface.
It helps novice coders quickly become productive in writing R and Python code. For those more familiar with writing code, DCA accelerates common data analysis tasks by auto-generating boilerplate code that can be further edited as needed for a project. DCA can autogenerate Python or R code for the following tasks:
-
Import data from Cloud stores like S3, Snowflake, or Redshift.
-
Transform data with operations like filtering and aggregation.
-
Create visualizations.
-
Create data apps.
-
Quickly and easily deploy data apps. DCA also allows you to create code snippets that can be shared for common tasks.
DCA ships standard in Domino 5.5. It can be installed manually in Domino 4.x or later.
Full documentation and installation instructions for DCA are available in the DCA documentation.
Asynchronous Model APIs
Domino Asynchronous Model APIs allow you to host complex inference processing involving workloads such as deep learning predictions, unstructured data transformations, and other similar executions that are computationally intensive and not suitable for synchronous HTTP endpoints.
An Asynchronous Model API queues incoming requests and processes them asynchronously while giving end users an interface to query the status of the processing and fetch the results when complete. For more details, see Long-running inferences and Asynchronous interfaces.
New Palantir Foundry data connector
Now you can access your Palantir Foundry data store from Domino, using a Starburst-powered connector.
An administrator must create the Palantir Foundry data source before users can access it.
-
When experiment management is enabled (by setting
com.cerebro.domino.workbench.experimentManagement.enabled
totrue
), the project menu takes on a new layout:-
Files becomes Code.
-
The Run section becomes Develop, and it includes the Code and Data pages.
-
The Materials section becomes Evaluate, and it now includes the Jobs page.
-
The Activity page becomes a tab on the Overview page.
-
The Scheduled Jobs page becomes a tab on the Jobs page.
5.4 project menu 5.5 project menu -
-
Git-based Projects are now a General-Availability (GA) feature.
-
Model monitoring improvements
-
Support for EKS version 1.24
This version requires the use of EKS optimized Amazon Linux AMI release 1.24.7-20221222 or later.
-
Keycloak upgraded to version 18
Custom Keycloak script providers (script mappers and script authenticators) are automatically migrated to a different format for compatibility with Keycloak version 18 and above.
As a result, you can no longer see or update script providers code in the Keycloak admin console. Domino automatically migrates existing code and stores it at
/opt/jboss/keycloak/standalone/deployments/keycloak-resources/script_providers.jar
in the Keycloak pod(s).See Javascript providers in the Keycloak documentation for further details about this format change. Any new script provider code should be packaged into a JAR file (multiple files can co-exist) and uploaded to the
/opt/jboss/keycloak/standalone/deployments/keycloak-resources/
folder. In case of using multiple Keycloak replica pods, it is enough to upload the file to one of them. This action does not require a Keycloak pod restart.
-
Changes to the User Session Count Limiter in Keycloak
You can select either Deny new session or Terminate oldest session as the desired behavior when configuring the User Session Count Limiter in Keycloak.
-
Improve system stability when running a large number (> 1000) of jobs concurrently. See Sizing the Domino platform: Large for more details on how to size resources when running workloads at this scale.
-
The
EnableLegacyJwtTooling
feature flag now defaults tofalse
for new deployments. When existing deployments are upgraded to 5.5.0, the value still defaults totrue
.
-
Docker bridge is not supported in Kubernetes 1.24 and above.
-
Domino Admin Toolkit is installed with Domino 5.5.0 and is now an always-on deployment.
As a result it is no longer necessary to use the Admin Toolkit CLI (
toolkit.sh
) to install it; it is automatically available even in air-gapped deployments.The Toolkit’s front-end UI is now available permanently at the URL https://<your-domino-url>/toolkit/
The sending of reports generated by Admin Toolkit now has an opt-in/opt-out feature, controlled through the Toolkit UI. The default is to opt in.
The Toolkit UI also includes a button to check for an automatic update to the latest version of the toolkit. For air-gapped deployments the Admin Toolkit CLI can be used to copy the new image to a private repository.
-
Adding model environment variables no longer restarts your model instances. The newly added environment variables take effect on the next model version start.
Updates to the <code>python-domino</code> library:
The DominoSparkOperator
test is updated for compute_cluster_properties
.
The minimum Python version required is now Python 3.7.
-
Metrics publishing is reduced to the minimal number of metrics essential to Domino Model Monitoring operation, to prevent prometheus-adapter OOM errors.
Please review your independent reliance on Kubernetes metrics at the cluster level and validate your upstream source, especially if you run Domino in a cluster environment shared with other applications. This change can be overridden via an upgrade/install path if desired. Please contact Domino Support for any questions or concerns via https://tickets.dominodatalab.com or support@dominodatalab.com.
-
Domino Model Monitor is now compatible with Kubernetes versions 1.21 and above; it no longer stops working after 90 days of uptime in these deployments and it is no longer necessary to periodically restart it.
-
You cannot view the latest raw file. In the navigation pane, go to Files and click a file to view its details. If you click View Latest Raw File, a blank page opens.
-
Running a large number of jobs concurrently (> 1000) can introduce system instability and result in some job failures. We now document how to properly size Domino resources in deployments that expect this level of load. See Sizing the Domino platform: Large.
-
When uploading a large file to the Azure blob store by syncing a workspace, you may encounter a Java Out of Memory error from Azure if the file/blob already exists. To work around this issue, use the Domino CLI to upload the file to the project.
-
Model Monitoring data sources aren’t validated. If you enter an invalid bucket name and attempt to save, the entry will go through. However, you won’t be able to see metrics for that entry because the name points to an invalid bucket.
-
When you use drag-and-drop to upload a large number of files at once, the upload might fail. The application displays many empty dialogs. Use the Domini CLI to upload large numbers of files instead. Domino 5.4.0 eliminates the empty dialogs and shows helpful information.