Project settings

Flow Display settings

Flow zones

Project wiki

Display the wiki of this project in the global (instance-wide) wikis list

Resources control

Explore / Prepare sample limit

These settings control the maximum size in memory of data samples taken for explore and prepare screens.
Note that this is also constrained by global settings.

Analysis / Prepare processed limit

These settings control the maximum size in memory of data during and after processing by a prepare script (either in analysis or a preparation recipe.
Note that this is further constrained by global (admin-level) settings.

Flow Build & Pipelines

Spark pipelines

Spark pipelines allow you to execute long data pipelines with multiple steps without always having to write the intermediate data and re-read it at the next steps.
You can read more in the

SQL pipelines

SQL pipelines allow you to execute long data pipelines with multiple steps without always having to write the intermediate data and re-read it at the next steps.
You can read more in the

Flow dependencies computation

This setting controls how Dataiku checks for changes before doing a smart rebuild of the Flow:
  • Enabled: Dataiku automatically checks if the input datasets have changed since the last build and skips running recipes for which no change occurred.
  • Enabled (but ignore non-ready sources): Same as Enabled but allows the Flow to proceed even if some input datasets are not ready. This is particularly useful if a recipe has inputs that are expected to not be built, such as when there is a loop in the Flow.
  • Disabled: Dataiku skips checking if the input datasets have changed and are ready. This is useful when computing if the input datasets have changed is costly, such as with large datasets made of many files. However, it means that Dataiku will always run all recipes, even if the input datasets haven't changed.
Note that this setting can be overwritten at the recipe level.
None

Change management

Automation

Scenarios can only be started manually.

Project tags

Modify global tag categories in Administration. Modify global tag categories in Administration.

Integrations

Dataiku Govern

Synchronization

These settings are only meaningful when Govern integration is enabled
Synchronize DSS items of this project on Dataiku Govern
Synced items successfully! ({{governIntegrationProjectSyncResult.data.syncedProjects}} {{ 'project' | plurify : governIntegrationProjectSyncResult.data.syncedProjects}}, {{governIntegrationProjectSyncResult.data.syncedDatasets}} datasets, {{governIntegrationProjectSyncResult.data.syncedBundles}} {{ 'bundle' | plurify : governIntegrationProjectSyncResult.data.syncedBundles}}, {{governIntegrationProjectSyncResult.data.syncedSavedModels}} {{ 'saved model' | plurify : governIntegrationProjectSyncResult.data.syncedSavedModels}}, {{governIntegrationProjectSyncResult.data.syncedRetrievableKnowledges}} {{ 'knowledge bank' | plurify : governIntegrationProjectSyncResult.data.syncedRetrievableKnowledges}}, {{governIntegrationProjectSyncResult.data.syncedSavedModelVersions}} {{ 'saved model version' | plurify : governIntegrationProjectSyncResult.data.syncedSavedModelVersions}})

Code envs

Default Python code env

Default R code env

Hadoop cluster selection

Use a cluster id, not name. Variables are allowed with a ${} syntax.
If the field above is blank or an undefined variable, use this cluster instead of the DSS global cluster.

Kubernetes cluster selection

Use a cluster id, not name. Variables are allowed with a ${} syntax.
If the field above is blank or an undefined variable, use this cluster instead of the DSS global cluster.

ACLs resync

These settings are only meaningful when multi-user-security is enabled

Engines & connections

Datasets creation settings

Use this connection by default for new managed output datasets when no proper contextual connection can be selected (for example, recipe from an "Uploaded" dataset).
Use this connection by default for new managed output datasets, even if a better contextual connection could be selected (for example, same as input)
Use this connection by default for uploaded datasets
Comma-separated list of ordered (most preferred to least preferred) storage formats.

Forbidden/preferred recipe engines

These settings supersede in this project the settings defined at the administration level

Recipe creation settings

Default value for new Hive-based recipes
Name By Origin Description
{{plugin.installedDesc.desc.meta.author}}
{{computePluginOrigin(plugin).toLocaleLowerCase() | capitalize}}

No plugin matches your filter.

Clear filters
There is no installed plugin with project-level presets on this Dataiku DSS instance.
This plugin has no presets yet.
This plugin does not have parameter sets with presets definable in projects.

Miscellaneous

Custom properties

For specific use cases

Documentation Portal

AI Types

AI Types are automatically computed badges that help you identify quickly which project or API Service contains Machine Learning, LLMs or Agents.

Project

Recompute badges based on AI Types in this project
Synced badges based on the project's AI Types successfully!

API Services

Recompute badges based on AI Types in this project's API services
Synced badges based on AI Types of API services successfully!