Concept | LLM connections#

Large language models are what make Generative AI possible. To use GenAI in Dataiku, you must first connect to a service providing access to LLMs.

Once an administrator sets up connections that make LLMs available to your organization, you can leverage LLMs from this connection with Dataiku’s:

Adding a connection#

Users with administrator rights can add and manage LLM connections in the connections settings of an instance. The settings location varies slightly depending on the instance type.

For self-managed instances, settings are in the waffle menu (Waffle icon.) > Administration > Connections > + New Connection > LLM Mesh.

Screenshot of the LLM connection settings in the Administration panel.

Types of connections#

LLMs are either provided as a service or as objects that your organization must manage. Your organization might decide to work with multiple providers, commercial services, or open-source models.

Built-in connections#

Dataiku provides connections out-of-the-box for both commercial services and open-source model providers. Connection requirements vary by provider.

Supported connections to GenAI services include:

Note

For open-source models, Dataiku provides serving capabilities, including containerized execution and integration with NVIDIA GPUs. Administrators only have to configure those capabilities, simplifying management of the serving infrastructure.

Custom connections#

You can also develop custom LLMs, or connect to your internal custom gateway, and use them via a custom LLM connection.

See also

For more details on the types of LLM connections and instructions to set them up, see the reference documentation.