Theia API Documentation v1.65.0
    Preparing search index...

    Module @theia/ai-openai


    theia-ext-logo

    ECLIPSE THEIA - OPEN AI EXTENSION


    The @theia/ai-openai integrates OpenAI's models with Theia AI. The OpenAI API key and the models to use can be configured via preferences. Alternatively the OpenAI API key can also be handed in via the OPENAI_API_KEY variable.

    The extension also supports OpenAI compatible models hosted on different end points. You can configure the end points via the ai-features.openAiCustom.customOpenAiModels preference:

    {
    model: string,
    url: string,
    id?: string,
    apiKey?: string | true,
    apiVersion?: string | true,
    developerMessageSettings?: 'user' | 'system' | 'developer' | 'mergeWithFollowingUserMessage' | 'skip',
    enableStreaming?: boolean
    }
    • model and url are mandatory attributes, indicating the end point and model to use
    • id is an optional attribute which is used in the UI to refer to this configuration
    • apiKey is either the key to access the API served at the given URL or true to use the global OpenAI API key. If not given 'no-key' will be used. The apiKey will be send as a Bearer Token in the authorization request.
    • apiVersion is either the api version to access the API served at the given URL in Azure or true to use the global OpenAI API version.
    • developerMessageSettings Controls the handling of system messages: user, system, and developer will be used as a role, mergeWithFollowingUserMessage will prefix the following user message with the system message or convert the system message to user message if the next message is not a user message. skip will just remove the system message. Defaulting to developer.
    • enableStreaming is a flag that indicates whether the streaming API shall be used or not. true by default.

    To use a custom OpenAI model hosted on Azure, the AzureOpenAI class needs to be used, as described in the openai-node docs.

    Requests to an OpenAI model hosted on Azure need an apiVersion. To configure a custom OpenAI model in Theia you therefore need to configure the apiVersion with the end point. Note that if you don't configure an apiVersion, the default OpenAI object is used for initialization and a connection to an Azure hosted OpenAI model will fail.

    An OpenAI model version deployed on Azure might not support the developer role. In that case it is possible to configure whether the developer role is supported or not via the developerMessageSettings option, e.g. setting it to system or user.

    The following snippet shows a possible configuration to access an OpenAI model hosted on Azure. The AZURE_OPENAI_API_BASE_URL needs to be given without the /chat/completions path and without the api-version parameter, e.g. https://<my_prefix>.openai.azure.com/openai/deployments/<my_deployment>

    {
    "ai-features.AiEnable.enableAI": true,
    "ai-features.openAiCustom.customOpenAiModels": [
    {
    "model": "gpt4o",
    "url": "<AZURE_OPENAI_API_BASE_URL>",
    "id": "azure-deployment",
    "apiKey": "<AZURE_OPENAI_API_KEY>",
    "apiVersion": "<AZURE_OPENAI_API_VERSION>",
    "developerMessageSettings": "system"
    }
    ],
    "ai-features.agentSettings": {
    "Universal": {
    "languageModelRequirements": [
    {
    "purpose": "chat",
    "identifier": "azure-deployment"
    }
    ]
    },
    "Orchestrator": {
    "languageModelRequirements": [
    {
    "purpose": "agent-selection",
    "identifier": "azure-deployment"
    }
    ]
    }
    }
    }

    "Theia" is a trademark of the Eclipse Foundation https://www.eclipse.org/theia

    Modules

    browser/openai-frontend-application-contribution
    browser/openai-frontend-module
    common
    common/openai-language-models-manager
    common/openai-preferences
    node/openai-backend-module
    node/openai-language-model
    node/openai-language-models-manager-impl
    node/openai-streaming-iterator