Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
22 changes: 22 additions & 0 deletions assets/src/generated/graphql.ts
Original file line number Diff line number Diff line change
Expand Up @@ -1011,6 +1011,8 @@ export type AnthropicSettings = {
__typename?: 'AnthropicSettings';
/** the anthropic model version to use */
model?: Maybe<Scalars['String']['output']>;
/** addditional models to support within the integrated ai proxy */
proxyModels?: Maybe<Array<Maybe<Scalars['String']['output']>>>;
/** the model to use for tool calls, which are less frequent and require more complex reasoning */
toolModel?: Maybe<Scalars['String']['output']>;
};
Expand All @@ -1020,6 +1022,8 @@ export type AnthropicSettingsAttributes = {
/** the model to use for vector embeddings */
embeddingModel?: InputMaybe<Scalars['String']['input']>;
model?: InputMaybe<Scalars['String']['input']>;
/** addditional models to support within the integrated ai proxy */
proxyModels?: InputMaybe<Array<InputMaybe<Scalars['String']['input']>>>;
/** the model to use for tool calls, which are less frequent and require more complex reasoning */
toolModel?: InputMaybe<Scalars['String']['input']>;
};
Expand Down Expand Up @@ -1337,6 +1341,8 @@ export type AzureOpenaiAttributes = {
endpoint: Scalars['String']['input'];
/** the exact model you wish to use */
model?: InputMaybe<Scalars['String']['input']>;
/** addditional models to support within the integrated ai proxy */
proxyModels?: InputMaybe<Array<InputMaybe<Scalars['String']['input']>>>;
/** the model to use for tool calls, which are less frequent and require more complex reasoning */
toolModel?: InputMaybe<Scalars['String']['input']>;
};
Expand All @@ -1351,6 +1357,8 @@ export type AzureOpenaiSettings = {
/** the endpoint of your azure openai version, should look like: https://{endpoint}/openai/deployments/{deployment-id} */
endpoint: Scalars['String']['output'];
model?: Maybe<Scalars['String']['output']>;
/** addditional models to support within the integrated ai proxy */
proxyModels?: Maybe<Array<Maybe<Scalars['String']['output']>>>;
/** the model to use for tool calls, which are less frequent and require more complex reasoning */
toolModel?: Maybe<Scalars['String']['output']>;
};
Expand Down Expand Up @@ -1402,6 +1410,8 @@ export type BedrockAiAttributes = {
embeddingModel?: InputMaybe<Scalars['String']['input']>;
/** the bedrock model id to use */
modelId: Scalars['String']['input'];
/** addditional models to support within the integrated ai proxy */
proxyModels?: InputMaybe<Array<InputMaybe<Scalars['String']['input']>>>;
/** the aws region the model is hosted in */
region?: InputMaybe<Scalars['String']['input']>;
/** the model to use for tool calls, which are less frequent and require more complex reasoning */
Expand All @@ -1417,6 +1427,8 @@ export type BedrockAiSettings = {
embeddingModel?: Maybe<Scalars['String']['output']>;
/** the bedrock model to use */
modelId: Scalars['String']['output'];
/** addditional models to support within the integrated ai proxy */
proxyModels?: Maybe<Array<Maybe<Scalars['String']['output']>>>;
/** the aws region the model is hosted in */
region?: Maybe<Scalars['String']['output']>;
/** the model to use for tool calls, which are less frequent and require more complex reasoning */
Expand Down Expand Up @@ -6227,6 +6239,8 @@ export type OllamaAttributes = {
/** the model to use for vector embeddings */
embeddingModel?: InputMaybe<Scalars['String']['input']>;
model: Scalars['String']['input'];
/** addditional models to support within the integrated ai proxy */
proxyModels?: InputMaybe<Array<InputMaybe<Scalars['String']['input']>>>;
/** the model to use for tool calls, which are less frequent and require more complex reasoning */
toolModel?: InputMaybe<Scalars['String']['input']>;
url: Scalars['String']['input'];
Expand All @@ -6251,6 +6265,8 @@ export type OpenaiSettings = {
embeddingModel?: Maybe<Scalars['String']['output']>;
/** the openai model version to use */
model?: Maybe<Scalars['String']['output']>;
/** addditional models to support within the integrated ai proxy */
proxyModels?: Maybe<Array<Maybe<Scalars['String']['output']>>>;
/** the model to use for tool calls, which are less frequent and require more complex reasoning */
toolModel?: Maybe<Scalars['String']['output']>;
};
Expand All @@ -6261,6 +6277,8 @@ export type OpenaiSettingsAttributes = {
/** the model to use for vector embeddings */
embeddingModel?: InputMaybe<Scalars['String']['input']>;
model?: InputMaybe<Scalars['String']['input']>;
/** addditional models to support within the integrated ai proxy */
proxyModels?: InputMaybe<Array<InputMaybe<Scalars['String']['input']>>>;
/** the model to use for tool calls, which are less frequent and require more complex reasoning */
toolModel?: InputMaybe<Scalars['String']['input']>;
};
Expand Down Expand Up @@ -13529,6 +13547,8 @@ export type VertexAiAttributes = {
model?: InputMaybe<Scalars['String']['input']>;
/** the gcp project id to use */
project: Scalars['String']['input'];
/** addditional models to support within the integrated ai proxy */
proxyModels?: InputMaybe<Array<InputMaybe<Scalars['String']['input']>>>;
/** optional service account json to auth to the GCP vertex apis */
serviceAccountJson?: InputMaybe<Scalars['String']['input']>;
/** the model to use for tool calls, which are less frequent and require more complex reasoning */
Expand All @@ -13546,6 +13566,8 @@ export type VertexAiSettings = {
model?: Maybe<Scalars['String']['output']>;
/** the gcp project id to use */
project: Scalars['String']['output'];
/** addditional models to support within the integrated ai proxy */
proxyModels?: Maybe<Array<Maybe<Scalars['String']['output']>>>;
/** the model to use for tool calls, which are less frequent and require more complex reasoning */
toolModel?: Maybe<Scalars['String']['output']>;
};
Expand Down
2 changes: 1 addition & 1 deletion charts/console-rapid/Chart.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -2,7 +2,7 @@ apiVersion: v2
name: console-rapid
description: rapid channel chart for the plural console (used for testing)
appVersion: 0.11.65
version: 0.3.175
version: 0.4.0
dependencies:
- name: kas
version: 0.3.1
Expand Down
Binary file modified charts/console-rapid/charts/controller-0.0.169.tgz
Binary file not shown.
Binary file modified charts/console-rapid/charts/kas-0.3.1.tgz
Binary file not shown.
Binary file modified charts/console/charts/controller-0.0.169.tgz
Binary file not shown.
Binary file modified charts/console/charts/kas-0.3.1.tgz
Binary file not shown.
Original file line number Diff line number Diff line change
Expand Up @@ -95,6 +95,12 @@ spec:
model:
description: Model is the LLM model name to use.
type: string
proxyModels:
description: ProxyModels are additional models to support
within our integrated ai proxy.
items:
type: string
type: array
tokenSecretRef:
description: |-
TokenSecretRef is a reference to the local secret holding the token to access
Expand Down Expand Up @@ -149,6 +155,12 @@ spec:
description: Model - the OpenAi model you wish to use. If
not specified, Plural will provide a default.
type: string
proxyModels:
description: ProxyModels are additional models to support
within the integrated ai proxy.
items:
type: string
type: array
tokenSecretRef:
description: |-
TokenSecretRef is a reference to the local secret holding the token to access
Expand Down Expand Up @@ -222,6 +234,12 @@ spec:
will use the openai compatible endpoint, so the model id
must be supported.
type: string
proxyModels:
description: ProxyModels are additional models to support
within the integrated ai proxy.
items:
type: string
type: array
region:
description: Region is the AWS region the model is hosted
in
Expand Down Expand Up @@ -393,6 +411,12 @@ spec:
model:
description: Model is the LLM model name to use.
type: string
proxyModels:
description: ProxyModels are additional models to support
within our integrated ai proxy.
items:
type: string
type: array
tokenSecretRef:
description: |-
TokenSecretRef is a reference to the local secret holding the token to access
Expand Down Expand Up @@ -630,6 +654,12 @@ spec:
project:
description: Project is the GCP project you'll be using
type: string
proxyModels:
description: ProxyModels are additional models to support
within the integrated ai proxy.
items:
type: string
type: array
serviceAccountJsonSecretRef:
description: ServiceAccountJsonSecretRef is a Service Account
json file stored w/in a kubernetes secret to use for authentication
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -639,6 +639,9 @@ spec:
items:
properties:
helm:
description: Helm is the Helm configuration to use for this
renderer, use if `path` points to a helm chart and you
want to override the default helm configuration.
properties:
ignoreHooks:
description: IgnoreHooks indicates whether to ignore
Expand All @@ -660,8 +663,10 @@ spec:
type: array
type: object
path:
description: Path is the path to the renderer works under.
type: string
type:
description: Type is the type of renderer to use.
enum:
- AUTO
- RAW
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -9955,9 +9955,7 @@ spec:
It is useful for projects with external modules or nested folder structures.
type: string
required:
- clusterRef
- git
- repositoryRef
- type
type: object
status:
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -515,6 +515,9 @@ spec:
items:
properties:
helm:
description: Helm is the Helm configuration to use for this
renderer, use if `path` points to a helm chart and you
want to override the default helm configuration.
properties:
ignoreHooks:
description: IgnoreHooks indicates whether to ignore
Expand All @@ -536,8 +539,10 @@ spec:
type: array
type: object
path:
description: Path is the path to the renderer works under.
type: string
type:
description: Type is the type of renderer to use.
enum:
- AUTO
- RAW
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -569,6 +569,9 @@ spec:
items:
properties:
helm:
description: Helm is the Helm configuration to use for this
renderer, use if `path` points to a helm chart and you
want to override the default helm configuration.
properties:
ignoreHooks:
description: IgnoreHooks indicates whether to ignore
Expand All @@ -590,8 +593,10 @@ spec:
type: array
type: object
path:
description: Path is the path to the renderer works under.
type: string
type:
description: Type is the type of renderer to use.
enum:
- AUTO
- RAW
Expand Down
60 changes: 60 additions & 0 deletions charts/controller/crds/deployments.plural.sh_servicecontexts.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -50,6 +50,51 @@ spec:
spec:
description: ServiceContextSpec defines the desired state of the ServiceContext.
properties:
configMapRef:
description: |-
ConfigMapRef references a ConfigMap containing configuration data to merge into the Configuration.
The keys and values from the ConfigMap will be merged into the Configuration JSON.
properties:
apiVersion:
description: API version of the referent.
type: string
fieldPath:
description: |-
If referring to a piece of an object instead of an entire object, this string
should contain a valid JSON/Go field access statement, such as desiredState.manifest.containers[2].
For example, if the object reference is to a container within a pod, this would take on a value like:
"spec.containers{name}" (where "name" refers to the name of the container that triggered
the event) or if no container name is specified "spec.containers[2]" (container with
index 2 in this pod). This syntax is chosen only to have some well-defined way of
referencing a part of an object.
type: string
kind:
description: |-
Kind of the referent.
More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#types-kinds
type: string
name:
description: |-
Name of the referent.
More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names
type: string
namespace:
description: |-
Namespace of the referent.
More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/namespaces/
type: string
resourceVersion:
description: |-
Specific resourceVersion to which this reference is made, if any.
More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency
type: string
uid:
description: |-
UID of the referent.
More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#uids
type: string
type: object
x-kubernetes-map-type: atomic
configuration:
description: |-
Configuration is a reusable configuration context that can include any JSON-compatible configuration data
Expand Down Expand Up @@ -129,6 +174,21 @@ spec:
example: 5m30s
type: string
type: object
secretRef:
description: |-
SecretRef references a Secret containing configuration data to merge into the Configuration.
The keys and values from the Secret will be merged into the Configuration JSON.
properties:
name:
description: name is unique within a namespace to reference a
secret resource.
type: string
namespace:
description: namespace defines the space within which the secret
name must be unique.
type: string
type: object
x-kubernetes-map-type: atomic
type: object
status:
properties:
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -558,6 +558,9 @@ spec:
items:
properties:
helm:
description: Helm is the Helm configuration to use for this
renderer, use if `path` points to a helm chart and you want
to override the default helm configuration.
properties:
ignoreHooks:
description: IgnoreHooks indicates whether to ignore Helm
Expand All @@ -579,8 +582,10 @@ spec:
type: array
type: object
path:
description: Path is the path to the renderer works under.
type: string
type:
description: Type is the type of renderer to use.
enum:
- AUTO
- RAW
Expand Down
Loading
Loading