Skip to content
Closed
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
159 changes: 24 additions & 135 deletions package-lock.json

Large diffs are not rendered by default.

2 changes: 1 addition & 1 deletion plugins/openai/package.json
Original file line number Diff line number Diff line change
Expand Up @@ -27,7 +27,7 @@
"openai": "^4.95.0"
},
"peerDependencies": {
"genkit": "^0.9.0 || ^1.0.0"
"genkit": "^1.19.3"
},
"devDependencies": {
"@jest/globals": "^29.7.0",
Expand Down
8 changes: 4 additions & 4 deletions plugins/openai/src/dalle.ts
Original file line number Diff line number Diff line change
Expand Up @@ -13,10 +13,11 @@
* See the License for the specific language governing permissions and
* limitations under the License.
*/
import type { GenerateRequest, GenerateResponseData, Genkit } from 'genkit';
import type { GenerateRequest, GenerateResponseData } from 'genkit';
import { GenerationCommonConfigSchema, Message, z } from 'genkit';
import type { ModelAction } from 'genkit/model';
import { modelRef } from 'genkit/model';
import { model } from 'genkit/plugin';
import type OpenAI from 'openai';
import {
type ImageGenerateParams,
Expand All @@ -32,7 +33,7 @@ export const DallE3ConfigSchema = GenerationCommonConfigSchema.extend({
});

export const dallE3 = modelRef({
name: 'openai/dall-e-3',
name: 'dall-e-3',
info: {
label: 'OpenAI - DALL-E 3',
supports: {
Expand Down Expand Up @@ -90,10 +91,9 @@ function toGenerateResponse(result: ImagesResponse): GenerateResponseData {
}

export function dallE3Model(
ai: Genkit,
client: OpenAI
): ModelAction<typeof DallE3ConfigSchema> {
return ai.defineModel<typeof DallE3ConfigSchema>(
return model<typeof DallE3ConfigSchema>(
{
name: dallE3.name,
...dallE3.info,
Expand Down
30 changes: 15 additions & 15 deletions plugins/openai/src/embedder.ts
Original file line number Diff line number Diff line change
Expand Up @@ -17,10 +17,10 @@
// import { defineEmbedder, embedderRef } from '@genkit-ai/ai/embedder';

import OpenAI from 'openai';
import type { Genkit } from 'genkit';
import { embedderRef, z } from 'genkit';
import { embedderRef as createEmbedderRef, z } from 'genkit';

import { type PluginOptions } from './index.js';
import { embedder } from 'genkit/plugin';

export const TextEmbeddingConfigSchema = z.object({
dimensions: z.number().optional(),
Expand All @@ -33,8 +33,8 @@ export type TextEmbeddingGeckoConfig = z.infer<

export const TextEmbeddingInputSchema = z.string();

export const textEmbedding3Small = embedderRef({
name: 'openai/text-embedding-3-small',
export const textEmbedding3Small = createEmbedderRef({
name: 'text-embedding-3-small',
configSchema: TextEmbeddingConfigSchema,
info: {
dimensions: 1536,
Expand All @@ -45,8 +45,8 @@ export const textEmbedding3Small = embedderRef({
},
});

export const textEmbedding3Large = embedderRef({
name: 'openai/text-embedding-3-large',
export const textEmbedding3Large = createEmbedderRef({
name: 'text-embedding-3-large',
configSchema: TextEmbeddingConfigSchema,
info: {
dimensions: 3072,
Expand All @@ -57,8 +57,8 @@ export const textEmbedding3Large = embedderRef({
},
});

export const textEmbeddingAda002 = embedderRef({
name: 'openai/text-embedding-ada-002',
export const textEmbeddingAda002 = createEmbedderRef({
name: 'text-embedding-ada-002',
configSchema: TextEmbeddingConfigSchema,
info: {
dimensions: 1536,
Expand All @@ -76,7 +76,6 @@ export const SUPPORTED_EMBEDDING_MODELS = {
};

export function openaiEmbedder(
ai: Genkit,
name: string,
options?: PluginOptions
) {
Expand All @@ -85,17 +84,18 @@ export function openaiEmbedder(
throw new Error(
'please pass in the API key or set the OPENAI_API_KEY environment variable'
);
const model = SUPPORTED_EMBEDDING_MODELS[name];
if (!model) throw new Error(`Unsupported model: ${name}`);
const modelRef = SUPPORTED_EMBEDDING_MODELS[name];
if (!modelRef) throw new Error(`Unsupported model: ${name}`);

const client = new OpenAI({ apiKey });
return ai.defineEmbedder(
return embedder(
{
info: model.info!,
info: modelRef.info!,
configSchema: TextEmbeddingConfigSchema,
name: model.name,
name: modelRef.name,
},
async (input, options) => {
async (request, _) => {
const { input, options } = request;
const embeddings = await client.embeddings.create({
model: name,
input: input.map((d) => d.text),
Expand Down
44 changes: 19 additions & 25 deletions plugins/openai/src/gpt.test.ts
Original file line number Diff line number Diff line change
Expand Up @@ -22,7 +22,7 @@ import type {
ChatCompletionRole,
} from 'openai/resources/index.mjs';
import type OpenAI from 'openai';
import type { GenerateRequest, Genkit, MessageData, Part, Role } from 'genkit';
import type { GenerateRequest, MessageData, Part, Role } from 'genkit';
import type { CandidateData } from 'genkit/model';

import {
Expand All @@ -39,9 +39,9 @@ import {
} from './gpt';
import type { OpenAiConfigSchema } from './gpt';

jest.mock('@genkit-ai/ai/model', () => ({
...jest.requireActual('@genkit-ai/ai/model'),
defineModel: jest.fn(),
jest.mock('genkit/plugin', () => ({
...jest.requireActual('genkit/plugin'),
model: jest.fn(() => ({})),
}));

describe('toOpenAIRole', () => {
Expand Down Expand Up @@ -1358,22 +1358,15 @@ describe('gptRunner', () => {
});

describe('gptModel', () => {
let ai: Genkit;

beforeEach(() => {
ai = {
defineModel: jest.fn(),
} as unknown as Genkit;
});

afterEach(() => {
jest.clearAllMocks();
});

it('should correctly define supported GPT models', () => {
jest.spyOn(ai, 'defineModel').mockImplementation((() => ({})) as any);
gptModel(ai, 'gpt-4o', {} as OpenAI);
expect(ai.defineModel).toHaveBeenCalledWith(
const { model } = jest.requireMock('genkit/plugin');

gptModel('gpt-4o', {} as OpenAI);
expect(model).toHaveBeenCalledWith(
{
name: gpt4o.name,
...gpt4o.info,
Expand All @@ -1384,29 +1377,30 @@ describe('gptModel', () => {
});

it('should correctly define gpt-4.1, gpt-4.1-mini, and gpt-4.1-nano', () => {
jest.spyOn(ai, 'defineModel').mockImplementation((() => ({})) as any);
gptModel(ai, 'gpt-4.1', {} as OpenAI);
expect(ai.defineModel).toHaveBeenCalledWith(
const { model } = jest.requireMock('genkit/plugin');

gptModel('gpt-4.1', {} as OpenAI);
expect(model).toHaveBeenCalledWith(
{
name: 'openai/gpt-4.1',
name: 'gpt-4.1',
...require('./gpt').gpt41.info,
configSchema: require('./gpt').gpt41.configSchema,
},
expect.any(Function)
);
gptModel(ai, 'gpt-4.1-mini', {} as OpenAI);
expect(ai.defineModel).toHaveBeenCalledWith(
gptModel('gpt-4.1-mini', {} as OpenAI);
expect(model).toHaveBeenCalledWith(
{
name: 'openai/gpt-4.1-mini',
name: 'gpt-4.1-mini',
...require('./gpt').gpt41Mini.info,
configSchema: require('./gpt').gpt41Mini.configSchema,
},
expect.any(Function)
);
gptModel(ai, 'gpt-4.1-nano', {} as OpenAI);
expect(ai.defineModel).toHaveBeenCalledWith(
gptModel('gpt-4.1-nano', {} as OpenAI);
expect(model).toHaveBeenCalledWith(
{
name: 'openai/gpt-4.1-nano',
name: 'gpt-4.1-nano',
...require('./gpt').gpt41Nano.info,
configSchema: require('./gpt').gpt41Nano.configSchema,
},
Expand Down
Loading