1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
|
- import fetch from 'node-fetch';
- import { OpenAiCompletionProvider, OpenAiChatCompletionProvider } from '../src/providers/openai';
- import { disableCache, enableCache } from '../src/cache.js';
- import { loadApiProvider } from '../src/providers.js';
- jest.mock('node-fetch', () => jest.fn());
- jest.mock('../src/esm.js');
- describe('providers', () => {
- afterEach(() => {
- jest.clearAllMocks();
- });
- test('OpenAiCompletionProvider callApi', async () => {
- const mockResponse = {
- json: jest.fn().mockResolvedValue({
- choices: [{ text: 'Test output' }],
- usage: { total_tokens: 10, prompt_tokens: 5, completion_tokens: 5 },
- }),
- };
- (fetch as unknown as jest.Mock).mockResolvedValue(mockResponse);
- const provider = new OpenAiCompletionProvider('text-davinci-003', 'test-api-key');
- const result = await provider.callApi('Test prompt');
- expect(fetch).toHaveBeenCalledTimes(1);
- expect(result.output).toBe('Test output');
- expect(result.tokenUsage).toEqual({ total: 10, prompt: 5, completion: 5 });
- });
- test('OpenAiChatCompletionProvider callApi', async () => {
- const mockResponse = {
- json: jest.fn().mockResolvedValue({
- choices: [{ message: { content: 'Test output' } }],
- usage: { total_tokens: 10, prompt_tokens: 5, completion_tokens: 5 },
- }),
- };
- (fetch as unknown as jest.Mock).mockResolvedValue(mockResponse);
- const provider = new OpenAiChatCompletionProvider('gpt-3.5-turbo', 'test-api-key');
- const result = await provider.callApi('Test prompt');
- expect(fetch).toHaveBeenCalledTimes(1);
- expect(result.output).toBe('Test output');
- expect(result.tokenUsage).toEqual({ total: 10, prompt: 5, completion: 5 });
- });
- test('OpenAiChatCompletionProvider callApi with cache disabled', async () => {
- disableCache();
- const mockResponse = {
- json: jest.fn().mockResolvedValue({
- choices: [{ message: { content: 'Test output' } }],
- usage: { total_tokens: 10, prompt_tokens: 5, completion_tokens: 5 },
- }),
- };
- (fetch as unknown as jest.Mock).mockResolvedValue(mockResponse);
- const provider = new OpenAiChatCompletionProvider('gpt-3.5-turbo', 'test-api-key');
- const result = await provider.callApi('Test prompt');
- expect(fetch).toHaveBeenCalledTimes(1);
- expect(result.output).toBe('Test output');
- expect(result.tokenUsage).toEqual({ total: 10, prompt: 5, completion: 5 });
- enableCache();
- });
- test('loadApiProvider with openai:chat', async () => {
- const provider = await loadApiProvider('openai:chat');
- expect(provider).toBeInstanceOf(OpenAiChatCompletionProvider);
- });
- test('loadApiProvider with openai:completion', async () => {
- const provider = await loadApiProvider('openai:completion');
- expect(provider).toBeInstanceOf(OpenAiCompletionProvider);
- });
- test('loadApiProvider with openai:chat:modelName', async () => {
- const provider = await loadApiProvider('openai:chat:gpt-3.5-turbo');
- expect(provider).toBeInstanceOf(OpenAiChatCompletionProvider);
- });
- test('loadApiProvider with openai:completion:modelName', async () => {
- const provider = await loadApiProvider('openai:completion:text-davinci-003');
- expect(provider).toBeInstanceOf(OpenAiCompletionProvider);
- });
- });
|