diff --git a/tests/anthropic.test.mjs b/tests/anthropic.test.mjs index 6faa6c4..4c2cca3 100644 --- a/tests/anthropic.test.mjs +++ b/tests/anthropic.test.mjs @@ -13,7 +13,7 @@ describe('Anthropic Test', () => { try { const message = await anthropic.messages.create({ model: "claude-3-opus-20240229", - max_tokens: 10, + max_tokens: 1, messages: [{ role: "user", content: "How to monitor LLM Applications in one sentence?" }], }); expect(message.type).to.equal('message'); @@ -28,7 +28,7 @@ describe('Anthropic Test', () => { DokuMetry.init({llm: anthropic, dokuUrl: process.env.DOKU_URL, apiKey: process.env.DOKU_TOKEN, environment: "dokumetry-testing", applicationName: "dokumetry-node-test", skipResp: false}); try { var stream = await anthropic.messages.create({ - max_tokens: 10, + max_tokens: 1, messages: [{ role: 'user', content: 'How to monitor LLM Applications in one sentence?' }], model: 'claude-3-opus-20240229', stream: true, diff --git a/tests/mistral.test.mjs b/tests/mistral.test.mjs index 6ada70b..2e15b9d 100644 --- a/tests/mistral.test.mjs +++ b/tests/mistral.test.mjs @@ -25,7 +25,7 @@ describe('Mistral Test', () => { const message = await client.chat({ model: 'mistral-large-latest', messages: [{role: 'user', content: 'What is LLM Observability?'}], - maxTokens: 10, + maxTokens: 1, }); expect(message.object).to.equal('chat.completion');