Register
Login
Resources
Docs Blog Datasets Glossary Case Studies Tutorials & Webinars
Product
Data Engine LLMs Platform Enterprise
Pricing Explore
Connect to our Discord channel

deepseek.md 3.7 KB

You have to be logged in to leave a comment. Sign In
sidebar_label
DeepSeek

DeepSeek

DeepSeek provides an OpenAI-compatible API for their language models, with specialized models for both general chat and advanced reasoning tasks. The DeepSeek provider is compatible with all the options provided by the OpenAI provider.

Setup

  1. Get an API key from the DeepSeek Platform
  2. Set DEEPSEEK_API_KEY environment variable or specify apiKey in your config

Configuration

Basic configuration example:

providers:
  - id: deepseek:deepseek-chat
    config:
      temperature: 0.7
      max_tokens: 4000
      apiKey: YOUR_DEEPSEEK_API_KEY

  - id: deepseek:deepseek-reasoner # DeepSeek-R1 model
    config:
      max_tokens: 8000

Configuration Options

  • temperature
  • max_tokens
  • top_p, presence_penalty, frequency_penalty
  • stream
  • showThinking - Control whether reasoning content is included in the output (default: true, applies to deepseek-reasoner model)

Available Models

:::note

The API model names are aliases that automatically point to the latest versions: deepseek-chat points to DeepSeek-V3-0324 and deepseek-reasoner points to DeepSeek-R1-0528.

:::

deepseek-chat (DeepSeek-V3)

  • General purpose model for conversations and content
  • 64K context window, 8K output tokens
  • Input: $0.07/1M (cache), $0.27/1M (no cache)
  • Output: $1.10/1M

deepseek-reasoner (DeepSeek-R1)

  • Specialized for reasoning and problem-solving
  • 64K context, 32K reasoning tokens, 8K output tokens
  • Input: $0.14/1M (cache), $0.55/1M (no cache)
  • Output: $2.19/1M
  • Supports showing or hiding reasoning content through the showThinking parameter

:::warning

The reasoning model does not support temperature, top_p, presence_penalty, frequency_penalty, logprobs, or top_logprobs parameters. Setting these parameters will not trigger an error but will have no effect.

:::

Example Usage

Here's an example comparing DeepSeek with OpenAI on reasoning tasks:

providers:
  - id: deepseek:deepseek-reasoner
    config:
      max_tokens: 8000
      showThinking: true # Include reasoning content in output (default)
  - id: openai:o-1
    config:
      temperature: 0.0

prompts:
  - 'Solve this step by step: {{math_problem}}'

tests:
  - vars:
      math_problem: 'What is the derivative of x^3 + 2x with respect to x?'

Controlling Reasoning Output

The DeepSeek-R1 model (deepseek-reasoner) includes detailed reasoning steps in its output. You can control whether this reasoning content is shown using the showThinking parameter:

providers:
  - id: deepseek:deepseek-reasoner
    config:
      showThinking: false # Hide reasoning content from output

When showThinking is set to true (default), the output includes both reasoning and the final answer in a standardized format:

Thinking: <reasoning content>

<final answer>

When set to false, only the final answer is included in the output. This is useful when you want better reasoning quality but don't want to expose the reasoning process to end users or in your assertions.

See our complete example that benchmarks it against OpenAI's o1 model on the MMLU reasoning tasks.

API Details

  • Base URL: https://api.deepseek.com/v1
  • OpenAI-compatible API format
  • Full API documentation

See Also

Tip!

Press p or to see the previous file or, n or to see the next file

Comments

Loading...