langchain-ai
langchain
BlogDocsChangelog

fix(openai): support reasoning_content in streaming delta chunks

#34152
Comparing
codeonbush:fix/openai-reasoning-content-streaming
(
82e6e5c
) with
master
(
12df938
)
CodSpeed Performance Gauge
0%
Untouched
6
Skipped
28

Benchmarks

Skipped (28)

test_create_chat_prompt_init_time
libs/partners/prompty/tests/unit_tests/test_standard.py
Skipped
311.9 µs*
test_exa_retriever_init_time
libs/partners/exa/tests/unit_tests/test_standard.py
Skipped
325.3 µs*
test_chroma_init_time
libs/partners/chroma/tests/unit_tests/test_standard.py
Skipped
57.2 ms*
test_nomic_embeddings_init_time
libs/partners/nomic/tests/unit_tests/test_standard.py
Skipped
1.5 ms*
test_init_time
libs/partners/ollama/tests/unit_tests/test_chat_models.py::TestChatOllama
Skipped
1.6 s*
test_init_time
libs/partners/fireworks/tests/unit_tests/test_standard.py::TestFireworksStandard
Skipped
6.6 s*
test_init_time
libs/partners/groq/tests/unit_tests/test_standard.py::TestGroqStandard
Skipped
1.6 s*
test_init_time
libs/partners/mistralai/tests/unit_tests/test_standard.py::TestMistralStandard
Skipped
9.2 ms*
test_init_time
libs/partners/deepseek/tests/unit_tests/test_chat_models.py::TestChatDeepSeekUnit
Skipped
1.6 s*
test_init_time
libs/partners/perplexity/tests/unit_tests/test_chat_models_standard.py::TestPerplexityStandard
Skipped
834.5 ms*
test_init_time
libs/partners/xai/tests/unit_tests/test_chat_models_standard.py::TestXAIStandard
Skipped
3.3 s*
test_qdrant_vectorstore_init_time
libs/partners/qdrant/tests/unit_tests/test_standard.py
Skipped
224.3 ms*
test_stream_time
libs/partners/anthropic/tests/integration_tests/test_standard.py::TestAnthropicStandard
Skipped
34.7 ms*
test_init_time
libs/partners/anthropic/tests/unit_tests/test_standard.py::TestAnthropicStandard
Skipped
807.8 µs*
test_init_time_with_client
libs/partners/anthropic/tests/unit_tests/test_standard.py
Skipped
2.3 ms*
test_import_time[HumanMessage]
libs/core/tests/benchmarks/test_imports.py
Skipped
245.5 ms*
test_import_time[BaseChatModel]
libs/core/tests/benchmarks/test_imports.py
Skipped
477.2 ms*
test_import_time[tool]
libs/core/tests/benchmarks/test_imports.py
Skipped
463.5 ms*
test_import_time[Document]
libs/core/tests/benchmarks/test_imports.py
Skipped
172.1 ms*
test_import_time[InMemoryRateLimiter]
libs/core/tests/benchmarks/test_imports.py
Skipped
162.3 ms*
test_import_time[PydanticOutputParser]
libs/core/tests/benchmarks/test_imports.py
Skipped
475.7 ms*
test_import_time[ChatPromptTemplate]
libs/core/tests/benchmarks/test_imports.py
Skipped
530.7 ms*
test_import_time[RunnableLambda]
libs/core/tests/benchmarks/test_imports.py
Skipped
444.5 ms*
test_import_time[CallbackManager]
libs/core/tests/benchmarks/test_imports.py
Skipped
414.4 ms*
test_import_time[Runnable]
libs/core/tests/benchmarks/test_imports.py
Skipped
445.9 ms*
test_import_time[LangChainTracer]
libs/core/tests/benchmarks/test_imports.py
Skipped
397.4 ms*
test_async_callbacks_in_sync
libs/core/tests/benchmarks/test_async_callbacks.py
Skipped
18.6 ms*
test_import_time[InMemoryVectorStore]
libs/core/tests/benchmarks/test_imports.py
Skipped
551 ms*

Passed

test_init_time
libs/partners/openai/tests/unit_tests/chat_models/test_responses_standard.py::TestOpenAIResponses
CodSpeed Performance Gauge
0%
12.3 ms*12.3 ms
test_stream_time
libs/partners/openai/tests/integration_tests/chat_models/test_base_standard.py::TestOpenAIStandard
CodSpeed Performance Gauge
0%
1.2 s*1.2 s
test_init_time
libs/partners/openai/tests/unit_tests/chat_models/test_base_standard.py::TestOpenAIStandard
CodSpeed Performance Gauge
0%
12.3 ms*12.2 ms
test_stream_time
libs/partners/openai/tests/integration_tests/chat_models/test_responses_standard.py::TestOpenAIStandard
CodSpeed Performance Gauge
0%
1.2 s*1.2 s
test_stream_time
libs/partners/openai/tests/integration_tests/chat_models/test_responses_standard.py::TestOpenAIResponses
CodSpeed Performance Gauge
0%
857.8 ms*857.3 ms
test_init_time
libs/partners/openai/tests/unit_tests/chat_models/test_azure_standard.py::TestOpenAIStandard
CodSpeed Performance Gauge
0%
1.7 s*1.7 s

Commits

Click on a commit to change the comparison range
Base
master
12df938
+0.14%
openai: support reasoning_content in streaming delta chunks This change adds support for the reasoning_content field in streaming delta chunks from the Chat Completions API. This field is used by: - LiteLLM when proxying Gemini models with reasoning_effort enabled - DeepSeek reasoning models (deepseek-reasoner) - vLLM with reasoning parser enabled - Other OpenAI-compatible providers that support reasoning/thinking output The reasoning_content is stored in additional_kwargs, consistent with how the Responses API handles reasoning content. Fixes #29513, #31326, #32981, #30580
82e6e5c
4 days ago
© 2025 CodSpeed Technology
Home Terms Privacy Docs