-
Notifications
You must be signed in to change notification settings - Fork 781
docs: add Qwen3 think/no_think config example (#1636) #1707
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
WalkthroughA new section was added to the configuration examples documentation, describing how to use Qwen3 models in Thinking and Non-Thinking modes. Additionally, a comprehensive YAML configuration template for WrenAI was introduced, detailing model, pipeline, and system settings for various AI tasks. Changes
Sequence Diagram(s)sequenceDiagram
participant User
participant WrenAI
participant OpenRouter/Qwen3
User->>WrenAI: Send prompt with /think or /no_think suffix
WrenAI->>OpenRouter/Qwen3: Route request to Qwen3 model (thinking or fast)
OpenRouter/Qwen3-->>WrenAI: Return model response
WrenAI-->>User: Deliver response (complex or fast)
Assessment against linked issues
Suggested labels
Poem
Thanks for using CodeRabbit! It's free for OSS, and your support helps us grow. If you like it, consider giving us a shout-out. 🪧 TipsChatThere are 3 ways to chat with CodeRabbit:
SupportNeed help? Create a ticket on our support page for assistance with any issues or questions. Note: Be mindful of the bot's finite context window. It's strongly recommended to break down tasks such as reading entire modules into smaller chunks. For a focused discussion, use review comments to chat about specific files and their changes, instead of using the PR comments. CodeRabbit Commands (Invoked using PR comments)
Other keywords and placeholders
CodeRabbit Configuration File (
|
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Actionable comments posted: 0
🧹 Nitpick comments (9)
wren-ai-service/docs/config_examples/config.qwen3.yaml (5)
1-7
: Fix enumeration mismatch in initial instructions
The intro comment says "3 steps basically" but then lists four steps (1 to 4). Please update either the step count or the enumerated list to avoid confusion.
8-58
: Enhance maintainability by minimizing repetition in LLM definitions
Theapi_base
,timeout
, and commonkwargs
blocks are repeated across each model entry. Consider extracting shared fields into YAML anchors/aliases to DRY up these definitions. This will make future updates easier and reduce the risk of inconsistencies.
81-87
: Caution onrecreate_index: true
By default,recreate_index: true
will drop and rebuild your Qdrant index on startup. In production, this may lead to data loss—consider adding a warning comment or making this setting optional.
89-182
: Consider DRY pipelines and highlight Qwen3-specific flows
The pipelines block is comprehensive but quite repetitive:
- Use YAML anchors for shared fields (
llm
,embedder
,document_store
,engine
).- Group or annotate Qwen3-specific pipelines (e.g., those using
qwen3-thinking
orqwen3-fast
) at the top to improve discoverability.
Also verify that omittingembedder
orengine
keys falls back to intended defaults.
183-205
: Settings section looks solid
Runtime parameters are clearly defined. Remember to setdevelopment: false
for production deployments.wren-ai-service/docs/config_examples/README.md (4)
1-3
: Tone down the admonition header
The "MUST READ!!!" header is overly forceful. Consider using a softer phrasing like "Please read carefully" to improve the reader experience.🧰 Tools
🪛 LanguageTool
[style] ~1-~1: Using many exclamation marks might seem excessive (in this case: 3 exclamation marks for a text that’s 1464 characters long)
Context: # MUST READ!!! Since these config files are examples,...(EN_EXCESSIVE_EXCLAMATION)
5-5
: Refine wording for contribution section
The phrase "We also definitely welcome your contribution" can be streamlined to "We welcome contributions" for clarity and conciseness.
7-7
: Consistent naming for Non-Thinking mode
The section header uses "No_Think" (underscore) while the body calls it "Non-Thinking Mode". Align on one term—e.g., "Non-Thinking"—for consistency.
29-36
: Specify fenced code block language
The usage example code block lacks a language tag. Add ```yaml after the backticks to enable proper syntax highlighting.🧰 Tools
🪛 markdownlint-cli2 (0.17.2)
30-30: Fenced code blocks should have a language specified
null(MD040, fenced-code-language)
📜 Review details
Configuration used: CodeRabbit UI
Review profile: CHILL
Plan: Pro
📒 Files selected for processing (2)
wren-ai-service/docs/config_examples/README.md
(1 hunks)wren-ai-service/docs/config_examples/config.qwen3.yaml
(1 hunks)
🧰 Additional context used
🪛 markdownlint-cli2 (0.17.2)
wren-ai-service/docs/config_examples/README.md
30-30: Fenced code blocks should have a language specified
null
(MD040, fenced-code-language)
🔇 Additional comments (5)
wren-ai-service/docs/config_examples/config.qwen3.yaml (5)
8-24
: Clarify the role of thedefault
alias
The first model entry usesalias: default
with identical settings toqwen3-thinking
. Confirm whetherdefault
is intended as a separate baseline or if it should be removed or renamed to avoid duplication.
25-46
: Approve Non-Thinking mode parameters
Theqwen3-fast
alias is correctly configured for non-thinking use: higher temperature and reducedtop_p
for faster, simpler outputs, with atext
response format.
47-58
: Validate JSON handling forqwen3-32b
You've setresponse_format.type: json_object
for the 32B model. Ensure that downstream pipelines and consumers can parse JSON objects; otherwise, document any required transformations.
59-69
: Embedder configuration is clear
The embedder section provides sensible defaults and points users to setOPENAI_API_KEY
.
71-79
: Engine definitions are properly specified
Bothwren_ui
andwren_ibis
endpoints are defined correctly with clear provider and URL settings.
@s2010 interesting! could you give me examples explaining adding |
In WrenAI configuration the system will automatically routes:
/no_think queries for Quick Responses → "Show me today's sales numbers /no_think"
The configuration I created supports this intelligent routing through the pipeline settings, where reasoning tasks like sql_generation_reasoning use the thinking model, while quick tasks like sql_answer use the fast model. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
thanks for contribution!
This PR adds a new configuration example (
config.qwen3.yaml
) under docs/config_examples for the Qwen3 “think” and “no_think” endpoints. It mirrors the existing DeepSeek example, swapping in the correctapi_base
and model names.Closes #1636
Summary by CodeRabbit
Documentation
OPENROUTER_API_KEY
environment variable for OpenRouter integration.New Features