haskell-config for nvim & ge_proton install script
This commit is contained in:
@ -6,6 +6,8 @@ return {
|
||||
-- add any opts here
|
||||
-- for example
|
||||
provider = "ollama",
|
||||
mode = "legacy",
|
||||
disable_tools = true,
|
||||
behaviour = {
|
||||
enable_cursor_planning_mode = true, -- enable cursor planning mode!
|
||||
},
|
||||
@ -18,6 +20,7 @@ return {
|
||||
max_completion_tokens = 40000, -- Increase this to include reasoning tokens (for reasoning models)
|
||||
stream = true,
|
||||
thinking = true,
|
||||
disable_tools = true,
|
||||
--system_prompt = "Enable deep thinking subroutine.",
|
||||
-- reasoning_effort = "high", -- low|medium|high, only used for reasoning models
|
||||
},
|
||||
@ -26,10 +29,11 @@ return {
|
||||
model = "qwen3:32b",
|
||||
max_completion_tokens = 40000,
|
||||
reasoning_effort = "high",
|
||||
disable_tools = true,
|
||||
},
|
||||
},
|
||||
rag_service = {
|
||||
enabled = true, -- Enables the RAG service
|
||||
enabled = false, -- Enables the RAG service
|
||||
host_mount = os.getenv("HOME"), -- Host mount path for the rag service
|
||||
provider = "ollama", -- The provider to use for RAG service (e.g. openai or ollama)
|
||||
llm_model = "cogito", -- The LLM model to use for RAG service
|
||||
|
Reference in New Issue
Block a user