refactor(ai): streaming chunks to avoid excessive redraw. (#73)
* perf(ai): token streaming with quick refactoring Signed-off-by: Aaron Pham <contact@aarnphm.xyz> * fix: window resize and AvanteSwitchProvider Signed-off-by: Aaron Pham <contact@aarnphm.xyz> * revert: config change Signed-off-by: Aaron Pham <contact@aarnphm.xyz> * chore: return early Signed-off-by: Aaron Pham <contact@aarnphm.xyz> --------- Signed-off-by: Aaron Pham <contact@aarnphm.xyz>
This commit is contained in:
		
							parent
							
								
									0fddfc7d8f
								
							
						
					
					
						commit
						5fa4f701dd
					
				@ -1,4 +1,3 @@
 | 
			
		||||
local fn = vim.fn
 | 
			
		||||
local api = vim.api
 | 
			
		||||
 | 
			
		||||
local curl = require("plenary.curl")
 | 
			
		||||
@ -6,10 +5,17 @@ local curl = require("plenary.curl")
 | 
			
		||||
local Utils = require("avante.utils")
 | 
			
		||||
local Config = require("avante.config")
 | 
			
		||||
local Tiktoken = require("avante.tiktoken")
 | 
			
		||||
local Dressing = require("avante.ui.dressing")
 | 
			
		||||
 | 
			
		||||
---@private
 | 
			
		||||
---@class AvanteAiBotInternal
 | 
			
		||||
local H = {}
 | 
			
		||||
 | 
			
		||||
---@class avante.AiBot
 | 
			
		||||
local M = {}
 | 
			
		||||
 | 
			
		||||
M.CANCEL_PATTERN = "AvanteAiBotEscape"
 | 
			
		||||
 | 
			
		||||
---@class EnvironmentHandler: table<[Provider], string>
 | 
			
		||||
local E = {
 | 
			
		||||
  ---@type table<Provider, string>
 | 
			
		||||
@ -20,7 +26,6 @@ local E = {
 | 
			
		||||
    deepseek = "DEEPSEEK_API_KEY",
 | 
			
		||||
    groq = "GROQ_API_KEY",
 | 
			
		||||
  },
 | 
			
		||||
  _once = false,
 | 
			
		||||
}
 | 
			
		||||
 | 
			
		||||
E = setmetatable(E, {
 | 
			
		||||
@ -29,30 +34,32 @@ E = setmetatable(E, {
 | 
			
		||||
    return os.getenv(E.env[k]) and true or false
 | 
			
		||||
  end,
 | 
			
		||||
})
 | 
			
		||||
E._once = false
 | 
			
		||||
 | 
			
		||||
--- return the environment variable name for the given provider
 | 
			
		||||
---@param provider? Provider
 | 
			
		||||
---@return string the envvar key
 | 
			
		||||
E.key = function(provider)
 | 
			
		||||
  provider = provider or Config.provider
 | 
			
		||||
  local var = E.env[provider]
 | 
			
		||||
  return type(var) == "table" and var[1] ---@cast var string
 | 
			
		||||
    or var
 | 
			
		||||
  return E.env[provider or Config.provider]
 | 
			
		||||
end
 | 
			
		||||
 | 
			
		||||
E.setup = function(var)
 | 
			
		||||
  local Dressing = require("avante.ui.dressing")
 | 
			
		||||
---@param provider? Provider
 | 
			
		||||
E.value = function(provider)
 | 
			
		||||
  return os.getenv(E.key(provider or Config.provider))
 | 
			
		||||
end
 | 
			
		||||
 | 
			
		||||
  if E._once then
 | 
			
		||||
    return
 | 
			
		||||
  end
 | 
			
		||||
--- intialize the environment variable for current neovim session.
 | 
			
		||||
--- This will only run once and spawn a UI for users to input the envvar.
 | 
			
		||||
---@param var Provider supported providers
 | 
			
		||||
---@param refresh? boolean
 | 
			
		||||
E.setup = function(var, refresh)
 | 
			
		||||
  refresh = refresh or false
 | 
			
		||||
 | 
			
		||||
  ---@param value string
 | 
			
		||||
  ---@return nil
 | 
			
		||||
  local function on_confirm(value)
 | 
			
		||||
    if value then
 | 
			
		||||
      vim.fn.setenv(var, value)
 | 
			
		||||
      E._once = true
 | 
			
		||||
    else
 | 
			
		||||
      if not E[Config.provider] then
 | 
			
		||||
        vim.notify_once("Failed to set " .. var .. ". Avante won't work as expected", vim.log.levels.WARN)
 | 
			
		||||
@ -60,6 +67,12 @@ E.setup = function(var)
 | 
			
		||||
    end
 | 
			
		||||
  end
 | 
			
		||||
 | 
			
		||||
  if refresh then
 | 
			
		||||
    vim.defer_fn(function()
 | 
			
		||||
      Dressing.initialize_input_buffer({ opts = { prompt = "Enter " .. var .. ": " }, on_confirm = on_confirm })
 | 
			
		||||
    end, 200)
 | 
			
		||||
  elseif not E._once then
 | 
			
		||||
    E._once = true
 | 
			
		||||
    api.nvim_create_autocmd({ "BufEnter", "BufWinEnter" }, {
 | 
			
		||||
      pattern = "*",
 | 
			
		||||
      once = true,
 | 
			
		||||
@ -77,6 +90,7 @@ E.setup = function(var)
 | 
			
		||||
            "ministarter",
 | 
			
		||||
            "TelescopePrompt",
 | 
			
		||||
            "gitcommit",
 | 
			
		||||
            "gitrebase",
 | 
			
		||||
          }
 | 
			
		||||
          if
 | 
			
		||||
            not vim.tbl_contains(exclude_buftypes, vim.bo.buftype)
 | 
			
		||||
@ -87,8 +101,11 @@ E.setup = function(var)
 | 
			
		||||
        end, 200)
 | 
			
		||||
      end,
 | 
			
		||||
    })
 | 
			
		||||
  end
 | 
			
		||||
end
 | 
			
		||||
 | 
			
		||||
------------------------------Prompt and type------------------------------
 | 
			
		||||
 | 
			
		||||
local system_prompt = [[
 | 
			
		||||
You are an excellent programming expert.
 | 
			
		||||
]]
 | 
			
		||||
@ -137,38 +154,49 @@ Replace lines: {{start_line}}-{{end_line}}
 | 
			
		||||
Remember: Accurate line numbers are CRITICAL. The range start_line to end_line must include ALL lines to be replaced, from the very first to the very last. Double-check every range before finalizing your response, paying special attention to the start_line to ensure it hasn't shifted down. Ensure that your line numbers perfectly match the original code structure without any overall shift.
 | 
			
		||||
]]
 | 
			
		||||
 | 
			
		||||
local function call_claude_api_stream(question, code_lang, code_content, selected_code_content, on_chunk, on_complete)
 | 
			
		||||
  local api_key = os.getenv(E.key("claude"))
 | 
			
		||||
---@class AvantePromptOptions: table<[string], string>
 | 
			
		||||
---@field question string
 | 
			
		||||
---@field code_lang string
 | 
			
		||||
---@field code_content string
 | 
			
		||||
---@field selected_code_content? string
 | 
			
		||||
---
 | 
			
		||||
---@alias AvanteAiMessageBuilder fun(opts: AvantePromptOptions): {role: "user" | "system", content: string | table<string, any>}[]
 | 
			
		||||
---
 | 
			
		||||
---@class AvanteCurlOutput: {url: string, body: table<string, any> | string, headers: table<string, string>}
 | 
			
		||||
---@alias AvanteCurlArgsBuilder fun(code_opts: AvantePromptOptions): AvanteCurlOutput
 | 
			
		||||
---
 | 
			
		||||
---@class ResponseParser
 | 
			
		||||
---@field event_state string
 | 
			
		||||
---@field on_chunk fun(chunk: string): any
 | 
			
		||||
---@field on_complete fun(err: string|nil): any
 | 
			
		||||
---@field on_error? fun(err_type: string): nil
 | 
			
		||||
---@alias AvanteAiResponseParser fun(data_stream: string, opts: ResponseParser): nil
 | 
			
		||||
 | 
			
		||||
  local tokens = Config.claude.max_tokens
 | 
			
		||||
  local headers = {
 | 
			
		||||
    ["Content-Type"] = "application/json",
 | 
			
		||||
    ["x-api-key"] = api_key,
 | 
			
		||||
    ["anthropic-version"] = "2023-06-01",
 | 
			
		||||
    ["anthropic-beta"] = "prompt-caching-2024-07-31",
 | 
			
		||||
  }
 | 
			
		||||
------------------------------Anthropic------------------------------
 | 
			
		||||
 | 
			
		||||
---@type AvanteAiMessageBuilder
 | 
			
		||||
H.make_claude_message = function(opts)
 | 
			
		||||
  local code_prompt_obj = {
 | 
			
		||||
    type = "text",
 | 
			
		||||
    text = string.format("<code>```%s\n%s```</code>", code_lang, code_content),
 | 
			
		||||
    text = string.format("<code>```%s\n%s```</code>", opts.code_lang, opts.code_content),
 | 
			
		||||
  }
 | 
			
		||||
 | 
			
		||||
  if Tiktoken.count(code_prompt_obj.text) > 1024 then
 | 
			
		||||
    code_prompt_obj.cache_control = { type = "ephemeral" }
 | 
			
		||||
  end
 | 
			
		||||
 | 
			
		||||
  if selected_code_content then
 | 
			
		||||
    code_prompt_obj.text = string.format("<code_context>```%s\n%s```</code_context>", code_lang, code_content)
 | 
			
		||||
  if opts.selected_code_content then
 | 
			
		||||
    code_prompt_obj.text = string.format("<code_context>```%s\n%s```</code_context>", opts.code_lang, opts.code_content)
 | 
			
		||||
  end
 | 
			
		||||
 | 
			
		||||
  local message_content = {
 | 
			
		||||
    code_prompt_obj,
 | 
			
		||||
  }
 | 
			
		||||
 | 
			
		||||
  if selected_code_content then
 | 
			
		||||
  if opts.selected_code_content then
 | 
			
		||||
    local selected_code_obj = {
 | 
			
		||||
      type = "text",
 | 
			
		||||
      text = string.format("<code>```%s\n%s```</code>", code_lang, selected_code_content),
 | 
			
		||||
      text = string.format("<code>```%s\n%s```</code>", opts.code_lang, opts.selected_code_content),
 | 
			
		||||
    }
 | 
			
		||||
 | 
			
		||||
    if Tiktoken.count(selected_code_obj.text) > 1024 then
 | 
			
		||||
@ -180,7 +208,7 @@ local function call_claude_api_stream(question, code_lang, code_content, selecte
 | 
			
		||||
 | 
			
		||||
  table.insert(message_content, {
 | 
			
		||||
    type = "text",
 | 
			
		||||
    text = string.format("<question>%s</question>", question),
 | 
			
		||||
    text = string.format("<question>%s</question>", opts.question),
 | 
			
		||||
  })
 | 
			
		||||
 | 
			
		||||
  local user_prompt = base_user_prompt
 | 
			
		||||
@ -196,220 +224,284 @@ local function call_claude_api_stream(question, code_lang, code_content, selecte
 | 
			
		||||
 | 
			
		||||
  table.insert(message_content, user_prompt_obj)
 | 
			
		||||
 | 
			
		||||
  local body = {
 | 
			
		||||
    model = Config.claude.model,
 | 
			
		||||
    system = system_prompt,
 | 
			
		||||
    messages = {
 | 
			
		||||
  return {
 | 
			
		||||
    {
 | 
			
		||||
      role = "user",
 | 
			
		||||
      content = message_content,
 | 
			
		||||
    },
 | 
			
		||||
    },
 | 
			
		||||
    stream = true,
 | 
			
		||||
    temperature = Config.claude.temperature,
 | 
			
		||||
    max_tokens = tokens,
 | 
			
		||||
  }
 | 
			
		||||
 | 
			
		||||
  local url = Utils.trim_suffix(Config.claude.endpoint, "/") .. "/v1/messages"
 | 
			
		||||
 | 
			
		||||
  curl.post(url, {
 | 
			
		||||
    ---@diagnostic disable-next-line: unused-local
 | 
			
		||||
    stream = function(err, data, job)
 | 
			
		||||
      if err then
 | 
			
		||||
        on_complete(err)
 | 
			
		||||
        return
 | 
			
		||||
      end
 | 
			
		||||
      if not data then
 | 
			
		||||
        return
 | 
			
		||||
      end
 | 
			
		||||
      for _, line in ipairs(vim.split(data, "\n")) do
 | 
			
		||||
        if line:sub(1, 6) ~= "data: " then
 | 
			
		||||
          return
 | 
			
		||||
        end
 | 
			
		||||
        vim.schedule(function()
 | 
			
		||||
          local success, parsed = pcall(fn.json_decode, line:sub(7))
 | 
			
		||||
          if not success then
 | 
			
		||||
            error("Error: failed to parse json: " .. parsed)
 | 
			
		||||
            return
 | 
			
		||||
          end
 | 
			
		||||
          if parsed and parsed.type == "content_block_delta" then
 | 
			
		||||
            on_chunk(parsed.delta.text)
 | 
			
		||||
          elseif parsed and parsed.type == "message_stop" then
 | 
			
		||||
            -- Stream request completed
 | 
			
		||||
            on_complete(nil)
 | 
			
		||||
          elseif parsed and parsed.type == "error" then
 | 
			
		||||
            -- Stream request completed
 | 
			
		||||
            on_complete(parsed)
 | 
			
		||||
          end
 | 
			
		||||
        end)
 | 
			
		||||
      end
 | 
			
		||||
    end,
 | 
			
		||||
    headers = headers,
 | 
			
		||||
    body = fn.json_encode(body),
 | 
			
		||||
  })
 | 
			
		||||
end
 | 
			
		||||
 | 
			
		||||
local function call_openai_api_stream(question, code_lang, code_content, selected_code_content, on_chunk, on_complete)
 | 
			
		||||
  local api_key = os.getenv(E.key("openai"))
 | 
			
		||||
---@type AvanteAiResponseParser
 | 
			
		||||
H.parse_claude_response = function(data_stream, opts)
 | 
			
		||||
  if opts.event_state == "content_block_delta" then
 | 
			
		||||
    local json = vim.json.decode(data_stream)
 | 
			
		||||
    opts.on_chunk(json.delta.text)
 | 
			
		||||
  elseif opts.event_state == "message_stop" then
 | 
			
		||||
    opts.on_complete(nil)
 | 
			
		||||
    return
 | 
			
		||||
  elseif opts.event_state == "error" then
 | 
			
		||||
    opts.on_complete(vim.json.decode(data_stream))
 | 
			
		||||
  end
 | 
			
		||||
end
 | 
			
		||||
 | 
			
		||||
---@type AvanteCurlArgsBuilder
 | 
			
		||||
H.make_claude_curl_args = function(code_opts)
 | 
			
		||||
  return {
 | 
			
		||||
    url = Utils.trim(Config.claude.endpoint, { suffix = "/" }) .. "/v1/messages",
 | 
			
		||||
    headers = {
 | 
			
		||||
      ["Content-Type"] = "application/json",
 | 
			
		||||
      ["x-api-key"] = E.value("claude"),
 | 
			
		||||
      ["anthropic-version"] = "2023-06-01",
 | 
			
		||||
      ["anthropic-beta"] = "prompt-caching-2024-07-31",
 | 
			
		||||
    },
 | 
			
		||||
    body = {
 | 
			
		||||
      model = Config.claude.model,
 | 
			
		||||
      system = system_prompt,
 | 
			
		||||
      stream = true,
 | 
			
		||||
      messages = H.make_claude_message(code_opts),
 | 
			
		||||
      temperature = Config.claude.temperature,
 | 
			
		||||
      max_tokens = Config.claude.max_tokens,
 | 
			
		||||
    },
 | 
			
		||||
  }
 | 
			
		||||
end
 | 
			
		||||
 | 
			
		||||
------------------------------OpenAI------------------------------
 | 
			
		||||
 | 
			
		||||
---@type AvanteAiMessageBuilder
 | 
			
		||||
H.make_openai_message = function(opts)
 | 
			
		||||
  local user_prompt = base_user_prompt
 | 
			
		||||
    .. "\n\nCODE:\n"
 | 
			
		||||
    .. "```"
 | 
			
		||||
    .. code_lang
 | 
			
		||||
    .. opts.code_lang
 | 
			
		||||
    .. "\n"
 | 
			
		||||
    .. code_content
 | 
			
		||||
    .. opts.code_content
 | 
			
		||||
    .. "\n```"
 | 
			
		||||
    .. "\n\nQUESTION:\n"
 | 
			
		||||
    .. question
 | 
			
		||||
    .. opts.question
 | 
			
		||||
 | 
			
		||||
  if selected_code_content then
 | 
			
		||||
  if opts.selected_code_content ~= nil then
 | 
			
		||||
    user_prompt = base_user_prompt
 | 
			
		||||
      .. "\n\nCODE CONTEXT:\n"
 | 
			
		||||
      .. "```"
 | 
			
		||||
      .. code_lang
 | 
			
		||||
      .. opts.code_lang
 | 
			
		||||
      .. "\n"
 | 
			
		||||
      .. code_content
 | 
			
		||||
      .. opts.code_content
 | 
			
		||||
      .. "\n```"
 | 
			
		||||
      .. "\n\nCODE:\n"
 | 
			
		||||
      .. "```"
 | 
			
		||||
      .. code_lang
 | 
			
		||||
      .. opts.code_lang
 | 
			
		||||
      .. "\n"
 | 
			
		||||
      .. selected_code_content
 | 
			
		||||
      .. opts.selected_code_content
 | 
			
		||||
      .. "\n```"
 | 
			
		||||
      .. "\n\nQUESTION:\n"
 | 
			
		||||
      .. question
 | 
			
		||||
      .. opts.question
 | 
			
		||||
  end
 | 
			
		||||
 | 
			
		||||
  local url, headers, body
 | 
			
		||||
  if Config.provider == "azure" then
 | 
			
		||||
    api_key = os.getenv(E.key("azure"))
 | 
			
		||||
  return {
 | 
			
		||||
    { role = "system", content = system_prompt },
 | 
			
		||||
    { role = "user", content = user_prompt },
 | 
			
		||||
  }
 | 
			
		||||
end
 | 
			
		||||
 | 
			
		||||
---@type AvanteAiResponseParser
 | 
			
		||||
H.parse_openai_response = function(data_stream, opts)
 | 
			
		||||
  if data_stream:match('"%[DONE%]":') then
 | 
			
		||||
    opts.on_complete(nil)
 | 
			
		||||
    return
 | 
			
		||||
  end
 | 
			
		||||
  if data_stream:match('"delta":') then
 | 
			
		||||
    local json = vim.json.decode(data_stream)
 | 
			
		||||
    if json.choices and json.choices[1] then
 | 
			
		||||
      local choice = json.choices[1]
 | 
			
		||||
      if choice.finish_reason == "stop" then
 | 
			
		||||
        opts.on_complete(nil)
 | 
			
		||||
      elseif choice.delta.content then
 | 
			
		||||
        opts.on_chunk(choice.delta.content)
 | 
			
		||||
      end
 | 
			
		||||
    end
 | 
			
		||||
  end
 | 
			
		||||
end
 | 
			
		||||
 | 
			
		||||
---@type AvanteCurlArgsBuilder
 | 
			
		||||
H.make_openai_curl_args = function(code_opts)
 | 
			
		||||
  return {
 | 
			
		||||
    url = Utils.trim(Config.openai.endpoint, { suffix = "/" }) .. "/v1/chat/completions",
 | 
			
		||||
    headers = {
 | 
			
		||||
      ["Content-Type"] = "application/json",
 | 
			
		||||
      ["Authorization"] = "Bearer " .. E.value("openai"),
 | 
			
		||||
    },
 | 
			
		||||
    body = {
 | 
			
		||||
      model = Config.openai.model,
 | 
			
		||||
      messages = H.make_openai_message(code_opts),
 | 
			
		||||
      temperature = Config.openai.temperature,
 | 
			
		||||
      max_tokens = Config.openai.max_tokens,
 | 
			
		||||
      stream = true,
 | 
			
		||||
    },
 | 
			
		||||
  }
 | 
			
		||||
end
 | 
			
		||||
 | 
			
		||||
------------------------------Azure------------------------------
 | 
			
		||||
 | 
			
		||||
---@type AvanteAiMessageBuilder
 | 
			
		||||
H.make_azure_message = H.make_openai_message
 | 
			
		||||
 | 
			
		||||
---@type AvanteAiResponseParser
 | 
			
		||||
H.parse_azure_response = H.parse_openai_response
 | 
			
		||||
 | 
			
		||||
---@type AvanteCurlArgsBuilder
 | 
			
		||||
H.make_azure_curl_args = function(code_opts)
 | 
			
		||||
  return {
 | 
			
		||||
    url = Config.azure.endpoint
 | 
			
		||||
      .. "/openai/deployments/"
 | 
			
		||||
      .. Config.azure.deployment
 | 
			
		||||
      .. "/chat/completions?api-version="
 | 
			
		||||
      .. Config.azure.api_version
 | 
			
		||||
      .. Config.azure.api_version,
 | 
			
		||||
    headers = {
 | 
			
		||||
      ["Content-Type"] = "application/json",
 | 
			
		||||
      ["api-key"] = api_key,
 | 
			
		||||
    }
 | 
			
		||||
    body = {
 | 
			
		||||
      messages = {
 | 
			
		||||
        { role = "system", content = system_prompt },
 | 
			
		||||
        { role = "user", content = user_prompt },
 | 
			
		||||
      ["api-key"] = E.value("azure"),
 | 
			
		||||
    },
 | 
			
		||||
    body = {
 | 
			
		||||
      messages = H.make_openai_message(code_opts),
 | 
			
		||||
      temperature = Config.azure.temperature,
 | 
			
		||||
      max_tokens = Config.azure.max_tokens,
 | 
			
		||||
      stream = true,
 | 
			
		||||
    },
 | 
			
		||||
  }
 | 
			
		||||
  elseif Config.provider == "deepseek" then
 | 
			
		||||
    api_key = os.getenv(E.key("deepseek"))
 | 
			
		||||
    url = Utils.trim_suffix(Config.deepseek.endpoint, "/") .. "/chat/completions"
 | 
			
		||||
end
 | 
			
		||||
 | 
			
		||||
------------------------------Deepseek------------------------------
 | 
			
		||||
 | 
			
		||||
---@type AvanteAiMessageBuilder
 | 
			
		||||
H.make_deepseek_message = H.make_openai_message
 | 
			
		||||
 | 
			
		||||
---@type AvanteAiResponseParser
 | 
			
		||||
H.parse_deepseek_response = H.parse_openai_response
 | 
			
		||||
 | 
			
		||||
---@type AvanteCurlArgsBuilder
 | 
			
		||||
H.make_deepseek_curl_args = function(code_opts)
 | 
			
		||||
  return {
 | 
			
		||||
    url = Utils.trim(Config.deepseek.endpoint, { suffix = "/" }) .. "/chat/completions",
 | 
			
		||||
    headers = {
 | 
			
		||||
      ["Content-Type"] = "application/json",
 | 
			
		||||
      ["Authorization"] = "Bearer " .. api_key,
 | 
			
		||||
    }
 | 
			
		||||
      ["Authorization"] = "Bearer " .. E.value("deepseek"),
 | 
			
		||||
    },
 | 
			
		||||
    body = {
 | 
			
		||||
      model = Config.deepseek.model,
 | 
			
		||||
      messages = {
 | 
			
		||||
        { role = "system", content = system_prompt },
 | 
			
		||||
        { role = "user", content = user_prompt },
 | 
			
		||||
      },
 | 
			
		||||
      messages = H.make_openai_message(code_opts),
 | 
			
		||||
      temperature = Config.deepseek.temperature,
 | 
			
		||||
      max_tokens = Config.deepseek.max_tokens,
 | 
			
		||||
      stream = true,
 | 
			
		||||
    },
 | 
			
		||||
  }
 | 
			
		||||
  elseif Config.provider == "groq" then
 | 
			
		||||
    api_key = os.getenv(E.key("groq"))
 | 
			
		||||
    url = Utils.trim_suffix(Config.groq.endpoint, "/") .. "/openai/v1/chat/completions"
 | 
			
		||||
end
 | 
			
		||||
 | 
			
		||||
------------------------------Grok------------------------------
 | 
			
		||||
 | 
			
		||||
---@type AvanteAiMessageBuilder
 | 
			
		||||
H.make_groq_message = H.make_openai_message
 | 
			
		||||
 | 
			
		||||
---@type AvanteAiResponseParser
 | 
			
		||||
H.parse_groq_response = H.parse_openai_response
 | 
			
		||||
 | 
			
		||||
---@type AvanteCurlArgsBuilder
 | 
			
		||||
H.make_groq_curl_args = function(code_opts)
 | 
			
		||||
  return {
 | 
			
		||||
    url = Utils.trim(Config.groq.endpoint, { suffix = "/" }) .. "/openai/v1/chat/completions",
 | 
			
		||||
    headers = {
 | 
			
		||||
      ["Content-Type"] = "application/json",
 | 
			
		||||
      ["Authorization"] = "Bearer " .. api_key,
 | 
			
		||||
    }
 | 
			
		||||
      ["Authorization"] = "Bearer " .. E.value("groq"),
 | 
			
		||||
    },
 | 
			
		||||
    body = {
 | 
			
		||||
      model = Config.groq.model,
 | 
			
		||||
      messages = {
 | 
			
		||||
        { role = "system", content = system_prompt },
 | 
			
		||||
        { role = "user", content = user_prompt },
 | 
			
		||||
      },
 | 
			
		||||
      messages = H.make_openai_message(code_opts),
 | 
			
		||||
      temperature = Config.groq.temperature,
 | 
			
		||||
      max_tokens = Config.groq.max_tokens,
 | 
			
		||||
      stream = true,
 | 
			
		||||
    }
 | 
			
		||||
  else
 | 
			
		||||
    url = Utils.trim_suffix(Config.openai.endpoint, "/") .. "/v1/chat/completions"
 | 
			
		||||
    headers = {
 | 
			
		||||
      ["Content-Type"] = "application/json",
 | 
			
		||||
      ["Authorization"] = "Bearer " .. api_key,
 | 
			
		||||
    }
 | 
			
		||||
    body = {
 | 
			
		||||
      model = Config.openai.model,
 | 
			
		||||
      messages = {
 | 
			
		||||
        { role = "system", content = system_prompt },
 | 
			
		||||
        { role = "user", content = user_prompt },
 | 
			
		||||
    },
 | 
			
		||||
      temperature = Config.openai.temperature,
 | 
			
		||||
      max_tokens = Config.openai.max_tokens,
 | 
			
		||||
      stream = true,
 | 
			
		||||
  }
 | 
			
		||||
  end
 | 
			
		||||
 | 
			
		||||
  curl.post(url, {
 | 
			
		||||
    ---@diagnostic disable-next-line: unused-local
 | 
			
		||||
    stream = function(err, data, job)
 | 
			
		||||
      if err then
 | 
			
		||||
        on_complete(err)
 | 
			
		||||
        return
 | 
			
		||||
      end
 | 
			
		||||
      if not data then
 | 
			
		||||
        return
 | 
			
		||||
      end
 | 
			
		||||
      for _, line in ipairs(vim.split(data, "\n")) do
 | 
			
		||||
        if line:sub(1, 6) ~= "data: " then
 | 
			
		||||
          return
 | 
			
		||||
        end
 | 
			
		||||
        vim.schedule(function()
 | 
			
		||||
          local piece = line:sub(7)
 | 
			
		||||
          local success, parsed = pcall(fn.json_decode, piece)
 | 
			
		||||
          if not success then
 | 
			
		||||
            if piece == "[DONE]" then
 | 
			
		||||
              on_complete(nil)
 | 
			
		||||
              return
 | 
			
		||||
            end
 | 
			
		||||
            error("Error: failed to parse json: " .. parsed)
 | 
			
		||||
            return
 | 
			
		||||
          end
 | 
			
		||||
          if parsed and parsed.choices and parsed.choices[1] then
 | 
			
		||||
            local choice = parsed.choices[1]
 | 
			
		||||
            if choice.finish_reason == "stop" then
 | 
			
		||||
              on_complete(nil)
 | 
			
		||||
            elseif choice.delta and choice.delta.content then
 | 
			
		||||
              on_chunk(choice.delta.content)
 | 
			
		||||
            end
 | 
			
		||||
          end
 | 
			
		||||
        end)
 | 
			
		||||
      end
 | 
			
		||||
    end,
 | 
			
		||||
    headers = headers,
 | 
			
		||||
    body = fn.json_encode(body),
 | 
			
		||||
  })
 | 
			
		||||
end
 | 
			
		||||
 | 
			
		||||
------------------------------Logic------------------------------
 | 
			
		||||
 | 
			
		||||
local group = vim.api.nvim_create_augroup("AvanteAiBot", { clear = true })
 | 
			
		||||
local active_job = nil
 | 
			
		||||
 | 
			
		||||
---@param question string
 | 
			
		||||
---@param code_lang string
 | 
			
		||||
---@param code_content string
 | 
			
		||||
---@param selected_content_content string | nil
 | 
			
		||||
---@param on_chunk fun(chunk: string): any
 | 
			
		||||
---@param on_complete fun(err: string|nil): any
 | 
			
		||||
function M.call_ai_api_stream(question, code_lang, code_content, selected_content_content, on_chunk, on_complete)
 | 
			
		||||
  if
 | 
			
		||||
    Config.provider == "openai"
 | 
			
		||||
    or Config.provider == "azure"
 | 
			
		||||
    or Config.provider == "deepseek"
 | 
			
		||||
    or Config.provider == "groq"
 | 
			
		||||
  then
 | 
			
		||||
    call_openai_api_stream(question, code_lang, code_content, selected_content_content, on_chunk, on_complete)
 | 
			
		||||
  elseif Config.provider == "claude" then
 | 
			
		||||
    call_claude_api_stream(question, code_lang, code_content, selected_content_content, on_chunk, on_complete)
 | 
			
		||||
M.invoke_llm_stream = function(question, code_lang, code_content, selected_content_content, on_chunk, on_complete)
 | 
			
		||||
  local provider = Config.provider
 | 
			
		||||
  local event_state = nil
 | 
			
		||||
 | 
			
		||||
  ---@type AvanteCurlOutput
 | 
			
		||||
  local spec = H["make_" .. provider .. "_curl_args"]({
 | 
			
		||||
    question = question,
 | 
			
		||||
    code_lang = code_lang,
 | 
			
		||||
    code_content = code_content,
 | 
			
		||||
    selected_code_content = selected_content_content,
 | 
			
		||||
  })
 | 
			
		||||
 | 
			
		||||
  ---@param line string
 | 
			
		||||
  local function parse_and_call(line)
 | 
			
		||||
    local event = line:match("^event: (.+)$")
 | 
			
		||||
    if event then
 | 
			
		||||
      event_state = event
 | 
			
		||||
      return
 | 
			
		||||
    end
 | 
			
		||||
    local data_match = line:match("^data: (.+)$")
 | 
			
		||||
    if data_match then
 | 
			
		||||
      H["parse_" .. provider .. "_response"](
 | 
			
		||||
        data_match,
 | 
			
		||||
        vim.deepcopy({ on_chunk = on_chunk, on_complete = on_complete, event_state = event_state }, true)
 | 
			
		||||
      )
 | 
			
		||||
    end
 | 
			
		||||
  end
 | 
			
		||||
 | 
			
		||||
  if active_job then
 | 
			
		||||
    active_job:shutdown()
 | 
			
		||||
    active_job = nil
 | 
			
		||||
  end
 | 
			
		||||
 | 
			
		||||
  active_job = curl.post(spec.url, {
 | 
			
		||||
    headers = spec.headers,
 | 
			
		||||
    body = vim.json.encode(spec.body),
 | 
			
		||||
    stream = function(err, data, _)
 | 
			
		||||
      if err then
 | 
			
		||||
        on_complete(err)
 | 
			
		||||
        return
 | 
			
		||||
      end
 | 
			
		||||
      if not data then
 | 
			
		||||
        return
 | 
			
		||||
      end
 | 
			
		||||
      vim.schedule(function()
 | 
			
		||||
        parse_and_call(data)
 | 
			
		||||
      end)
 | 
			
		||||
    end,
 | 
			
		||||
    on_error = function(err)
 | 
			
		||||
      on_complete(err)
 | 
			
		||||
    end,
 | 
			
		||||
    callback = function(_)
 | 
			
		||||
      active_job = nil
 | 
			
		||||
    end,
 | 
			
		||||
  })
 | 
			
		||||
 | 
			
		||||
  api.nvim_create_autocmd("User", {
 | 
			
		||||
    group = group,
 | 
			
		||||
    pattern = M.CANCEL_PATTERN,
 | 
			
		||||
    callback = function()
 | 
			
		||||
      if active_job then
 | 
			
		||||
        active_job:shutdown()
 | 
			
		||||
        vim.notify("LLM request cancelled", vim.log.levels.DEBUG)
 | 
			
		||||
        active_job = nil
 | 
			
		||||
      end
 | 
			
		||||
    end,
 | 
			
		||||
  })
 | 
			
		||||
 | 
			
		||||
  return active_job
 | 
			
		||||
end
 | 
			
		||||
 | 
			
		||||
function M.setup()
 | 
			
		||||
@ -417,6 +509,38 @@ function M.setup()
 | 
			
		||||
  if not has then
 | 
			
		||||
    E.setup(E.key())
 | 
			
		||||
  end
 | 
			
		||||
 | 
			
		||||
  M.commands()
 | 
			
		||||
end
 | 
			
		||||
 | 
			
		||||
---@param provider Provider
 | 
			
		||||
function M.refresh(provider)
 | 
			
		||||
  local has = E[provider]
 | 
			
		||||
  if not has then
 | 
			
		||||
    E.setup(E.key(provider), true)
 | 
			
		||||
  else
 | 
			
		||||
    vim.notify_once("Switch to provider: " .. provider, vim.log.levels.INFO)
 | 
			
		||||
  end
 | 
			
		||||
  require("avante").setup({ provider = provider })
 | 
			
		||||
end
 | 
			
		||||
 | 
			
		||||
M.commands = function()
 | 
			
		||||
  api.nvim_create_user_command("AvanteSwitchProvider", function(args)
 | 
			
		||||
    local cmd = vim.trim(args.args or "")
 | 
			
		||||
    M.refresh(cmd)
 | 
			
		||||
  end, {
 | 
			
		||||
    nargs = 1,
 | 
			
		||||
    desc = "avante: switch provider",
 | 
			
		||||
    complete = function(_, line)
 | 
			
		||||
      if line:match("^%s*AvanteSwitchProvider %w") then
 | 
			
		||||
        return {}
 | 
			
		||||
      end
 | 
			
		||||
      local prefix = line:match("^%s*AvanteSwitchProvider (%w*)") or ""
 | 
			
		||||
      return vim.tbl_filter(function(key)
 | 
			
		||||
        return key:find(prefix) == 1
 | 
			
		||||
      end, vim.tbl_keys(E.env))
 | 
			
		||||
    end,
 | 
			
		||||
  })
 | 
			
		||||
end
 | 
			
		||||
 | 
			
		||||
return M
 | 
			
		||||
 | 
			
		||||
@ -68,7 +68,7 @@ M.defaults = {
 | 
			
		||||
    },
 | 
			
		||||
  },
 | 
			
		||||
  windows = {
 | 
			
		||||
    wrap_line = true,
 | 
			
		||||
    wrap_line = true, -- similar to vim.o.wrap
 | 
			
		||||
    width = 30, -- default % based on available width
 | 
			
		||||
  },
 | 
			
		||||
  --- @class AvanteConflictUserConfig
 | 
			
		||||
 | 
			
		||||
@ -143,14 +143,11 @@ function M._init(id)
 | 
			
		||||
  return M
 | 
			
		||||
end
 | 
			
		||||
 | 
			
		||||
M.open = function()
 | 
			
		||||
  M._init(api.nvim_get_current_tabpage())._get(false):open()
 | 
			
		||||
end
 | 
			
		||||
 | 
			
		||||
M.toggle = function()
 | 
			
		||||
  local sidebar = M._get()
 | 
			
		||||
  if not sidebar then
 | 
			
		||||
    M.open()
 | 
			
		||||
    M._init(api.nvim_get_current_tabpage())
 | 
			
		||||
    M.current.sidebar:open()
 | 
			
		||||
    return true
 | 
			
		||||
  end
 | 
			
		||||
 | 
			
		||||
 | 
			
		||||
@ -141,6 +141,15 @@ function Sidebar:intialize()
 | 
			
		||||
      mode = { "n" },
 | 
			
		||||
      key = "q",
 | 
			
		||||
      handler = function()
 | 
			
		||||
        api.nvim_exec_autocmds("User", { pattern = AiBot.CANCEL_PATTERN })
 | 
			
		||||
        self.renderer:close()
 | 
			
		||||
      end,
 | 
			
		||||
    },
 | 
			
		||||
    {
 | 
			
		||||
      mode = { "n" },
 | 
			
		||||
      key = "<Esc>",
 | 
			
		||||
      handler = function()
 | 
			
		||||
        api.nvim_exec_autocmds("User", { pattern = AiBot.CANCEL_PATTERN })
 | 
			
		||||
        self.renderer:close()
 | 
			
		||||
      end,
 | 
			
		||||
    },
 | 
			
		||||
@ -171,10 +180,16 @@ function Sidebar:intialize()
 | 
			
		||||
    api.nvim_create_autocmd("VimResized", {
 | 
			
		||||
      group = self.augroup,
 | 
			
		||||
      callback = function()
 | 
			
		||||
        if not self.view:is_open() then
 | 
			
		||||
          return
 | 
			
		||||
        end
 | 
			
		||||
        local new_layout = Config.get_renderer_layout_options()
 | 
			
		||||
        vim.api.nvim_win_set_width(self.view.win, new_layout.width)
 | 
			
		||||
        vim.api.nvim_win_set_height(self.view.win, new_layout.height)
 | 
			
		||||
        self.renderer:set_size({ width = new_layout.width, height = new_layout.height })
 | 
			
		||||
        vim.defer_fn(function()
 | 
			
		||||
          vim.cmd("AvanteRefresh")
 | 
			
		||||
        end, 200)
 | 
			
		||||
      end,
 | 
			
		||||
    })
 | 
			
		||||
 | 
			
		||||
@ -228,9 +243,25 @@ function Sidebar:is_focused()
 | 
			
		||||
end
 | 
			
		||||
 | 
			
		||||
---@param content string concatenated content of the buffer
 | 
			
		||||
---@param opts? {focus?: boolean, scroll?: boolean, callback?: fun(): nil} whether to focus the result view
 | 
			
		||||
---@param opts? {focus?: boolean, stream?: boolean, scroll?: boolean, callback?: fun(): nil} whether to focus the result view
 | 
			
		||||
function Sidebar:update_content(content, opts)
 | 
			
		||||
  opts = vim.tbl_deep_extend("force", { focus = true, scroll = true, callback = nil }, opts or {})
 | 
			
		||||
  opts = vim.tbl_deep_extend("force", { focus = true, scroll = true, stream = false, callback = nil }, opts or {})
 | 
			
		||||
  if opts.stream then
 | 
			
		||||
    vim.schedule(function()
 | 
			
		||||
      api.nvim_set_option_value("modifiable", true, { buf = self.view.buf })
 | 
			
		||||
      local current_window = vim.api.nvim_get_current_win()
 | 
			
		||||
      local cursor_position = vim.api.nvim_win_get_cursor(current_window)
 | 
			
		||||
      local row, col = cursor_position[1], cursor_position[2]
 | 
			
		||||
 | 
			
		||||
      local lines = vim.split(content, "\n")
 | 
			
		||||
 | 
			
		||||
      vim.api.nvim_put(lines, "c", true, true)
 | 
			
		||||
 | 
			
		||||
      local num_lines = #lines
 | 
			
		||||
      local last_line_length = #lines[num_lines]
 | 
			
		||||
      vim.api.nvim_win_set_cursor(current_window, { row + num_lines - 1, col + last_line_length })
 | 
			
		||||
    end)
 | 
			
		||||
  else
 | 
			
		||||
    vim.defer_fn(function()
 | 
			
		||||
      api.nvim_set_option_value("modifiable", true, { buf = self.view.buf })
 | 
			
		||||
      api.nvim_buf_set_lines(self.view.buf, 0, -1, false, vim.split(content, "\n"))
 | 
			
		||||
@ -256,6 +287,7 @@ function Sidebar:update_content(content, opts)
 | 
			
		||||
        end)
 | 
			
		||||
      end
 | 
			
		||||
    end, 0)
 | 
			
		||||
  end
 | 
			
		||||
  return self
 | 
			
		||||
end
 | 
			
		||||
 | 
			
		||||
@ -638,7 +670,7 @@ function Sidebar:render()
 | 
			
		||||
 | 
			
		||||
    local filetype = api.nvim_get_option_value("filetype", { buf = self.code.buf })
 | 
			
		||||
 | 
			
		||||
    AiBot.call_ai_api_stream(
 | 
			
		||||
    AiBot.invoke_llm_stream(
 | 
			
		||||
      request,
 | 
			
		||||
      filetype,
 | 
			
		||||
      content_with_line_numbers,
 | 
			
		||||
@ -646,7 +678,7 @@ function Sidebar:render()
 | 
			
		||||
      function(chunk)
 | 
			
		||||
        signal.is_loading = true
 | 
			
		||||
        full_response = full_response .. chunk
 | 
			
		||||
        self:update_content(content_prefix .. full_response)
 | 
			
		||||
        self:update_content(chunk, { stream = true, scroll = false })
 | 
			
		||||
        vim.schedule(function()
 | 
			
		||||
          vim.cmd("redraw")
 | 
			
		||||
        end)
 | 
			
		||||
 | 
			
		||||
@ -1,8 +1,17 @@
 | 
			
		||||
local Range = require("avante.range")
 | 
			
		||||
local SelectionResult = require("avante.selection_result")
 | 
			
		||||
local M = {}
 | 
			
		||||
function M.trim_suffix(str, suffix)
 | 
			
		||||
  return string.gsub(str, suffix .. "$", "")
 | 
			
		||||
---@param str string
 | 
			
		||||
---@param opts? {suffix?: string, prefix?: string}
 | 
			
		||||
function M.trim(str, opts)
 | 
			
		||||
  if not opts then
 | 
			
		||||
    return str
 | 
			
		||||
  end
 | 
			
		||||
  if opts.suffix then
 | 
			
		||||
    return str:sub(-1) == opts.suffix and str:sub(1, -2) or str
 | 
			
		||||
  elseif opts.prefix then
 | 
			
		||||
    return str:sub(1, 1) == opts.prefix and str:sub(2) or str
 | 
			
		||||
  end
 | 
			
		||||
end
 | 
			
		||||
function M.trim_line_number_prefix(line)
 | 
			
		||||
  return line:gsub("^L%d+: ", "")
 | 
			
		||||
 | 
			
		||||
@ -42,14 +42,17 @@ function View:setup(split_command, size)
 | 
			
		||||
  api.nvim_set_option_value("foldcolumn", "0", { win = self.win })
 | 
			
		||||
  api.nvim_set_option_value("number", false, { win = self.win })
 | 
			
		||||
  api.nvim_set_option_value("relativenumber", false, { win = self.win })
 | 
			
		||||
  api.nvim_set_option_value("winfixwidth", true, { win = self.win })
 | 
			
		||||
  api.nvim_set_option_value("list", false, { win = self.win })
 | 
			
		||||
  api.nvim_set_option_value("wrap", Config.windows.wrap_line, { win = self.win })
 | 
			
		||||
  api.nvim_set_option_value("winhl", "", { win = self.win })
 | 
			
		||||
  api.nvim_set_option_value("linebreak", true, { win = self.win }) -- only has effect when wrap=true
 | 
			
		||||
  api.nvim_set_option_value("breakindent", true, { win = self.win }) -- only has effect when wrap=true
 | 
			
		||||
 | 
			
		||||
  -- buffer stuff
 | 
			
		||||
  xpcall(function()
 | 
			
		||||
    api.nvim_buf_set_name(self.buf, RESULT_BUF_NAME)
 | 
			
		||||
  end, function(err) end)
 | 
			
		||||
  end, function(_) end)
 | 
			
		||||
 | 
			
		||||
  return self
 | 
			
		||||
end
 | 
			
		||||
 | 
			
		||||
		Loading…
	
	
			
			x
			
			
		
	
		Reference in New Issue
	
	Block a user