package main import ( "bufio" "bytes" "encoding/json" "io" "net/http" "github.com/gin-gonic/gin" ) type OpenAIRequest struct { Model string `json:"model"` Messages []struct { Role string `json:"role"` Content string `json:"content"` } `json:"messages"` Stream bool `json:"stream"` } type OpenAIResponse struct { ID string `json:"id"` Object string `json:"object"` Created int64 `json:"created"` Model string `json:"model"` Choices []OpenAIChoice `json:"choices"` } type OpenAIChoice struct { Index int `json:"index"` Delta OpenAIDelta `json:"delta"` Logprobs interface{} `json:"logprobs"` FinishReason *string `json:"finish_reason"` } type OpenAIDelta struct { Role string `json:"role,omitempty"` Content string `json:"content,omitempty"` } type OpenAINonStreamResponse struct { ID string `json:"id"` Object string `json:"object"` Created int64 `json:"created"` Model string `json:"model"` Choices []OpenAINonStreamChoice `json:"choices"` } type OpenAINonStreamChoice struct { Index int `json:"index"` Message OpenAIDelta `json:"message"` FinishReason *string `json:"finish_reason"` } type DuckDuckGoResponse struct { Role string `json:"role"` Message string `json:"message"` Created int64 `json:"created"` ID string `json:"id"` Action string `json:"action"` Model string `json:"model"` } func chatWithDuckDuckGo(c *gin.Context, messages []struct { Role string `json:"role"` Content string `json:"content"` }, stream bool) { userAgent := "Mozilla/5.0 (X11; Ubuntu; Linux x86_64; rv:123.0) Gecko/20100101 Firefox/123.0" headers := map[string]string{ "User-Agent": userAgent, "Accept": "text/event-stream", "Accept-Language": "de,en-US;q=0.7,en;q=0.3", "Accept-Encoding": "gzip, deflate, br", "Referer": "https://duckduckgo.com/", "Content-Type": "application/json", "Origin": "https://duckduckgo.com", "Connection": "keep-alive", "Cookie": "dcm=1", "Sec-Fetch-Dest": "empty", "Sec-Fetch-Mode": "cors", "Sec-Fetch-Site": "same-origin", "Pragma": "no-cache", "TE": "trailers", } statusURL := "https://duckduckgo.com/duckchat/v1/status" chatURL := "https://duckduckgo.com/duckchat/v1/chat" // get vqd_4 req, err := http.NewRequest("GET", statusURL, nil) if err != nil { c.JSON(http.StatusInternalServerError, gin.H{"error": err.Error()}) return } req.Header.Set("x-vqd-accept", "1") for key, value := range headers { req.Header.Set(key, value) } resp, err := http.DefaultClient.Do(req) if err != nil { c.JSON(http.StatusInternalServerError, gin.H{"error": err.Error()}) return } defer resp.Body.Close() vqd4 := resp.Header.Get("x-vqd-4") payload := map[string]interface{}{ "model": "gpt-3.5-turbo-0125", "messages": messages, } payloadBytes, err := json.Marshal(payload) if err != nil { c.JSON(http.StatusInternalServerError, gin.H{"error": err.Error()}) return } req, err = http.NewRequest("POST", chatURL, bytes.NewBuffer(payloadBytes)) if err != nil { c.JSON(http.StatusInternalServerError, gin.H{"error": err.Error()}) return } req.Header.Set("x-vqd-4", vqd4) for key, value := range headers { req.Header.Set(key, value) } resp, err = http.DefaultClient.Do(req) if err != nil { c.JSON(http.StatusInternalServerError, gin.H{"error": err.Error()}) return } defer resp.Body.Close() reader := bufio.NewReader(resp.Body) c.Header("Content-Type", "text/event-stream") c.Header("Cache-Control", "no-cache") c.Header("Connection", "keep-alive") c.Header("Transfer-Encoding", "chunked") flusher, _ := c.Writer.(http.Flusher) var response OpenAIResponse var nonStreamResponse OpenAINonStreamResponse response.Choices = make([]OpenAIChoice, 1) nonStreamResponse.Choices = make([]OpenAINonStreamChoice, 1) var responseContent string for { line, err := reader.ReadBytes('\n') if err != nil { if err == io.EOF { break } c.JSON(http.StatusInternalServerError, gin.H{"error": err.Error()}) return } if bytes.HasPrefix(line, []byte("data: ")) { chunk := line[6:] if bytes.HasPrefix(chunk, []byte("[DONE]")) { if !stream { nonStreamResponse.Choices[0].Message.Content = responseContent nonStreamResponse.Choices[0].Message.Role = "assistant" nonStreamResponse.Choices[0].FinishReason = new(string) *nonStreamResponse.Choices[0].FinishReason = "stop" c.JSON(http.StatusOK, nonStreamResponse) return } else { stopData := OpenAIResponse{ ID: "chatcmpl-9HOzx2PhnYCLPxQ3Dpa2OKoqR2lgl", Object: "chat.completion", Created: 1713934697, Model: "gpt-3.5-turbo-0125", Choices: []OpenAIChoice{ { Index: 0, FinishReason: stringPtr("stop"), }, }, } stopDataBytes, _ := json.Marshal(stopData) c.Data(http.StatusOK, "application/json", []byte("data: ")) c.Data(http.StatusOK, "application/json", stopDataBytes) c.Data(http.StatusOK, "application/json", []byte("\n\n")) flusher.Flush() c.Data(http.StatusOK, "application/json", []byte("data: [DONE]\n\n")) flusher.Flush() return } } var data DuckDuckGoResponse decoder := json.NewDecoder(bytes.NewReader(chunk)) decoder.UseNumber() err = decoder.Decode(&data) if err != nil { c.JSON(http.StatusInternalServerError, gin.H{"error": err.Error()}) return } response.ID = data.ID response.Object = "chat.completion" response.Created = data.Created response.Model = data.Model nonStreamResponse.ID = data.ID nonStreamResponse.Object = "chat.completion" nonStreamResponse.Created = data.Created nonStreamResponse.Model = data.Model responseContent += data.Message if stream { response.Choices[0].Delta.Content = data.Message responseBytes, err := json.Marshal(response) if err != nil { c.JSON(http.StatusInternalServerError, gin.H{"error": err.Error()}) return } c.Data(http.StatusOK, "application/json", append(append([]byte("data: "), responseBytes...), []byte("\n\n")...)) flusher.Flush() response.Choices[0].Delta.Content = "" } } } } func stringPtr(s string) *string { return &s } func main() { gin.SetMode(gin.ReleaseMode) r := gin.Default() r.GET("/", func(c *gin.Context) { c.JSON(http.StatusOK, gin.H{ "message": "Thankyou", }) }) r.OPTIONS("/v1/chat/completions", func(c *gin.Context) { c.JSON(http.StatusOK, gin.H{ "message": "ok", }) }) r.POST("/v1/chat/completions", func(c *gin.Context) { var req OpenAIRequest if err := c.ShouldBindJSON(&req); err != nil { c.JSON(http.StatusBadRequest, gin.H{"error": err.Error()}) return } // only support user role for i := range req.Messages { if req.Messages[i].Role == "system" { req.Messages[i].Role = "user" } } // set model to gpt-3.5-turbo-0125 req.Model = "gpt-3.5-turbo-0125" chatWithDuckDuckGo(c, req.Messages, req.Stream) }) r.GET("/v1/models", func(c *gin.Context) { c.JSON(http.StatusOK, gin.H{ "object": "list", "data": []gin.H{ { "id": "gpt-3.5-turbo-0125", "object": "model", "created": 1692901427, "owned_by": "system", }, }, }) }) r.Run(":3456") }