You don’t need to set up an environment, just have an overseas vps, preferably a vps in an area supported by openai, download the [executive file] (./bin/api_proxy) in the bin and run it directly, the simplest api proxy method, the most important The most important thing is to support SSE, so that the client can respond more quickly when requesting, and also provides the source code of golang, which can be improved by itself if it needs to be customized.
./api_proxy -daemon -port 9000 # It is best to open the daemon process mode
Python use cases:
import os
import openai
openai.api_key = YOUR-API-KEY
openai.api_base = "http://host:port/v1" # Be sure to add v1
for resp in openai.ChatCompletion.create(
model="gpt-3.5-turbo",
messages=[
{"role": "user", "content": "Bubble Sort"}
],
stream = True): # Streaming output, support SSE
if 'content' in resp.choices[0].delta:
print(resp.choices[0].delta.content, end="", flush=True) # flush prints in time
JS use case, Take https://www.npmjs.com/package/chatgpt as an example
chatApi= new gpt.ChatGPTAPI({
apiKey: 'sk.....:<proxy_key write here>',
apiBaseUrl: "http://host:port", // delivery proxy address
});
(Recommend:)An example of using ChatGPT-Next-Web on the server, after setting the key, you can set the code password to access, and the api uses the current proxy, which is very easy to use, refer to the webpage https://gpt.nephen.cn/.
docker pull nephen2023/chatgpt-next-web:v1.7.1
docker run -d -p 3000:3000 -e OPENAI_API_KEY="" -e CODE="" -e BASE_URL="ip:port" -e PROTOCOL="http" nephen2023/chatgpt-next-web:v1.7.1