summaryrefslogtreecommitdiffstats
path: root/g4f/Provider/FreeChatgpt.py
diff options
context:
space:
mode:
authorTekky <98614666+xtekky@users.noreply.github.com>2024-07-11 20:03:13 +0200
committerGitHub <noreply@github.com>2024-07-11 20:03:13 +0200
commit31edda55f75b85651998bc0a77f22e538accaa1f (patch)
treecfb769c3569449780c9edc5073eb47180e60655d /g4f/Provider/FreeChatgpt.py
parentMerge pull request #2097 from gamelist1990/main (diff)
parentUpdate g4f/models.py (diff)
downloadgpt4free-31edda55f75b85651998bc0a77f22e538accaa1f.tar
gpt4free-31edda55f75b85651998bc0a77f22e538accaa1f.tar.gz
gpt4free-31edda55f75b85651998bc0a77f22e538accaa1f.tar.bz2
gpt4free-31edda55f75b85651998bc0a77f22e538accaa1f.tar.lz
gpt4free-31edda55f75b85651998bc0a77f22e538accaa1f.tar.xz
gpt4free-31edda55f75b85651998bc0a77f22e538accaa1f.tar.zst
gpt4free-31edda55f75b85651998bc0a77f22e538accaa1f.zip
Diffstat (limited to 'g4f/Provider/FreeChatgpt.py')
-rw-r--r--g4f/Provider/FreeChatgpt.py97
1 files changed, 56 insertions, 41 deletions
diff --git a/g4f/Provider/FreeChatgpt.py b/g4f/Provider/FreeChatgpt.py
index 3fb247c7..7d8c1d10 100644
--- a/g4f/Provider/FreeChatgpt.py
+++ b/g4f/Provider/FreeChatgpt.py
@@ -1,17 +1,27 @@
from __future__ import annotations
-
import json
-from aiohttp import ClientSession, ClientTimeout
-
+from aiohttp import ClientSession
from ..typing import AsyncResult, Messages
from .base_provider import AsyncGeneratorProvider, ProviderModelMixin
-from ..requests.raise_for_status import raise_for_status
+from .helper import format_prompt
+
class FreeChatgpt(AsyncGeneratorProvider, ProviderModelMixin):
- url = "https://free.chatgpt.org.uk"
+ url = "https://chat.chatgpt.org.uk"
+ api_endpoint = "/api/openai/v1/chat/completions"
working = True
- supports_message_history = True
- default_model = "google-gemini-pro"
+ supports_gpt_35_turbo = True
+ default_model = 'gpt-3.5-turbo'
+ models = [
+ 'gpt-3.5-turbo',
+ 'SparkDesk-v1.1',
+ 'deepseek-coder',
+ 'deepseek-chat',
+ 'Qwen2-7B-Instruct',
+ 'glm4-9B-chat',
+ 'chatglm3-6B',
+ 'Yi-1.5-9B-Chat',
+ ]
@classmethod
async def create_async_generator(
@@ -19,45 +29,50 @@ class FreeChatgpt(AsyncGeneratorProvider, ProviderModelMixin):
model: str,
messages: Messages,
proxy: str = None,
- timeout: int = 120,
**kwargs
) -> AsyncResult:
headers = {
- "Accept": "application/json, text/event-stream",
- "Content-Type":"application/json",
- "Accept-Encoding": "gzip, deflate, br",
- "Accept-Language": "en-US,en;q=0.5",
- "Host":"free.chatgpt.org.uk",
- "Referer":f"{cls.url}/",
- "Origin":f"{cls.url}",
- "Sec-Fetch-Dest": "empty",
- "Sec-Fetch-Mode": "cors",
- "Sec-Fetch-Site": "same-origin",
- "User-Agent": "Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/120.0.0.0 Safari/537.36",
+ "accept": "application/json, text/event-stream",
+ "accept-language": "en-US,en;q=0.9",
+ "content-type": "application/json",
+ "dnt": "1",
+ "origin": cls.url,
+ "referer": f"{cls.url}/",
+ "sec-ch-ua": '"Not/A)Brand";v="8", "Chromium";v="126"',
+ "sec-ch-ua-mobile": "?0",
+ "sec-ch-ua-platform": '"Linux"',
+ "sec-fetch-dest": "empty",
+ "sec-fetch-mode": "cors",
+ "sec-fetch-site": "same-origin",
+ "user-agent": "Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/126.0.0.0 Safari/537.36",
}
- async with ClientSession(headers=headers, timeout=ClientTimeout(timeout)) as session:
+ async with ClientSession(headers=headers) as session:
+ prompt = format_prompt(messages)
data = {
- "messages": messages,
+ "messages": [
+ {"role": "system", "content": "\nYou are ChatGPT, a large language model trained by OpenAI.\nKnowledge cutoff: 2021-09\nCurrent model: gpt-3.5-turbo\nCurrent time: Thu Jul 04 2024 21:35:59 GMT+0300 (Eastern European Summer Time)\nLatex inline: \\(x^2\\) \nLatex block: $$e=mc^2$$\n\n"},
+ {"role": "user", "content": prompt}
+ ],
"stream": True,
- "model": cls.get_model(""),
- "temperature": kwargs.get("temperature", 0.5),
- "presence_penalty": kwargs.get("presence_penalty", 0),
- "frequency_penalty": kwargs.get("frequency_penalty", 0),
- "top_p": kwargs.get("top_p", 1)
+ "model": model,
+ "temperature": 0.5,
+ "presence_penalty": 0,
+ "frequency_penalty": 0,
+ "top_p": 1
}
- async with session.post(f'{cls.url}/api/openai/v1/chat/completions', json=data, proxy=proxy) as response:
- await raise_for_status(response)
- started = False
+ async with session.post(f"{cls.url}{cls.api_endpoint}", json=data, proxy=proxy) as response:
+ response.raise_for_status()
+ accumulated_text = ""
async for line in response.content:
- if line.startswith(b"data: [DONE]"):
- break
- elif line.startswith(b"data: "):
- line = json.loads(line[6:])
- if(line["choices"]==[]):
- continue
- chunk = line["choices"][0]["delta"].get("content")
- if chunk:
- started = True
- yield chunk
- if not started:
- raise RuntimeError("Empty response") \ No newline at end of file
+ if line:
+ line_str = line.decode().strip()
+ if line_str == "data: [DONE]":
+ yield accumulated_text
+ break
+ elif line_str.startswith("data: "):
+ try:
+ chunk = json.loads(line_str[6:])
+ delta_content = chunk.get("choices", [{}])[0].get("delta", {}).get("content", "")
+ accumulated_text += delta_content
+ except json.JSONDecodeError:
+ pass