summaryrefslogtreecommitdiffstats
path: root/g4f/Provider/nexra/NexraChatGptV2.py
diff options
context:
space:
mode:
authorTekky <98614666+xtekky@users.noreply.github.com>2024-10-30 09:57:55 +0100
committerTekky <98614666+xtekky@users.noreply.github.com>2024-10-30 09:57:55 +0100
commit1443c60cc86f7f02cc6c7a4b2a31d6b1dad66a26 (patch)
tree435d831df4ad5c18839cfaa647f23e5be035cdd6 /g4f/Provider/nexra/NexraChatGptV2.py
parentimplement direct import of `Client` without using `g4f.client` (diff)
parentMerge pull request #2304 from kqlio67/main (diff)
downloadgpt4free-1443c60cc86f7f02cc6c7a4b2a31d6b1dad66a26.tar
gpt4free-1443c60cc86f7f02cc6c7a4b2a31d6b1dad66a26.tar.gz
gpt4free-1443c60cc86f7f02cc6c7a4b2a31d6b1dad66a26.tar.bz2
gpt4free-1443c60cc86f7f02cc6c7a4b2a31d6b1dad66a26.tar.lz
gpt4free-1443c60cc86f7f02cc6c7a4b2a31d6b1dad66a26.tar.xz
gpt4free-1443c60cc86f7f02cc6c7a4b2a31d6b1dad66a26.tar.zst
gpt4free-1443c60cc86f7f02cc6c7a4b2a31d6b1dad66a26.zip
Diffstat (limited to 'g4f/Provider/nexra/NexraChatGptV2.py')
-rw-r--r--g4f/Provider/nexra/NexraChatGptV2.py92
1 files changed, 0 insertions, 92 deletions
diff --git a/g4f/Provider/nexra/NexraChatGptV2.py b/g4f/Provider/nexra/NexraChatGptV2.py
deleted file mode 100644
index 1ff42705..00000000
--- a/g4f/Provider/nexra/NexraChatGptV2.py
+++ /dev/null
@@ -1,92 +0,0 @@
-from __future__ import annotations
-
-import json
-import requests
-
-from ...typing import CreateResult, Messages
-from ..base_provider import ProviderModelMixin, AbstractProvider
-from ..helper import format_prompt
-
-class NexraChatGptV2(AbstractProvider, ProviderModelMixin):
- label = "Nexra ChatGPT v2"
- url = "https://nexra.aryahcr.cc/documentation/chatgpt/en"
- api_endpoint = "https://nexra.aryahcr.cc/api/chat/complements"
- working = True
- supports_stream = True
-
- default_model = 'chatgpt'
- models = [default_model]
- model_aliases = {"gpt-4": "chatgpt"}
-
- @classmethod
- def get_model(cls, model: str) -> str:
- if model in cls.models:
- return model
- elif model in cls.model_aliases:
- return cls.model_aliases[model]
- else:
- return cls.default_model
-
- @classmethod
- def create_completion(
- cls,
- model: str,
- messages: Messages,
- stream: bool,
- proxy: str = None,
- markdown: bool = False,
- **kwargs
- ) -> CreateResult:
- model = cls.get_model(model)
-
- headers = {
- 'Content-Type': 'application/json'
- }
-
- data = {
- "messages": [
- {
- "role": "user",
- "content": format_prompt(messages)
- }
- ],
- "stream": stream,
- "markdown": markdown,
- "model": model
- }
-
- response = requests.post(cls.api_endpoint, headers=headers, json=data, stream=stream)
-
- if stream:
- return cls.process_streaming_response(response)
- else:
- return cls.process_non_streaming_response(response)
-
- @classmethod
- def process_non_streaming_response(cls, response):
- if response.status_code == 200:
- try:
- content = response.text.lstrip('')
- data = json.loads(content)
- return data.get('message', '')
- except json.JSONDecodeError:
- return "Error: Unable to decode JSON response"
- else:
- return f"Error: {response.status_code}"
-
- @classmethod
- def process_streaming_response(cls, response):
- full_message = ""
- for line in response.iter_lines(decode_unicode=True):
- if line:
- try:
- line = line.lstrip('')
- data = json.loads(line)
- if data.get('finish'):
- break
- message = data.get('message', '')
- if message:
- yield message[len(full_message):]
- full_message = message
- except json.JSONDecodeError:
- pass