diff options
author | Heiner Lohaus <heiner@lohaus.eu> | 2023-09-26 01:02:02 +0200 |
---|---|---|
committer | Heiner Lohaus <heiner@lohaus.eu> | 2023-09-26 01:02:02 +0200 |
commit | 72c3ff7a25630582844cdd6a82055fdd2aff6c60 (patch) | |
tree | 1a31086e1f8240f9bd67d3a16afc54753a8f5f93 /g4f | |
parent | "create_async" support for BaseProvider, (diff) | |
download | gpt4free-72c3ff7a25630582844cdd6a82055fdd2aff6c60.tar gpt4free-72c3ff7a25630582844cdd6a82055fdd2aff6c60.tar.gz gpt4free-72c3ff7a25630582844cdd6a82055fdd2aff6c60.tar.bz2 gpt4free-72c3ff7a25630582844cdd6a82055fdd2aff6c60.tar.lz gpt4free-72c3ff7a25630582844cdd6a82055fdd2aff6c60.tar.xz gpt4free-72c3ff7a25630582844cdd6a82055fdd2aff6c60.tar.zst gpt4free-72c3ff7a25630582844cdd6a82055fdd2aff6c60.zip |
Diffstat (limited to 'g4f')
-rw-r--r-- | g4f/Provider/Vercel.py | 30 | ||||
-rw-r--r-- | g4f/Provider/base_provider.py | 31 | ||||
-rw-r--r-- | g4f/models.py | 7 |
3 files changed, 50 insertions, 18 deletions
diff --git a/g4f/Provider/Vercel.py b/g4f/Provider/Vercel.py index ca124fec..4102c07b 100644 --- a/g4f/Provider/Vercel.py +++ b/g4f/Provider/Vercel.py @@ -18,7 +18,13 @@ class Vercel(BaseProvider): def create_completion( model: str, messages: list[dict[str, str]], - stream: bool, **kwargs ) -> CreateResult: + stream: bool, + **kwargs + ) -> CreateResult: + if not model: + model = "gpt-3.5-turbo" + elif model not in model_info: + raise ValueError(f"Model are not supported: {model}") headers = { 'authority' : 'sdk.vercel.ai', @@ -26,7 +32,7 @@ class Vercel(BaseProvider): 'accept-language' : 'en,fr-FR;q=0.9,fr;q=0.8,es-ES;q=0.7,es;q=0.6,en-US;q=0.5,am;q=0.4,de;q=0.3', 'cache-control' : 'no-cache', 'content-type' : 'application/json', - 'custom-encoding' : AntiBotToken(), + 'custom-encoding' : get_anti_bot_token(), 'origin' : 'https://sdk.vercel.ai', 'pragma' : 'no-cache', 'referer' : 'https://sdk.vercel.ai/', @@ -48,22 +54,20 @@ class Vercel(BaseProvider): 'playgroundId': str(uuid.uuid4()), 'chatIndex' : 0} | model_info[model]['default_params'] - server_error = True - retries = 0 max_retries = kwargs.get('max_retries', 20) - - while server_error and not retries > max_retries: + for i in range(max_retries): response = requests.post('https://sdk.vercel.ai/api/generate', headers=headers, json=json_data, stream=True) + try: + response.raise_for_status() + except: + continue + for token in response.iter_content(chunk_size=8): + yield token.decode() + break - for token in response.iter_content(chunk_size=2046): - if token != b'Internal Server Error': - server_error = False - yield (token.decode()) - - retries += 1 -def AntiBotToken() -> str: +def get_anti_bot_token() -> str: headers = { 'authority' : 'sdk.vercel.ai', 'accept' : '*/*', diff --git a/g4f/Provider/base_provider.py b/g4f/Provider/base_provider.py index ea81502f..e8a54f78 100644 --- a/g4f/Provider/base_provider.py +++ b/g4f/Provider/base_provider.py @@ -1,7 +1,9 @@ from __future__ import annotations import asyncio -from asyncio import SelectorEventLoop +import functools +from asyncio import SelectorEventLoop, AbstractEventLoop +from concurrent.futures import ThreadPoolExecutor from abc import ABC, abstractmethod import browser_cookie3 @@ -27,6 +29,31 @@ class BaseProvider(ABC): ) -> CreateResult: raise NotImplementedError() + @classmethod + async def create_async( + cls, + model: str, + messages: list[dict[str, str]], + *, + loop: AbstractEventLoop = None, + executor: ThreadPoolExecutor = None, + **kwargs + ) -> str: + if not loop: + loop = asyncio.get_event_loop() + + partial_func = functools.partial( + cls.create_completion, + model, + messages, + False, + **kwargs + ) + response = await loop.run_in_executor( + executor, + partial_func + ) + return "".join(response) @classmethod @property @@ -127,7 +154,7 @@ def create_event_loop() -> SelectorEventLoop: except RuntimeError: return SelectorEventLoop() raise RuntimeError( - 'Use "create_async" instead of "create" function in a async loop.') + 'Use "create_async" instead of "create" function in a running event loop.') _cookies = {} diff --git a/g4f/models.py b/g4f/models.py index 5cf8d9e9..23fc8e65 100644 --- a/g4f/models.py +++ b/g4f/models.py @@ -17,6 +17,7 @@ from .Provider import ( Wewordle, Yqcloud, AItianhu, + AItianhuSpace, Aichat, Myshell, ) @@ -38,7 +39,7 @@ default = Model( Wewordle, # Responds with markdown Yqcloud, # Answers short questions in chinese ChatBase, # Don't want to answer creatively - DeepAi, ChatgptLogin, ChatgptAi, Aivvm, GptGo, AItianhu, Aichat, Myshell, + DeepAi, ChatgptLogin, ChatgptAi, Aivvm, GptGo, AItianhu, AItianhuSpace, Aichat, Myshell, ]) ) @@ -47,7 +48,7 @@ gpt_35_turbo = Model( name = 'gpt-3.5-turbo', base_provider = 'openai', best_provider = RetryProvider([ - DeepAi, ChatgptLogin, ChatgptAi, Aivvm, GptGo, AItianhu, Aichat, Myshell, + DeepAi, ChatgptLogin, ChatgptAi, Aivvm, GptGo, AItianhu, Aichat, AItianhuSpace, Myshell, ]) ) @@ -55,7 +56,7 @@ gpt_4 = Model( name = 'gpt-4', base_provider = 'openai', best_provider = RetryProvider([ - Aivvm, Myshell + Aivvm, Myshell, AItianhuSpace, ]) ) |