From 3bece242045b885c71788f735cf77ac6187d2472 Mon Sep 17 00:00:00 2001 From: kqlio67 Date: Mon, 2 Sep 2024 11:55:51 +0300 Subject: Update models.py: Refine provider list and model support --- g4f/Provider/TeachAnything.py | 76 +++++++++++++++++++++++++++++++++++++++++++ g4f/Provider/__init__.py | 6 ++-- g4f/models.py | 8 +++-- 3 files changed, 84 insertions(+), 6 deletions(-) create mode 100644 g4f/Provider/TeachAnything.py diff --git a/g4f/Provider/TeachAnything.py b/g4f/Provider/TeachAnything.py new file mode 100644 index 00000000..3d34293f --- /dev/null +++ b/g4f/Provider/TeachAnything.py @@ -0,0 +1,76 @@ +from __future__ import annotations + +from typing import Any, Dict + +from aiohttp import ClientSession, ClientTimeout + +from ..typing import AsyncResult, Messages +from .base_provider import AsyncGeneratorProvider, ProviderModelMixin +from .helper import format_prompt + + +class TeachAnything(AsyncGeneratorProvider, ProviderModelMixin): + url = "https://www.teach-anything.com" + api_endpoint = "/api/generate" + working = True + default_model = "llama-3.1-70b" + + @classmethod + async def create_async_generator( + cls, + model: str, + messages: Messages, + proxy: str | None = None, + **kwargs: Any + ) -> AsyncResult: + headers = cls._get_headers() + + async with ClientSession(headers=headers) as session: + prompt = format_prompt(messages) + data = {"prompt": prompt} + + timeout = ClientTimeout(total=60) + + async with session.post( + f"{cls.url}{cls.api_endpoint}", + json=data, + proxy=proxy, + timeout=timeout + ) as response: + response.raise_for_status() + buffer = b"" + async for chunk in response.content.iter_any(): + buffer += chunk + try: + decoded = buffer.decode('utf-8') + yield decoded + buffer = b"" + except UnicodeDecodeError: + # If we can't decode, we'll wait for more data + continue + + # Handle any remaining data in the buffer + if buffer: + try: + yield buffer.decode('utf-8', errors='replace') + except Exception as e: + print(f"Error decoding final buffer: {e}") + + @staticmethod + def _get_headers() -> Dict[str, str]: + return { + "accept": "*/*", + "accept-language": "en-US,en;q=0.9", + "content-type": "application/json", + "dnt": "1", + "origin": "https://www.teach-anything.com", + "priority": "u=1, i", + "referer": "https://www.teach-anything.com/", + "sec-ch-ua": '"Not/A)Brand";v="8", "Chromium";v="126"', + "sec-ch-ua-mobile": "?0", + "sec-ch-ua-platform": '"Linux"', + "sec-fetch-dest": "empty", + "sec-fetch-mode": "cors", + "sec-fetch-site": "same-origin", + "user-agent": "Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/126.0.0.0 Safari/537.36" + } diff --git a/g4f/Provider/__init__.py b/g4f/Provider/__init__.py index dc8acfbe..e5234685 100644 --- a/g4f/Provider/__init__.py +++ b/g4f/Provider/__init__.py @@ -17,8 +17,8 @@ from .Bing import Bing from .BingCreateImages import BingCreateImages from .Blackbox import Blackbox from .ChatGot import ChatGot -from .Chatgpt4o import Chatgpt4o from .Chatgpt4Online import Chatgpt4Online +from .Chatgpt4o import Chatgpt4o from .ChatgptFree import ChatgptFree from .DDG import DDG from .DeepInfra import DeepInfra @@ -33,7 +33,6 @@ from .GigaChat import GigaChat from .GptTalkRu import GptTalkRu from .HuggingChat import HuggingChat from .HuggingFace import HuggingFace -from .HuggingFace import HuggingFace from .Koala import Koala from .Liaobots import Liaobots from .LiteIcoding import LiteIcoding @@ -47,11 +46,12 @@ from .Ollama import Ollama from .PerplexityLabs import PerplexityLabs from .Pi import Pi from .Pizzagpt import Pizzagpt -from .Upstage import Upstage from .Reka import Reka from .Replicate import Replicate from .ReplicateHome import ReplicateHome from .Rocks import Rocks +from .TeachAnything import TeachAnything +from .Upstage import Upstage from .Vercel import Vercel from .WhiteRabbitNeo import WhiteRabbitNeo from .You import You diff --git a/g4f/models.py b/g4f/models.py index 6ea2d2fd..4176f0a2 100644 --- a/g4f/models.py +++ b/g4f/models.py @@ -4,12 +4,13 @@ from dataclasses import dataclass from .Provider import IterListProvider, ProviderType from .Provider import ( + AiChatOnline, Allyfy, Bing, Blackbox, ChatGot, - Chatgpt4o, Chatgpt4Online, + Chatgpt4o, ChatgptFree, DDG, DeepInfra, @@ -35,6 +36,7 @@ from .Provider import ( Reka, Replicate, ReplicateHome, + TeachAnything, Upstage, You, ) @@ -106,7 +108,7 @@ gpt_4o_mini = Model( name = 'gpt-4o-mini', base_provider = 'openai', best_provider = IterListProvider([ - DDG, Liaobots, You, FreeNetfly, MagickPenAsk, MagickPenChat, Pizzagpt, ChatgptFree, OpenaiChat, Koala, + DDG, Liaobots, You, FreeNetfly, MagickPenAsk, MagickPenChat, Pizzagpt, ChatgptFree, AiChatOnline, OpenaiChat, Koala, ]) ) @@ -163,7 +165,7 @@ llama_3_1_8b = Model( llama_3_1_70b = Model( name = "llama-3.1-70b", base_provider = "meta", - best_provider = IterListProvider([DDG, HuggingChat, FreeGpt, Blackbox, HuggingFace]) + best_provider = IterListProvider([DDG, HuggingChat, FreeGpt, Blackbox, TeachAnything, HuggingFace]) ) llama_3_1_405b = Model( -- cgit v1.2.3