mirror of
https://github.com/xtekky/gpt4free.git
synced 2024-12-23 19:11:48 +03:00
6ce493d4df
* Fix api streaming, fix AsyncClient, Improve Client class, Some providers fixes, Update models list, Fix some tests, Update model list in Airforce provid er, Add OpenAi image generation url to api, Fix reload and debug in api arguments, Fix websearch in gui * Fix Cloadflare and Pi and AmigoChat provider * Fix conversation support in DDG provider, Add cloudflare bypass with nodriver * Fix unittests without curl_cffi
76 lines
2.9 KiB
Python
76 lines
2.9 KiB
Python
from __future__ import annotations
|
|
|
|
import json
|
|
|
|
from ..typing import AsyncResult, Messages, Cookies
|
|
from .base_provider import AsyncGeneratorProvider, format_prompt
|
|
from ..requests import StreamSession, get_args_from_nodriver, raise_for_status, merge_cookies
|
|
|
|
class Pi(AsyncGeneratorProvider):
|
|
url = "https://pi.ai/talk"
|
|
working = True
|
|
supports_stream = True
|
|
default_model = "pi"
|
|
models = [default_model]
|
|
_headers: dict = None
|
|
_cookies: Cookies = {}
|
|
|
|
@classmethod
|
|
async def create_async_generator(
|
|
cls,
|
|
model: str,
|
|
messages: Messages,
|
|
stream: bool,
|
|
proxy: str = None,
|
|
timeout: int = 180,
|
|
conversation_id: str = None,
|
|
**kwargs
|
|
) -> AsyncResult:
|
|
if cls._headers is None:
|
|
args = await get_args_from_nodriver(cls.url, proxy=proxy, timeout=timeout)
|
|
cls._cookies = args.get("cookies", {})
|
|
cls._headers = args.get("headers")
|
|
async with StreamSession(headers=cls._headers, cookies=cls._cookies, proxy=proxy) as session:
|
|
if not conversation_id:
|
|
conversation_id = await cls.start_conversation(session)
|
|
prompt = format_prompt(messages)
|
|
else:
|
|
prompt = messages[-1]["content"]
|
|
answer = cls.ask(session, prompt, conversation_id)
|
|
async for line in answer:
|
|
if "text" in line:
|
|
yield line["text"]
|
|
|
|
@classmethod
|
|
async def start_conversation(cls, session: StreamSession) -> str:
|
|
async with session.post('https://pi.ai/api/chat/start', data="{}", headers={
|
|
'accept': 'application/json',
|
|
'x-api-version': '3'
|
|
}) as response:
|
|
await raise_for_status(response)
|
|
return (await response.json())['conversations'][0]['sid']
|
|
|
|
async def get_chat_history(session: StreamSession, conversation_id: str):
|
|
params = {
|
|
'conversation': conversation_id,
|
|
}
|
|
async with session.get('https://pi.ai/api/chat/history', params=params) as response:
|
|
await raise_for_status(response)
|
|
return await response.json()
|
|
|
|
@classmethod
|
|
async def ask(cls, session: StreamSession, prompt: str, conversation_id: str):
|
|
json_data = {
|
|
'text': prompt,
|
|
'conversation': conversation_id,
|
|
'mode': 'BASE',
|
|
}
|
|
async with session.post('https://pi.ai/api/chat', json=json_data) as response:
|
|
await raise_for_status(response)
|
|
cls._cookies = merge_cookies(cls._cookies, response)
|
|
async for line in response.iter_lines():
|
|
if line.startswith(b'data: {"text":'):
|
|
yield json.loads(line.split(b'data: ')[1])
|
|
elif line.startswith(b'data: {"title":'):
|
|
yield json.loads(line.split(b'data: ')[1])
|