mirror of
https://github.com/xtekky/gpt4free.git
synced 2024-11-24 09:48:15 +03:00
commit
99cd4088e5
62
testing/aiservice/AiService.py
Normal file
62
testing/aiservice/AiService.py
Normal file
@ -0,0 +1,62 @@
|
||||
import os,sys
|
||||
import requests
|
||||
# from ...typing import get_type_hints
|
||||
|
||||
url = "https://aiservice.vercel.app/api/chat/answer"
|
||||
model = ['gpt-3.5-turbo']
|
||||
supports_stream = False
|
||||
needs_auth = False
|
||||
|
||||
|
||||
def _create_completion(model: str, messages: list, stream: bool, **kwargs):
|
||||
base = ''
|
||||
for message in messages:
|
||||
base += '%s: %s\n' % (message['role'], message['content'])
|
||||
base += 'assistant:'
|
||||
|
||||
headers = {
|
||||
"accept": "*/*",
|
||||
"content-type": "text/plain;charset=UTF-8",
|
||||
"sec-fetch-dest": "empty",
|
||||
"sec-fetch-mode": "cors",
|
||||
"sec-fetch-site": "same-origin",
|
||||
"Referer": "https://aiservice.vercel.app/chat",
|
||||
}
|
||||
data = {
|
||||
"input": base
|
||||
}
|
||||
response = requests.post(url, headers=headers, json=data)
|
||||
if response.status_code == 200:
|
||||
_json = response.json()
|
||||
yield _json['data']
|
||||
else:
|
||||
print(f"Error Occurred::{response.status_code}")
|
||||
return None
|
||||
|
||||
|
||||
|
||||
# params = f'g4f.Providers.{os.path.basename(__file__)[:-3]} supports: ' + \
|
||||
# '(%s)' % ', '.join(
|
||||
# [f"{name}: {get_type_hints(_create_completion)[name].__name__}" for name in _create_completion.__code__.co_varnames[:_create_completion.__code__.co_argcount]])
|
||||
|
||||
|
||||
# Temporary For ChatCompletion Class
|
||||
class ChatCompletion:
|
||||
@staticmethod
|
||||
def create(model: str, messages: list, provider: None or str, stream: bool = False, auth: str = False, **kwargs):
|
||||
kwargs['auth'] = auth
|
||||
|
||||
if provider and needs_auth and not auth:
|
||||
print(
|
||||
f'ValueError: {provider} requires authentication (use auth="cookie or token or jwt ..." param)', file=sys.stderr)
|
||||
sys.exit(1)
|
||||
|
||||
try:
|
||||
return (_create_completion(model, messages, stream, **kwargs)
|
||||
if stream else ''.join(_create_completion(model, messages, stream, **kwargs)))
|
||||
except TypeError as e:
|
||||
print(e)
|
||||
arg: str = str(e).split("'")[1]
|
||||
print(
|
||||
f"ValueError: {provider} does not support '{arg}' argument", file=sys.stderr)
|
||||
sys.exit(1)
|
2
testing/aiservice/README.md
Normal file
2
testing/aiservice/README.md
Normal file
@ -0,0 +1,2 @@
|
||||
https://github.com/xtekky/gpt4free/issues/40#issuecomment-1629152431
|
||||
probably gpt-3.5
|
30
testing/aiservice/testing.py
Normal file
30
testing/aiservice/testing.py
Normal file
@ -0,0 +1,30 @@
|
||||
from AiService import ChatCompletion
|
||||
|
||||
# Test 1
|
||||
response = ChatCompletion.create(model="gpt-3.5-turbo",
|
||||
provider="AiService",
|
||||
stream=False,
|
||||
messages=[{'role': 'user', 'content': 'who are you?'}])
|
||||
|
||||
print(response)
|
||||
|
||||
# Test 2
|
||||
response = ChatCompletion.create(model="gpt-3.5-turbo",
|
||||
provider="AiService",
|
||||
stream=False,
|
||||
messages=[{'role': 'user', 'content': 'what you can do?'}])
|
||||
|
||||
print(response)
|
||||
|
||||
|
||||
# Test 3
|
||||
response = ChatCompletion.create(model="gpt-3.5-turbo",
|
||||
provider="AiService",
|
||||
stream=False,
|
||||
messages=[
|
||||
{'role': 'user', 'content': 'now your name is Bob'},
|
||||
{'role': 'assistant', 'content': 'Hello Im Bob, you asistant'},
|
||||
{'role': 'user', 'content': 'what your name again?'},
|
||||
])
|
||||
|
||||
print(response)
|
Loading…
Reference in New Issue
Block a user