mirror of
https://github.com/xtekky/gpt4free.git
synced 2024-11-26 09:57:24 +03:00
quora (poe) [gpt-4/3.5] create bot feature
added create bot feature quora.Model.create new model names: { 'sage' : 'capybara', 'gpt-4' : 'beaver', 'claude-v1.2' : 'a2_2', 'claude-instant-v1.0' : 'a2', 'gpt-3.5-turbo' : 'chinchilla' }
This commit is contained in:
parent
b68da4c114
commit
081dfff8b4
52
README.md
52
README.md
@ -6,7 +6,7 @@ This repository provides reverse-engineered language models from various sources
|
||||
|
||||
## To-Do List
|
||||
|
||||
- [ ] implement poe.com create bot feature (4)
|
||||
- [x] implement poe.com create bot feature (4) AVAILABLE NOW
|
||||
- [ ] poe.com chat history management (3)
|
||||
- [x] renaming the 'poe' module to 'quora' (2)
|
||||
- [x] add you.com api (1)
|
||||
@ -47,31 +47,41 @@ These sites will be reverse engineered but need account access:
|
||||
### Example: `quora (poe)` (use like openai pypi package) - GPT-4 <a name="example-poe"></a>
|
||||
|
||||
```python
|
||||
# Import quora (poe)
|
||||
# quora model names: (use left key as argument)
|
||||
models = {
|
||||
'sage' : 'capybara',
|
||||
'gpt-4' : 'beaver',
|
||||
'claude-v1.2' : 'a2_2',
|
||||
'claude-instant-v1.0' : 'a2',
|
||||
'gpt-3.5-turbo' : 'chinchilla'
|
||||
}
|
||||
```
|
||||
|
||||
#### !! new: bot creation
|
||||
|
||||
```python
|
||||
# import quora (poe) package
|
||||
import quora
|
||||
|
||||
# quora.Account.create
|
||||
# quora.Completion.create
|
||||
# quora.StreamCompletion.create
|
||||
# create account
|
||||
# make shure to set enable_bot_creation to True
|
||||
token = quora.Account.create(logging = True, enable_bot_creation=True)
|
||||
|
||||
[...]
|
||||
model = quora.Model.create(
|
||||
token = token,
|
||||
model = 'gpt-3.5-turbo', # or claude-instant-v1.0
|
||||
system_prompt = 'you are ChatGPT a large language model ...'
|
||||
)
|
||||
|
||||
```
|
||||
print(model.name) # gptx....
|
||||
|
||||
#### Create Token (3-6s)
|
||||
```python
|
||||
token = quora.Account.create(logging = True)
|
||||
print('token', token)
|
||||
```
|
||||
|
||||
#### Streaming Response
|
||||
```python
|
||||
|
||||
for response in quora.StreamingCompletion.create(model = 'gpt-4',
|
||||
prompt = 'hello world',
|
||||
token = token):
|
||||
# streaming response
|
||||
for response in quora.StreamingCompletion.create(
|
||||
custom_model = model.name,
|
||||
prompt ='hello world',
|
||||
token = token):
|
||||
|
||||
print(response.completion.choices[0].text, end="", flush=True)
|
||||
print(response.completion.choices[0].text)
|
||||
```
|
||||
|
||||
#### Normal Response:
|
||||
@ -84,8 +94,6 @@ response = quora.Completion.create(model = 'gpt-4',
|
||||
print(response.completion.choices[0].text)
|
||||
```
|
||||
|
||||
|
||||
|
||||
### Example: `t3nsor` (use like openai pypi package) <a name="example-t3nsor"></a>
|
||||
|
||||
```python
|
||||
|
@ -3,11 +3,12 @@ from quora.mail import Mail
|
||||
from requests import Session
|
||||
from re import search, findall
|
||||
from json import loads
|
||||
from time import sleep, time
|
||||
from time import sleep
|
||||
from pathlib import Path
|
||||
from random import choice
|
||||
from random import choice, choices, randint
|
||||
from string import ascii_letters, digits
|
||||
from urllib import parse
|
||||
|
||||
|
||||
class PoeResponse:
|
||||
|
||||
class Completion:
|
||||
@ -48,8 +49,88 @@ class PoeResponse:
|
||||
def json(self) -> dict:
|
||||
return self.response_dict
|
||||
|
||||
|
||||
class ModelResponse:
|
||||
def __init__(self, json_response: dict) -> None:
|
||||
self.id = json_response['data']['poeBotCreate']['bot']['id']
|
||||
self.name = json_response['data']['poeBotCreate']['bot']['displayName']
|
||||
self.limit = json_response['data']['poeBotCreate']['bot']['messageLimit']['dailyLimit']
|
||||
self.deleted = json_response['data']['poeBotCreate']['bot']['deletionState']
|
||||
|
||||
class Model:
|
||||
def create(
|
||||
token: str,
|
||||
model: str = 'gpt-3.5-turbo', # claude-instant
|
||||
system_prompt: str = 'You are ChatGPT a large language model developed by Openai. Answer as consisely as possible',
|
||||
description: str = 'gpt-3.5 language model from openai, skidded by poe.com',
|
||||
handle: str = None) -> ModelResponse:
|
||||
|
||||
models = {
|
||||
'gpt-3.5-turbo' : 'chinchilla',
|
||||
'claude-instant-v1.0': 'a2'
|
||||
}
|
||||
|
||||
if not handle:
|
||||
handle = f'gptx{randint(1111111, 9999999)}'
|
||||
|
||||
client = Session()
|
||||
client.cookies['p-b'] = token
|
||||
|
||||
settings = client.get('https://poe.com/api/settings').json()
|
||||
|
||||
client.headers = {
|
||||
"host" : "poe.com",
|
||||
"origin" : "https://poe.com",
|
||||
"referer" : "https://poe.com/",
|
||||
"content-type" : "application/json",
|
||||
"poe-formkey" : settings['formkey'],
|
||||
"poe-tchannel" : settings['tchannelData']['channel'],
|
||||
"user-agent" : "Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_7) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/112.0.0.0 Safari/537.36",
|
||||
"connection" : "keep-alive",
|
||||
"sec-ch-ua" : "\"Chromium\";v=\"112\", \"Google Chrome\";v=\"112\", \"Not:A-Brand\";v=\"99\"",
|
||||
"sec-ch-ua-mobile" : "?0",
|
||||
"sec-ch-ua-platform": "\"macOS\"",
|
||||
"sec-fetch-site" : "same-origin",
|
||||
"sec-fetch-mode" : "cors",
|
||||
"sec-fetch-dest" : "empty",
|
||||
"accept" : "*/*",
|
||||
"accept-encoding" : "gzip, deflate, br",
|
||||
"accept-language" : "en-GB,en-US;q=0.9,en;q=0.8",
|
||||
}
|
||||
|
||||
response = client.post("https://poe.com/api/gql_POST", json = {
|
||||
'queryName': 'CreateBotMain_poeBotCreate_Mutation',
|
||||
'variables': {
|
||||
'model' : models[model],
|
||||
'handle' : handle,
|
||||
'prompt' : system_prompt,
|
||||
'isPromptPublic' : True,
|
||||
'introduction' : '',
|
||||
'description' : description,
|
||||
'profilePictureUrl' : 'https://qph.fs.quoracdn.net/main-qimg-24e0b480dcd946e1cc6728802c5128b6',
|
||||
'apiUrl' : None,
|
||||
'apiKey' : ''.join(choices(ascii_letters + digits, k = 32)),
|
||||
'isApiBot' : False,
|
||||
'hasLinkification' : False,
|
||||
'hasMarkdownRendering' : False,
|
||||
'hasSuggestedReplies' : False,
|
||||
'isPrivateBot' : False
|
||||
},
|
||||
'query': 'mutation CreateBotMain_poeBotCreate_Mutation(\n $model: String!\n $handle: String!\n $prompt: String!\n $isPromptPublic: Boolean!\n $introduction: String!\n $description: String!\n $profilePictureUrl: String\n $apiUrl: String\n $apiKey: String\n $isApiBot: Boolean\n $hasLinkification: Boolean\n $hasMarkdownRendering: Boolean\n $hasSuggestedReplies: Boolean\n $isPrivateBot: Boolean\n) {\n poeBotCreate(model: $model, handle: $handle, promptPlaintext: $prompt, isPromptPublic: $isPromptPublic, introduction: $introduction, description: $description, profilePicture: $profilePictureUrl, apiUrl: $apiUrl, apiKey: $apiKey, isApiBot: $isApiBot, hasLinkification: $hasLinkification, hasMarkdownRendering: $hasMarkdownRendering, hasSuggestedReplies: $hasSuggestedReplies, isPrivateBot: $isPrivateBot) {\n status\n bot {\n id\n ...BotHeader_bot\n }\n }\n}\n\nfragment BotHeader_bot on Bot {\n displayName\n messageLimit {\n dailyLimit\n }\n ...BotImage_bot\n ...BotLink_bot\n ...IdAnnotation_node\n ...botHelpers_useViewerCanAccessPrivateBot\n ...botHelpers_useDeletion_bot\n}\n\nfragment BotImage_bot on Bot {\n displayName\n ...botHelpers_useDeletion_bot\n ...BotImage_useProfileImage_bot\n}\n\nfragment BotImage_useProfileImage_bot on Bot {\n image {\n __typename\n ... on LocalBotImage {\n localName\n }\n ... on UrlBotImage {\n url\n }\n }\n ...botHelpers_useDeletion_bot\n}\n\nfragment BotLink_bot on Bot {\n displayName\n}\n\nfragment IdAnnotation_node on Node {\n __isNode: __typename\n id\n}\n\nfragment botHelpers_useDeletion_bot on Bot {\n deletionState\n}\n\nfragment botHelpers_useViewerCanAccessPrivateBot on Bot {\n isPrivateBot\n viewerIsCreator\n}\n',
|
||||
})
|
||||
|
||||
if not 'success' in response.text:
|
||||
raise Exception('''
|
||||
Bot creation Failed
|
||||
!! Important !!
|
||||
Bot creation was not enabled on this account
|
||||
please use: quora.Account.create with enable_bot_creation set to True
|
||||
''')
|
||||
|
||||
return ModelResponse(response.json())
|
||||
|
||||
class Account:
|
||||
def create(proxy: None or str = None, logging: bool = False):
|
||||
def create(proxy: None or str = None, logging: bool = False, enable_bot_creation: bool = False):
|
||||
|
||||
client = Session()
|
||||
client.proxies = {
|
||||
@ -133,6 +214,13 @@ class Account:
|
||||
with open(Path(__file__).resolve().parent / 'cookies.txt', 'a') as f:
|
||||
f.write(f'{token}\n')
|
||||
|
||||
if enable_bot_creation:
|
||||
client.post("https://poe.com/api/gql_POST", json = {
|
||||
"queryName": "UserProfileConfigurePreviewModal_markMultiplayerNuxCompleted_Mutation",
|
||||
"variables": {},
|
||||
"query": "mutation UserProfileConfigurePreviewModal_markMultiplayerNuxCompleted_Mutation {\n markMultiplayerNuxCompleted {\n viewer {\n hasCompletedMultiplayerNux\n id\n }\n }\n}\n"
|
||||
})
|
||||
|
||||
return token
|
||||
|
||||
def get():
|
||||
@ -142,17 +230,20 @@ class Account:
|
||||
class StreamingCompletion:
|
||||
def create(
|
||||
model : str = 'gpt-4',
|
||||
custom_model : str = None,
|
||||
prompt: str = 'hello world',
|
||||
token : str = ''):
|
||||
|
||||
models = {
|
||||
'sage' : 'capybara',
|
||||
'gpt-4' : 'beaver',
|
||||
'claude+': 'a2_2',
|
||||
'claude' : 'a2',
|
||||
'gpt-3.5': 'chinchilla'
|
||||
'claude-v1.2' : 'a2_2',
|
||||
'claude-instant-v1.0' : 'a2',
|
||||
'gpt-3.5-turbo' : 'chinchilla'
|
||||
}
|
||||
|
||||
_model = models[model] if not custom_model else custom_model
|
||||
|
||||
client = PoeClient(token)
|
||||
|
||||
for chunk in client.send_message(models[model], prompt):
|
||||
@ -161,7 +252,7 @@ class StreamingCompletion:
|
||||
'id' : chunk["messageId"],
|
||||
'object' : 'text_completion',
|
||||
'created': chunk['creationTime'],
|
||||
'model' : models[model],
|
||||
'model' : _model,
|
||||
'choices': [{
|
||||
'text' : chunk["text_new"],
|
||||
'index' : 0,
|
||||
@ -178,17 +269,20 @@ class StreamingCompletion:
|
||||
class Completion:
|
||||
def create(
|
||||
model : str = 'gpt-4',
|
||||
custom_model : str = None,
|
||||
prompt: str = 'hello world',
|
||||
token : str = ''):
|
||||
|
||||
models = {
|
||||
'sage' : 'capybara',
|
||||
'gpt-4' : 'beaver',
|
||||
'claude+': 'a2_2',
|
||||
'claude' : 'a2',
|
||||
'gpt-3.5': 'chinchilla'
|
||||
'claude-v1.2' : 'a2_2',
|
||||
'claude-instant-v1.0' : 'a2',
|
||||
'gpt-3.5-turbo' : 'chinchilla'
|
||||
}
|
||||
|
||||
_model = models[model] if not custom_model else custom_model
|
||||
|
||||
client = PoeClient(token)
|
||||
|
||||
for chunk in client.send_message(models[model], prompt):
|
||||
@ -198,7 +292,7 @@ class Completion:
|
||||
'id' : chunk["messageId"],
|
||||
'object' : 'text_completion',
|
||||
'created': chunk['creationTime'],
|
||||
'model' : models[model],
|
||||
'model' : _model,
|
||||
'choices': [{
|
||||
'text' : chunk["text"],
|
||||
'index' : 0,
|
||||
|
@ -8,3 +8,8 @@ pUEbtxobN_QUSpLIR8RGww==
|
||||
9_dUWxKkHHhpQRSvCvBk2Q==
|
||||
UV45rvGwUwi2qV9QdIbMcw==
|
||||
cVIN0pK1Wx-F7zCdUxlYqA==
|
||||
UP2wQVds17VFHh6IfCQFrA==
|
||||
18eKr0ME2Tzifdfqat38Aw==
|
||||
FNgKEpc2r-XqWe0rHBfYpg==
|
||||
juCAh6kB0sUpXHvKik2woA==
|
||||
nBvuNYRLaE4xE4HuzBPiIQ==
|
||||
|
@ -1,3 +1,3 @@
|
||||
websocket-client
|
||||
requests
|
||||
tls-client
|
||||
tls-client
|
18
testing/quora_test_2.py
Normal file
18
testing/quora_test_2.py
Normal file
@ -0,0 +1,18 @@
|
||||
import quora
|
||||
|
||||
token = quora.Account.create(logging = True, enable_bot_creation=True)
|
||||
|
||||
model = quora.Model.create(
|
||||
token = token,
|
||||
model = 'gpt-3.5-turbo', # or claude-instant-v1.0
|
||||
system_prompt = 'you are ChatGPT a large language model ...'
|
||||
)
|
||||
|
||||
print(model.name)
|
||||
|
||||
for response in quora.StreamingCompletion.create(
|
||||
custom_model = model.name,
|
||||
prompt ='hello world',
|
||||
token = token):
|
||||
|
||||
print(response.completion.choices[0].text)
|
Loading…
Reference in New Issue
Block a user