mirror of
https://github.com/xtekky/gpt4free.git
synced 2024-10-03 19:57:15 +03:00
Compare commits
3 Commits
410b46d308
...
8b73fa926d
Author | SHA1 | Date | |
---|---|---|---|
|
8b73fa926d | ||
|
80503af104 | ||
|
cba121daac |
@ -23,6 +23,7 @@ class Blackbox(AsyncGeneratorProvider, ProviderModelMixin):
|
||||
'gemini-1.5-flash',
|
||||
"llama-3.1-8b",
|
||||
'llama-3.1-70b',
|
||||
'llama-3.1-405b',
|
||||
'ImageGenerationLV45LJp',
|
||||
'gpt-4o',
|
||||
'gemini-pro',
|
||||
@ -38,6 +39,7 @@ class Blackbox(AsyncGeneratorProvider, ProviderModelMixin):
|
||||
"gemini-1.5-flash": {'mode': True, 'id': 'Gemini'},
|
||||
"llama-3.1-8b": {'mode': True, 'id': "llama-3.1-8b"},
|
||||
'llama-3.1-70b': {'mode': True, 'id': "llama-3.1-70b"},
|
||||
'llama-3.1-405b': {'mode': True, 'id': "llama-3.1-405b"},
|
||||
}
|
||||
|
||||
userSelectedModel = {
|
||||
|
@ -136,7 +136,6 @@ class ChatGpt(AbstractProvider, ProviderModelMixin):
|
||||
response = session.post('https://chatgpt.com/backend-anon/sentinel/chat-requirements',
|
||||
headers=headers, json={'p': pow_req})
|
||||
|
||||
# Додана перевірка на статус відповідей, якщо "незвична активність"
|
||||
if response.status_code != 200:
|
||||
print(f"Request failed with status: {response.status_code}")
|
||||
print(f"Response content: {response.content}")
|
||||
@ -199,7 +198,6 @@ class ChatGpt(AbstractProvider, ProviderModelMixin):
|
||||
},
|
||||
}
|
||||
|
||||
# Додав паузу між запитами для уникнення блокувань через частоту
|
||||
time.sleep(2)
|
||||
|
||||
response = session.post('https://chatgpt.com/backend-anon/conversation',
|
||||
@ -209,10 +207,10 @@ class ChatGpt(AbstractProvider, ProviderModelMixin):
|
||||
for line in response.iter_lines():
|
||||
if line:
|
||||
decoded_line = line.decode()
|
||||
print(f"Received line: {decoded_line}") # Debugging output
|
||||
print(f"Received line: {decoded_line}")
|
||||
if decoded_line.startswith('data:'):
|
||||
json_string = decoded_line[6:] # Extract JSON part after 'data:'
|
||||
if json_string.strip(): # Check if there's actual content
|
||||
json_string = decoded_line[6:]
|
||||
if json_string.strip():
|
||||
try:
|
||||
data = json.loads(json_string)
|
||||
except json.JSONDecodeError as e:
|
||||
|
@ -206,7 +206,7 @@ llama_3_1_70b = Model(
|
||||
llama_3_1_405b = Model(
|
||||
name = "llama-3.1-405b",
|
||||
base_provider = "Meta Llama",
|
||||
best_provider = IterListProvider([DeepInfraChat, Airforce])
|
||||
best_provider = IterListProvider([DeepInfraChat, Blackbox, Airforce])
|
||||
)
|
||||
|
||||
llama_3_1 = Model(
|
||||
|
Loading…
Reference in New Issue
Block a user