mirror of
https://github.com/xtekky/gpt4free.git
synced 2024-12-25 12:16:17 +03:00
~ | g4f v-0.1.6.1
some bug fixes
This commit is contained in:
parent
b569189764
commit
f74ca10fbd
@ -2,7 +2,7 @@
|
|||||||
|
|
||||||
By using this repository or any code related to it, you agree to the [legal notice](./LEGAL_NOTICE.md). The author is not responsible for any copies, forks, reuploads made by other users, or anything else related to gpt4free. This is the author's only account and repository. To prevent impersonation or irresponsible actions, please comply with the GNU GPL license this Repository uses.
|
By using this repository or any code related to it, you agree to the [legal notice](./LEGAL_NOTICE.md). The author is not responsible for any copies, forks, reuploads made by other users, or anything else related to gpt4free. This is the author's only account and repository. To prevent impersonation or irresponsible actions, please comply with the GNU GPL license this Repository uses.
|
||||||
|
|
||||||
- latest pypi version: [`0.1.5.9`](https://pypi.org/project/g4f/0.1.5.9):
|
- latest pypi version: [`0.1.6.1`](https://pypi.org/project/g4f/0.1.6.1):
|
||||||
```sh
|
```sh
|
||||||
pip install -U g4f
|
pip install -U g4f
|
||||||
```
|
```
|
||||||
|
@ -20,7 +20,7 @@ models = {
|
|||||||
class Aivvm(BaseProvider):
|
class Aivvm(BaseProvider):
|
||||||
url = 'https://chat.aivvm.com'
|
url = 'https://chat.aivvm.com'
|
||||||
supports_stream = True
|
supports_stream = True
|
||||||
working = True
|
working = False
|
||||||
supports_gpt_35_turbo = True
|
supports_gpt_35_turbo = True
|
||||||
supports_gpt_4 = True
|
supports_gpt_4 = True
|
||||||
|
|
||||||
|
@ -11,7 +11,7 @@ from .base_provider import AsyncGeneratorProvider, format_prompt
|
|||||||
|
|
||||||
class H2o(AsyncGeneratorProvider):
|
class H2o(AsyncGeneratorProvider):
|
||||||
url = "https://gpt-gm.h2o.ai"
|
url = "https://gpt-gm.h2o.ai"
|
||||||
working = True
|
working = False
|
||||||
model = "h2oai/h2ogpt-gm-oasst1-en-2048-falcon-40b-v1"
|
model = "h2oai/h2ogpt-gm-oasst1-en-2048-falcon-40b-v1"
|
||||||
|
|
||||||
@classmethod
|
@classmethod
|
||||||
|
@ -5,7 +5,7 @@ from .Provider import BaseProvider
|
|||||||
from .typing import Messages, CreateResult, Union
|
from .typing import Messages, CreateResult, Union
|
||||||
from .debug import logging
|
from .debug import logging
|
||||||
|
|
||||||
version = '0.1.5.9'
|
version = '0.1.6.1'
|
||||||
version_check = True
|
version_check = True
|
||||||
|
|
||||||
def check_pypi_version() -> None:
|
def check_pypi_version() -> None:
|
||||||
|
@ -27,7 +27,6 @@ from .Provider import (
|
|||||||
Bing,
|
Bing,
|
||||||
You,
|
You,
|
||||||
H2o,
|
H2o,
|
||||||
Aivvm
|
|
||||||
)
|
)
|
||||||
|
|
||||||
@dataclass(unsafe_hash=True)
|
@dataclass(unsafe_hash=True)
|
||||||
@ -64,7 +63,7 @@ gpt_35_turbo = Model(
|
|||||||
name = 'gpt-3.5-turbo',
|
name = 'gpt-3.5-turbo',
|
||||||
base_provider = 'openai',
|
base_provider = 'openai',
|
||||||
best_provider = RetryProvider([
|
best_provider = RetryProvider([
|
||||||
Aivvm, ChatgptLogin, ChatgptAi, GptGo, AItianhu, Aichat, AItianhuSpace, Myshell, Aibn, FreeGpt, Ylokh
|
ChatgptLogin, ChatgptAi, GptGo, AItianhu, Aichat, AItianhuSpace, Myshell, Aibn, FreeGpt, Ylokh
|
||||||
])
|
])
|
||||||
)
|
)
|
||||||
|
|
||||||
@ -72,7 +71,7 @@ gpt_4 = Model(
|
|||||||
name = 'gpt-4',
|
name = 'gpt-4',
|
||||||
base_provider = 'openai',
|
base_provider = 'openai',
|
||||||
best_provider = RetryProvider([
|
best_provider = RetryProvider([
|
||||||
Aivvm, Bing
|
Bing
|
||||||
])
|
])
|
||||||
)
|
)
|
||||||
|
|
||||||
@ -167,31 +166,31 @@ gpt_35_turbo_16k = Model(
|
|||||||
gpt_35_turbo_16k_0613 = Model(
|
gpt_35_turbo_16k_0613 = Model(
|
||||||
name = 'gpt-3.5-turbo-16k-0613',
|
name = 'gpt-3.5-turbo-16k-0613',
|
||||||
base_provider = 'openai',
|
base_provider = 'openai',
|
||||||
best_provider = Aivvm
|
best_provider = gpt_35_turbo.best_provider
|
||||||
)
|
)
|
||||||
|
|
||||||
gpt_35_turbo_0613 = Model(
|
gpt_35_turbo_0613 = Model(
|
||||||
name = 'gpt-3.5-turbo-0613',
|
name = 'gpt-3.5-turbo-0613',
|
||||||
base_provider = 'openai',
|
base_provider = 'openai',
|
||||||
best_provider = Aivvm
|
best_provider = gpt_35_turbo.best_provider
|
||||||
)
|
)
|
||||||
|
|
||||||
gpt_4_0613 = Model(
|
gpt_4_0613 = Model(
|
||||||
name = 'gpt-4-0613',
|
name = 'gpt-4-0613',
|
||||||
base_provider = 'openai',
|
base_provider = 'openai',
|
||||||
best_provider = Aivvm
|
best_provider = gpt_4.best_provider
|
||||||
)
|
)
|
||||||
|
|
||||||
gpt_4_32k = Model(
|
gpt_4_32k = Model(
|
||||||
name = 'gpt-4-32k',
|
name = 'gpt-4-32k',
|
||||||
base_provider = 'openai',
|
base_provider = 'openai',
|
||||||
best_provider = Aivvm
|
best_provider = gpt_4.best_provider
|
||||||
)
|
)
|
||||||
|
|
||||||
gpt_4_32k_0613 = Model(
|
gpt_4_32k_0613 = Model(
|
||||||
name = 'gpt-4-32k-0613',
|
name = 'gpt-4-32k-0613',
|
||||||
base_provider = 'openai',
|
base_provider = 'openai',
|
||||||
best_provider = Aivvm
|
best_provider = gpt_4.best_provider
|
||||||
)
|
)
|
||||||
|
|
||||||
text_ada_001 = Model(
|
text_ada_001 = Model(
|
||||||
|
2
setup.py
2
setup.py
@ -14,7 +14,7 @@ with open("requirements.txt") as f:
|
|||||||
with open("etc/interference/requirements.txt") as f:
|
with open("etc/interference/requirements.txt") as f:
|
||||||
api_required = f.read().splitlines()
|
api_required = f.read().splitlines()
|
||||||
|
|
||||||
VERSION = '0.1.5.9'
|
VERSION = '0.1.6.1'
|
||||||
DESCRIPTION = (
|
DESCRIPTION = (
|
||||||
"The official gpt4free repository | various collection of powerful language models"
|
"The official gpt4free repository | various collection of powerful language models"
|
||||||
)
|
)
|
||||||
|
Loading…
Reference in New Issue
Block a user