mirror of
https://github.com/xtekky/gpt4free.git
synced 2024-12-11 07:07:04 +03:00
Update (g4f/models.py g4f/Provider/ReplicateHome.py)
This commit is contained in:
parent
2574101570
commit
330d920e4e
@ -17,7 +17,13 @@ class ReplicateHome(AsyncGeneratorProvider, ProviderModelMixin):
|
|||||||
supports_system_message = True
|
supports_system_message = True
|
||||||
supports_message_history = True
|
supports_message_history = True
|
||||||
|
|
||||||
default_model = 'meta/meta-llama-3-70b-instruct'
|
default_model = 'yorickvp/llava-13b'
|
||||||
|
|
||||||
|
image_models = [
|
||||||
|
'stability-ai/stable-diffusion-3',
|
||||||
|
'bytedance/sdxl-lightning-4step',
|
||||||
|
'playgroundai/playground-v2.5-1024px-aesthetic',
|
||||||
|
]
|
||||||
|
|
||||||
text_models = [
|
text_models = [
|
||||||
'meta/meta-llama-3-70b-instruct',
|
'meta/meta-llama-3-70b-instruct',
|
||||||
@ -26,35 +32,31 @@ class ReplicateHome(AsyncGeneratorProvider, ProviderModelMixin):
|
|||||||
'yorickvp/llava-13b',
|
'yorickvp/llava-13b',
|
||||||
]
|
]
|
||||||
|
|
||||||
image_models = [
|
|
||||||
'black-forest-labs/flux-schnell',
|
|
||||||
'stability-ai/stable-diffusion-3',
|
|
||||||
'bytedance/sdxl-lightning-4step',
|
|
||||||
'playgroundai/playground-v2.5-1024px-aesthetic',
|
|
||||||
]
|
|
||||||
|
|
||||||
models = text_models + image_models
|
models = text_models + image_models
|
||||||
|
|
||||||
model_aliases = {
|
model_aliases = {
|
||||||
"flux-schnell": "black-forest-labs/flux-schnell",
|
# image_models
|
||||||
"sd-3": "stability-ai/stable-diffusion-3",
|
"sd-3": "stability-ai/stable-diffusion-3",
|
||||||
"sdxl": "bytedance/sdxl-lightning-4step",
|
"sdxl": "bytedance/sdxl-lightning-4step",
|
||||||
"playground-v2.5": "playgroundai/playground-v2.5-1024px-aesthetic",
|
"playground-v2.5": "playgroundai/playground-v2.5-1024px-aesthetic",
|
||||||
"llama-3-70b": "meta/meta-llama-3-70b-instruct",
|
|
||||||
"mixtral-8x7b": "mistralai/mixtral-8x7b-instruct-v0.1",
|
# text_models
|
||||||
"gemma-2b": "google-deepmind/gemma-2b-it",
|
"gemma-2b": "google-deepmind/gemma-2b-it",
|
||||||
"llava-13b": "yorickvp/llava-13b",
|
"llava-13b": "yorickvp/llava-13b",
|
||||||
}
|
}
|
||||||
|
|
||||||
model_versions = {
|
model_versions = {
|
||||||
"meta/meta-llama-3-70b-instruct": "fbfb20b472b2f3bdd101412a9f70a0ed4fc0ced78a77ff00970ee7a2383c575d",
|
# image_models
|
||||||
"mistralai/mixtral-8x7b-instruct-v0.1": "5d78bcd7a992c4b793465bcdcf551dc2ab9668d12bb7aa714557a21c1e77041c",
|
|
||||||
"google-deepmind/gemma-2b-it": "dff94eaf770e1fc211e425a50b51baa8e4cac6c39ef074681f9e39d778773626",
|
|
||||||
"yorickvp/llava-13b": "80537f9eead1a5bfa72d5ac6ea6414379be41d4d4f6679fd776e9535d1eb58bb",
|
|
||||||
'black-forest-labs/flux-schnell': "f2ab8a5bfe79f02f0789a146cf5e73d2a4ff2684a98c2b303d1e1ff3814271db",
|
|
||||||
'stability-ai/stable-diffusion-3': "527d2a6296facb8e47ba1eaf17f142c240c19a30894f437feee9b91cc29d8e4f",
|
'stability-ai/stable-diffusion-3': "527d2a6296facb8e47ba1eaf17f142c240c19a30894f437feee9b91cc29d8e4f",
|
||||||
'bytedance/sdxl-lightning-4step': "5f24084160c9089501c1b3545d9be3c27883ae2239b6f412990e82d4a6210f8f",
|
'bytedance/sdxl-lightning-4step': "5f24084160c9089501c1b3545d9be3c27883ae2239b6f412990e82d4a6210f8f",
|
||||||
'playgroundai/playground-v2.5-1024px-aesthetic': "a45f82a1382bed5c7aeb861dac7c7d191b0fdf74d8d57c4a0e6ed7d4d0bf7d24",
|
'playgroundai/playground-v2.5-1024px-aesthetic': "a45f82a1382bed5c7aeb861dac7c7d191b0fdf74d8d57c4a0e6ed7d4d0bf7d24",
|
||||||
|
|
||||||
|
# text_models
|
||||||
|
"google-deepmind/gemma-2b-it": "dff94eaf770e1fc211e425a50b51baa8e4cac6c39ef074681f9e39d778773626",
|
||||||
|
"yorickvp/llava-13b": "80537f9eead1a5bfa72d5ac6ea6414379be41d4d4f6679fd776e9535d1eb58bb",
|
||||||
|
|
||||||
}
|
}
|
||||||
|
|
||||||
@classmethod
|
@classmethod
|
||||||
|
@ -164,12 +164,6 @@ llama_3_8b = Model(
|
|||||||
best_provider = IterListProvider([Cloudflare])
|
best_provider = IterListProvider([Cloudflare])
|
||||||
)
|
)
|
||||||
|
|
||||||
llama_3_70b = Model(
|
|
||||||
name = "llama-3-70b",
|
|
||||||
base_provider = "Meta Llama",
|
|
||||||
best_provider = IterListProvider([ReplicateHome])
|
|
||||||
)
|
|
||||||
|
|
||||||
# llama 3.1
|
# llama 3.1
|
||||||
llama_3_1_8b = Model(
|
llama_3_1_8b = Model(
|
||||||
name = "llama-3.1-8b",
|
name = "llama-3.1-8b",
|
||||||
@ -212,7 +206,7 @@ mistral_7b = Model(
|
|||||||
mixtral_8x7b = Model(
|
mixtral_8x7b = Model(
|
||||||
name = "mixtral-8x7b",
|
name = "mixtral-8x7b",
|
||||||
base_provider = "Mistral",
|
base_provider = "Mistral",
|
||||||
best_provider = IterListProvider([DDG, ReplicateHome])
|
best_provider = DDG
|
||||||
)
|
)
|
||||||
|
|
||||||
mistral_nemo = Model(
|
mistral_nemo = Model(
|
||||||
@ -279,7 +273,7 @@ gemini = Model(
|
|||||||
gemma_2b = Model(
|
gemma_2b = Model(
|
||||||
name = 'gemma-2b',
|
name = 'gemma-2b',
|
||||||
base_provider = 'Google',
|
base_provider = 'Google',
|
||||||
best_provider = IterListProvider([ReplicateHome])
|
best_provider = ReplicateHome
|
||||||
)
|
)
|
||||||
|
|
||||||
|
|
||||||
@ -583,12 +577,6 @@ flux_4o = Model(
|
|||||||
|
|
||||||
)
|
)
|
||||||
|
|
||||||
flux_schnell = Model(
|
|
||||||
name = 'flux-schnell',
|
|
||||||
base_provider = 'Flux AI',
|
|
||||||
best_provider = ReplicateHome
|
|
||||||
|
|
||||||
)
|
|
||||||
|
|
||||||
|
|
||||||
### Other ###
|
### Other ###
|
||||||
@ -635,7 +623,6 @@ class ModelUtils:
|
|||||||
|
|
||||||
# llama-3
|
# llama-3
|
||||||
'llama-3-8b': llama_3_8b,
|
'llama-3-8b': llama_3_8b,
|
||||||
'llama-3-70b': llama_3_70b,
|
|
||||||
|
|
||||||
# llama-3.1
|
# llama-3.1
|
||||||
'llama-3.1-8b': llama_3_1_8b,
|
'llama-3.1-8b': llama_3_1_8b,
|
||||||
@ -802,7 +789,6 @@ class ModelUtils:
|
|||||||
'flux-disney': flux_disney,
|
'flux-disney': flux_disney,
|
||||||
'flux-pixel': flux_pixel,
|
'flux-pixel': flux_pixel,
|
||||||
'flux-4o': flux_4o,
|
'flux-4o': flux_4o,
|
||||||
'flux-schnell': flux_schnell,
|
|
||||||
|
|
||||||
|
|
||||||
### Other ###
|
### Other ###
|
||||||
|
Loading…
Reference in New Issue
Block a user