From 58c0f39714b813b031674e8b6e8fcf63f7e70a7b Mon Sep 17 00:00:00 2001 From: Peter Tripp Date: Fri, 30 Aug 2024 18:57:50 +0000 Subject: [PATCH] OpenAI: Fix GPT-4. Only include max_tokens when max_output_tokens provided (#17168) - Fixed GPT-4 breakage (incorrect `max_output_tokens` handling). --- crates/open_ai/src/open_ai.rs | 6 +----- 1 file changed, 1 insertion(+), 5 deletions(-) diff --git a/crates/open_ai/src/open_ai.rs b/crates/open_ai/src/open_ai.rs index ecb0828ee6..6be5327c04 100644 --- a/crates/open_ai/src/open_ai.rs +++ b/crates/open_ai/src/open_ai.rs @@ -120,14 +120,10 @@ impl Model { pub fn max_output_tokens(&self) -> Option { match self { - Self::ThreePointFiveTurbo => Some(4096), - Self::Four => Some(8192), - Self::FourTurbo => Some(4096), - Self::FourOmni => Some(4096), - Self::FourOmniMini => Some(16384), Self::Custom { max_output_tokens, .. } => *max_output_tokens, + _ => None, } } }