diff --git a/+llms/+internal/callOpenAIChatAPI.m b/+llms/+internal/callOpenAIChatAPI.m index 5259fdf..975de52 100644 --- a/+llms/+internal/callOpenAIChatAPI.m +++ b/+llms/+internal/callOpenAIChatAPI.m @@ -131,7 +131,7 @@ nvpOptions = keys(dict); if strcmp(nvp.ModelName,'gpt-4-vision-preview') - nvpOptions(ismember(nvpOptions,["MaxNumTokens","StopSequences"])) = []; + nvpOptions(ismember(nvpOptions,"StopSequences")) = []; end for opt = nvpOptions.' diff --git a/.gitignore b/.gitignore index 1286e03..e853794 100644 --- a/.gitignore +++ b/.gitignore @@ -1,3 +1,4 @@ *.env *.asv +*.mat startup.m diff --git a/examples/ExampleEmbeddings.mlx b/examples/ExampleEmbeddings.mlx new file mode 100644 index 0000000..c687ce7 Binary files /dev/null and b/examples/ExampleEmbeddings.mlx differ diff --git a/examples/ExampleGPT4Vision.mlx b/examples/ExampleGPT4Vision.mlx index 2143c12..a653dc2 100644 Binary files a/examples/ExampleGPT4Vision.mlx and b/examples/ExampleGPT4Vision.mlx differ diff --git a/openAIChat.m b/openAIChat.m index fc15fbb..54db2a0 100644 --- a/openAIChat.m +++ b/openAIChat.m @@ -209,8 +209,8 @@ % reproducible responses % % Currently, GPT-4 Turbo with vision does not support the message.name - % parameter, functions/tools, response_format parameter, stop - % sequences, and max_tokens + % parameter, functions/tools, response_format parameter, and stop + % sequences. It also has a low MaxNumTokens default, which can be overridden. arguments this (1,1) openAIChat @@ -221,11 +221,6 @@ nvp.Seed {mustBeIntegerOrEmpty(nvp.Seed)} = [] end - if nvp.MaxNumTokens ~= Inf && strcmp(this.ModelName,'gpt-4-vision-preview') - error("llms:invalidOptionForModel", ... - llms.utils.errorMessageCatalog.getMessage("llms:invalidOptionForModel", "MaxNumTokens", this.ModelName)); - end - toolChoice = convertToolChoice(this, nvp.ToolChoice); if ~isempty(nvp.ToolChoice) && strcmp(this.ModelName,'gpt-4-vision-preview') error("llms:invalidOptionForModel", ...