From 906eb70ca15b7cea2470e9571ad7113416ddccfe Mon Sep 17 00:00:00 2001 From: Timothy Carambat Date: Thu, 12 Sep 2024 13:13:47 -0700 Subject: [PATCH] bump Perplexity models (#2275) --- server/utils/AiProviders/perplexity/models.js | 55 ++++--------------- .../perplexity/scripts/chat_models.txt | 13 +---- .../AiProviders/perplexity/scripts/parse.mjs | 2 +- 3 files changed, 14 insertions(+), 56 deletions(-) diff --git a/server/utils/AiProviders/perplexity/models.js b/server/utils/AiProviders/perplexity/models.js index 64bd2cffae..5a71aac187 100644 --- a/server/utils/AiProviders/perplexity/models.js +++ b/server/utils/AiProviders/perplexity/models.js @@ -1,59 +1,29 @@ const MODELS = { - "llama-3-sonar-small-32k-online": { - id: "llama-3-sonar-small-32k-online", - name: "llama-3-sonar-small-32k-online", - maxLength: 28000, - }, - "llama-3-sonar-small-32k-chat": { - id: "llama-3-sonar-small-32k-chat", - name: "llama-3-sonar-small-32k-chat", - maxLength: 32768, - }, - "llama-3-sonar-large-32k-online": { - id: "llama-3-sonar-large-32k-online", - name: "llama-3-sonar-large-32k-online", - maxLength: 28000, - }, - "llama-3-sonar-large-32k-chat": { - id: "llama-3-sonar-large-32k-chat", - name: "llama-3-sonar-large-32k-chat", - maxLength: 32768, - }, "llama-3.1-sonar-small-128k-online": { id: "llama-3.1-sonar-small-128k-online", name: "llama-3.1-sonar-small-128k-online", maxLength: 127072, }, - "llama-3.1-sonar-small-128k-chat": { - id: "llama-3.1-sonar-small-128k-chat", - name: "llama-3.1-sonar-small-128k-chat", - maxLength: 131072, - }, "llama-3.1-sonar-large-128k-online": { id: "llama-3.1-sonar-large-128k-online", name: "llama-3.1-sonar-large-128k-online", maxLength: 127072, }, + "llama-3.1-sonar-huge-128k-online": { + id: "llama-3.1-sonar-huge-128k-online", + name: "llama-3.1-sonar-huge-128k-online", + maxLength: 127072, + }, + "llama-3.1-sonar-small-128k-chat": { + id: "llama-3.1-sonar-small-128k-chat", + name: "llama-3.1-sonar-small-128k-chat", + maxLength: 131072, + }, "llama-3.1-sonar-large-128k-chat": { id: "llama-3.1-sonar-large-128k-chat", name: "llama-3.1-sonar-large-128k-chat", maxLength: 131072, }, - "llama-3-8b-instruct": { - id: "llama-3-8b-instruct", - name: "llama-3-8b-instruct", - maxLength: 8192, - }, - "llama-3-70b-instruct": { - id: "llama-3-70b-instruct", - name: "llama-3-70b-instruct", - maxLength: 8192, - }, - "mixtral-8x7b-instruct": { - id: "mixtral-8x7b-instruct", - name: "mixtral-8x7b-instruct", - maxLength: 16384, - }, "llama-3.1-8b-instruct": { id: "llama-3.1-8b-instruct", name: "llama-3.1-8b-instruct", @@ -64,11 +34,6 @@ const MODELS = { name: "llama-3.1-70b-instruct", maxLength: 131072, }, - "llama-3.1-sonar-huge-128k-chat": { - id: "llama-3.1-sonar-huge-128k-chat", - name: "llama-3.1-sonar-huge-128k-chat", - maxLength: 127072, - }, }; module.exports.MODELS = MODELS; diff --git a/server/utils/AiProviders/perplexity/scripts/chat_models.txt b/server/utils/AiProviders/perplexity/scripts/chat_models.txt index ec9081efe9..fc3ab5b6f5 100644 --- a/server/utils/AiProviders/perplexity/scripts/chat_models.txt +++ b/server/utils/AiProviders/perplexity/scripts/chat_models.txt @@ -1,16 +1,9 @@ | Model | Parameter Count | Context Length | Model Type | | :---------------------------------- | :-------------- | :------------- | :-------------- | -| `llama-3-sonar-small-32k-online` | 8B | 28,000 | Chat Completion | -| `llama-3-sonar-small-32k-chat` | 8B | 32,768 | Chat Completion | -| `llama-3-sonar-large-32k-online` | 70B | 28,000 | Chat Completion | -| `llama-3-sonar-large-32k-chat` | 70B | 32,768 | Chat Completion | | `llama-3.1-sonar-small-128k-online` | 8B | 127,072 | Chat Completion | -| `llama-3.1-sonar-small-128k-chat` | 8B | 131,072 | Chat Completion | | `llama-3.1-sonar-large-128k-online` | 70B | 127,072 | Chat Completion | +| `llama-3.1-sonar-huge-128k-online` | 405B | 127,072 | Chat Completion | +| `llama-3.1-sonar-small-128k-chat` | 8B | 131,072 | Chat Completion | | `llama-3.1-sonar-large-128k-chat` | 70B | 131,072 | Chat Completion | -| `llama-3-8b-instruct` | 8B | 8,192 | Chat Completion | -| `llama-3-70b-instruct` | 70B | 8,192 | Chat Completion | -| `mixtral-8x7b-instruct` | 8x7B | 16,384 | Chat Completion | | `llama-3.1-8b-instruct` | 8B | 131,072 | Chat Completion | -| `llama-3.1-70b-instruct` | 70B | 131,072 | Chat Completion | -| `llama-3.1-sonar-huge-128k-chat` | 405B | 127,072 | Chat Completion | \ No newline at end of file +| `llama-3.1-70b-instruct` | 70B | 131,072 | Chat Completion | \ No newline at end of file diff --git a/server/utils/AiProviders/perplexity/scripts/parse.mjs b/server/utils/AiProviders/perplexity/scripts/parse.mjs index e4f27276f0..3cb1158834 100644 --- a/server/utils/AiProviders/perplexity/scripts/parse.mjs +++ b/server/utils/AiProviders/perplexity/scripts/parse.mjs @@ -8,7 +8,7 @@ // copy outputs into the export in ../models.js // Update the date below if you run this again because Perplexity added new models. -// Last Collected: Jul 31, 2024 +// Last Collected: Sept 12, 2024 import fs from "fs";