Skip to content

Commit

Permalink
Merge pull request #770 from airtai/dev
Browse files Browse the repository at this point in the history
Dev
  • Loading branch information
kumaranvpl authored Nov 12, 2024
2 parents 319611f + 6bffc00 commit 96789bb
Show file tree
Hide file tree
Showing 7 changed files with 32 additions and 35 deletions.
2 changes: 1 addition & 1 deletion auth_callout/package.json
Original file line number Diff line number Diff line change
Expand Up @@ -6,7 +6,7 @@
"packages/*"
],
"devDependencies": {
"tsx": "^4.19.1",
"tsx": "^4.19.2",
"typescript": "^5.6.3",
"vite": "^5.4.9",
"vite-plugin-checker": "^0.8.0"
Expand Down
2 changes: 1 addition & 1 deletion auth_callout/packages/auth-service/package.json
Original file line number Diff line number Diff line change
Expand Up @@ -8,7 +8,7 @@
"type": "module",
"main": "./lib/index.js",
"devDependencies": {
"@types/node": "^22.7.7"
"@types/node": "^22.8.7"
},
"dependencies": {
"@prisma/client": "^5.21.1",
Expand Down
32 changes: 16 additions & 16 deletions auth_callout/pnpm-lock.yaml

Some generated files are not rendered by default. Learn more about how customized files appear on GitHub.

6 changes: 3 additions & 3 deletions fastagency_studio/io/ionats.py
Original file line number Diff line number Diff line change
Expand Up @@ -234,15 +234,15 @@ def start_chat() -> Optional[list[dict[str, Any]]]: # type: ignore [return]
task = asyncio.create_task(async_start_chat()) # type: ignore
background_tasks.add(task)

def callback(t: asyncio.Task[Any]) -> None:
async def callback(t: asyncio.Task[Any]) -> None:
try:
background_tasks.discard(t)
syncify(iostream.subscriber.close)()
await iostream.subscriber.close()
except Exception as e:
logger.error(f"Error in callback: {e}")
logger.error(traceback.format_exc())

task.add_done_callback(callback)
task.add_done_callback(lambda t: asyncio.create_task(callback(t)))

except Exception as e:
logger.error(f"Error in handling initiate chat: {e}")
Expand Down
19 changes: 8 additions & 11 deletions fastagency_studio/models/llms/together.py
Original file line number Diff line number Diff line change
Expand Up @@ -12,47 +12,44 @@
"TogetherAI",
]

# retrieve the models from the API on Oct 24, 2024
# retrieve the models from the API on Nov 11, 2024
together_model_string = {
"Code Llama Instruct (34B)": "togethercomputer/CodeLlama-34b-Instruct",
"Upstage SOLAR Instruct v1 (11B)": "upstage/SOLAR-10.7B-Instruct-v1.0",
"Meta Llama 3.1 70B Instruct Turbo": "meta-llama/Meta-Llama-3.1-70B-Instruct-Turbo",
"Llama3 8B Chat HF INT4": "togethercomputer/Llama-3-8b-chat-hf-int4",
"Gemma Instruct (2B)": "google/gemma-2b-it",
"MythoMax-L2 (13B)": "Gryphe/MythoMax-L2-13b",
"Mistral (7B) Instruct": "mistralai/Mistral-7B-Instruct-v0.1",
"Mistral (7B) Instruct v0.2": "mistralai/Mistral-7B-Instruct-v0.2",
"Meta Llama 3.1 405B Instruct Turbo": "meta-llama/Meta-Llama-3.1-405B-Instruct-Lite-Pro",
"DeepSeek LLM Chat (67B)": "deepseek-ai/deepseek-llm-67b-chat",
"Qwen 2 Instruct (72B)": "Qwen/Qwen2-72B-Instruct",
"Togethercomputer Llama3 8B Instruct Int8": "togethercomputer/Llama-3-8b-chat-hf-int8",
"Mistral (7B) Instruct v0.3": "mistralai/Mistral-7B-Instruct-v0.3",
"LLaMA-2 Chat (13B)": "togethercomputer/llama-2-13b-chat",
"Meta Llama 3 70B Instruct Turbo": "meta-llama/Meta-Llama-3-70B-Instruct-Turbo",
"Meta Llama 3 70B Instruct Lite": "meta-llama/Meta-Llama-3-70B-Instruct-Lite",
"Meta Llama 3.2 90B Vision Instruct Turbo": "meta-llama/Llama-3.2-90B-Vision-Instruct-Turbo",
"Gemma-2 Instruct (27B)": "google/gemma-2-27b-it",
"Meta Llama 3 8B Instruct Reference": "meta-llama/Llama-3-8b-chat-hf",
"Qwen2.5 72B Instruct Turbo": "Qwen/Qwen2.5-72B-Instruct-Turbo",
"Mixtral-8x7B Instruct v0.1": "mistralai/Mixtral-8x7B-Instruct-v0.1",
"WizardLM-2 (8x22B)": "microsoft/WizardLM-2-8x22B",
"Meta Llama 3.2 90B Vision Instruct Turbo": "meta-llama/Llama-3.2-90B-Vision-Instruct-Turbo",
"Meta Llama 3.2 11B Vision Instruct Turbo": "meta-llama/Llama-3.2-11B-Vision-Instruct-Turbo",
"LLaVa-Next (Mistral-7B)": "llava-hf/llava-v1.6-mistral-7b-hf",
"Meta Llama 3.1 405B Instruct Turbo": "meta-llama/Meta-Llama-3.1-405B-Instruct-Lite-Pro",
"DBRX Instruct": "databricks/dbrx-instruct",
"Nous Hermes 2 - Mixtral 8x7B-DPO ": "NousResearch/Nous-Hermes-2-Mixtral-8x7B-DPO",
"Meta Llama 3 70B Instruct Reference": "meta-llama/Llama-3-70b-chat-hf",
"Qwen2.5 7B Instruct Turbo": "Qwen/Qwen2.5-7B-Instruct-Turbo",
"Meta Llama 3.2 3B Instruct Turbo": "meta-llama/Llama-3.2-3B-Instruct-Turbo",
"Qwen2.5 72B Instruct Turbo": "Qwen/Qwen2.5-72B-Instruct-Turbo",
"Meta Llama 3 8B Instruct Turbo": "meta-llama/Meta-Llama-3-8B-Instruct-Turbo",
"Meta Llama 3 8B Instruct Lite": "meta-llama/Meta-Llama-3-8B-Instruct-Lite",
"Meta Llama 3.1 8B Instruct Turbo": "meta-llama/Meta-Llama-3.1-8B-Instruct-Turbo",
"Qwen 2 Instruct (72B)": "Qwen/Qwen2-72B-Instruct",
"WizardLM-2 (8x22B)": "microsoft/WizardLM-2-8x22B",
"Mixtral-8x22B Instruct v0.1": "mistralai/Mixtral-8x22B-Instruct-v0.1",
"Gryphe MythoMax L2 Lite (13B)": "Gryphe/MythoMax-L2-13b-Lite",
"Qwen 1.5 Chat (110B)": "Qwen/Qwen1.5-110B-Chat",
"Meta Llama 3.1 70B Instruct Turbo": "albert/meta-llama-3-1-70b-instruct-turbo",
"LLaMA-2 Chat (7B)": "togethercomputer/llama-2-7b-chat",
"Qwen 1.5 Chat (72B)": "Qwen/Qwen1.5-72B-Chat",
"Meta Llama Vision Free": "meta-llama/Llama-Vision-Free",
"Gemma-2 Instruct (9B)": "google/gemma-2-9b-it",
"Mixtral-8x22B Instruct v0.1": "mistralai/Mixtral-8x22B-Instruct-v0.1",
"WizardLM v1.2 (13B)": "WizardLM/WizardLM-13B-V1.2",
"Koala (7B)": "togethercomputer/Koala-7B",
"Qwen 2 Instruct (1.5B)": "Qwen/Qwen2-1.5B-Instruct",
Expand Down
4 changes: 2 additions & 2 deletions pyproject.toml
Original file line number Diff line number Diff line change
Expand Up @@ -47,8 +47,8 @@ dependencies = [
"fastagency @ git+https://github.com/airtai/fastagency.git@main",
"faststream[nats]>=0.5.10,<0.6",
"pydantic>=2.3,<3",
"fastapi==0.115.2",
"prisma>=0.13.1,<0.15",
"fastapi==0.115.4",
"prisma>=0.13.1,<0.16",
"asyncer==0.0.8",
"markdownify==0.13.1", # Needed by autogen.WebSurferAgent but not included
"httpx==0.27.2",
Expand Down
2 changes: 1 addition & 1 deletion tests/app/test_openai_extensively.py
Original file line number Diff line number Diff line change
Expand Up @@ -336,7 +336,7 @@ async def test_validate_toolbox_with_invalid_openapi_spec(self) -> None:

@pytest.mark.asyncio
async def test_validate_toolbox_with_yaml_openapi_spec(self) -> None:
invalid_url = "https://raw.githubusercontent.com/OAI/OpenAPI-Specification/main/examples/v3.0/petstore.yaml"
invalid_url = "https://raw.githubusercontent.com/OAI/learn.openapis.org/refs/heads/main/examples/v3.0/petstore.yaml"

openapi_auth = OpenAPIAuth(
name="openapi_auth_secret",
Expand Down

0 comments on commit 96789bb

Please sign in to comment.