From 0ed2d2c7d296da0dceef04be6ef02994e6eb992a Mon Sep 17 00:00:00 2001 From: hex Date: Mon, 16 Dec 2024 10:04:08 +0200 Subject: [PATCH] Add llama-3.3-70b-instruct support --- llm_perplexity.py | 1 + pyproject.toml | 2 +- 2 files changed, 2 insertions(+), 1 deletion(-) diff --git a/llm_perplexity.py b/llm_perplexity.py index 93b7e21..f1c2ffb 100644 --- a/llm_perplexity.py +++ b/llm_perplexity.py @@ -14,6 +14,7 @@ def register_models(register): register(Perplexity("llama-3.1-sonar-large-128k-chat"), aliases=("sonar-large-chat",)) register(Perplexity("llama-3.1-70b-instruct")) register(Perplexity("llama-3.1-8b-instruct")) + register(Perplexity("llama-3.3-70b-instruct")) class PerplexityOptions(llm.Options): diff --git a/pyproject.toml b/pyproject.toml index 7c51be3..a9aab74 100644 --- a/pyproject.toml +++ b/pyproject.toml @@ -1,6 +1,6 @@ [project] name = "llm-perplexity" -version = "2024.11.0" +version = "2024.12.0" description = "LLM access to pplx-api 3 by Perplexity Labs" readme = "README.md" authors = [{name = "hex"}]