-
Notifications
You must be signed in to change notification settings - Fork 1.3k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Gibberish or blank completions in both swagger interface and vscode #1277
Labels
Comments
I tried looking into this problem, but I can't reproduce it in my environment. I can get the response properly with the Tabby/CodeLlama-13B model, both 0.7.0 and 0.6.0 tested. Completion request:
Completion response:
Health check response
output of nvidia-smi
|
I compiled it from source today and it is working well. Thanks. And please merge #1286 ASAP. A single line of code killed my whole day :D |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Describe the bug
I'm trying to run tabby via following script:
docker run -it --gpus all -p 8080:8080 -v $HOME/.tabby:/data tabbyml/tabby serve --model TabbyML/CodeLlama-13B --device cuda
I have 4 x 3090 machine. Even if I limit tabby to one GPU, I'm getting following:
Request:
Response (200):
Events log:
{"ts":1706002279828,"event":{"completion":{"completion_id":"cmpl-f79069d0-fa5f-41b4-aa06-50eb7015409f","language":"python","prompt":"<PRE> def fib(n):\n <SUF>\n return fib(n - 1) + fib(n - 2) <MID>","segments":{"prefix":"def fib(n):\n ","suffix":"\n return fib(n - 1) + fib(n - 2)","clipboard":null},"choices":[{"index":0,"text":"fte▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅▅"}],"user":null}}}
Information about your version
Please provide output of
tabby --version
tabby 0.7.0
Information about your GPU
Please provide output of
nvidia-smi
I have also tried with indexing a github repository
here is completion event log after:
{"ts":1706005076127,"event":{"completion":{"completion_id":"cmpl-6b807cf1-422d-4428-a2e1-aae93ea89c75","language":"python","prompt":"<PRE> # Path: samples/apps/autogen-studio/autogenstudio/chatmanager.py\n# print(\"Modified files: \", len(modified_files))\n#\n# Path: samples/apps/autogen-studio/autogenstudio/chatmanager.py\n# Message(\n# user_id=message.user_id,\n# root_msg_id=message.root_msg_id,\n# role=\"assistant\",\n# content=output,\n# metadata=json.dumps(metadata),\n# session_id=message.session_id,\n# )\n#\n# Path: samples/apps/autogen-studio/autogenstudio/utils/dbutils.py\n# sqlite3.connect(self.path, check_same_thread=False, **kwargs)\n#\n# Path: samples/apps/autogen-studio/autogenstudio/utils/dbutils.py\n# def reset_db(self):\n# \"\"\"\n# Reset the database by deleting the database file and creating a new one.\n# \"\"\"\n# print(\"resetting db\")\n# if os.path.exists(self.path):\n# os.remove(self.path)\n# self.init_db(path=self.path)\n teachability.add_to_agent(teachable_agent)\n\n return teachable_agent\n\n\ndef interact_freely_with_user():\n \"\"\"Starts a free-form chat between the user and a teachable agent.\"\"\"\n\n # Create the agents.\n print(colored(\"\\nLoading previous memory (if any) from disk.\", \"light_cyan\"))\n teachable_agent = create_teachable_agent(reset_db=False)\n user = UserProxyAgent(\"user\", human_input_mode=\"ALWAYS\")\n\n # Start the chat.\n teachable_agent.initiate_chat(user, message=\"Greetings, I'm a teachable user assistant! What's on your mind today?\")\n\n\nif __name__ == \"__main__\":\n \"\"\"Lets the user test a teachable agent interactively.\"\"\"\n <SUF>\n <MID>","segments":{"prefix":" teachability.add_to_agent(teachable_agent)\n\n return teachable_agent\n\n\ndef interact_freely_with_user():\n \"\"\"Starts a free-form chat between the user and a teachable agent.\"\"\"\n\n # Create the agents.\n print(colored(\"\\nLoading previous memory (if any) from disk.\", \"light_cyan\"))\n teachable_agent = create_teachable_agent(reset_db=False)\n user = UserProxyAgent(\"user\", human_input_mode=\"ALWAYS\")\n\n # Start the chat.\n teachable_agent.initiate_chat(user, message=\"Greetings, I'm a teachable user assistant! What's on your mind today?\")\n\n\nif __name__ == \"__main__\":\n \"\"\"Lets the user test a teachable agent interactively.\"\"\"\n ","suffix":"","clipboard":null},"choices":[{"index":0,"text":"#ogormormormormormormormormormormormormormormormormormormormormormormormormormormormormormormormormormormormormormormormormormormormormormormormormormormormormormormormormormormormormormormormormormormormormormormormormormormormormormormormormormormscore agrprimeprimeprimeprimeprimeprimeprimeprimeprimeprimeprimeprimeprimeprimeprimeprimeprimeprimeprimeprimeprimeprimeprimeprimeprimeprimeprimeprimeprimeprimeprimeprimeprimeprimeprimeprimeprimeprimeprimeprimeprimeprime"}],"user":null}}}
The text was updated successfully, but these errors were encountered: