Files
AMD-SHARK-Studio/apps/language_models
Abhishek Varma 310d5d0a49 Fix llama2 13b crashing + add spec file for CLI execution of Llama (#1797)
* [Llama2] Add a fix for Llama2 13B downloading/crashing

-- This commit fixes downloading/crashing of llama2 13B on wrong
   .mlir file.
-- Also adds support for downloading vmfb from shark_tank in CLI.

Signed-off-by: Abhishek Varma <abhishek@nod-labs.com>

* [llama2] Add a spec file to run Llama/Vicuna CLI exe

-- This commit adds a spec file to run Llama/Vicuna CLI exe.

Signed-off-by: Abhishek Varma <abhishek@nod-labs.com>

---------

Signed-off-by: Abhishek Varma <abhishek@nod-labs.com>
2023-08-25 09:36:09 -05:00
..
2023-07-19 23:10:23 +05:30

CodeGen Setup using SHARK-server

Setup Server

  • clone SHARK and setup the venv
  • host the server using python apps/stable_diffusion/web/index.py --api --server_port=<PORT>
  • default server address is http://0.0.0.0:8080

Setup Client

  1. fauxpilot-vscode (VSCode Extension):
  • Code for the extension can be found here
  • PreReq: VSCode extension (will need nodejs and npm to compile and run the extension)
  • Compile and Run the extension on VSCode (press F5 on VSCode), this opens a new VSCode window with the extension running
  • Open VSCode settings, search for fauxpilot in settings and modify server : http://<IP>:<PORT>, Model : codegen , Max Lines : 30
  1. Others (REST API curl, OpenAI Python bindings) as shown here
  • using Github Copilot VSCode extension with SHARK-server needs more work to be functional.