Files
AMD-SHARK-Studio/apps/language_models
Jakub Kuderski 2da31c4109 [vicuna.py] Rework benchmark statistics calculation (#1992)
- Move statistics out of the main loop
- Add 'end-to-end' numbers
- Switch the main display unit from s to ms
- Start measuring time at 0

The new print format looks like this:
```
Number of iterations: 5
Num tokens: 1 (prompt), 512 (generated), 513 (total)
Prefill: avg. 0.01 ms (stdev 0.00), avg. 97.99 tokens/s
Decode: avg. 4840.44 ms (stdev 28.80), avg. 97.99 tokens/s
Decode end-2-end: avg. 85.78 tokens/s (w/o prompt), avg. 95.98 (w/ prompt)
```
2023-11-23 12:04:03 -05:00
..
2023-11-15 10:04:07 -08:00
2023-07-19 23:10:23 +05:30

CodeGen Setup using SHARK-server

Setup Server

  • clone SHARK and setup the venv
  • host the server using python apps/stable_diffusion/web/index.py --api --server_port=<PORT>
  • default server address is http://0.0.0.0:8080

Setup Client

  1. fauxpilot-vscode (VSCode Extension):
  • Code for the extension can be found here
  • PreReq: VSCode extension (will need nodejs and npm to compile and run the extension)
  • Compile and Run the extension on VSCode (press F5 on VSCode), this opens a new VSCode window with the extension running
  • Open VSCode settings, search for fauxpilot in settings and modify server : http://<IP>:<PORT>, Model : codegen , Max Lines : 30
  1. Others (REST API curl, OpenAI Python bindings) as shown here
  • using Github Copilot VSCode extension with SHARK-server needs more work to be functional.