Commit Graph

  • 4a44620e9e better handling if ICL examples are missing Alex O'Connell 2024-04-08 21:13:10 -04:00
  • 1577950137 more tests, fix missing default options, and load ICL as utf8 Alex O'Connell 2024-04-08 20:58:07 -04:00
  • c436498c2e write some tests Alex O'Connell 2024-04-08 00:36:27 -04:00
  • 9c3a3db696 more random settings Alex O'Connell 2024-04-07 22:27:13 -04:00
  • 6a89045796 don't filter out all ICL examples Alex O'Connell 2024-04-07 17:07:17 -04:00
  • 7609894136 hook up temp, top_k, and context length for all backends that support it + warn for exceeding the context length Alex O'Connell 2024-04-07 14:38:11 -04:00
  • 670e6d8625 type checking + comments Alex O'Connell 2024-04-07 12:28:17 -04:00
  • 27044a8fae refactor agents to separate file Alex O'Connell 2024-04-07 12:07:28 -04:00
  • 3eaa89c683 skip the step if the package is already installed + prioritize local file Alex O'Connell 2024-04-06 23:14:14 -04:00
  • 5def7669f0 add recommended models Alex O'Connell 2024-04-06 22:55:52 -04:00
  • d3817a4f8f Merge branch 'main' into develop Alex O'Connell 2024-04-06 20:17:57 -04:00
  • da0591723f Merge pull request #102 from acon96/release/v0.2.11 v0.2.11 Alex O'Connell 2024-04-07 00:17:17 +00:00
  • 5afc5014eb Release v0.2.11 Alex O'Connell 2024-04-06 20:09:49 -04:00
  • 3b2377b5b3 update setup docs for new wheel install Alex O'Connell 2024-04-06 18:44:32 -04:00
  • d5e69e09fd clean up todo Alex O'Connell 2024-04-06 18:32:37 -04:00
  • bbe4f61da8 mark todo done Alex O'Connell 2024-04-06 18:25:36 -04:00
  • 793b36f215 Merge branch 'feature/prime-kv-cache' into develop Alex O'Connell 2024-04-06 18:24:18 -04:00
  • 6d1ffaca6d deprecate addon Alex O'Connell 2024-04-06 18:12:22 -04:00
  • f226dda4fc Use github actions to build wheels Alex O'Connell 2024-04-06 18:02:42 -04:00
  • 3574887c33 fix startup errors Alex O'Connell 2024-04-01 23:38:35 -04:00
  • 4058a42ee5 clean up UI + hook up other llama.cpp settings Alex O'Connell 2024-03-31 17:08:52 -04:00
  • fdd9f1bc67 wire up llama cpp runtime options Alex O'Connell 2024-03-31 11:08:07 -04:00
  • 0d559b63be add cooldown logic Alex O'Connell 2024-03-31 10:55:33 -04:00
  • 47417aaee8 fix on update Alex O'Connell 2024-03-30 23:58:35 -04:00
  • 3ee8135064 better prompt caching Alex O'Connell 2024-03-30 23:04:52 -04:00
  • 9c2892532f track cache refreshes Alex O'Connell 2024-03-30 16:54:31 -04:00
  • 0008f6bb5e hook up prompt caching to config flow Alex O'Connell 2024-03-30 16:50:24 -04:00
  • aa3f82d56f add homeassistant to requirements Alex O'Connell 2024-03-30 14:50:43 -04:00
  • 10b6862583 working prompt caching prototype Alex O'Connell 2024-03-30 12:56:28 -04:00
  • 4fea8e6908 try priming kv cache Alex O'Connell 2024-03-25 18:29:17 -04:00
  • 19ae895f96 quickstart guide typos Alex O'Connell 2024-03-24 11:27:26 -04:00
  • ed622e7bbd Merge commit 'd9d42b50cd49e786649f1f00d239e8dce719807f' into develop Alex O'Connell 2024-03-24 11:18:19 -04:00
  • d9d42b50cd update manifest version v0.2.10 Alex O'Connell 2024-03-24 11:17:45 -04:00
  • 0244e79003 Merge pull request #98 from acon96/release/v0.2.10 Alex O'Connell 2024-03-24 15:15:06 +00:00
  • 1c2cbc63d6 Release v0.2.10 Alex O'Connell 2024-03-24 11:07:10 -04:00
  • bf04cc3e6e remove default host since it didn't make sense Alex O'Connell 2024-03-24 11:02:49 -04:00
  • 46e1c4fc1d reset model_config when starting the configflow Alex O'Connell 2024-03-24 10:56:50 -04:00
  • b555e5a502 fix typo in validate Alex O'Connell 2024-03-24 07:25:46 -04:00
  • ac7b71ca4f Make initial configuration easier + rewrite quickstart guide Alex O'Connell 2024-03-24 00:04:32 -04:00
  • 0a6b41d5bc support many checkpoints in evaluate Alex O'Connell 2024-03-23 10:09:42 -04:00
  • d51c172c07 Small fixes to brightness adjustment and EOS token removal (#95) Isabella Nightshade 2024-03-22 23:29:34 +11:00
  • 0e618a50a2 properly class variables Alex O'Connell 2024-03-21 20:16:27 -04:00
  • 89042302a7 Merge branch 'main' into develop Alex O'Connell 2024-03-20 23:27:14 -04:00
  • 5f6e3dd4f6 fix component metadata Alex O'Connell 2024-03-20 23:27:07 -04:00
  • 1ab0d82045 Merge pull request #94 from acon96/release/v0.2.9 v0.2.9 Alex O'Connell 2024-03-21 03:24:08 +00:00
  • 7637c758a5 more docs updates + code cleanup Alex O'Connell 2024-03-20 23:22:07 -04:00
  • 93dacffa20 update todo list Alex O'Connell 2024-03-20 23:16:55 -04:00
  • a2eb8d9b99 changelog update Alex O'Connell 2024-03-20 23:13:24 -04:00
  • 7bd072623c add updated llama-cpp-python x64 wheels Alex O'Connell 2024-03-20 23:07:24 -04:00
  • c67759e16f experiment notes Alex O'Connell 2024-03-20 23:05:22 -04:00
  • fa31682c51 working version of in context examples Alex O'Connell 2024-03-20 23:03:31 -04:00
  • 0f1c773bff Fix model download + llama-cpp-python install steps Alex O'Connell 2024-03-20 21:41:04 -04:00
  • f1659893d7 start working on dpo for the datasets Alex O'Connell 2024-03-19 21:31:34 -04:00
  • b9d394f860 actually fix cover service names Alex O'Connell 2024-03-16 13:50:04 -04:00
  • 4978901412 start working on icl examples Alex O'Connell 2024-03-07 18:14:44 -05:00
  • 9b71a1860b fix "cover" type system call names Alex O'Connell 2024-03-06 20:40:25 -05:00
  • 4f6ed08be9 split out service call argument allow list + properly parse rgb color arguments Alex O'Connell 2024-03-06 17:57:41 -05:00
  • 41b7ceae57 Add Ollama options to docs Alex O'Connell 2024-03-05 21:52:37 -05:00
  • 263b21151f Merge branch 'main' into develop Alex O'Connell 2024-03-05 21:47:38 -05:00
  • 316459b6cd Merge pull request #85 from acon96/release/v0.2.8 v0.2.8 Alex O'Connell 2024-03-05 21:47:02 -05:00
  • 841beb5e77 Release v0.2.8 Alex O'Connell 2024-03-05 21:46:25 -05:00
  • 569f6e848a properly handle colons in ollama model names Alex O'Connell 2024-03-05 17:39:41 -05:00
  • 5a51c8eb7b Merge branch 'main' into develop Alex O'Connell 2024-03-04 22:58:50 -05:00
  • 474901a367 Merge pull request #82 from acon96/release/v0.2.7 v0.2.7 Alex O'Connell 2024-03-04 22:57:41 -05:00
  • 0f3d518826 Bump version number + change notes Alex O'Connell 2024-03-04 22:56:48 -05:00
  • 1c5414b8af fix ollama keep alive properly + check if model exists for ollama too Alex O'Connell 2024-03-04 22:40:22 -05:00
  • c13d706879 bump text-generation-webui version Alex O'Connell 2024-03-04 20:56:04 -05:00
  • 6cc3f47096 fix deprecated configflow behavior Alex O'Connell 2024-03-02 22:20:55 -05:00
  • 7b0b021b59 handle voice assistant aliases as duplicate devices Alex O'Connell 2024-03-01 22:59:38 -05:00
  • b50919340b Ollama keep_alive + better docs Alex O'Connell 2024-03-01 22:05:57 -05:00
  • b197632b3e Update llama-cpp-python and text-generation-webui Alex O'Connell 2024-02-25 17:41:00 -05:00
  • 869b91b6ab add model references in the setup guide Alex O'Connell 2024-02-22 21:30:50 -05:00
  • 32338242f2 Merge pull request #68 from acon96/feature/dataset-customization Alex O'Connell 2024-02-22 21:24:27 -05:00
  • fc171ac04d Merge branch 'develop' into feature/dataset-customization Alex O'Connell 2024-02-22 21:20:59 -05:00
  • d22ef571b8 update readme Alex O'Connell 2024-02-22 21:20:20 -05:00
  • 5dec71eae2 finalize new model version Alex O'Connell 2024-02-22 21:12:38 -05:00
  • 4910a34f11 Create issue templates Alex O'Connell 2024-02-18 11:49:22 -05:00
  • c285e3c6a9 instructions for adding personas Alex O'Connell 2024-02-17 23:13:45 -05:00
  • a78e57031c cleanup + table of contents Alex O'Connell 2024-02-17 23:06:21 -05:00
  • b2836bc250 tweak readme structure Alex O'Connell 2024-02-17 23:05:53 -05:00
  • ccf2c2c293 docs: Format README for better usability (#50) Zeno Jiricek 2024-02-18 14:34:46 +10:30
  • 19ed596614 resolve conflicts Alex O'Connell 2024-02-17 23:04:19 -05:00
  • 43510d1d7c finish up dataset changes for new device types Alex O'Connell 2024-02-17 21:16:56 -05:00
  • 984cf2c0a3 add vacuum, todo, and timer device names Alex O'Connell 2024-02-17 19:55:19 -05:00
  • 6cc3b13c2a Merge branch 'develop' into feature/dataset-customization Alex O'Connell 2024-02-17 19:55:07 -05:00
  • e192a8aee6 Add additional data for new entity types and services (#67) colino17 2024-02-17 20:49:58 -04:00
  • 3f3fb21cd9 Handle python 3.12 upgrade smoother + make utils file Alex O'Connell 2024-02-16 23:44:48 -05:00
  • bcd67aef37 start working on new entities Alex O'Connell 2024-02-16 23:21:22 -05:00
  • 9c6d64f373 add note to readme about wheels Alex O'Connell 2024-02-16 18:30:17 -05:00
  • 5edb66c7f2 build python 3.12 wheels too Alex O'Connell 2024-02-16 18:24:33 -05:00
  • fdfea02e1d typo Alex O'Connell 2024-02-15 00:04:45 -05:00
  • 5660666bcc Bump included llama-cpp-python to 0.2.42 Alex O'Connell 2024-02-14 08:42:40 -05:00
  • 4e873a873c Merge branch 'develop' into feature/dataset-customization Alex O'Connell 2024-02-13 20:23:25 -05:00
  • 411276408b train new models based on stablelm + properly add new response types Alex O'Connell 2024-02-13 20:21:51 -05:00
  • 73c2934e20 Update text-generation-webui addon version Alex O'Connell 2024-02-11 16:42:33 -05:00
  • cc1b0740f3 Fix Model Prompting.md (#60) jds11111 2024-02-10 16:57:03 -06:00
  • 1da51c9fb6 remove broken training tweaks Alex O'Connell 2024-02-10 10:04:58 -05:00
  • e0fa7805e4 Merge branch 'main' into develop Alex O'Connell 2024-02-08 20:49:31 -05:00
  • fd9dc2e23a Merge pull request #58 from acon96/release/v0.2.6 v0.2.6 Alex O'Connell 2024-02-08 20:48:47 -05:00
  • 3f22bf3ace Release v0.2.6 Alex O'Connell 2024-02-08 20:47:05 -05:00