* Add mypy check in pre-commit
* Add mypy check in pre-commit
* bug fix
* CI fix
* add python version matrix for checking types in CI
* update github actions
* revert some accidental changes
* revert some accidental changes
* revert some accidental changes
* revert some accidental changes
---------
Co-authored-by: Eric Zhu <ekzhu@users.noreply.github.com>
Co-authored-by: Chi Wang <wang.chi@microsoft.com>
* add doc about effects for capabilities
* remove unnecessary imports
* improve doc
* test
* test location
* polish
* improve import of colored
* termcolor
* termcolor
* use pull request for openai test
* Initialization of fsm blog
* Pre-commit pass
* Pinyin name correction
* Updated title
* Updated title
* Added summary of notebook and note about DAG
* Update index.mdx
* Update website/blog/2024-02-11-FSM-GroupChat/index.mdx
Co-authored-by: Chi Wang <wang.chi@microsoft.com>
* Update website/blog/2024-02-11-FSM-GroupChat/index.mdx
Co-authored-by: Chi Wang <wang.chi@microsoft.com>
* Removed outdated filename and followed SoM notebook format
* Pre-commit pass
* Update notebook link
* Corrected link to rendered notebook
* Removed agentchat_hierarchy_flow_using_select_speaker and modified agentchat_groupchat_finite_state_machine
* Add notebook/agentchat_groupchat_finite_state_machine.ipynb to paths
---------
Co-authored-by: freedeaths <register917@gmail.com>
Co-authored-by: Chi Wang <wang.chi@microsoft.com>
* implement redis cache mode, if redis_url is set in the llm_config then
it will try to use this. also adds a test to validate both the existing
and the redis cache behavior.
* PR feedback, add unit tests
* more PR feedback, move the new style cache to a context manager
* Update agent_chat.md
* more PR feedback, remove tests from contrib and have them run with the normal jobs
* doc
* updated
* Update website/docs/Use-Cases/agent_chat.md
Co-authored-by: Chi Wang <wang.chi@microsoft.com>
* update docs
* update docs; let openaiwrapper to use cache object
* typo
* Update website/docs/Use-Cases/enhanced_inference.md
Co-authored-by: Chi Wang <wang.chi@microsoft.com>
* save previous client cache and reset it after send/a_send
* a_run_chat
---------
Co-authored-by: Vijay Ramesh <vijay@regrello.com>
Co-authored-by: Eric Zhu <ekzhu@users.noreply.github.com>
Co-authored-by: Chi Wang <wang.chi@microsoft.com>
* Initial commit.
* Disable LLM response caching.
* Add teachability option to setup.py
* Modify test to use OAI_CONFIG_LIST as suggested in the docs.
* Expand unit test.
* Complete unit test.
* Add filter_dict
* details
* AnalysisAgent
* details
* More documentation and debug output.
* Support retrieval of any number of relevant memos, including zero.
* More robust analysis separator.
* cleanup
* teach_config
* refactoring
* For robustness, allow more flexibility on memo storage and retrieval.
* de-dupe the retrieved memos.
* Simplify AnalysisAgent. The unit tests now pass with gpt-3.5
* comments
* Add a verbosity level to control analyzer messages.
* refactoring
* comments
* Persist memory on disk.
* cleanup
* Use markdown to format retrieved memos.
* Use markdown in TextAnalyzerAgent
* Add another verbosity level.
* clean up logging
* notebook
* minor edits
* cleanup
* linter fixes
* Skip tests that fail to import openai
* Address reviewer feedback.
* lint
* refactoring
* Improve wording
* Improve code coverage.
* lint
* Use llm_config to control caching.
* lowercase notebook name
* Sort out the parameters passed through to ConversableAgent, and supply full docstrings for the others.
* lint
* Allow TextAnalyzerAgent to be given a different llm_config than TeachableAgent.
* documentation
* Modifications to run openai workflow.
* Test on just python 3.10.
Replace agent with agent teachable_agent as recommended.
* Test on python 3.9 instead of 3.10.
* Remove space from name -> teachableagent
---------
Co-authored-by: Li Jiang <bnujli@gmail.com>
Co-authored-by: Chi Wang <wang.chi@microsoft.com>
* update colab link
* typo
* upload file instruction
* update system message and notebooks
* update notebooks
* notebook test
* aoai api version and exclusion
* gpt-3.5-turbo
* dict check
* change model for test
* endpoints, cache_path and func description update
* model list
* gitter -> discord
* response filter
* rewrite implement based on the filter
* multi responses
* abs path
* code handling
* option to not use docker
* context
* eval_only -> raise_error
* notebook
* utils
* utils
* separate tests
* test
* test
* test
* test
* test
* test
* test
* test
* **config in test()
* test
* test
* filename
* improve max_valid_n and doc
* Update README.md
Co-authored-by: Li Jiang <lijiang1@microsoft.com>
* add support for chatgpt
* notebook
* newline at end of file
* chatgpt notebook
* ChatGPT in Azure
* doc
* math
* warning, timeout, log file name
* handle import error
* doc update; default value
* paper
* doc
* docstr
* eval_func
* add a test func in completion
* update notebook
* update math notebook
* improve notebok
* lint and handle exception
* flake8
* exception in test
* add agg_method
* NameError
* refactor
* Update flaml/integrations/oai/completion.py
Co-authored-by: Chi Wang <wang.chi@microsoft.com>
* Update flaml/integrations/oai/completion.py
Co-authored-by: Chi Wang <wang.chi@microsoft.com>
* add example
* merge files from oai_eval_test
* Revert "merge files from oai_eval_test"
This reverts commit 1e6a550f913bb94df6e9680934ccb7175d00702e.
* merge
* save results to notebook_output
* update version and cache
* update doc
* save nb cell results to file
* fix typo in model name
* code improvements
* improve docstr
* docstr
* docstr on the Returns of test
---------
Co-authored-by: Chi Wang <wang.chi@microsoft.com>
Co-authored-by: Li Jiang <lijiang1@microsoft.com>
Co-authored-by: Susan Xueqing Liu <liususan091219@users.noreply.github.com>
* add cost budget; move loc of make_dir
* support openai completion
* install pytest in workflow
* skip openai test
* test openai
* path for docs rebuild
* install datasets
* signal
* notebook
* notebook in workflow
* optional arguments and special params
* key -> k
* improve readability
* assumption
* optimize for model selection
* larger range of max_tokens
* notebook
* python package workflow
* skip on win