- Refactored language_prompt to handle Chinese characters
- Updated model_token_limit_error message formatting
- Modified function signature for chat_completion_stream
- Updated commit/command.yml to use "en" language code
- Updated commit/commit.py to handle model's maximum context length error
- Updated commit/zh/command.yml to use "zh" language code
- Updated libs/llm_api/openai.py to ignore tokens limit error
Closesdevchat-ai/devchat#189
- Simplify and clarify user prompts and output messages
- Remove unnecessary print statements and text
- Handle commit abort and errors more gracefully
- Improve error logging in OpenAI chat stream handler
- Update the signature of the chat_completion_stream function
- Remove unnecessary line breaks and indents
- Improve code readability and maintainability
- Reformatted multiline strings for clarity in commit.py
- Applied consistent quotation marks for string literals in commit.py
- Introduced newlines between rpc_call decorator functions in services.py
- Add stream_out parameter to chat_completion_no_stream function
- Print delta content when stream_out is True
- Ensure seamless streaming of chat completion response
- Added new LLM API functions for non-streaming chat completions
- Introduced UI utility functions for input/output interaction
- Implemented multi-select and single-select UI components