Commit graph

44 commits

Author SHA1 Message Date
Tal
dcb7b66fd7
Update pr_agent/algo/ai_handlers/litellm_ai_handler.py
Co-authored-by: codiumai-pr-agent-pro[bot] <151058649+codiumai-pr-agent-pro[bot]@users.noreply.github.com>
2024-10-19 11:34:57 +03:00
mrT23
b7437147af
fix: correct model type extraction for O1 model handling in litellm_ai_handler.py 2024-10-19 11:32:45 +03:00
Tal
e6c56c7355
Update pr_agent/algo/ai_handlers/litellm_ai_handler.py
Co-authored-by: codiumai-pr-agent-pro[bot] <151058649+codiumai-pr-agent-pro[bot]@users.noreply.github.com>
2024-10-09 08:56:31 +03:00
mrT23
727b08fde3
feat: add support for O1 model by combining system and user prompts in litellm_ai_handler 2024-10-09 08:53:34 +03:00
mrT23
8d82cb2e04
f string 2024-09-15 08:50:24 +03:00
mrT23
8f943a0d44
fix: update error logging messages and system prompt handling in litellm_ai_handler.py 2024-09-15 08:07:59 +03:00
woung717
578d7c69f8
fix: change deprecated timeout parameter for litellm 2024-08-29 21:45:48 +09:00
MarkRx
8aa76a0ac5 Add and document abilty to use LiteLLM Logging Observability tools 2024-08-19 15:45:47 -04:00
mrT23
aa87bc60f6
Rename 'add_callbacks' to 'add_litellm_callbacks' for clarity in litellm_ai_handler 2024-08-17 09:20:30 +03:00
mrT23
c76aabc71e
Add callback functionality to litellm_ai_handler for enhanced logging and metadata capture 2024-08-17 09:15:05 +03:00
mrT23
e7e3970874
Add error handling for empty system prompt in litellm_ai_handler and type conversion in utils.py 2024-08-13 16:26:32 +03:00
Hussam.lawen
70da871876
lower OpenAI errors to warnings 2024-08-12 12:27:48 +03:00
Tal
38c38ec280
Update pr_agent/algo/ai_handlers/litellm_ai_handler.py
Co-authored-by: codiumai-pr-agent-pro[bot] <151058649+codiumai-pr-agent-pro[bot]@users.noreply.github.com>
2024-07-27 18:03:35 +03:00
Tal
3904eebf85
Update pr_agent/algo/ai_handlers/litellm_ai_handler.py
Co-authored-by: codiumai-pr-agent-pro[bot] <151058649+codiumai-pr-agent-pro[bot]@users.noreply.github.com>
2024-07-27 18:02:57 +03:00
mrT23
3067afbcb3
Update seed handling: log fixed seed usage; adjust default seed and temperature in config 2024-07-27 17:50:59 +03:00
mrT23
7eadb45c09
Refactor seed handling logic in litellm_ai_handler to improve readability and error checking 2024-07-27 17:23:42 +03:00
mrT23
ac247dbc2c
Add end-to-end tests for GitHub, GitLab, and Bitbucket apps; update temperature setting usage across tools 2024-07-27 17:19:32 +03:00
Tal
20d9d8ad07
Update pr_agent/algo/ai_handlers/litellm_ai_handler.py
Co-authored-by: codiumai-pr-agent-pro[bot] <151058649+codiumai-pr-agent-pro[bot]@users.noreply.github.com>
2024-07-04 12:26:23 +03:00
mrT23
f3c80891f8
sonnet-3.5 2024-07-04 12:23:36 +03:00
mrT23
bf5673912d
APITimeoutError 2024-06-29 11:30:15 +03:00
BrianTeeman
5268a84bcc
repetition_penalty
Correct the spelling of this variable.

Fix spelling errors now will prevent issues going forward where people have to misspell something on purpose
2024-06-16 17:28:30 +01:00
Kamakura
b4f0ad948f Update Python code formatting, configuration loading, and local model additions
1. Code Formatting:
   - Standardized Python code formatting across multiple files to align with PEP 8 guidelines. This includes adjustments to whitespace, line breaks, and inline comments.

2. Configuration Loader Enhancements:
   - Enhanced the `get_settings` function in `config_loader.py` to provide more robust handling of settings retrieval. Added detailed documentation to improve code maintainability and clarity.

3. Model Addition in __init__.py:
   - Added a new model "ollama/llama3" with a token limit to the MAX_TOKENS dictionary in `__init__.py` to support new AI capabilities and configurations.
2024-06-03 23:58:31 +08:00
Tal
fae6cab2a7
Merge pull request #877 from randy-tsukemen/support-groq-llama3
Add Groq Llama3 support
2024-04-22 11:41:12 +03:00
Randy, Huang
0a53f09a7f Add GROQ.KEY support in LiteLLMAIHandler 2024-04-21 15:21:45 +09:00
Randy, Huang
7a9e73702d Fix duplicate assignment of replicate_key in LiteLLMAIHandler 2024-04-21 14:47:25 +09:00
mrT23
92ef2b4464
ask 2024-04-14 14:09:58 +03:00
mrT23
4683a29819
s 2024-04-14 12:34:14 +03:00
mrT23
8f0f08006f
s 2024-04-14 12:00:19 +03:00
gregoryboue
501b059575
feat: allows ollama usage
Fix https://github.com/Codium-ai/pr-agent/issues/657
2024-04-02 11:01:45 +02:00
Tal
dd83b196b4
Merge pull request #781 from koid/feature/support-bedrock-claude3
added support for bedrock/claude3
2024-03-18 08:03:29 +02:00
mrT23
669e076938
Enhance AI handler logging and add main PR language attribute to AI handler in various tools 2024-03-16 13:52:02 +02:00
mrT23
74345284bd
Enhance AI handler logging and add main PR language attribute to AI handler in various tools 2024-03-16 13:47:44 +02:00
koid
f94a0fd704 add Claude3Config 2024-03-13 11:24:51 +09:00
koid
1ed2cd064a add config litellm.drop_params 2024-03-13 11:20:02 +09:00
mrT23
0690f2bbfd
Refactor litellm_ai_handler.py and update requirements.txt
- Replace retry library with tenacity for better exception handling
- Add verbosity level checks for logging prompts and AI responses
- Add support for HuggingFace API base and repetition penalty in chat completion
- Update requirements.txt with tenacity library
2024-03-06 12:13:54 +02:00
mrT23
1c856a7d41
upgrade litellm 2024-03-06 08:06:59 +02:00
mrT23
dad3d3429f
artifact 2024-02-25 10:45:15 +02:00
mrT23
984a2888ae
Refactor logging statements for better readability and debugging 2024-02-25 10:04:04 +02:00
mrT23
8252b98bf5
Refactor logging statements for better readability and debugging 2024-02-25 10:01:53 +02:00
mrT23
34e421f79b
Refactor logging statements for better readability and debugging 2024-02-25 09:58:58 +02:00
mrT23
877796b539
Refactor logging statements for better readability and debugging 2024-02-25 09:46:07 +02:00
zmeir
ba3f22d81e Move logging to a central location for all AI Handlers 2024-01-04 16:22:22 +02:00
Brian Pham
557b39ec87 Merge branch 'base-ai-handler' into abstract-BaseAiHandler 2023-12-14 07:44:13 +08:00
Brian Pham
7eb2e769cf Move ai handlers to specific folder 2023-12-12 23:03:38 +08:00
Renamed from pr_agent/algo/litellm_ai_handler.py (Browse further)