Merge branch 'main' into pr/2046

# Conflicts:
#	requirements.txt
This commit is contained in:
ofir-frd 2025-10-11 16:50:46 +03:00
commit 40633d7a9e
4 changed files with 19 additions and 11 deletions

View file

@ -39,6 +39,13 @@ Generate the token and add it to .secret.toml under `bitbucket_server` section
bearer_token = "<your key>" bearer_token = "<your key>"
``` ```
Don't forget to also set the URL of your Bitbucket Server instance (either in `.secret.toml` or in `configuration.toml`):
```toml
[bitbucket_server]
url = "<full URL to your Bitbucket instance, e.g.: https://git.bitbucket.com>"
```
### Run it as CLI ### Run it as CLI
Modify `configuration.toml`: Modify `configuration.toml`:
@ -47,6 +54,8 @@ Modify `configuration.toml`:
git_provider="bitbucket_server" git_provider="bitbucket_server"
``` ```
and pass the Pull request URL: and pass the Pull request URL:
```shell ```shell

View file

@ -329,12 +329,13 @@ async def retry_with_fallback_models(f: Callable, model_type: ModelType = ModelT
) )
get_settings().set("openai.deployment_id", deployment_id) get_settings().set("openai.deployment_id", deployment_id)
return await f(model) return await f(model)
except: except Exception as e:
get_logger().warning( get_logger().warning(
f"Failed to generate prediction with {model}" f"Failed to generate prediction with {model}",
artifact={"error": e},
) )
if i == len(all_models) - 1: # If it's the last iteration if i == len(all_models) - 1: # If it's the last iteration
raise Exception(f"Failed to generate prediction with any model of {all_models}") raise Exception(f"Failed to generate prediction with any model of {all_models}") from e
def _get_all_models(model_type: ModelType = ModelType.REGULAR) -> List[str]: def _get_all_models(model_type: ModelType = ModelType.REGULAR) -> List[str]:

View file

@ -376,7 +376,6 @@ def get_main_pr_language(languages, files) -> str:
break break
except Exception as e: except Exception as e:
get_logger().exception(f"Failed to get main language: {e}") get_logger().exception(f"Failed to get main language: {e}")
pass
## old approach: ## old approach:
# most_common_extension = max(set(extension_list), key=extension_list.count) # most_common_extension = max(set(extension_list), key=extension_list.count)
@ -401,7 +400,6 @@ def get_main_pr_language(languages, files) -> str:
except Exception as e: except Exception as e:
get_logger().exception(e) get_logger().exception(e)
pass
return main_language_str return main_language_str

View file

@ -1,19 +1,19 @@
aiohttp==3.10.2 aiohttp==3.12.15
anthropic>=0.69.0 anthropic>=0.69.0
#anthropic[vertex]==0.47.1 #anthropic[vertex]==0.47.1
atlassian-python-api==3.41.4 atlassian-python-api==3.41.4
azure-devops==7.1.0b3 azure-devops==7.1.0b3
azure-identity==1.15.0 azure-identity==1.25.0
boto3==1.33.6 boto3==1.40.45
certifi==2024.8.30 certifi==2024.8.30
dynaconf==3.2.4 dynaconf==3.2.4
fastapi==0.115.6 fastapi==0.118.0
GitPython==3.1.41 GitPython==3.1.41
google-cloud-aiplatform==1.38.0 google-cloud-aiplatform==1.38.0
google-generativeai==0.8.3 google-generativeai==0.8.3
google-cloud-storage==2.10.0 google-cloud-storage==2.10.0
Jinja2==3.1.2 Jinja2==3.1.6
litellm==1.73.6 litellm==1.77.7
loguru==0.7.2 loguru==0.7.2
msrest==0.7.1 msrest==0.7.1
openai>=1.55.3 openai>=1.55.3