From 5f999b03cfca2488e2b582772dd57a73719c24c2 Mon Sep 17 00:00:00 2001 From: "renovate[bot]" <29139614+renovate[bot]@users.noreply.github.com> Date: Fri, 25 Jul 2025 08:42:59 -0300 Subject: [PATCH 1/2] Update all non-major dependencies (#1086) Co-authored-by: renovate[bot] <29139614+renovate[bot]@users.noreply.github.com> --- pyproject.toml | 6 +++--- 1 file changed, 3 insertions(+), 3 deletions(-) diff --git a/pyproject.toml b/pyproject.toml index ced0b2e0..bb1cf1ff 100644 --- a/pyproject.toml +++ b/pyproject.toml @@ -51,11 +51,11 @@ get-hashes = 'codemodder.scripts.get_hashes:main' [project.optional-dependencies] semgrep = [ - "semgrep>=1.128,<1.129", + "semgrep>=1.130,<1.131", ] test = [ "azure-ai-inference>=1.0.0b1,<2.0", - "coverage>=7.9,<7.10", + "coverage>=7.10,<7.11", "coverage-threshold~=0.4", "defusedxml==0.7.1", "types-defusedxml==0.7.0.20250708", @@ -86,7 +86,7 @@ test = [ "fickling~=0.1.0,>=0.1.3", "graphql-server~=3.0.0b7", "unidiff>=0.7.5", - "semgrep>=1.128,<1.129", + "semgrep>=1.130,<1.131", ] complexity = [ "radon==6.0.*", From f0a3f1bb03b5e553f71253b2bee6e66f9ea835ea Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?Andr=C3=A9=20C=2E=20Silva?= <12188364+andrecsilva@users.noreply.github.com> Date: Wed, 6 Aug 2025 10:19:09 -0300 Subject: [PATCH 2/2] Fixed prompt tokens not being accounted in CodeTF v2 to v3 conversion (#1087) Fixed prompt tokens not being accounted --- src/codemodder/codetf/v3/codetf.py | 1 + 1 file changed, 1 insertion(+) diff --git a/src/codemodder/codetf/v3/codetf.py b/src/codemodder/codetf/v3/codetf.py index d60e140c..f06a3fbd 100644 --- a/src/codemodder/codetf/v3/codetf.py +++ b/src/codemodder/codetf/v3/codetf.py @@ -158,6 +158,7 @@ def from_v2_aimetadata(ai_metadata: AIMetadatav2) -> AIMetadata: models=[ai_metadata.model] if ai_metadata.model else None, total_tokens=ai_metadata.tokens, completion_tokens=ai_metadata.completion_tokens, + prompt_tokens=ai_metadata.prompt_tokens, )