Skip to content
Open
Show file tree
Hide file tree
Changes from 35 commits
Commits
Show all changes
45 commits
Select commit Hold shift + click to select a range
720b203
add detector corpora test workflow and script
mustansir14 Apr 29, 2026
942b25f
only run once per PR, make comment descriptive, add handling for manu…
mustansir14 Apr 29, 2026
27b8867
comment out types to see result on all commits
mustansir14 Apr 29, 2026
e360e6d
uncomment types
mustansir14 Apr 29, 2026
1aae080
remove table from comment
mustansir14 Apr 29, 2026
0a78ecc
comment out types
mustansir14 Apr 29, 2026
b9d8506
Phase 0: add explicit pipefail and capture trufflehog stderr
shahzadhaider1 Apr 29, 2026
26c1c03
Phase 1: differential diffing PR vs main
shahzadhaider1 Apr 29, 2026
f46e86c
DEMO: loosen Stripe regex (will revert)
shahzadhaider1 Apr 29, 2026
021e8c3
DEMO: loosen JDBC regex (will revert)
shahzadhaider1 Apr 29, 2026
420ec56
Phase 1 fix: add --allow-verification-overlap, fix no-diff detection
shahzadhaider1 Apr 29, 2026
b0c3d28
revert jdbc detector change
shahzadhaider1 Apr 29, 2026
25f08fc
Phase 2: detector scoping, new-detector handling, blast radius, statu…
shahzadhaider1 Apr 29, 2026
735522b
DEMO: loosen JDBC + add fictional acmevault detector
shahzadhaider1 Apr 29, 2026
7b44c92
Phase 2 fix: harden corpus byte counting against early trufflehog exit
shahzadhaider1 May 1, 2026
e0e33bc
Phase 3a (1/3): add hack/extract-keywords for detector keyword intros…
shahzadhaider1 May 2, 2026
a93890e
Phase 3a (2/3): add Layer 1 keyword corpus builder + workflow integra…
shahzadhaider1 May 2, 2026
24fdf36
Phase 4 complete - Heatmap visualization
shahzadhaider1 May 2, 2026
c20a5dd
Phase 4 rework (1/2): emit heatmap-grid.json sidecar from render_heat…
shahzadhaider1 May 2, 2026
80f6747
Phase 4 rework (2/2): replace data-URL embed with emoji-bucketed Mark…
shahzadhaider1 May 2, 2026
02ae97b
Phase 5 complete - Polish
shahzadhaider1 May 2, 2026
5186d12
cleanup, enable verification
mustansir14 May 4, 2026
bbaa4af
fix bug
mustansir14 May 4, 2026
648ae6a
optimizations
mustansir14 May 4, 2026
b56b46b
cache keywords corpus
mustansir14 May 4, 2026
3284602
rewrite comment message
mustansir14 May 4, 2026
2421212
cache github api corpus per keyword
mustansir14 May 4, 2026
6ba2661
cleanup
mustansir14 May 4, 2026
a00d129
remove github corpus
mustansir14 May 4, 2026
8a20a97
revert changes for testing
mustansir14 May 4, 2026
30c98bb
move Configure AWS credentials step to run only when detector changes…
mustansir14 May 4, 2026
0d852e6
revert unnecessary changes
mustansir14 May 4, 2026
624cfbe
cleanup + bugbot fixes
mustansir14 May 4, 2026
f222e57
run test with bigger (30gb) dataset, loosen jdbc regex
mustansir14 May 4, 2026
d0d94a2
optimizations
mustansir14 May 4, 2026
dacf850
bugbot fixes
mustansir14 May 4, 2026
1cf22a8
revert jdbc changes and bugbot fix
mustansir14 May 4, 2026
5792a09
run only on regex and/or keywords change
mustansir14 May 4, 2026
032b2a8
bugbot fixes
mustansir14 May 4, 2026
c53fbcb
Merge branch 'main' into hackathon/detector-tests-in-ci
mustansir14 May 4, 2026
88c63e6
bugbot fix
mustansir14 May 4, 2026
f7f08e6
Merge branch 'hackathon/detector-tests-in-ci' of mustansir:trufflesec…
mustansir14 May 4, 2026
5a6e1a2
incorporate brad's comments, loosen jdbc regex to run a test to ensur…
mustansir14 May 5, 2026
5b722a2
revert test changes
mustansir14 May 5, 2026
6c3bbae
fix misleading bench skipped message
mustansir14 May 5, 2026
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
263 changes: 263 additions & 0 deletions .github/workflows/detector-corpora-test.yml
Original file line number Diff line number Diff line change
@@ -0,0 +1,263 @@
name: Corpora Test

on:
workflow_dispatch:
pull_request:
paths:
- 'pkg/detectors/**'
- 'pkg/engine/defaults/defaults.go'
- '.github/workflows/detector-corpora-test.yml'
- 'scripts/test/detector_corpora_test.sh'
- 'scripts/test/diff_corpora_results.py'
- 'scripts/test/detect_changed_detectors.sh'
Comment on lines +5 to +12
Copy link
Copy Markdown
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Since the CPU work required by a single run of this workflow is pretty expensive (30+ minutes), is this something we want running automatically on pull requests (as it looks like it does here), or only as an opt-in workflow?

Copy link
Copy Markdown
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Good question. The workflow is scoped to only trigger when a PR modifies regex patterns or Keywords() in a detector. Purely structural changes (verification logic, redaction, comments, etc.) are filtered out and skip the bench entirely. In practice this means it only runs on PRs that actually affect match behavior.

We also don't currently merge a detector without running this test manually, so automating it in CI seems like the right call. it ensures the check never gets skipped and gives reviewers the data they need without having to ask for it.


env:
DATASETS: |
s3://trufflehog-corpora-datasets/contents.2025-11-04.jsonl.zstd
s3://trufflehog-corpora-datasets/contents.jsonl.zstd

jobs:
Copy link
Copy Markdown
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Note, some of the actions used here are old versions. Also, you might consider pinning the action versions used here to reduce risk of possible supply-chain attacks.

zizmor is helpful: https://docs.zizmor.sh/

Copy link
Copy Markdown
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Thanks for this. Really helpful! I'll do the needful

corpora-test:
if: ${{ github.repository == 'trufflesecurity/trufflehog' && !github.event.pull_request.head.repo.fork }}
runs-on: ubuntu-latest
permissions:
contents: read
pull-requests: write
steps:
- name: Checkout code
uses: actions/checkout@v4
with:
fetch-depth: 0

- name: Install Go
uses: actions/setup-go@v5
with:
go-version: "1.25"

- name: Install dependencies
run: sudo apt-get install -y zstd jq

- name: Resolve merge-base
id: merge_base
shell: bash
run: |
set -o pipefail
git fetch --no-tags --prune origin main
MERGE_BASE=$(git merge-base origin/main HEAD)
echo "Merge base: $MERGE_BASE"
echo "sha=$MERGE_BASE" >> "$GITHUB_OUTPUT"

# Determine which detectors changed in this PR. The PR build scopes its
# scan to the full set; the main build excludes detectors that don't
# exist there yet (new detectors). If the set is empty, the workflow
# short-circuits with a skip comment — scoping is the entire point of
# Phase 2, falling back to scan-all defeats it.
- name: Detect changed detectors
id: detect
shell: bash
env:
BASE_REF: ${{ steps.merge_base.outputs.sha }}
run: |
set -o pipefail
chmod +x scripts/test/detect_changed_detectors.sh
PR_CSV=$(./scripts/test/detect_changed_detectors.sh --pr-csv || true)
MAIN_CSV=$(./scripts/test/detect_changed_detectors.sh --main-csv || true)
NEW_LIST=$(./scripts/test/detect_changed_detectors.sh --new-only || true)
NEW_CSV=$(echo "$NEW_LIST" | paste -sd, -)
echo "PR detectors: $PR_CSV"
echo "Main detectors: $MAIN_CSV"
echo "New detectors: $NEW_CSV"
echo "pr_csv=$PR_CSV" >> "$GITHUB_OUTPUT"
echo "main_csv=$MAIN_CSV" >> "$GITHUB_OUTPUT"
echo "new_csv=$NEW_CSV" >> "$GITHUB_OUTPUT"
if [[ -n "$PR_CSV" ]]; then
echo "any_changed=true" >> "$GITHUB_OUTPUT"
else
echo "any_changed=false" >> "$GITHUB_OUTPUT"
fi

# Sticky comment: find any prior detector-bench comment on the PR by
# the marker substring and update it in place. The marker — kept in
# sync with STICKY_COMMENT_MARKER in scripts/test/diff_corpora_results.py —
# has to appear in BOTH the skip body and the diff body so the same
# comment flips between them as iterative pushes change which path
# fires. Skip body is only posted on pull_request events; workflow_dispatch
# runs with no changed detectors silently finish without posting.
- name: Find existing skip comment
if: steps.detect.outputs.any_changed != 'true' && github.event_name == 'pull_request'
id: find_skip_comment
uses: peter-evans/find-comment@v3
with:
issue-number: ${{ github.event.pull_request.number }}
comment-author: 'github-actions[bot]'
body-includes: '<!-- detector-bench -->'

- name: Post or update skip comment
if: steps.detect.outputs.any_changed != 'true' && github.event_name == 'pull_request'
uses: peter-evans/create-or-update-comment@v4
with:
comment-id: ${{ steps.find_skip_comment.outputs.comment-id }}
issue-number: ${{ github.event.pull_request.number }}
edit-mode: replace
body: |
<!-- detector-bench -->
## Corpora Test Results

No detector source files changed in this PR. Bench skipped.

- name: Configure AWS credentials
if: steps.detect.outputs.any_changed == 'true'
uses: aws-actions/configure-aws-credentials@v4
with:
aws-access-key-id: ${{ secrets.AWS_ACCESS_KEY }}
aws-secret-access-key: ${{ secrets.AWS_SECRET_ACCESS_KEY }}
aws-region: us-east-1

# Cache the main scan results by merge-base + scoped detector set.
# On subsequent pushes to the same PR without a rebase, both are
# identical, so the main scan (35 GB of S3 streaming + trufflehog) is
# skipped entirely.
- name: Restore main scan cache
id: main_scan_cache
if: steps.detect.outputs.any_changed == 'true' && steps.detect.outputs.main_csv != ''
uses: actions/cache/restore@v4
with:
path: /tmp/results-main.jsonl
key: main-scan-v1-${{ steps.merge_base.outputs.sha }}-${{ steps.detect.outputs.main_csv }}

# Two independent builds run in parallel:
# A) prepare main worktree → build main binary (git I/O then CPU)
# Skipped on main scan cache hit — binary is not needed.
# B) build PR binary (CPU, no dependencies)
- name: Build binaries
if: steps.detect.outputs.any_changed == 'true'
shell: bash
env:
MERGE_BASE: ${{ steps.merge_base.outputs.sha }}
MAIN_SCAN_CACHE_HIT: ${{ steps.main_scan_cache.outputs.cache-hit }}
run: |
set -o pipefail

# Chain A: prepare worktree, then build main binary.
# Skipped when main scan results are already cached.
if [[ "$MAIN_SCAN_CACHE_HIT" != 'true' ]]; then
(
git worktree add /tmp/trufflehog-main-src "$MERGE_BASE"
cd /tmp/trufflehog-main-src
CGO_ENABLED=0 go build -o /tmp/trufflehog-main .
) &
PID_MAIN_BUILD=$!
fi

# Chain B: build PR binary (no dependencies).
CGO_ENABLED=0 go build -o /tmp/trufflehog-pr . &
PID_PR_BUILD=$!

[[ -n "${PID_MAIN_BUILD:-}" ]] && { wait $PID_MAIN_BUILD || { echo "Main binary build failed" >&2; exit 1; }; }
wait $PID_PR_BUILD || { echo "PR binary build failed" >&2; exit 1; }
Comment thread
cursor[bot] marked this conversation as resolved.

# PR and main scans share a single S3 stream per dataset file, teed to
# both binaries simultaneously. The main side is skipped on a cache hit
# (results already in /tmp/results-main.jsonl) or when main_csv is empty
# (PR adds only new detectors — no overlap with main).
- name: Run corpora tests
if: steps.detect.outputs.any_changed == 'true'
shell: bash
env:
PR_CSV: ${{ steps.detect.outputs.pr_csv }}
MAIN_CSV: ${{ steps.detect.outputs.main_csv }}
MAIN_SCAN_CACHE_HIT: ${{ steps.main_scan_cache.outputs.cache-hit }}
run: |
set -o pipefail
files=()
while IFS= read -r dataset; do
[[ -z "$dataset" ]] && continue
files+=("$dataset")
done <<< "$DATASETS"

export TRUFFLEHOG_BIN=/tmp/trufflehog-pr
export OUTPUT_JSONL=/tmp/results-pr.jsonl
export STDERR_FILE=/tmp/corpora-stderr-pr.txt
export INCLUDE_DETECTORS="$PR_CSV"

if [[ -n "$MAIN_CSV" && "$MAIN_SCAN_CACHE_HIT" != 'true' ]]; then
# Dual-binary: single S3 download teed to both PR and main binaries.
export TRUFFLEHOG_BIN_MAIN=/tmp/trufflehog-main
export OUTPUT_JSONL_MAIN=/tmp/results-main.jsonl
export INCLUDE_DETECTORS_MAIN="$MAIN_CSV"
elif [[ -z "$MAIN_CSV" ]]; then
echo "No overlapping detectors in main; skipping main scan."
: > /tmp/results-main.jsonl
else
echo "Main scan cache hit; skipping main scan."
fi

./scripts/test/detector_corpora_test.sh "${files[@]}" \
|| { echo "Corpora scan failed" >&2; exit 1; }

- name: Save main scan cache
if: steps.detect.outputs.any_changed == 'true' && steps.detect.outputs.main_csv != '' && steps.main_scan_cache.outputs.cache-hit != 'true'
uses: actions/cache/save@v4
with:
path: /tmp/results-main.jsonl
key: main-scan-v1-${{ steps.merge_base.outputs.sha }}-${{ steps.detect.outputs.main_csv }}

- name: Diff results
if: steps.detect.outputs.any_changed == 'true'
shell: bash
env:
CHANGED: ${{ steps.detect.outputs.pr_csv }}
NEW_DETECTORS: ${{ steps.detect.outputs.new_csv }}
run: |
set -o pipefail
python3 scripts/test/diff_corpora_results.py \
/tmp/results-main.jsonl /tmp/results-pr.jsonl \
--changed-detectors="$CHANGED" \
--new-detectors="$NEW_DETECTORS" \
> /tmp/diff-report.md
cat /tmp/diff-report.md

# workflow_dispatch runs don't carry an issue context, so resolve the
# PR number by branch lookup. pull_request events fall through to the
# event's issue number. Output feeds the find/update pair below.
- name: Resolve PR number
if: steps.detect.outputs.any_changed == 'true'
id: resolve_pr
uses: actions/github-script@v7
with:
script: |
let issue_number;
if (context.eventName === 'workflow_dispatch') {
const pulls = await github.rest.pulls.list({
owner: context.repo.owner,
repo: context.repo.repo,
head: `${context.repo.owner}:${context.ref.replace('refs/heads/', '')}`,
state: 'open',
});
if (pulls.data.length === 0) {
core.setFailed(`No open PR found for branch ${context.ref}`);
return;
}
issue_number = pulls.data[0].number;
} else {
issue_number = context.issue.number;
}
core.setOutput('issue_number', issue_number);

- name: Find existing diff comment
if: steps.detect.outputs.any_changed == 'true'
id: find_diff_comment
uses: peter-evans/find-comment@v3
with:
issue-number: ${{ steps.resolve_pr.outputs.issue_number }}
comment-author: 'github-actions[bot]'
body-includes: '<!-- detector-bench -->'

- name: Post or update diff comment
if: steps.detect.outputs.any_changed == 'true'
uses: peter-evans/create-or-update-comment@v4
with:
comment-id: ${{ steps.find_diff_comment.outputs.comment-id }}
issue-number: ${{ steps.resolve_pr.outputs.issue_number }}
edit-mode: replace
body-path: /tmp/diff-report.md
4 changes: 4 additions & 0 deletions .gitignore
Original file line number Diff line number Diff line change
Expand Up @@ -10,3 +10,7 @@ tmp/go-test.json
.captain/detectors/quarantines.yaml
.captain/detectors/flakes.yaml
.vscode

# Python
__pycache__/
*.pyc
3 changes: 2 additions & 1 deletion pkg/detectors/jdbc/jdbc.go
Original file line number Diff line number Diff line change
Expand Up @@ -53,7 +53,8 @@ var (
// Matches typical JDBC connection strings.
// The terminal character class additionally excludes () and & to avoid
// capturing surrounding delimiters (e.g. "(jdbc:…)" or "…&user=x&").
keyPat = regexp.MustCompile(`(?i)jdbc:[\w]{3,10}:[^\s"'<>,{}[\]]{10,511}[^\s"'<>,{}[\]()&]`)
// TODO: revert before merging — regex intentionally loosened to trigger corpora test CI on this PR.
keyPat = regexp.MustCompile(`(?i)[\w]{3,10}:[^\s"'<>,{}[\]]{10,511}[^\s"'<>,{}[\]()&]`)
Comment thread
cursor[bot] marked this conversation as resolved.
Outdated
)

// Keywords are used for efficiently pre-filtering chunks.
Expand Down
Loading
Loading