From a48134be3ee5f10276a189dd3cbbcdb42ada9b78 Mon Sep 17 00:00:00 2001 From: Felix Chapman Date: Tue, 23 Dec 2025 18:25:29 +0000 Subject: [PATCH 1/6] Add new parameter list-files for when there are multiple files in a report --- README.md | 8 +- __tests__/report/get-report.test.ts | 145 +++++++++++++++++++++++++++- action.yml | 8 ++ src/main.ts | 10 +- src/report/get-report.ts | 18 +++- 5 files changed, 182 insertions(+), 7 deletions(-) diff --git a/README.md b/README.md index 38b2285..a39f83d 100644 --- a/README.md +++ b/README.md @@ -184,6 +184,12 @@ jobs: # none list-tests: 'all' + # Limits which test result files are listed: + # all + # failed + # none + list-files: 'all' + # Limits number of created annotations with error message and stack trace captured during test execution. # Must be less or equal to 50. max-annotations: '10' @@ -444,7 +450,7 @@ Support for Swift test results in xUnit format is experimental - should work but Unfortunately, there are some known issues and limitations caused by GitHub API: - Test report (i.e. build summary) is Markdown text. No custom styling or HTML is possible. -- Maximum report size is 65535 bytes. Input parameters `list-suites` and `list-tests` will be automatically adjusted if max size is exceeded. +- Maximum report size is 65535 bytes. Input parameters `list-suites`, `list-tests`, and `list-files` will be automatically adjusted if max size is exceeded. - Test report can't reference any additional files (e.g. screenshots). You can use `actions/upload-artifact@v4` to upload them and inspect them manually. - Check Runs are created for specific commit SHA. It's not possible to specify under which workflow test report should belong if more workflows are running for the same SHA. Thanks to this GitHub "feature" it's possible your test report will appear in an unexpected place in GitHub UI. diff --git a/__tests__/report/get-report.test.ts b/__tests__/report/get-report.test.ts index 30e1e89..6543208 100644 --- a/__tests__/report/get-report.test.ts +++ b/__tests__/report/get-report.test.ts @@ -1,4 +1,5 @@ -import {getBadge, DEFAULT_OPTIONS, ReportOptions} from '../../src/report/get-report.js' +import {DEFAULT_OPTIONS, getBadge, getReport, ReportOptions} from '../../src/report/get-report.js' +import {TestCaseResult, TestGroupResult, TestRunResult, TestSuiteResult} from '../../src/test-results' describe('getBadge', () => { describe('URI encoding with special characters', () => { @@ -131,3 +132,145 @@ describe('getBadge', () => { }) }) }) + +describe('getReport', () => { + // Helper function to create test results + function createTestResult(path: string, passed: number, failed: number, skipped: number): TestRunResult { + const tests: TestCaseResult[] = [] + for (let i = 0; i < passed; i++) { + tests.push(new TestCaseResult(`passed-test-${i}`, 'success', 100)) + } + for (let i = 0; i < failed; i++) { + tests.push(new TestCaseResult(`failed-test-${i}`, 'failed', 100, { + details: 'Test failed', + message: 'Assertion error' + })) + } + for (let i = 0; i < skipped; i++) { + tests.push(new TestCaseResult(`skipped-test-${i}`, 'skipped', 0)) + } + + const group = new TestGroupResult('test-group', tests) + const suite = new TestSuiteResult('test-suite', [group]) + return new TestRunResult(path, [suite]) + } + + describe('list-files parameter', () => { + const results = [ + createTestResult('passing-file.spec.ts', 5, 0, 0), + createTestResult('failing-file.spec.ts', 3, 2, 1), + createTestResult('passing-with-skipped-file.spec.ts', 8, 0, 2) + ] + + it('shows all files when list-files is "all"', () => { + const report = getReport(results, { + ...DEFAULT_OPTIONS, + listFiles: 'all', + listSuites: 'none', + listTests: 'none' + }) + + expect(report).toContain('passing-file.spec.ts') + expect(report).toContain('failing-file.spec.ts') + expect(report).toContain('passing-with-skipped-file.spec.ts') + }) + + it('shows only failed files when list-files is "failed"', () => { + const report = getReport(results, { + ...DEFAULT_OPTIONS, + listFiles: 'failed', + listSuites: 'none', + listTests: 'none' + }) + + expect(report).not.toContain('passing-file.spec.ts') + expect(report).toContain('failing-file.spec.ts') + expect(report).not.toContain('passing-with-skipped-file.spec.ts') + }) + + it('shows no file details when list-files is "none"', () => { + const report = getReport(results, { + ...DEFAULT_OPTIONS, + listFiles: 'none', + listSuites: 'none', + listTests: 'none' + }) + + // Should still have badge + expect(report).toContain('![') + // Should not have file names in detail sections + expect(report).not.toContain('passing-file.spec.ts') + expect(report).not.toContain('failing-file.spec.ts') + expect(report).not.toContain('passing-with-skipped-file.spec.ts') + }) + + it('includes summary table even with list-files "none"', () => { + const report = getReport(results, { + ...DEFAULT_OPTIONS, + listFiles: 'none', + listSuites: 'all', + listTests: 'none' + }) + + // Badge should still be present + expect(report).toContain('![') + expect(report).toContain('badge') + // File names should not be present + expect(report).not.toContain('passing-file.spec.ts') + expect(report).not.toContain('failing-file.spec.ts') + expect(report).not.toContain('passing-with-skipped-file.spec.ts') + }) + + it('works correctly with list-suites and list-tests when list-files is "failed"', () => { + const report = getReport(results, { + ...DEFAULT_OPTIONS, + listFiles: 'failed', + listSuites: 'all', + listTests: 'all' + }) + + expect(report).not.toContain('passing-file.spec.ts') + expect(report).toContain('failing-file.spec.ts') + expect(report).not.toContain('passing-with-skipped-file.spec.ts') + // Should show suite details for the failed file + expect(report).toContain('test-suite') + }) + + it('filters correctly when all files pass and list-files is "failed"', () => { + const allPassingResults = [ + createTestResult('passing-file-1.spec.ts', 5, 0, 0), + createTestResult('passing-file-2.spec.ts', 8, 0, 2) + ] + + const report = getReport(allPassingResults, { + ...DEFAULT_OPTIONS, + listFiles: 'failed', + listSuites: 'all', + listTests: 'none' + }) + + expect(report).not.toContain('passing-file-1.spec.ts') + expect(report).not.toContain('passing-file-2.spec.ts') + // Badge should still be present + expect(report).toContain('![') + expect(report).toContain('badge') + }) + + it('filters correctly when all files fail and list-files is "failed"', () => { + const allFailingResults = [ + createTestResult('failing-file-1.spec.ts', 0, 5, 0), + createTestResult('failing-file-2.spec.ts', 1, 2, 1) + ] + + const report = getReport(allFailingResults, { + ...DEFAULT_OPTIONS, + listFiles: 'failed', + listSuites: 'all', + listTests: 'none' + }) + + expect(report).toContain('failing-file-1.spec.ts') + expect(report).toContain('failing-file-2.spec.ts') + }) + }) +}) diff --git a/action.yml b/action.yml index be606d4..577b0c3 100644 --- a/action.yml +++ b/action.yml @@ -54,6 +54,14 @@ inputs: - none required: false default: 'all' + list-files: + description: | + Limits which test result files are listed. Supported options: + - all + - failed + - none + required: false + default: 'all' max-annotations: description: | Limits number of created annotations with error message and stack trace captured during test execution. diff --git a/src/main.ts b/src/main.ts index fc6ea16..6705acb 100644 --- a/src/main.ts +++ b/src/main.ts @@ -44,6 +44,7 @@ class TestReporter { readonly reporter = core.getInput('reporter', {required: true}) readonly listSuites = core.getInput('list-suites', {required: true}) as 'all' | 'failed' | 'none' readonly listTests = core.getInput('list-tests', {required: true}) as 'all' | 'failed' | 'none' + readonly listFiles = core.getInput('list-files', {required: true}) as 'all' | 'failed' | 'none' readonly maxAnnotations = parseInt(core.getInput('max-annotations', {required: true})) readonly failOnError = core.getInput('fail-on-error', {required: true}) === 'true' readonly failOnEmpty = core.getInput('fail-on-empty', {required: true}) === 'true' @@ -71,6 +72,11 @@ class TestReporter { return } + if (this.listFiles !== 'all' && this.listFiles !== 'failed' && this.listFiles !== 'none') { + core.setFailed(`Input parameter 'list-files' has invalid value`) + return + } + if (this.collapsed !== 'auto' && this.collapsed !== 'always' && this.collapsed !== 'never') { core.setFailed(`Input parameter 'collapsed' has invalid value`) return @@ -177,7 +183,7 @@ class TestReporter { } } - const {listSuites, listTests, slugPrefix, onlySummary, useActionsSummary, badgeTitle, reportTitle, collapsed} = this + const {listSuites, listTests, slugPrefix, listFiles, onlySummary, useActionsSummary, badgeTitle, reportTitle, collapsed} = this const passed = results.reduce((sum, tr) => sum + tr.passed, 0) const failed = results.reduce((sum, tr) => sum + tr.failed, 0) @@ -192,6 +198,7 @@ class TestReporter { listSuites, listTests, slugPrefix, + listFiles, baseUrl, onlySummary, useActionsSummary, @@ -224,6 +231,7 @@ class TestReporter { listSuites, listTests, slugPrefix, + listFiles, baseUrl, onlySummary, useActionsSummary, diff --git a/src/report/get-report.ts b/src/report/get-report.ts index 325e28c..c6fc185 100644 --- a/src/report/get-report.ts +++ b/src/report/get-report.ts @@ -11,7 +11,8 @@ const MAX_ACTIONS_SUMMARY_LENGTH = 1048576 export interface ReportOptions { listSuites: 'all' | 'failed' | 'none' listTests: 'all' | 'failed' | 'none' - slugPrefix: string + slugPrefix: string; + listFiles: 'all' | 'failed' | 'none' baseUrl: string onlySummary: boolean useActionsSummary: boolean @@ -24,6 +25,7 @@ export const DEFAULT_OPTIONS: ReportOptions = { listSuites: 'all', listTests: 'all', slugPrefix: '', + listFiles: 'all', baseUrl: '', onlySummary: false, useActionsSummary: true, @@ -173,8 +175,16 @@ function getTestRunsReport(testRuns: TestRunResult[], options: ReportOptions): s sections.push(` `) } - if (testRuns.length > 0 || options.onlySummary) { - const tableData = testRuns + // Filter test runs based on list-files option + const filteredTestRuns = + options.listFiles === 'failed' + ? testRuns.filter(tr => tr.result === 'failed') + : options.listFiles === 'none' + ? [] + : testRuns + + if (filteredTestRuns.length > 0 || options.onlySummary) { + const tableData = filteredTestRuns .map((tr, originalIndex) => ({tr, originalIndex})) .filter(({tr}) => tr.passed > 0 || tr.failed > 0 || tr.skipped > 0) .map(({tr, originalIndex}) => { @@ -197,7 +207,7 @@ function getTestRunsReport(testRuns: TestRunResult[], options: ReportOptions): s } if (options.onlySummary === false) { - const suitesReports = testRuns.map((tr, i) => getSuitesReport(tr, i, options)).flat() + const suitesReports = filteredTestRuns.map((tr, i) => getSuitesReport(tr, i, options)).flat() sections.push(...suitesReports) } From ba6f1d11e2db25ec4dcbe7d39e4628d10fb5e8ff Mon Sep 17 00:00:00 2001 From: Felix Chapman Date: Wed, 24 Dec 2025 11:24:30 +0000 Subject: [PATCH 2/6] Correct README to indicate list-files is not automatically adjusted --- README.md | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/README.md b/README.md index a39f83d..81d04a5 100644 --- a/README.md +++ b/README.md @@ -450,7 +450,7 @@ Support for Swift test results in xUnit format is experimental - should work but Unfortunately, there are some known issues and limitations caused by GitHub API: - Test report (i.e. build summary) is Markdown text. No custom styling or HTML is possible. -- Maximum report size is 65535 bytes. Input parameters `list-suites`, `list-tests`, and `list-files` will be automatically adjusted if max size is exceeded. +- Maximum report size is 65535 bytes. Input parameters `list-suites` and `list-tests` will be automatically adjusted if max size is exceeded. - Test report can't reference any additional files (e.g. screenshots). You can use `actions/upload-artifact@v4` to upload them and inspect them manually. - Check Runs are created for specific commit SHA. It's not possible to specify under which workflow test report should belong if more workflows are running for the same SHA. Thanks to this GitHub "feature" it's possible your test report will appear in an unexpected place in GitHub UI. From 45c8d46681ab4f9ee9dd01180e5773fc95017b92 Mon Sep 17 00:00:00 2001 From: Jozef Izso Date: Sat, 25 Apr 2026 12:13:19 +0200 Subject: [PATCH 3/6] Resolve list-files conflicts on current main --- __tests__/report/get-report.test.ts | 12 +++++++----- dist/index.js | 22 ++++++++++++++++++---- src/main.ts | 12 +++++++++++- src/report/get-report.ts | 2 +- 4 files changed, 37 insertions(+), 11 deletions(-) diff --git a/__tests__/report/get-report.test.ts b/__tests__/report/get-report.test.ts index 6543208..6908c6c 100644 --- a/__tests__/report/get-report.test.ts +++ b/__tests__/report/get-report.test.ts @@ -1,5 +1,5 @@ import {DEFAULT_OPTIONS, getBadge, getReport, ReportOptions} from '../../src/report/get-report.js' -import {TestCaseResult, TestGroupResult, TestRunResult, TestSuiteResult} from '../../src/test-results' +import {TestCaseResult, TestGroupResult, TestRunResult, TestSuiteResult} from '../../src/test-results.js' describe('getBadge', () => { describe('URI encoding with special characters', () => { @@ -141,10 +141,12 @@ describe('getReport', () => { tests.push(new TestCaseResult(`passed-test-${i}`, 'success', 100)) } for (let i = 0; i < failed; i++) { - tests.push(new TestCaseResult(`failed-test-${i}`, 'failed', 100, { - details: 'Test failed', - message: 'Assertion error' - })) + tests.push( + new TestCaseResult(`failed-test-${i}`, 'failed', 100, { + details: 'Test failed', + message: 'Assertion error' + }) + ) } for (let i = 0; i < skipped; i++) { tests.push(new TestCaseResult(`skipped-test-${i}`, 'skipped', 0)) diff --git a/dist/index.js b/dist/index.js index c8865b0..576e7ab 100644 --- a/dist/index.js +++ b/dist/index.js @@ -56954,6 +56954,7 @@ const DEFAULT_OPTIONS = { listSuites: 'all', listTests: 'all', slugPrefix: '', + listFiles: 'all', baseUrl: '', onlySummary: false, useActionsSummary: true, @@ -57074,8 +57075,14 @@ function getTestRunsReport(testRuns, options) { sections.push(`
Expand for details`); sections.push(` `); } - if (testRuns.length > 0 || options.onlySummary) { - const tableData = testRuns + // Filter test runs based on list-files option + const filteredTestRuns = options.listFiles === 'failed' + ? testRuns.filter(tr => tr.result === 'failed') + : options.listFiles === 'none' + ? [] + : testRuns; + if (filteredTestRuns.length > 0 || options.onlySummary) { + const tableData = filteredTestRuns .map((tr, originalIndex) => ({ tr, originalIndex })) .filter(({ tr }) => tr.passed > 0 || tr.failed > 0 || tr.skipped > 0) .map(({ tr, originalIndex }) => { @@ -57092,7 +57099,7 @@ function getTestRunsReport(testRuns, options) { sections.push(resultsTable); } if (options.onlySummary === false) { - const suitesReports = testRuns.map((tr, i) => getSuitesReport(tr, i, options)).flat(); + const suitesReports = filteredTestRuns.map((tr, i) => getSuitesReport(tr, i, options)).flat(); sections.push(...suitesReports); } if (shouldCollapse) { @@ -58949,6 +58956,7 @@ class TestReporter { reporter = getInput('reporter', { required: true }); listSuites = getInput('list-suites', { required: true }); listTests = getInput('list-tests', { required: true }); + listFiles = getInput('list-files', { required: true }); maxAnnotations = parseInt(getInput('max-annotations', { required: true })); failOnError = getInput('fail-on-error', { required: true }) === 'true'; failOnEmpty = getInput('fail-on-empty', { required: true }) === 'true'; @@ -58972,6 +58980,10 @@ class TestReporter { setFailed(`Input parameter 'list-tests' has invalid value`); return; } + if (this.listFiles !== 'all' && this.listFiles !== 'failed' && this.listFiles !== 'none') { + setFailed(`Input parameter 'list-files' has invalid value`); + return; + } if (this.collapsed !== 'auto' && this.collapsed !== 'always' && this.collapsed !== 'never') { setFailed(`Input parameter 'collapsed' has invalid value`); return; @@ -59056,7 +59068,7 @@ class TestReporter { throw error; } } - const { listSuites, listTests, slugPrefix, onlySummary, useActionsSummary, badgeTitle, reportTitle, collapsed } = this; + const { listSuites, listTests, slugPrefix, listFiles, onlySummary, useActionsSummary, badgeTitle, reportTitle, collapsed } = this; const passed = results.reduce((sum, tr) => sum + tr.passed, 0); const failed = results.reduce((sum, tr) => sum + tr.failed, 0); const skipped = results.reduce((sum, tr) => sum + tr.skipped, 0); @@ -59067,6 +59079,7 @@ class TestReporter { listSuites, listTests, slugPrefix, + listFiles, baseUrl, onlySummary, useActionsSummary, @@ -59096,6 +59109,7 @@ class TestReporter { listSuites, listTests, slugPrefix, + listFiles, baseUrl, onlySummary, useActionsSummary, diff --git a/src/main.ts b/src/main.ts index 6705acb..eb1b3a5 100644 --- a/src/main.ts +++ b/src/main.ts @@ -183,7 +183,17 @@ class TestReporter { } } - const {listSuites, listTests, slugPrefix, listFiles, onlySummary, useActionsSummary, badgeTitle, reportTitle, collapsed} = this + const { + listSuites, + listTests, + slugPrefix, + listFiles, + onlySummary, + useActionsSummary, + badgeTitle, + reportTitle, + collapsed + } = this const passed = results.reduce((sum, tr) => sum + tr.passed, 0) const failed = results.reduce((sum, tr) => sum + tr.failed, 0) diff --git a/src/report/get-report.ts b/src/report/get-report.ts index c6fc185..52f8645 100644 --- a/src/report/get-report.ts +++ b/src/report/get-report.ts @@ -11,7 +11,7 @@ const MAX_ACTIONS_SUMMARY_LENGTH = 1048576 export interface ReportOptions { listSuites: 'all' | 'failed' | 'none' listTests: 'all' | 'failed' | 'none' - slugPrefix: string; + slugPrefix: string listFiles: 'all' | 'failed' | 'none' baseUrl: string onlySummary: boolean From d7f15447e633efd2285e381ef4a3deb0100f3485 Mon Sep 17 00:00:00 2001 From: Jozef Izso Date: Sat, 25 Apr 2026 13:04:58 +0200 Subject: [PATCH 4/6] Tighten list-files report tests --- __tests__/report/get-report.test.ts | 16 ++++++++-------- 1 file changed, 8 insertions(+), 8 deletions(-) diff --git a/__tests__/report/get-report.test.ts b/__tests__/report/get-report.test.ts index 6908c6c..182469e 100644 --- a/__tests__/report/get-report.test.ts +++ b/__tests__/report/get-report.test.ts @@ -172,6 +172,7 @@ describe('getReport', () => { listTests: 'none' }) + expect(report).toContain('|Report|Passed|Failed|Skipped|Time|') expect(report).toContain('passing-file.spec.ts') expect(report).toContain('failing-file.spec.ts') expect(report).toContain('passing-with-skipped-file.spec.ts') @@ -185,6 +186,7 @@ describe('getReport', () => { listTests: 'none' }) + expect(report).toContain('|Report|Passed|Failed|Skipped|Time|') expect(report).not.toContain('passing-file.spec.ts') expect(report).toContain('failing-file.spec.ts') expect(report).not.toContain('passing-with-skipped-file.spec.ts') @@ -198,26 +200,24 @@ describe('getReport', () => { listTests: 'none' }) - // Should still have badge expect(report).toContain('![') - // Should not have file names in detail sections + expect(report).not.toContain('|Report|Passed|Failed|Skipped|Time|') expect(report).not.toContain('passing-file.spec.ts') expect(report).not.toContain('failing-file.spec.ts') expect(report).not.toContain('passing-with-skipped-file.spec.ts') }) - it('includes summary table even with list-files "none"', () => { + it('shows an empty summary table when list-files is "none" and only-summary is enabled', () => { const report = getReport(results, { ...DEFAULT_OPTIONS, listFiles: 'none', listSuites: 'all', + onlySummary: true, listTests: 'none' }) - // Badge should still be present expect(report).toContain('![') - expect(report).toContain('badge') - // File names should not be present + expect(report).toContain('|Report|Passed|Failed|Skipped|Time|') expect(report).not.toContain('passing-file.spec.ts') expect(report).not.toContain('failing-file.spec.ts') expect(report).not.toContain('passing-with-skipped-file.spec.ts') @@ -231,6 +231,7 @@ describe('getReport', () => { listTests: 'all' }) + expect(report).toContain('|Report|Passed|Failed|Skipped|Time|') expect(report).not.toContain('passing-file.spec.ts') expect(report).toContain('failing-file.spec.ts') expect(report).not.toContain('passing-with-skipped-file.spec.ts') @@ -253,9 +254,8 @@ describe('getReport', () => { expect(report).not.toContain('passing-file-1.spec.ts') expect(report).not.toContain('passing-file-2.spec.ts') - // Badge should still be present expect(report).toContain('![') - expect(report).toContain('badge') + expect(report).not.toContain('|Report|Passed|Failed|Skipped|Time|') }) it('filters correctly when all files fail and list-files is "failed"', () => { From e05b863b43f247da6aa3fe54d89138e16f576d23 Mon Sep 17 00:00:00 2001 From: Jozef Izso Date: Sat, 25 Apr 2026 13:17:09 +0200 Subject: [PATCH 5/6] Suppress empty list-files summary table --- __tests__/report/get-report.test.ts | 4 ++-- dist/index.js | 28 ++++++++++++++-------------- src/report/get-report.ts | 28 ++++++++++++++-------------- 3 files changed, 30 insertions(+), 30 deletions(-) diff --git a/__tests__/report/get-report.test.ts b/__tests__/report/get-report.test.ts index 182469e..560f849 100644 --- a/__tests__/report/get-report.test.ts +++ b/__tests__/report/get-report.test.ts @@ -207,7 +207,7 @@ describe('getReport', () => { expect(report).not.toContain('passing-with-skipped-file.spec.ts') }) - it('shows an empty summary table when list-files is "none" and only-summary is enabled', () => { + it('does not show an empty summary table when list-files is "none" and only-summary is enabled', () => { const report = getReport(results, { ...DEFAULT_OPTIONS, listFiles: 'none', @@ -217,7 +217,7 @@ describe('getReport', () => { }) expect(report).toContain('![') - expect(report).toContain('|Report|Passed|Failed|Skipped|Time|') + expect(report).not.toContain('|Report|Passed|Failed|Skipped|Time|') expect(report).not.toContain('passing-file.spec.ts') expect(report).not.toContain('failing-file.spec.ts') expect(report).not.toContain('passing-with-skipped-file.spec.ts') diff --git a/dist/index.js b/dist/index.js index 576e7ab..2847890 100644 --- a/dist/index.js +++ b/dist/index.js @@ -57081,20 +57081,20 @@ function getTestRunsReport(testRuns, options) { : options.listFiles === 'none' ? [] : testRuns; - if (filteredTestRuns.length > 0 || options.onlySummary) { - const tableData = filteredTestRuns - .map((tr, originalIndex) => ({ tr, originalIndex })) - .filter(({ tr }) => tr.passed > 0 || tr.failed > 0 || tr.skipped > 0) - .map(({ tr, originalIndex }) => { - const time = formatTime(tr.time); - const name = tr.path; - const addr = options.baseUrl + makeRunSlug(originalIndex, options).link; - const nameLink = markdown_utils_link(name, addr); - const passed = tr.passed > 0 ? `${tr.passed} ${Icon.success}` : ''; - const failed = tr.failed > 0 ? `${tr.failed} ${Icon.fail}` : ''; - const skipped = tr.skipped > 0 ? `${tr.skipped} ${Icon.skip}` : ''; - return [nameLink, passed, failed, skipped, time]; - }); + const tableData = filteredTestRuns + .map((tr, originalIndex) => ({ tr, originalIndex })) + .filter(({ tr }) => tr.passed > 0 || tr.failed > 0 || tr.skipped > 0) + .map(({ tr, originalIndex }) => { + const time = formatTime(tr.time); + const name = tr.path; + const addr = options.baseUrl + makeRunSlug(originalIndex, options).link; + const nameLink = markdown_utils_link(name, addr); + const passed = tr.passed > 0 ? `${tr.passed} ${Icon.success}` : ''; + const failed = tr.failed > 0 ? `${tr.failed} ${Icon.fail}` : ''; + const skipped = tr.skipped > 0 ? `${tr.skipped} ${Icon.skip}` : ''; + return [nameLink, passed, failed, skipped, time]; + }); + if (tableData.length > 0) { const resultsTable = table(['Report', 'Passed', 'Failed', 'Skipped', 'Time'], [Align.Left, Align.Right, Align.Right, Align.Right, Align.Right], ...tableData); sections.push(resultsTable); } diff --git a/src/report/get-report.ts b/src/report/get-report.ts index 52f8645..cac5596 100644 --- a/src/report/get-report.ts +++ b/src/report/get-report.ts @@ -183,21 +183,21 @@ function getTestRunsReport(testRuns: TestRunResult[], options: ReportOptions): s ? [] : testRuns - if (filteredTestRuns.length > 0 || options.onlySummary) { - const tableData = filteredTestRuns - .map((tr, originalIndex) => ({tr, originalIndex})) - .filter(({tr}) => tr.passed > 0 || tr.failed > 0 || tr.skipped > 0) - .map(({tr, originalIndex}) => { - const time = formatTime(tr.time) - const name = tr.path - const addr = options.baseUrl + makeRunSlug(originalIndex, options).link - const nameLink = link(name, addr) - const passed = tr.passed > 0 ? `${tr.passed} ${Icon.success}` : '' - const failed = tr.failed > 0 ? `${tr.failed} ${Icon.fail}` : '' - const skipped = tr.skipped > 0 ? `${tr.skipped} ${Icon.skip}` : '' - return [nameLink, passed, failed, skipped, time] - }) + const tableData = filteredTestRuns + .map((tr, originalIndex) => ({tr, originalIndex})) + .filter(({tr}) => tr.passed > 0 || tr.failed > 0 || tr.skipped > 0) + .map(({tr, originalIndex}) => { + const time = formatTime(tr.time) + const name = tr.path + const addr = options.baseUrl + makeRunSlug(originalIndex, options).link + const nameLink = link(name, addr) + const passed = tr.passed > 0 ? `${tr.passed} ${Icon.success}` : '' + const failed = tr.failed > 0 ? `${tr.failed} ${Icon.fail}` : '' + const skipped = tr.skipped > 0 ? `${tr.skipped} ${Icon.skip}` : '' + return [nameLink, passed, failed, skipped, time] + }) + if (tableData.length > 0) { const resultsTable = table( ['Report', 'Passed', 'Failed', 'Skipped', 'Time'], [Align.Left, Align.Right, Align.Right, Align.Right, Align.Right], From 528bb1a4e4ccc68f5ba254d5c58e0184343719e1 Mon Sep 17 00:00:00 2001 From: Jozef Izso Date: Sat, 25 Apr 2026 13:31:01 +0200 Subject: [PATCH 6/6] Add changelog for `list-files` input --- CHANGELOG.md | 3 +++ 1 file changed, 3 insertions(+) diff --git a/CHANGELOG.md b/CHANGELOG.md index 20a8c2b..3b107e0 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -1,5 +1,8 @@ # Changelog +## 3.1.0 +* Feature: Add `list-files` input to control test report file listing https://github.com/dorny/test-reporter/pull/773 + ## 3.0.0 * Feature: Use NodeJS 24 LTS as default runtime https://github.com/dorny/test-reporter/pull/738