diff --git a/.github/workflows/check-dist.yml b/.github/workflows/check-dist.yml index e571511..c994dfa 100644 --- a/.github/workflows/check-dist.yml +++ b/.github/workflows/check-dist.yml @@ -21,10 +21,10 @@ jobs: runs-on: ubuntu-latest steps: - - uses: actions/checkout@v4 + - uses: actions/checkout@v6 - name: Set Node.js - uses: actions/setup-node@v4 + uses: actions/setup-node@v6 with: node-version-file: '.nvmrc' @@ -46,7 +46,7 @@ jobs: id: diff # If index.js was different than expected, upload the expected version as an artifact - - uses: actions/upload-artifact@v4 + - uses: actions/upload-artifact@v5 if: ${{ failure() && steps.diff.conclusion == 'failure' }} with: name: dist diff --git a/.github/workflows/ci.yml b/.github/workflows/ci.yml index a8a86e4..aeec96d 100644 --- a/.github/workflows/ci.yml +++ b/.github/workflows/ci.yml @@ -13,8 +13,8 @@ jobs: name: Build & Test runs-on: ubuntu-latest steps: - - uses: actions/checkout@v4 - - uses: actions/setup-node@v4 + - uses: actions/checkout@v6 + - uses: actions/setup-node@v6 with: node-version-file: '.nvmrc' - run: npm ci @@ -24,8 +24,8 @@ jobs: - run: npm test - name: Upload test results - if: success() || failure() - uses: actions/upload-artifact@v4 + if: ${{ !cancelled() }} + uses: actions/upload-artifact@v5 with: name: test-results path: __tests__/__results__/*.xml diff --git a/.github/workflows/manual-run.yml b/.github/workflows/manual-run.yml index f8b1ee0..c1875a9 100644 --- a/.github/workflows/manual-run.yml +++ b/.github/workflows/manual-run.yml @@ -8,14 +8,14 @@ jobs: runs-on: ubuntu-latest steps: - - uses: actions/checkout@v4 + - uses: actions/checkout@v6 - run: npm ci - run: npm run build - run: npm test - name: Create test report uses: ./ - if: success() || failure() + if: ${{ !cancelled() }} with: name: JEST Tests path: __tests__/__results__/*.xml diff --git a/.github/workflows/test-report.yml b/.github/workflows/test-report.yml index e3f9555..11b266a 100644 --- a/.github/workflows/test-report.yml +++ b/.github/workflows/test-report.yml @@ -11,7 +11,7 @@ jobs: name: Workflow test runs-on: ubuntu-latest steps: - - uses: actions/checkout@v4 + - uses: actions/checkout@v6 - uses: ./ with: artifact: test-results diff --git a/.markdownlint.json b/.markdownlint.json new file mode 100644 index 0000000..3f71d87 --- /dev/null +++ b/.markdownlint.json @@ -0,0 +1,13 @@ +{ + "blanks-around-headings": false, + "blanks-around-lists": false, + "blanks-around-tables": false, + "blanks-around-fences": false, + "no-bare-urls": false, + "line-length": false, + "ul-style": false, + "no-inline-html": false, + "no-multiple-blanks": { + "maximum": 3 + } +} diff --git a/CHANGELOG.md b/CHANGELOG.md index d47e76e..b0badd5 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -1,5 +1,30 @@ # Changelog +## 2.3.0 +* Feature: Add Python support with `python-xunit` reporter (pytest) https://github.com/dorny/test-reporter/pull/643 +* Feature: Add pytest traceback parsing and `directory-mapping` option https://github.com/dorny/test-reporter/pull/238 +* Performance: Update sax.js to fix large XML file parsing https://github.com/dorny/test-reporter/pull/681 +* Documentation: Complete documentation for all supported reporters https://github.com/dorny/test-reporter/pull/691 +* Security: Bump js-yaml and mocha in /reports/mocha (fixes prototype pollution) https://github.com/dorny/test-reporter/pull/682 + +## 2.2.0 +* Feature: Add collapsed option to control report summary visibility https://github.com/dorny/test-reporter/pull/664 +* Fix badge encoding for values including underscore and hyphens https://github.com/dorny/test-reporter/pull/672 +* Fix missing `report-title` attribute in action definition https://github.com/dorny/test-reporter/pull/637 +* Refactor variable names to fix shadowing issues https://github.com/dorny/test-reporter/pull/630 + +## 2.1.1 +* Fix error when a TestMethod element does not have a className attribute in a trx file https://github.com/dorny/test-reporter/pull/623 +* Add stack trace from trx to summary https://github.com/dorny/test-reporter/pull/615 +* List only failed tests https://github.com/dorny/test-reporter/pull/606 +* Add type definitions to `github-utils.ts` https://github.com/dorny/test-reporter/pull/604 +* Avoid split on undefined https://github.com/dorny/test-reporter/pull/258 +* Return links to summary report https://github.com/dorny/test-reporter/pull/588 +* Add step summary short summary https://github.com/dorny/test-reporter/pull/589 +* Fix for empty TRX TestDefinitions https://github.com/dorny/test-reporter/pull/582 +* Increase step summary limit to 1MiB https://github.com/dorny/test-reporter/pull/581 +* Fix input description for list options https://github.com/dorny/test-reporter/pull/572 + ## 2.1.0 * Feature: Add summary title https://github.com/dorny/test-reporter/pull/568 * Feature: Add Golang test parser https://github.com/dorny/test-reporter/pull/571 diff --git a/README.md b/README.md index 676d21e..d900926 100644 --- a/README.md +++ b/README.md @@ -9,7 +9,7 @@ This [Github Action](https://github.com/features/actions) displays test results ✔️ Provides final `conclusion` and counts of `passed`, `failed` and `skipped` tests as output parameters **How it looks:** -|![](assets/fluent-validation-report.png)|![](assets/provider-error-summary.png)|![](assets/provider-error-details.png)|![](assets/mocha-groups.png)| +|![Summary showing test run with all tests passed, including details such as test file names, number of passed, failed, and skipped tests, and execution times. The interface is dark-themed and displays a green badge indicating 3527 passed and 4 skipped tests.](assets/fluent-validation-report.png)|![Summary showing test run with a failed unit test. The summary uses a dark background and highlights errors in red for quick identification.](assets/provider-error-summary.png)|![GitHub Actions annotation showing details of a failed unit test with a detailed error message, stack trace, and code annotation.](assets/provider-error-details.png)|![Test cases written in Mocha framework with a list of expectations for each test case. The table format and color-coded badges help users quickly assess test suite health.](assets/mocha-groups.png)| |:--:|:--:|:--:|:--:| **Supported languages / frameworks:** @@ -19,6 +19,8 @@ This [Github Action](https://github.com/features/actions) displays test results - Go / [go test](https://pkg.go.dev/testing) - Java / [JUnit](https://junit.org/) - JavaScript / [JEST](https://jestjs.io/) / [Mocha](https://mochajs.org/) +- Python / [pytest](https://docs.pytest.org/en/stable/) / [unittest](https://docs.python.org/3/library/unittest.html) +- Ruby / [RSpec](https://rspec.info/) - Swift / xUnit For more information see [Supported formats](#supported-formats) section. @@ -50,7 +52,7 @@ jobs: - name: Test Report uses: dorny/test-reporter@v2 - if: success() || failure() # run this step even if previous step failed + if: ${{ !cancelled() }} # run this step even if previous step failed with: name: JEST Tests # Name of the check run which will be created path: reports/jest-*.xml # Path to test results @@ -79,7 +81,7 @@ jobs: - run: npm ci # install packages - run: npm test # run tests (configured to use jest-junit reporter) - uses: actions/upload-artifact@v4 # upload test results - if: success() || failure() # run this step even if previous step failed + if: ${{ !cancelled() }} # run this step even if previous step failed with: name: test-results path: jest-junit.xml @@ -145,7 +147,9 @@ jobs: # java-junit # jest-junit # mocha-json + # python-xunit # rspec-json + # swift-xunit reporter: '' # Allows you to generate only the summary. @@ -253,6 +257,20 @@ Supported testing frameworks: For more information see [dotnet test](https://docs.microsoft.com/en-us/dotnet/core/tools/dotnet-test#examples) +
+ dotnet-nunit + +Test execution must be configured to generate [NUnit3](https://docs.nunit.org/articles/nunit/technical-notes/usage/Test-Result-XML-Format.html) XML test results. +Install the [NUnit3TestAdapter](https://www.nuget.org/packages/NUnit3TestAdapter) package (required; it registers the `nunit` logger for `dotnet test`), then run tests with: + +`dotnet test --logger "nunit;LogFileName=test-results.xml"` + +Supported testing frameworks: +- [NUnit](https://nunit.org/) + +For more information see [dotnet test](https://docs.microsoft.com/en-us/dotnet/core/tools/dotnet-test#examples) +
+
flutter-json @@ -349,6 +367,41 @@ Before version [v9.1.0](https://github.com/mochajs/mocha/releases/tag/v9.1.0), M Please update Mocha to version [v9.1.0](https://github.com/mochajs/mocha/releases/tag/v9.1.0) or above if you encounter this issue.
+
+ python-xunit (Experimental) + +Support for Python test results in xUnit format is experimental - should work but it was not extensively tested. + +For **pytest** support, configure [JUnit XML output](https://docs.pytest.org/en/stable/how-to/output.html#creating-junitxml-format-files) and run with the `--junit-xml` option, which also lets you specify the output path for test results. + +```shell +pytest --junit-xml=test-report.xml +``` + +For **unittest** support, use a test runner that outputs the JUnit report format, such as [unittest-xml-reporting](https://pypi.org/project/unittest-xml-reporting/). +
+ +
+ rspec-json + +[RSpec](https://rspec.info/) testing framework support requires the usage of JSON formatter. +You can configure RSpec to output JSON format by using the `--format json` option and redirecting to a file: + +```shell +rspec --format json --out rspec-results.json +``` + +Or configure it in `.rspec` file: +``` +--format json +--out rspec-results.json +``` + +For more information see: +- [RSpec documentation](https://rspec.info/) +- [RSpec Formatters](https://relishapp.com/rspec/rspec-core/docs/formatters) +
+
swift-xunit (Experimental) diff --git a/__tests__/__outputs__/dart-json.md b/__tests__/__outputs__/dart-json.md index 9805d20..2b7ec09 100644 --- a/__tests__/__outputs__/dart-json.md +++ b/__tests__/__outputs__/dart-json.md @@ -1,7 +1,7 @@ ![Tests failed](https://img.shields.io/badge/tests-1%20passed%2C%204%20failed%2C%201%20skipped-critical) |Report|Passed|Failed|Skipped|Time| |:---|---:|---:|---:|---:| -|fixtures/dart-json.json|1 ✅|4 ❌|1 ⚪|4s| +|[fixtures/dart-json.json](#user-content-r0)|1 ✅|4 ❌|1 ⚪|4s| ## ❌ fixtures/dart-json.json **6** tests were completed in **4s** with **1** passed, **4** failed and **1** skipped. |Test suite|Passed|Failed|Skipped|Time| diff --git a/__tests__/__outputs__/dotnet-nunit.md b/__tests__/__outputs__/dotnet-nunit.md index e33c66d..a0985f5 100644 --- a/__tests__/__outputs__/dotnet-nunit.md +++ b/__tests__/__outputs__/dotnet-nunit.md @@ -1,7 +1,7 @@ ![Tests failed](https://img.shields.io/badge/tests-3%20passed%2C%205%20failed%2C%201%20skipped-critical) |Report|Passed|Failed|Skipped|Time| |:---|---:|---:|---:|---:| -|fixtures/dotnet-nunit.xml|3 ✅|5 ❌|1 ⚪|230ms| +|[fixtures/dotnet-nunit.xml](#user-content-r0)|3 ✅|5 ❌|1 ⚪|230ms| ## ❌ fixtures/dotnet-nunit.xml **9** tests were completed in **230ms** with **3** passed, **5** failed and **1** skipped. |Test suite|Passed|Failed|Skipped|Time| diff --git a/__tests__/__outputs__/dotnet-trx-only-failed.md b/__tests__/__outputs__/dotnet-trx-only-failed.md new file mode 100644 index 0000000..5eeaec5 --- /dev/null +++ b/__tests__/__outputs__/dotnet-trx-only-failed.md @@ -0,0 +1,34 @@ +![Tests failed](https://img.shields.io/badge/tests-5%20passed%2C%205%20failed%2C%201%20skipped-critical) +|Report|Passed|Failed|Skipped|Time| +|:---|---:|---:|---:|---:| +|[fixtures/dotnet-trx.trx](#user-content-r0)|5 ✅|5 ❌|1 ⚪|1s| +## ❌ fixtures/dotnet-trx.trx +**11** tests were completed in **1s** with **5** passed, **5** failed and **1** skipped. +|Test suite|Passed|Failed|Skipped|Time| +|:---|---:|---:|---:|---:| +|[DotnetTests.XUnitTests.CalculatorTests](#user-content-r0s0)|5 ✅|5 ❌|1 ⚪|118ms| +### ❌ DotnetTests.XUnitTests.CalculatorTests +``` +❌ Exception_In_TargetTest + System.DivideByZeroException : Attempted to divide by zero. + at DotnetTests.Unit.Calculator.Div(Int32 a, Int32 b) in C:\Users\Michal\Workspace\dorny\test-reporter\reports\dotnet\DotnetTests.Unit\Calculator.cs:line 9 + at DotnetTests.XUnitTests.CalculatorTests.Exception_In_TargetTest() in C:\Users\Michal\Workspace\dorny\test-reporter\reports\dotnet\DotnetTests.XUnitTests\CalculatorTests.cs:line 33 +❌ Exception_In_Test + System.Exception : Test + at DotnetTests.XUnitTests.CalculatorTests.Exception_In_Test() in C:\Users\Michal\Workspace\dorny\test-reporter\reports\dotnet\DotnetTests.XUnitTests\CalculatorTests.cs:line 39 +❌ Failing_Test + Assert.Equal() Failure + Expected: 3 + Actual: 2 + at DotnetTests.XUnitTests.CalculatorTests.Failing_Test() in C:\Users\Michal\Workspace\dorny\test-reporter\reports\dotnet\DotnetTests.XUnitTests\CalculatorTests.cs:line 27 +❌ Is_Even_Number(i: 3) + Assert.True() Failure + Expected: True + Actual: False + at DotnetTests.XUnitTests.CalculatorTests.Is_Even_Number(Int32 i) in C:\Users\Michal\Workspace\dorny\test-reporter\reports\dotnet\DotnetTests.XUnitTests\CalculatorTests.cs:line 59 +❌ Should be even number(i: 3) + Assert.True() Failure + Expected: True + Actual: False + at DotnetTests.XUnitTests.CalculatorTests.Theory_With_Custom_Name(Int32 i) in C:\Users\Michal\Workspace\dorny\test-reporter\reports\dotnet\DotnetTests.XUnitTests\CalculatorTests.cs:line 67 +``` \ No newline at end of file diff --git a/__tests__/__outputs__/dotnet-trx.md b/__tests__/__outputs__/dotnet-trx.md index eedab46..40bcaf2 100644 --- a/__tests__/__outputs__/dotnet-trx.md +++ b/__tests__/__outputs__/dotnet-trx.md @@ -1,7 +1,7 @@ ![Tests failed](https://img.shields.io/badge/tests-5%20passed%2C%205%20failed%2C%201%20skipped-critical) |Report|Passed|Failed|Skipped|Time| |:---|---:|---:|---:|---:| -|fixtures/dotnet-trx.trx|5 ✅|5 ❌|1 ⚪|1s| +|[fixtures/dotnet-trx.trx](#user-content-r0)|5 ✅|5 ❌|1 ⚪|1s| ## ❌ fixtures/dotnet-trx.trx **11** tests were completed in **1s** with **5** passed, **5** failed and **1** skipped. |Test suite|Passed|Failed|Skipped|Time| @@ -12,23 +12,29 @@ ✅ Custom Name ❌ Exception_In_TargetTest System.DivideByZeroException : Attempted to divide by zero. + at DotnetTests.Unit.Calculator.Div(Int32 a, Int32 b) in C:\Users\Michal\Workspace\dorny\test-reporter\reports\dotnet\DotnetTests.Unit\Calculator.cs:line 9 + at DotnetTests.XUnitTests.CalculatorTests.Exception_In_TargetTest() in C:\Users\Michal\Workspace\dorny\test-reporter\reports\dotnet\DotnetTests.XUnitTests\CalculatorTests.cs:line 33 ❌ Exception_In_Test System.Exception : Test + at DotnetTests.XUnitTests.CalculatorTests.Exception_In_Test() in C:\Users\Michal\Workspace\dorny\test-reporter\reports\dotnet\DotnetTests.XUnitTests\CalculatorTests.cs:line 39 ❌ Failing_Test Assert.Equal() Failure Expected: 3 Actual: 2 + at DotnetTests.XUnitTests.CalculatorTests.Failing_Test() in C:\Users\Michal\Workspace\dorny\test-reporter\reports\dotnet\DotnetTests.XUnitTests\CalculatorTests.cs:line 27 ✅ Is_Even_Number(i: 2) ❌ Is_Even_Number(i: 3) Assert.True() Failure Expected: True Actual: False + at DotnetTests.XUnitTests.CalculatorTests.Is_Even_Number(Int32 i) in C:\Users\Michal\Workspace\dorny\test-reporter\reports\dotnet\DotnetTests.XUnitTests\CalculatorTests.cs:line 59 ✅ Passing_Test ✅ Should be even number(i: 2) ❌ Should be even number(i: 3) Assert.True() Failure Expected: True Actual: False + at DotnetTests.XUnitTests.CalculatorTests.Theory_With_Custom_Name(Int32 i) in C:\Users\Michal\Workspace\dorny\test-reporter\reports\dotnet\DotnetTests.XUnitTests\CalculatorTests.cs:line 67 ⚪ Skipped_Test ✅ Timeout_Test ``` \ No newline at end of file diff --git a/__tests__/__outputs__/dotnet-xunitv3.md b/__tests__/__outputs__/dotnet-xunitv3.md new file mode 100644 index 0000000..f27f9e2 --- /dev/null +++ b/__tests__/__outputs__/dotnet-xunitv3.md @@ -0,0 +1,26 @@ +![Tests failed](https://img.shields.io/badge/tests-1%20passed%2C%203%20failed-critical) +|Report|Passed|Failed|Skipped|Time| +|:---|---:|---:|---:|---:| +|[fixtures/dotnet-xunitv3.trx](#user-content-r0)|1 ✅|3 ❌||267ms| +## ❌ fixtures/dotnet-xunitv3.trx +**4** tests were completed in **267ms** with **1** passed, **3** failed and **0** skipped. +|Test suite|Passed|Failed|Skipped|Time| +|:---|---:|---:|---:|---:| +|[DotnetTests.XUnitV3Tests.FixtureTests](#user-content-r0s0)|1 ✅|1 ❌||18ms| +|[Unclassified](#user-content-r0s1)||2 ❌||0ms| +### ❌ DotnetTests.XUnitV3Tests.FixtureTests +``` +❌ Failing_Test + Assert.Null() Failure: Value is not null + Expected: null + Actual: Fixture { } + at DotnetTests.XUnitV3Tests.FixtureTests.Failing_Test() in /_/reports/dotnet/DotnetTests.XUnitV3Tests/FixtureTests.cs:line 25 + at System.RuntimeMethodHandle.InvokeMethod(Object target, Void** arguments, Signature sig, Boolean isConstructor) + at System.Reflection.MethodBaseInvoker.InvokeWithNoArgs(Object obj, BindingFlags invokeAttr) +✅ Passing_Test +``` +### ❌ Unclassified +``` +❌ [Test Class Cleanup Failure (DotnetTests.XUnitV3Tests.FixtureTests.Failing_Test)] +❌ [Test Class Cleanup Failure (DotnetTests.XUnitV3Tests.FixtureTests.Passing_Test)] +``` \ No newline at end of file diff --git a/__tests__/__outputs__/fluent-validation-test-results.md b/__tests__/__outputs__/fluent-validation-test-results.md index 9337a62..83d0503 100644 --- a/__tests__/__outputs__/fluent-validation-test-results.md +++ b/__tests__/__outputs__/fluent-validation-test-results.md @@ -3,7 +3,7 @@ |Report|Passed|Failed|Skipped|Time| |:---|---:|---:|---:|---:| -|fixtures/external/FluentValidation.Tests.trx|803 ✅||1 ⚪|4s| +|[fixtures/external/FluentValidation.Tests.trx](#user-content-r0)|803 ✅||1 ⚪|4s| ## ✅ fixtures/external/FluentValidation.Tests.trx **804** tests were completed in **4s** with **803** passed, **0** failed and **1** skipped. |Test suite|Passed|Failed|Skipped|Time| diff --git a/__tests__/__outputs__/golang-json.md b/__tests__/__outputs__/golang-json.md index b3640d8..8b63704 100644 --- a/__tests__/__outputs__/golang-json.md +++ b/__tests__/__outputs__/golang-json.md @@ -1,7 +1,7 @@ ![Tests failed](https://img.shields.io/badge/tests-5%20passed%2C%206%20failed%2C%201%20skipped-critical) |Report|Passed|Failed|Skipped|Time| |:---|---:|---:|---:|---:| -|fixtures/golang-json.json|5 ✅|6 ❌|1 ⚪|6s| +|[fixtures/golang-json.json](#user-content-r0)|5 ✅|6 ❌|1 ⚪|6s| ## ❌ fixtures/golang-json.json **12** tests were completed in **6s** with **5** passed, **6** failed and **1** skipped. |Test suite|Passed|Failed|Skipped|Time| diff --git a/__tests__/__outputs__/jest-junit-eslint.md b/__tests__/__outputs__/jest-junit-eslint.md index d3ad9b9..5ebb57e 100644 --- a/__tests__/__outputs__/jest-junit-eslint.md +++ b/__tests__/__outputs__/jest-junit-eslint.md @@ -3,7 +3,7 @@ |Report|Passed|Failed|Skipped|Time| |:---|---:|---:|---:|---:| -|fixtures/jest-junit-eslint.xml|1 ✅|||0ms| +|[fixtures/jest-junit-eslint.xml](#user-content-r0)|1 ✅|||0ms| ## ✅ fixtures/jest-junit-eslint.xml **1** tests were completed in **0ms** with **1** passed, **0** failed and **0** skipped. |Test suite|Passed|Failed|Skipped|Time| diff --git a/__tests__/__outputs__/jest-junit.md b/__tests__/__outputs__/jest-junit.md index ed5a174..951256f 100644 --- a/__tests__/__outputs__/jest-junit.md +++ b/__tests__/__outputs__/jest-junit.md @@ -1,7 +1,7 @@ ![Tests failed](https://img.shields.io/badge/tests-1%20passed%2C%204%20failed%2C%201%20skipped-critical) |Report|Passed|Failed|Skipped|Time| |:---|---:|---:|---:|---:| -|fixtures/jest-junit.xml|1 ✅|4 ❌|1 ⚪|1s| +|[fixtures/jest-junit.xml](#user-content-r0)|1 ✅|4 ❌|1 ⚪|1s| ## ❌ fixtures/jest-junit.xml **6** tests were completed in **1s** with **1** passed, **4** failed and **1** skipped. |Test suite|Passed|Failed|Skipped|Time| diff --git a/__tests__/__outputs__/jest-react-component-test-results.md b/__tests__/__outputs__/jest-react-component-test-results.md index d71db4c..1365818 100644 --- a/__tests__/__outputs__/jest-react-component-test-results.md +++ b/__tests__/__outputs__/jest-react-component-test-results.md @@ -3,7 +3,7 @@ |Report|Passed|Failed|Skipped|Time| |:---|---:|---:|---:|---:| -|fixtures/external/jest/jest-react-component-test-results.xml|1 ✅|||1000ms| +|[fixtures/external/jest/jest-react-component-test-results.xml](#user-content-r0)|1 ✅|||1000ms| ## ✅ fixtures/external/jest/jest-react-component-test-results.xml **1** tests were completed in **1000ms** with **1** passed, **0** failed and **0** skipped. |Test suite|Passed|Failed|Skipped|Time| diff --git a/__tests__/__outputs__/jest-test-results.md b/__tests__/__outputs__/jest-test-results.md index 25dd567..cfbc169 100644 --- a/__tests__/__outputs__/jest-test-results.md +++ b/__tests__/__outputs__/jest-test-results.md @@ -1,7 +1,7 @@ ![Tests failed](https://img.shields.io/badge/tests-4207%20passed%2C%202%20failed%2C%2030%20skipped-critical) |Report|Passed|Failed|Skipped|Time| |:---|---:|---:|---:|---:| -|fixtures/external/jest/jest-test-results.xml|4207 ✅|2 ❌|30 ⚪|166s| +|[fixtures/external/jest/jest-test-results.xml](#user-content-r0)|4207 ✅|2 ❌|30 ⚪|166s| ## ❌ fixtures/external/jest/jest-test-results.xml **4239** tests were completed in **166s** with **4207** passed, **2** failed and **30** skipped. |Test suite|Passed|Failed|Skipped|Time| diff --git a/__tests__/__outputs__/junit-with-message.md b/__tests__/__outputs__/junit-with-message.md index 988d8fc..634a402 100644 --- a/__tests__/__outputs__/junit-with-message.md +++ b/__tests__/__outputs__/junit-with-message.md @@ -1,7 +1,7 @@ ![Tests failed](https://img.shields.io/badge/tests-1%20failed-critical) |Report|Passed|Failed|Skipped|Time| |:---|---:|---:|---:|---:| -|fixtures/junit-with-message.xml||1 ❌||1ms| +|[fixtures/junit-with-message.xml](#user-content-r0)||1 ❌||1ms| ## ❌ fixtures/junit-with-message.xml **1** tests were completed in **1ms** with **0** passed, **1** failed and **0** skipped. |Test suite|Passed|Failed|Skipped|Time| diff --git a/__tests__/__outputs__/mocha-json.md b/__tests__/__outputs__/mocha-json.md index 50419e1..875f881 100644 --- a/__tests__/__outputs__/mocha-json.md +++ b/__tests__/__outputs__/mocha-json.md @@ -1,7 +1,7 @@ ![Tests failed](https://img.shields.io/badge/tests-1%20passed%2C%204%20failed%2C%201%20skipped-critical) |Report|Passed|Failed|Skipped|Time| |:---|---:|---:|---:|---:| -|fixtures/mocha-json.json|1 ✅|4 ❌|1 ⚪|12ms| +|[fixtures/mocha-json.json](#user-content-r0)|1 ✅|4 ❌|1 ⚪|12ms| ## ❌ fixtures/mocha-json.json **6** tests were completed in **12ms** with **1** passed, **4** failed and **1** skipped. |Test suite|Passed|Failed|Skipped|Time| diff --git a/__tests__/__outputs__/mocha-test-results.md b/__tests__/__outputs__/mocha-test-results.md index 8831d7b..4a6e2f6 100644 --- a/__tests__/__outputs__/mocha-test-results.md +++ b/__tests__/__outputs__/mocha-test-results.md @@ -3,7 +3,7 @@ |Report|Passed|Failed|Skipped|Time| |:---|---:|---:|---:|---:| -|fixtures/external/mocha/mocha-test-results.json|833 ✅||6 ⚪|6s| +|[fixtures/external/mocha/mocha-test-results.json](#user-content-r0)|833 ✅||6 ⚪|6s| ## ✅ fixtures/external/mocha/mocha-test-results.json **839** tests were completed in **6s** with **833** passed, **0** failed and **6** skipped. |Test suite|Passed|Failed|Skipped|Time| diff --git a/__tests__/__outputs__/provider-test-results.md b/__tests__/__outputs__/provider-test-results.md index b2216ad..172f070 100644 --- a/__tests__/__outputs__/provider-test-results.md +++ b/__tests__/__outputs__/provider-test-results.md @@ -1,7 +1,7 @@ ![Tests failed](https://img.shields.io/badge/tests-268%20passed%2C%201%20failed-critical) |Report|Passed|Failed|Skipped|Time| |:---|---:|---:|---:|---:| -|fixtures/external/flutter/provider-test-results.json|268 ✅|1 ❌||0ms| +|[fixtures/external/flutter/provider-test-results.json](#user-content-r0)|268 ✅|1 ❌||0ms| ## ❌ fixtures/external/flutter/provider-test-results.json **269** tests were completed in **0ms** with **268** passed, **1** failed and **0** skipped. |Test suite|Passed|Failed|Skipped|Time| diff --git a/__tests__/__outputs__/pulsar-test-results-no-merge.md b/__tests__/__outputs__/pulsar-test-results-no-merge.md index b1738a4..de5a9b6 100644 --- a/__tests__/__outputs__/pulsar-test-results-no-merge.md +++ b/__tests__/__outputs__/pulsar-test-results-no-merge.md @@ -1,7 +1,7 @@ ![Tests failed](https://img.shields.io/badge/tests-1%20failed%2C%201%20skipped-critical) |Report|Passed|Failed|Skipped|Time| |:---|---:|---:|---:|---:| -|fixtures/external/java/TEST-org.apache.pulsar.AddMissingPatchVersionTest.xml||1 ❌|1 ⚪|116ms| +|[fixtures/external/java/TEST-org.apache.pulsar.AddMissingPatchVersionTest.xml](#user-content-r0)||1 ❌|1 ⚪|116ms| ## ❌ fixtures/external/java/TEST-org.apache.pulsar.AddMissingPatchVersionTest.xml **2** tests were completed in **116ms** with **0** passed, **1** failed and **1** skipped. |Test suite|Passed|Failed|Skipped|Time| diff --git a/__tests__/__outputs__/pulsar-test-results.md b/__tests__/__outputs__/pulsar-test-results.md index a9a5290..aaaa82e 100644 --- a/__tests__/__outputs__/pulsar-test-results.md +++ b/__tests__/__outputs__/pulsar-test-results.md @@ -1,7 +1,7 @@ ![Tests failed](https://img.shields.io/badge/tests-793%20passed%2C%201%20failed%2C%2014%20skipped-critical) |Report|Passed|Failed|Skipped|Time| |:---|---:|---:|---:|---:| -|fixtures/external/java/pulsar-test-report.xml|793 ✅|1 ❌|14 ⚪|2127s| +|[fixtures/external/java/pulsar-test-report.xml](#user-content-r0)|793 ✅|1 ❌|14 ⚪|2127s| ## ❌ fixtures/external/java/pulsar-test-report.xml **808** tests were completed in **2127s** with **793** passed, **1** failed and **14** skipped. |Test suite|Passed|Failed|Skipped|Time| diff --git a/__tests__/__outputs__/python-xunit-pytest.md b/__tests__/__outputs__/python-xunit-pytest.md new file mode 100644 index 0000000..7b13e28 --- /dev/null +++ b/__tests__/__outputs__/python-xunit-pytest.md @@ -0,0 +1,26 @@ +![Tests failed](https://img.shields.io/badge/tests-6%20passed%2C%202%20failed%2C%202%20skipped-critical) +|Report|Passed|Failed|Skipped|Time| +|:---|---:|---:|---:|---:| +|[fixtures/python-xunit-pytest.xml](#user-content-r0)|6 ✅|2 ❌|2 ⚪|19ms| +## ❌ fixtures/python-xunit-pytest.xml +**10** tests were completed in **19ms** with **6** passed, **2** failed and **2** skipped. +|Test suite|Passed|Failed|Skipped|Time| +|:---|---:|---:|---:|---:| +|[pytest](#user-content-r0s0)|6 ✅|2 ❌|2 ⚪|19ms| +### ❌ pytest +``` +tests.test_lib + ✅ test_always_pass + ✅ test_with_subtests + ✅ test_parameterized[param1] + ✅ test_parameterized[param2] + ⚪ test_always_skip + ❌ test_always_fail + assert False + ⚪ test_expected_failure + ❌ test_error + Exception: error + ✅ test_with_record_property +custom_classname + ✅ test_with_record_xml_attribute +``` \ No newline at end of file diff --git a/__tests__/__outputs__/python-xunit-unittest.md b/__tests__/__outputs__/python-xunit-unittest.md new file mode 100644 index 0000000..230d186 --- /dev/null +++ b/__tests__/__outputs__/python-xunit-unittest.md @@ -0,0 +1,23 @@ +![Tests failed](https://img.shields.io/badge/tests-4%20passed%2C%202%20failed%2C%202%20skipped-critical) +|Report|Passed|Failed|Skipped|Time| +|:---|---:|---:|---:|---:| +|[fixtures/python-xunit-unittest.xml](#user-content-r0)|4 ✅|2 ❌|2 ⚪|1ms| +## ❌ fixtures/python-xunit-unittest.xml +**8** tests were completed in **1ms** with **4** passed, **2** failed and **2** skipped. +|Test suite|Passed|Failed|Skipped|Time| +|:---|---:|---:|---:|---:| +|[TestAcme-20251114214921](#user-content-r0s0)|4 ✅|2 ❌|2 ⚪|1ms| +### ❌ TestAcme-20251114214921 +``` +TestAcme + ✅ test_always_pass + ✅ test_parameterized_0_param1 + ✅ test_parameterized_1_param2 + ✅ test_with_subtests + ❌ test_always_fail + AssertionError: failed + ❌ test_error + Exception: error + ⚪ test_always_skip + ⚪ test_expected_failure +``` \ No newline at end of file diff --git a/__tests__/__outputs__/rspec-json.md b/__tests__/__outputs__/rspec-json.md index 7444608..d64cf44 100644 --- a/__tests__/__outputs__/rspec-json.md +++ b/__tests__/__outputs__/rspec-json.md @@ -1,7 +1,7 @@ ![Tests failed](https://img.shields.io/badge/tests-1%20passed%2C%201%20failed%2C%201%20skipped-critical) |Report|Passed|Failed|Skipped|Time| |:---|---:|---:|---:|---:| -|fixtures/rspec-json.json|1 ✅|1 ❌|1 ⚪|0ms| +|[fixtures/rspec-json.json](#user-content-r0)|1 ✅|1 ❌|1 ⚪|0ms| ## ❌ fixtures/rspec-json.json **3** tests were completed in **0ms** with **1** passed, **1** failed and **1** skipped. |Test suite|Passed|Failed|Skipped|Time| diff --git a/__tests__/__outputs__/silent-notes-test-results.md b/__tests__/__outputs__/silent-notes-test-results.md index 34f5bab..e3abc49 100644 --- a/__tests__/__outputs__/silent-notes-test-results.md +++ b/__tests__/__outputs__/silent-notes-test-results.md @@ -3,7 +3,7 @@ |Report|Passed|Failed|Skipped|Time| |:---|---:|---:|---:|---:| -|fixtures/external/SilentNotes.trx|67 ✅||12 ⚪|1s| +|[fixtures/external/SilentNotes.trx](#user-content-r0)|67 ✅||12 ⚪|1s| ## ✅ fixtures/external/SilentNotes.trx **79** tests were completed in **1s** with **67** passed, **0** failed and **12** skipped. |Test suite|Passed|Failed|Skipped|Time| diff --git a/__tests__/__outputs__/swift-xunit.md b/__tests__/__outputs__/swift-xunit.md index 6f9ed46..b001151 100644 --- a/__tests__/__outputs__/swift-xunit.md +++ b/__tests__/__outputs__/swift-xunit.md @@ -1,7 +1,7 @@ ![Tests failed](https://img.shields.io/badge/tests-2%20passed%2C%201%20failed-critical) |Report|Passed|Failed|Skipped|Time| |:---|---:|---:|---:|---:| -|fixtures/swift-xunit.xml|2 ✅|1 ❌||220ms| +|[fixtures/swift-xunit.xml](#user-content-r0)|2 ✅|1 ❌||220ms| ## ❌ fixtures/swift-xunit.xml **3** tests were completed in **220ms** with **2** passed, **1** failed and **0** skipped. |Test suite|Passed|Failed|Skipped|Time| diff --git a/__tests__/__snapshots__/dart-json.test.ts.snap b/__tests__/__snapshots__/dart-json.test.ts.snap index a499822..88a7349 100644 --- a/__tests__/__snapshots__/dart-json.test.ts.snap +++ b/__tests__/__snapshots__/dart-json.test.ts.snap @@ -1,4 +1,4 @@ -// Jest Snapshot v1, https://goo.gl/fbAQLP +// Jest Snapshot v1, https://jestjs.io/docs/snapshot-testing exports[`dart-json tests matches report snapshot 1`] = ` TestRunResult { diff --git a/__tests__/__snapshots__/dotnet-nunit.test.ts.snap b/__tests__/__snapshots__/dotnet-nunit.test.ts.snap index 60d55f2..529f702 100644 --- a/__tests__/__snapshots__/dotnet-nunit.test.ts.snap +++ b/__tests__/__snapshots__/dotnet-nunit.test.ts.snap @@ -1,4 +1,4 @@ -// Jest Snapshot v1, https://goo.gl/fbAQLP +// Jest Snapshot v1, https://jestjs.io/docs/snapshot-testing exports[`dotnet-nunit tests report from ./reports/dotnet test results matches snapshot 1`] = ` TestRunResult { diff --git a/__tests__/__snapshots__/dotnet-trx.test.ts.snap b/__tests__/__snapshots__/dotnet-trx.test.ts.snap index 1ca07eb..b9d272d 100644 --- a/__tests__/__snapshots__/dotnet-trx.test.ts.snap +++ b/__tests__/__snapshots__/dotnet-trx.test.ts.snap @@ -1,6 +1,6 @@ -// Jest Snapshot v1, https://goo.gl/fbAQLP +// Jest Snapshot v1, https://jestjs.io/docs/snapshot-testing -exports[`dotnet-trx tests matches report snapshot 1`] = ` +exports[`dotnet-trx tests matches dotnet-trx report snapshot 1`] = ` TestRunResult { "path": "fixtures/dotnet-trx.trx", "suites": [ @@ -21,7 +21,9 @@ TestRunResult { at DotnetTests.Unit.Calculator.Div(Int32 a, Int32 b) in C:\\Users\\Michal\\Workspace\\dorny\\test-reporter\\reports\\dotnet\\DotnetTests.Unit\\Calculator.cs:line 9 at DotnetTests.XUnitTests.CalculatorTests.Exception_In_TargetTest() in C:\\Users\\Michal\\Workspace\\dorny\\test-reporter\\reports\\dotnet\\DotnetTests.XUnitTests\\CalculatorTests.cs:line 33", "line": 9, - "message": "System.DivideByZeroException : Attempted to divide by zero.", + "message": "System.DivideByZeroException : Attempted to divide by zero. + at DotnetTests.Unit.Calculator.Div(Int32 a, Int32 b) in C:\\Users\\Michal\\Workspace\\dorny\\test-reporter\\reports\\dotnet\\DotnetTests.Unit\\Calculator.cs:line 9 + at DotnetTests.XUnitTests.CalculatorTests.Exception_In_TargetTest() in C:\\Users\\Michal\\Workspace\\dorny\\test-reporter\\reports\\dotnet\\DotnetTests.XUnitTests\\CalculatorTests.cs:line 33", "path": "DotnetTests.Unit/Calculator.cs", }, "name": "Exception_In_TargetTest", @@ -33,7 +35,8 @@ TestRunResult { "details": "System.Exception : Test at DotnetTests.XUnitTests.CalculatorTests.Exception_In_Test() in C:\\Users\\Michal\\Workspace\\dorny\\test-reporter\\reports\\dotnet\\DotnetTests.XUnitTests\\CalculatorTests.cs:line 39", "line": 39, - "message": "System.Exception : Test", + "message": "System.Exception : Test + at DotnetTests.XUnitTests.CalculatorTests.Exception_In_Test() in C:\\Users\\Michal\\Workspace\\dorny\\test-reporter\\reports\\dotnet\\DotnetTests.XUnitTests\\CalculatorTests.cs:line 39", "path": "DotnetTests.XUnitTests/CalculatorTests.cs", }, "name": "Exception_In_Test", @@ -49,7 +52,8 @@ Actual: 2 "line": 27, "message": "Assert.Equal() Failure Expected: 3 -Actual: 2", +Actual: 2 + at DotnetTests.XUnitTests.CalculatorTests.Failing_Test() in C:\\Users\\Michal\\Workspace\\dorny\\test-reporter\\reports\\dotnet\\DotnetTests.XUnitTests\\CalculatorTests.cs:line 27", "path": "DotnetTests.XUnitTests/CalculatorTests.cs", }, "name": "Failing_Test", @@ -71,7 +75,8 @@ Actual: False "line": 59, "message": "Assert.True() Failure Expected: True -Actual: False", +Actual: False + at DotnetTests.XUnitTests.CalculatorTests.Is_Even_Number(Int32 i) in C:\\Users\\Michal\\Workspace\\dorny\\test-reporter\\reports\\dotnet\\DotnetTests.XUnitTests\\CalculatorTests.cs:line 59", "path": "DotnetTests.XUnitTests/CalculatorTests.cs", }, "name": "Is_Even_Number(i: 3)", @@ -99,7 +104,213 @@ Actual: False "line": 67, "message": "Assert.True() Failure Expected: True -Actual: False", +Actual: False + at DotnetTests.XUnitTests.CalculatorTests.Theory_With_Custom_Name(Int32 i) in C:\\Users\\Michal\\Workspace\\dorny\\test-reporter\\reports\\dotnet\\DotnetTests.XUnitTests\\CalculatorTests.cs:line 67", + "path": "DotnetTests.XUnitTests/CalculatorTests.cs", + }, + "name": "Should be even number(i: 3)", + "result": "failed", + "time": 0.6537000000000001, + }, + TestCaseResult { + "error": undefined, + "name": "Skipped_Test", + "result": "skipped", + "time": 1, + }, + TestCaseResult { + "error": undefined, + "name": "Timeout_Test", + "result": "success", + "time": 108.42580000000001, + }, + ], + }, + ], + "name": "DotnetTests.XUnitTests.CalculatorTests", + "totalTime": undefined, + }, + ], + "totalTime": 1116, +} +`; + +exports[`dotnet-trx tests matches dotnet-xunitv3 report snapshot 1`] = ` +TestRunResult { + "path": "fixtures/dotnet-xunitv3.trx", + "suites": [ + TestSuiteResult { + "groups": [ + TestGroupResult { + "name": null, + "tests": [ + TestCaseResult { + "error": { + "details": "Assert.Null() Failure: Value is not null +Expected: null +Actual: Fixture { } + at DotnetTests.XUnitV3Tests.FixtureTests.Failing_Test() in /_/reports/dotnet/DotnetTests.XUnitV3Tests/FixtureTests.cs:line 25 + at System.RuntimeMethodHandle.InvokeMethod(Object target, Void** arguments, Signature sig, Boolean isConstructor) + at System.Reflection.MethodBaseInvoker.InvokeWithNoArgs(Object obj, BindingFlags invokeAttr)", + "line": 25, + "message": "Assert.Null() Failure: Value is not null +Expected: null +Actual: Fixture { } + at DotnetTests.XUnitV3Tests.FixtureTests.Failing_Test() in /_/reports/dotnet/DotnetTests.XUnitV3Tests/FixtureTests.cs:line 25 + at System.RuntimeMethodHandle.InvokeMethod(Object target, Void** arguments, Signature sig, Boolean isConstructor) + at System.Reflection.MethodBaseInvoker.InvokeWithNoArgs(Object obj, BindingFlags invokeAttr)", + "path": "DotnetTests.XUnitV3Tests/FixtureTests.cs", + }, + "name": "Failing_Test", + "result": "failed", + "time": 17.0545, + }, + TestCaseResult { + "error": undefined, + "name": "Passing_Test", + "result": "success", + "time": 0.8786, + }, + ], + }, + ], + "name": "DotnetTests.XUnitV3Tests.FixtureTests", + "totalTime": undefined, + }, + TestSuiteResult { + "groups": [ + TestGroupResult { + "name": null, + "tests": [ + TestCaseResult { + "error": undefined, + "name": "[Test Class Cleanup Failure (DotnetTests.XUnitV3Tests.FixtureTests.Failing_Test)]", + "result": "failed", + "time": 0, + }, + TestCaseResult { + "error": undefined, + "name": "[Test Class Cleanup Failure (DotnetTests.XUnitV3Tests.FixtureTests.Passing_Test)]", + "result": "failed", + "time": 0, + }, + ], + }, + ], + "name": "Unclassified", + "totalTime": undefined, + }, + ], + "totalTime": 267, +} +`; + +exports[`dotnet-trx tests matches report snapshot (only failed tests) 1`] = ` +TestRunResult { + "path": "fixtures/dotnet-trx.trx", + "suites": [ + TestSuiteResult { + "groups": [ + TestGroupResult { + "name": null, + "tests": [ + TestCaseResult { + "error": undefined, + "name": "Custom Name", + "result": "success", + "time": 0.1371, + }, + TestCaseResult { + "error": { + "details": "System.DivideByZeroException : Attempted to divide by zero. + at DotnetTests.Unit.Calculator.Div(Int32 a, Int32 b) in C:\\Users\\Michal\\Workspace\\dorny\\test-reporter\\reports\\dotnet\\DotnetTests.Unit\\Calculator.cs:line 9 + at DotnetTests.XUnitTests.CalculatorTests.Exception_In_TargetTest() in C:\\Users\\Michal\\Workspace\\dorny\\test-reporter\\reports\\dotnet\\DotnetTests.XUnitTests\\CalculatorTests.cs:line 33", + "line": 9, + "message": "System.DivideByZeroException : Attempted to divide by zero. + at DotnetTests.Unit.Calculator.Div(Int32 a, Int32 b) in C:\\Users\\Michal\\Workspace\\dorny\\test-reporter\\reports\\dotnet\\DotnetTests.Unit\\Calculator.cs:line 9 + at DotnetTests.XUnitTests.CalculatorTests.Exception_In_TargetTest() in C:\\Users\\Michal\\Workspace\\dorny\\test-reporter\\reports\\dotnet\\DotnetTests.XUnitTests\\CalculatorTests.cs:line 33", + "path": "DotnetTests.Unit/Calculator.cs", + }, + "name": "Exception_In_TargetTest", + "result": "failed", + "time": 0.8377, + }, + TestCaseResult { + "error": { + "details": "System.Exception : Test + at DotnetTests.XUnitTests.CalculatorTests.Exception_In_Test() in C:\\Users\\Michal\\Workspace\\dorny\\test-reporter\\reports\\dotnet\\DotnetTests.XUnitTests\\CalculatorTests.cs:line 39", + "line": 39, + "message": "System.Exception : Test + at DotnetTests.XUnitTests.CalculatorTests.Exception_In_Test() in C:\\Users\\Michal\\Workspace\\dorny\\test-reporter\\reports\\dotnet\\DotnetTests.XUnitTests\\CalculatorTests.cs:line 39", + "path": "DotnetTests.XUnitTests/CalculatorTests.cs", + }, + "name": "Exception_In_Test", + "result": "failed", + "time": 2.5175, + }, + TestCaseResult { + "error": { + "details": "Assert.Equal() Failure +Expected: 3 +Actual: 2 + at DotnetTests.XUnitTests.CalculatorTests.Failing_Test() in C:\\Users\\Michal\\Workspace\\dorny\\test-reporter\\reports\\dotnet\\DotnetTests.XUnitTests\\CalculatorTests.cs:line 27", + "line": 27, + "message": "Assert.Equal() Failure +Expected: 3 +Actual: 2 + at DotnetTests.XUnitTests.CalculatorTests.Failing_Test() in C:\\Users\\Michal\\Workspace\\dorny\\test-reporter\\reports\\dotnet\\DotnetTests.XUnitTests\\CalculatorTests.cs:line 27", + "path": "DotnetTests.XUnitTests/CalculatorTests.cs", + }, + "name": "Failing_Test", + "result": "failed", + "time": 3.8697, + }, + TestCaseResult { + "error": undefined, + "name": "Is_Even_Number(i: 2)", + "result": "success", + "time": 0.0078, + }, + TestCaseResult { + "error": { + "details": "Assert.True() Failure +Expected: True +Actual: False + at DotnetTests.XUnitTests.CalculatorTests.Is_Even_Number(Int32 i) in C:\\Users\\Michal\\Workspace\\dorny\\test-reporter\\reports\\dotnet\\DotnetTests.XUnitTests\\CalculatorTests.cs:line 59", + "line": 59, + "message": "Assert.True() Failure +Expected: True +Actual: False + at DotnetTests.XUnitTests.CalculatorTests.Is_Even_Number(Int32 i) in C:\\Users\\Michal\\Workspace\\dorny\\test-reporter\\reports\\dotnet\\DotnetTests.XUnitTests\\CalculatorTests.cs:line 59", + "path": "DotnetTests.XUnitTests/CalculatorTests.cs", + }, + "name": "Is_Even_Number(i: 3)", + "result": "failed", + "time": 0.41409999999999997, + }, + TestCaseResult { + "error": undefined, + "name": "Passing_Test", + "result": "success", + "time": 0.1365, + }, + TestCaseResult { + "error": undefined, + "name": "Should be even number(i: 2)", + "result": "success", + "time": 0.0097, + }, + TestCaseResult { + "error": { + "details": "Assert.True() Failure +Expected: True +Actual: False + at DotnetTests.XUnitTests.CalculatorTests.Theory_With_Custom_Name(Int32 i) in C:\\Users\\Michal\\Workspace\\dorny\\test-reporter\\reports\\dotnet\\DotnetTests.XUnitTests\\CalculatorTests.cs:line 67", + "line": 67, + "message": "Assert.True() Failure +Expected: True +Actual: False + at DotnetTests.XUnitTests.CalculatorTests.Theory_With_Custom_Name(Int32 i) in C:\\Users\\Michal\\Workspace\\dorny\\test-reporter\\reports\\dotnet\\DotnetTests.XUnitTests\\CalculatorTests.cs:line 67", "path": "DotnetTests.XUnitTests/CalculatorTests.cs", }, "name": "Should be even number(i: 3)", diff --git a/__tests__/__snapshots__/golang-json.test.ts.snap b/__tests__/__snapshots__/golang-json.test.ts.snap index 75c6de1..bd28d4a 100644 --- a/__tests__/__snapshots__/golang-json.test.ts.snap +++ b/__tests__/__snapshots__/golang-json.test.ts.snap @@ -1,4 +1,4 @@ -// Jest Snapshot v1, https://goo.gl/fbAQLP +// Jest Snapshot v1, https://jestjs.io/docs/snapshot-testing exports[`golang-json tests report from ./reports/dotnet test results matches snapshot 1`] = ` TestRunResult { diff --git a/__tests__/__snapshots__/java-junit.test.ts.snap b/__tests__/__snapshots__/java-junit.test.ts.snap index 341b092..38daca9 100644 --- a/__tests__/__snapshots__/java-junit.test.ts.snap +++ b/__tests__/__snapshots__/java-junit.test.ts.snap @@ -1,4 +1,4 @@ -// Jest Snapshot v1, https://goo.gl/fbAQLP +// Jest Snapshot v1, https://jestjs.io/docs/snapshot-testing exports[`java-junit tests report from apache/pulsar single suite test results matches snapshot 1`] = ` TestRunResult { diff --git a/__tests__/__snapshots__/jest-junit.test.ts.snap b/__tests__/__snapshots__/jest-junit.test.ts.snap index eb20dfe..eca0092 100644 --- a/__tests__/__snapshots__/jest-junit.test.ts.snap +++ b/__tests__/__snapshots__/jest-junit.test.ts.snap @@ -1,4 +1,4 @@ -// Jest Snapshot v1, https://goo.gl/fbAQLP +// Jest Snapshot v1, https://jestjs.io/docs/snapshot-testing exports[`jest-junit tests parsing ESLint report without timing information works - PR #134 1`] = ` TestRunResult { diff --git a/__tests__/__snapshots__/mocha-json.test.ts.snap b/__tests__/__snapshots__/mocha-json.test.ts.snap index 7038239..4a1448c 100644 --- a/__tests__/__snapshots__/mocha-json.test.ts.snap +++ b/__tests__/__snapshots__/mocha-json.test.ts.snap @@ -1,4 +1,4 @@ -// Jest Snapshot v1, https://goo.gl/fbAQLP +// Jest Snapshot v1, https://jestjs.io/docs/snapshot-testing exports[`mocha-json tests report from ./reports/mocha-json test results matches snapshot 1`] = ` TestRunResult { diff --git a/__tests__/__snapshots__/python-xunit.test.ts.snap b/__tests__/__snapshots__/python-xunit.test.ts.snap new file mode 100644 index 0000000..f325c84 --- /dev/null +++ b/__tests__/__snapshots__/python-xunit.test.ts.snap @@ -0,0 +1,192 @@ +// Jest Snapshot v1, https://jestjs.io/docs/snapshot-testing + +exports[`python-xunit pytest report report from python test results matches snapshot 1`] = ` +TestRunResult { + "path": "fixtures/python-xunit-pytest.xml", + "suites": [ + TestSuiteResult { + "groups": [ + TestGroupResult { + "name": "tests.test_lib", + "tests": [ + TestCaseResult { + "error": undefined, + "name": "test_always_pass", + "result": "success", + "time": 2, + }, + TestCaseResult { + "error": undefined, + "name": "test_with_subtests", + "result": "success", + "time": 5, + }, + TestCaseResult { + "error": undefined, + "name": "test_parameterized[param1]", + "result": "success", + "time": 0, + }, + TestCaseResult { + "error": undefined, + "name": "test_parameterized[param2]", + "result": "success", + "time": 0, + }, + TestCaseResult { + "error": undefined, + "name": "test_always_skip", + "result": "skipped", + "time": 0, + }, + TestCaseResult { + "error": { + "details": "def test_always_fail(): + > assert False + E assert False + + tests/test_lib.py:25: AssertionError + ", + "line": undefined, + "message": "assert False", + "path": undefined, + }, + "name": "test_always_fail", + "result": "failed", + "time": 0, + }, + TestCaseResult { + "error": undefined, + "name": "test_expected_failure", + "result": "skipped", + "time": 0, + }, + TestCaseResult { + "error": { + "details": "def test_error(): + > raise Exception("error") + E Exception: error + + tests/test_lib.py:32: Exception + ", + "line": undefined, + "message": "Exception: error", + "path": undefined, + }, + "name": "test_error", + "result": "failed", + "time": 0, + }, + TestCaseResult { + "error": undefined, + "name": "test_with_record_property", + "result": "success", + "time": 0, + }, + ], + }, + TestGroupResult { + "name": "custom_classname", + "tests": [ + TestCaseResult { + "error": undefined, + "name": "test_with_record_xml_attribute", + "result": "success", + "time": 0, + }, + ], + }, + ], + "name": "pytest", + "totalTime": 19, + }, + ], + "totalTime": undefined, +} +`; + +exports[`python-xunit unittest report report from python test results matches snapshot 1`] = ` +TestRunResult { + "path": "fixtures/python-xunit-unittest.xml", + "suites": [ + TestSuiteResult { + "groups": [ + TestGroupResult { + "name": "TestAcme", + "tests": [ + TestCaseResult { + "error": undefined, + "name": "test_always_pass", + "result": "success", + "time": 0, + }, + TestCaseResult { + "error": undefined, + "name": "test_parameterized_0_param1", + "result": "success", + "time": 1, + }, + TestCaseResult { + "error": undefined, + "name": "test_parameterized_1_param2", + "result": "success", + "time": 0, + }, + TestCaseResult { + "error": undefined, + "name": "test_with_subtests", + "result": "success", + "time": 0, + }, + TestCaseResult { + "error": { + "details": "Traceback (most recent call last): + File "/Users/foo/Projects/python-test/tests/test_lib.py", line 24, in test_always_fail + self.fail("failed") +AssertionError: failed +", + "line": undefined, + "message": "AssertionError: failed", + "path": undefined, + }, + "name": "test_always_fail", + "result": "failed", + "time": 0, + }, + TestCaseResult { + "error": { + "details": "Traceback (most recent call last): + File "/Users/foo/Projects/python-test/tests/test_lib.py", line 31, in test_error + raise Exception("error") +Exception: error +", + "line": undefined, + "message": "Exception: error", + "path": undefined, + }, + "name": "test_error", + "result": "failed", + "time": 0, + }, + TestCaseResult { + "error": undefined, + "name": "test_always_skip", + "result": "skipped", + "time": 0, + }, + TestCaseResult { + "error": undefined, + "name": "test_expected_failure", + "result": "skipped", + "time": 0, + }, + ], + }, + ], + "name": "TestAcme-20251114214921", + "totalTime": 1, + }, + ], + "totalTime": 1, +} +`; diff --git a/__tests__/__snapshots__/rspec-json.test.ts.snap b/__tests__/__snapshots__/rspec-json.test.ts.snap index cc14bfb..51c1943 100644 --- a/__tests__/__snapshots__/rspec-json.test.ts.snap +++ b/__tests__/__snapshots__/rspec-json.test.ts.snap @@ -1,4 +1,4 @@ -// Jest Snapshot v1, https://goo.gl/fbAQLP +// Jest Snapshot v1, https://jestjs.io/docs/snapshot-testing exports[`rspec-json tests report from ./reports/rspec-json test results matches snapshot 1`] = ` TestRunResult { diff --git a/__tests__/__snapshots__/swift-xunit.test.ts.snap b/__tests__/__snapshots__/swift-xunit.test.ts.snap index ae34deb..bddc6ea 100644 --- a/__tests__/__snapshots__/swift-xunit.test.ts.snap +++ b/__tests__/__snapshots__/swift-xunit.test.ts.snap @@ -1,4 +1,4 @@ -// Jest Snapshot v1, https://goo.gl/fbAQLP +// Jest Snapshot v1, https://jestjs.io/docs/snapshot-testing exports[`swift-xunit tests report from swift test results matches snapshot 1`] = ` TestRunResult { diff --git a/__tests__/dotnet-trx.test.ts b/__tests__/dotnet-trx.test.ts index e7f83ee..a7aead6 100644 --- a/__tests__/dotnet-trx.test.ts +++ b/__tests__/dotnet-trx.test.ts @@ -3,7 +3,7 @@ import * as path from 'path' import {DotnetTrxParser} from '../src/parsers/dotnet-trx/dotnet-trx-parser' import {ParseOptions} from '../src/test-parser' -import {DEFAULT_OPTIONS, getReport} from '../src/report/get-report' +import {DEFAULT_OPTIONS, getReport, ReportOptions} from '../src/report/get-report' import {normalizeFilePath} from '../src/utils/path-utils' describe('dotnet-trx tests', () => { @@ -39,9 +39,34 @@ describe('dotnet-trx tests', () => { expect(result.result).toBe('success') }) - it('matches report snapshot', async () => { + it.each([['dotnet-trx'], ['dotnet-xunitv3']])('matches %s report snapshot', async reportName => { + const fixturePath = path.join(__dirname, 'fixtures', `${reportName}.trx`) + const outputPath = path.join(__dirname, '__outputs__', `${reportName}.md`) + const filePath = normalizeFilePath(path.relative(__dirname, fixturePath)) + const fileContent = fs.readFileSync(fixturePath, {encoding: 'utf8'}) + + const opts: ParseOptions = { + parseErrors: true, + trackedFiles: [ + 'DotnetTests.Unit/Calculator.cs', + 'DotnetTests.XUnitTests/CalculatorTests.cs', + 'DotnetTests.XUnitV3Tests/FixtureTests.cs' + ] + //workDir: 'C:/Users/Michal/Workspace/dorny/test-check/reports/dotnet/' + } + + const parser = new DotnetTrxParser(opts) + const result = await parser.parse(filePath, fileContent) + expect(result).toMatchSnapshot() + + const report = getReport([result]) + fs.mkdirSync(path.dirname(outputPath), {recursive: true}) + fs.writeFileSync(outputPath, report) + }) + + it('matches report snapshot (only failed tests)', async () => { const fixturePath = path.join(__dirname, 'fixtures', 'dotnet-trx.trx') - const outputPath = path.join(__dirname, '__outputs__', 'dotnet-trx.md') + const outputPath = path.join(__dirname, '__outputs__', 'dotnet-trx-only-failed.md') const filePath = normalizeFilePath(path.relative(__dirname, fixturePath)) const fileContent = fs.readFileSync(fixturePath, {encoding: 'utf8'}) @@ -55,7 +80,12 @@ describe('dotnet-trx tests', () => { const result = await parser.parse(filePath, fileContent) expect(result).toMatchSnapshot() - const report = getReport([result]) + const reportOptions: ReportOptions = { + ...DEFAULT_OPTIONS, + listSuites: 'all', + listTests: 'failed' + } + const report = getReport([result], reportOptions) fs.mkdirSync(path.dirname(outputPath), {recursive: true}) fs.writeFileSync(outputPath, report) }) diff --git a/__tests__/fixtures/dotnet-xunitv3.trx b/__tests__/fixtures/dotnet-xunitv3.trx new file mode 100644 index 0000000..6bd9c25 --- /dev/null +++ b/__tests__/fixtures/dotnet-xunitv3.trx @@ -0,0 +1,60 @@ + + + + + + + + + + + Assert.Null() Failure: Value is not null +Expected: null +Actual: Fixture { } + at DotnetTests.XUnitV3Tests.FixtureTests.Failing_Test() in /_/reports/dotnet/DotnetTests.XUnitV3Tests/FixtureTests.cs:line 25 + at System.RuntimeMethodHandle.InvokeMethod(Object target, Void** arguments, Signature sig, Boolean isConstructor) + at System.Reflection.MethodBaseInvoker.InvokeWithNoArgs(Object obj, BindingFlags invokeAttr) + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + Exit code indicates failure: '2'. Please refer to https://aka.ms/testingplatform/exitcodes for more information. + + + + \ No newline at end of file diff --git a/__tests__/fixtures/python-xunit-pytest.xml b/__tests__/fixtures/python-xunit-pytest.xml new file mode 100644 index 0000000..fcb044a --- /dev/null +++ b/__tests__/fixtures/python-xunit-pytest.xml @@ -0,0 +1,42 @@ + + + + + + + + + + + + /Users/mike/Projects/python-test/tests/test_lib.py:20: skipped + + + + def test_always_fail(): + > assert False + E assert False + + tests/test_lib.py:25: AssertionError + + + + + + + def test_error(): + > raise Exception("error") + E Exception: error + + tests/test_lib.py:32: Exception + + + + + + + + + + diff --git a/__tests__/fixtures/python-xunit-unittest.xml b/__tests__/fixtures/python-xunit-unittest.xml new file mode 100644 index 0000000..ecc67d4 --- /dev/null +++ b/__tests__/fixtures/python-xunit-unittest.xml @@ -0,0 +1,27 @@ + + + + + + + + + + + + + + + + + + + diff --git a/__tests__/jest-junit.test.ts b/__tests__/jest-junit.test.ts index f4b8335..912ebde 100644 --- a/__tests__/jest-junit.test.ts +++ b/__tests__/jest-junit.test.ts @@ -207,4 +207,143 @@ describe('jest-junit tests', () => { // Report should have the title as the first line expect(report).toMatch(/^# My Custom Title\n/) }) + + it('report can be collapsed when configured', async () => { + const fixturePath = path.join(__dirname, 'fixtures', 'jest-junit.xml') + const filePath = normalizeFilePath(path.relative(__dirname, fixturePath)) + const fileContent = fs.readFileSync(fixturePath, {encoding: 'utf8'}) + + const opts: ParseOptions = { + parseErrors: true, + trackedFiles: [] + } + + const parser = new JestJunitParser(opts) + const result = await parser.parse(filePath, fileContent) + const report = getReport([result], { + ...DEFAULT_OPTIONS, + collapsed: 'always' + }) + // Report should include collapsible details + expect(report).toContain('
Expand for details') + expect(report).toContain('
') + }) + + it('report is not collapsed when configured to never', async () => { + const fixturePath = path.join(__dirname, 'fixtures', 'jest-junit.xml') + const filePath = normalizeFilePath(path.relative(__dirname, fixturePath)) + const fileContent = fs.readFileSync(fixturePath, {encoding: 'utf8'}) + + const opts: ParseOptions = { + parseErrors: true, + trackedFiles: [] + } + + const parser = new JestJunitParser(opts) + const result = await parser.parse(filePath, fileContent) + const report = getReport([result], { + ...DEFAULT_OPTIONS, + collapsed: 'never' + }) + // Report should not include collapsible details + expect(report).not.toContain('
Expand for details') + expect(report).not.toContain('
') + }) + + it('report auto-collapses when all tests pass', async () => { + // Test with a fixture that has all passing tests (no failures) + const fixturePath = path.join(__dirname, 'fixtures', 'jest-junit-eslint.xml') + const filePath = normalizeFilePath(path.relative(__dirname, fixturePath)) + const fileContent = fs.readFileSync(fixturePath, {encoding: 'utf8'}) + + const opts: ParseOptions = { + parseErrors: true, + trackedFiles: [] + } + + const parser = new JestJunitParser(opts) + const result = await parser.parse(filePath, fileContent) + + // Verify this fixture has no failures + expect(result.failed).toBe(0) + + const report = getReport([result], { + ...DEFAULT_OPTIONS, + collapsed: 'auto' + }) + + // Should collapse when all tests pass + expect(report).toContain('
Expand for details') + expect(report).toContain('
') + }) + + it('report does not auto-collapse when tests fail', async () => { + // Test with a fixture that has failing tests + const fixturePath = path.join(__dirname, 'fixtures', 'jest-junit.xml') + const filePath = normalizeFilePath(path.relative(__dirname, fixturePath)) + const fileContent = fs.readFileSync(fixturePath, {encoding: 'utf8'}) + + const opts: ParseOptions = { + parseErrors: true, + trackedFiles: [] + } + + const parser = new JestJunitParser(opts) + const result = await parser.parse(filePath, fileContent) + + // Verify this fixture has failures + expect(result.failed).toBeGreaterThan(0) + + const report = getReport([result], { + ...DEFAULT_OPTIONS, + collapsed: 'auto' + }) + + // Should not collapse when there are failures + expect(report).not.toContain('
Expand for details') + expect(report).not.toContain('
') + }) + + it('report includes the short summary', async () => { + const fixturePath = path.join(__dirname, 'fixtures', 'jest-junit.xml') + const filePath = normalizeFilePath(path.relative(__dirname, fixturePath)) + const fileContent = fs.readFileSync(fixturePath, {encoding: 'utf8'}) + + const opts: ParseOptions = { + parseErrors: true, + trackedFiles: [] + } + + const parser = new JestJunitParser(opts) + const result = await parser.parse(filePath, fileContent) + const shortSummary = '1 passed, 4 failed and 1 skipped' + const report = getReport([result], DEFAULT_OPTIONS, shortSummary) + // Report should have the title as the first line + expect(report).toMatch(/^## 1 passed, 4 failed and 1 skipped\n/) + }) + + it('report includes a custom report title and short summary', async () => { + const fixturePath = path.join(__dirname, 'fixtures', 'jest-junit.xml') + const filePath = normalizeFilePath(path.relative(__dirname, fixturePath)) + const fileContent = fs.readFileSync(fixturePath, {encoding: 'utf8'}) + + const opts: ParseOptions = { + parseErrors: true, + trackedFiles: [] + } + + const parser = new JestJunitParser(opts) + const result = await parser.parse(filePath, fileContent) + const shortSummary = '1 passed, 4 failed and 1 skipped' + const report = getReport( + [result], + { + ...DEFAULT_OPTIONS, + reportTitle: 'My Custom Title' + }, + shortSummary + ) + // Report should have the title as the first line + expect(report).toMatch(/^# My Custom Title\n## 1 passed, 4 failed and 1 skipped\n/) + }) }) diff --git a/__tests__/python-xunit.test.ts b/__tests__/python-xunit.test.ts new file mode 100644 index 0000000..c1550a4 --- /dev/null +++ b/__tests__/python-xunit.test.ts @@ -0,0 +1,93 @@ +import * as fs from 'fs' +import * as path from 'path' + +import {PythonXunitParser} from '../src/parsers/python-xunit/python-xunit-parser' +import {ParseOptions} from '../src/test-parser' +import {DEFAULT_OPTIONS, getReport} from '../src/report/get-report' +import {normalizeFilePath} from '../src/utils/path-utils' + +const defaultOpts: ParseOptions = { + parseErrors: true, + trackedFiles: [] +} + +describe('python-xunit unittest report', () => { + const fixturePath = path.join(__dirname, 'fixtures', 'python-xunit-unittest.xml') + const filePath = normalizeFilePath(path.relative(__dirname, fixturePath)) + const fileContent = fs.readFileSync(fixturePath, {encoding: 'utf8'}) + const outputPath = path.join(__dirname, '__outputs__', 'python-xunit-unittest.md') + + it('report from python test results matches snapshot', async () => { + const trackedFiles = ['tests/test_lib.py'] + const opts: ParseOptions = { + ...defaultOpts, + trackedFiles + } + + const parser = new PythonXunitParser(opts) + const result = await parser.parse(filePath, fileContent) + expect(result).toMatchSnapshot() + + const report = getReport([result]) + fs.mkdirSync(path.dirname(outputPath), {recursive: true}) + fs.writeFileSync(outputPath, report) + }) + + it('report does not include a title by default', async () => { + const parser = new PythonXunitParser(defaultOpts) + const result = await parser.parse(filePath, fileContent) + const report = getReport([result]) + // Report should have the badge as the first line + expect(report).toMatch(/^!\[Tests failed]/) + }) + + it.each([ + ['empty string', ''], + ['space', ' '], + ['tab', '\t'], + ['newline', '\n'] + ])('report does not include a title when configured value is %s', async (_, reportTitle) => { + const parser = new PythonXunitParser(defaultOpts) + const result = await parser.parse(filePath, fileContent) + const report = getReport([result], { + ...DEFAULT_OPTIONS, + reportTitle + }) + // Report should have the badge as the first line + expect(report).toMatch(/^!\[Tests failed]/) + }) + + it('report includes a custom report title', async () => { + const parser = new PythonXunitParser(defaultOpts) + const result = await parser.parse(filePath, fileContent) + const report = getReport([result], { + ...DEFAULT_OPTIONS, + reportTitle: 'My Custom Title' + }) + // Report should have the title as the first line + expect(report).toMatch(/^# My Custom Title\n/) + }) +}) + +describe('python-xunit pytest report', () => { + const fixturePath = path.join(__dirname, 'fixtures', 'python-xunit-pytest.xml') + const filePath = normalizeFilePath(path.relative(__dirname, fixturePath)) + const fileContent = fs.readFileSync(fixturePath, {encoding: 'utf8'}) + const outputPath = path.join(__dirname, '__outputs__', 'python-xunit-pytest.md') + + it('report from python test results matches snapshot', async () => { + const trackedFiles = ['tests/test_lib.py'] + const opts: ParseOptions = { + ...defaultOpts, + trackedFiles + } + + const parser = new PythonXunitParser(opts) + const result = await parser.parse(filePath, fileContent) + expect(result).toMatchSnapshot() + + const report = getReport([result]) + fs.mkdirSync(path.dirname(outputPath), {recursive: true}) + fs.writeFileSync(outputPath, report) + }) +}) diff --git a/__tests__/report/get-report.test.ts b/__tests__/report/get-report.test.ts new file mode 100644 index 0000000..670b0ad --- /dev/null +++ b/__tests__/report/get-report.test.ts @@ -0,0 +1,120 @@ +import {getBadge, DEFAULT_OPTIONS, ReportOptions} from '../../src/report/get-report' + +describe('getBadge', () => { + describe('URI encoding with special characters', () => { + it('generates correct URI with simple badge title', () => { + const options: ReportOptions = { + ...DEFAULT_OPTIONS, + badgeTitle: 'tests' + } + const badge = getBadge(5, 0, 1, options) + expect(badge).toBe('![Tests passed successfully](https://img.shields.io/badge/tests-5%20passed%2C%201%20skipped-success)') + }) + + it('handles badge title with single hyphen', () => { + const options: ReportOptions = { + ...DEFAULT_OPTIONS, + badgeTitle: 'unit-tests' + } + const badge = getBadge(3, 0, 0, options) + // The hyphen in the badge title should be encoded as -- + expect(badge).toBe('![Tests passed successfully](https://img.shields.io/badge/unit--tests-3%20passed-success)') + }) + + it('handles badge title with multiple hyphens', () => { + const options: ReportOptions = { + ...DEFAULT_OPTIONS, + badgeTitle: 'integration-api-tests' + } + const badge = getBadge(10, 0, 0, options) + // All hyphens in the title should be encoded as -- + expect(badge).toBe('![Tests passed successfully](https://img.shields.io/badge/integration--api--tests-10%20passed-success)') + }) + + it('handles badge title with multiple underscores', () => { + const options: ReportOptions = { + ...DEFAULT_OPTIONS, + badgeTitle: 'my_integration_test' + } + const badge = getBadge(10, 0, 0, options) + // All underscores in the title should be encoded as __ + expect(badge).toBe('![Tests passed successfully](https://img.shields.io/badge/my__integration__test-10%20passed-success)') + }) + + it('handles badge title with version format containing hyphen', () => { + const options: ReportOptions = { + ...DEFAULT_OPTIONS, + badgeTitle: 'MariaDb 12.0-ubi database tests' + } + const badge = getBadge(1, 0, 0, options) + // The hyphen in "12.0-ubi" should be encoded as -- + expect(badge).toBe('![Tests passed successfully](https://img.shields.io/badge/MariaDb%2012.0--ubi%20database%20tests-1%20passed-success)') + }) + + it('handles badge title with dots and hyphens', () => { + const options: ReportOptions = { + ...DEFAULT_OPTIONS, + badgeTitle: 'v1.2.3-beta-test' + } + const badge = getBadge(4, 1, 0, options) + expect(badge).toBe('![Tests failed](https://img.shields.io/badge/v1.2.3--beta--test-4%20passed%2C%201%20failed-critical)') + }) + + it('preserves structural hyphens between label and message', () => { + const options: ReportOptions = { + ...DEFAULT_OPTIONS, + badgeTitle: 'test-suite' + } + const badge = getBadge(2, 3, 1, options) + // The URI should have literal hyphens separating title-message-color + expect(badge).toBe('![Tests failed](https://img.shields.io/badge/test--suite-2%20passed%2C%203%20failed%2C%201%20skipped-critical)') + }) + }) + + describe('generates test outcome as color name for imgshields', () => { + it('uses success color when all tests pass', () => { + const options: ReportOptions = {...DEFAULT_OPTIONS} + const badge = getBadge(5, 0, 0, options) + expect(badge).toContain('-success)') + }) + + it('uses critical color when tests fail', () => { + const options: ReportOptions = {...DEFAULT_OPTIONS} + const badge = getBadge(5, 2, 0, options) + expect(badge).toContain('-critical)') + }) + + it('uses yellow color when no tests found', () => { + const options: ReportOptions = {...DEFAULT_OPTIONS} + const badge = getBadge(0, 0, 0, options) + expect(badge).toContain('-yellow)') + }) + }) + + describe('badge message composition', () => { + it('includes only passed count when no failures or skips', () => { + const options: ReportOptions = {...DEFAULT_OPTIONS} + const badge = getBadge(5, 0, 0, options) + expect(badge).toBe('![Tests passed successfully](https://img.shields.io/badge/tests-5%20passed-success)') + }) + + it('includes passed and failed counts', () => { + const options: ReportOptions = {...DEFAULT_OPTIONS} + const badge = getBadge(5, 2, 0, options) + expect(badge).toBe('![Tests failed](https://img.shields.io/badge/tests-5%20passed%2C%202%20failed-critical)') + }) + + it('includes passed, failed and skipped counts', () => { + const options: ReportOptions = {...DEFAULT_OPTIONS} + const badge = getBadge(5, 2, 1, options) + expect(badge).toBe('![Tests failed](https://img.shields.io/badge/tests-5%20passed%2C%202%20failed%2C%201%20skipped-critical)') + }) + + it('uses "none" message when no tests', () => { + const options: ReportOptions = {...DEFAULT_OPTIONS} + const badge = getBadge(0, 0, 0, options) + expect(badge).toBe('![Tests passed successfully](https://img.shields.io/badge/tests-none-yellow)') + }) + }) +}) + diff --git a/__tests__/utils/parse-utils.test.ts b/__tests__/utils/parse-utils.test.ts index 83689ef..0f02867 100644 --- a/__tests__/utils/parse-utils.test.ts +++ b/__tests__/utils/parse-utils.test.ts @@ -32,6 +32,6 @@ describe('parseNetDuration', () => { }) it('throws when string has invalid format', () => { - expect(() => parseNetDuration('12:34:56 not a duration')).toThrowError(/^Invalid format/) + expect(() => parseNetDuration('12:34:56 not a duration')).toThrow(/^Invalid format/) }) }) diff --git a/action.yml b/action.yml index ec4772f..530435c 100644 --- a/action.yml +++ b/action.yml @@ -1,6 +1,5 @@ name: Test Reporter -description: | - Shows test results in GitHub UI: .NET (xUnit, NUnit, MSTest), Dart, Flutter, Java (JUnit), JavaScript (JEST, Mocha) +description: Displays test results from popular testing frameworks directly in GitHub author: Michal Dorner inputs: artifact: @@ -29,9 +28,11 @@ inputs: - dotnet-nunit - dotnet-trx - flutter-json + - golang-json - java-junit - jest-junit - mocha-json + - python-xunit - rspec-json - swift-xunit required: true @@ -68,6 +69,10 @@ inputs: working-directory: description: Relative path under $GITHUB_WORKSPACE where the repository was checked out required: false + report-title: + description: Title for the test report summary + required: false + default: '' only-summary: description: | Allows you to generate only the summary. @@ -85,6 +90,14 @@ inputs: description: Customize badge title required: false default: 'tests' + collapsed: + description: | + Controls whether test report details are collapsed or expanded. Supported options: + - auto: Collapse only if all tests pass (default behavior) + - always: Always collapse the report details + - never: Always expand the report details + required: false + default: 'auto' token: description: GitHub Access Token required: false diff --git a/dist/index.js b/dist/index.js index 9cb9528..42cb52e 100644 --- a/dist/index.js +++ b/dist/index.js @@ -277,6 +277,7 @@ const golang_json_parser_1 = __nccwpck_require__(5162); const java_junit_parser_1 = __nccwpck_require__(8342); const jest_junit_parser_1 = __nccwpck_require__(1042); const mocha_json_parser_1 = __nccwpck_require__(5402); +const python_xunit_parser_1 = __nccwpck_require__(6578); const rspec_json_parser_1 = __nccwpck_require__(9768); const swift_xunit_parser_1 = __nccwpck_require__(7330); const path_utils_1 = __nccwpck_require__(9132); @@ -309,6 +310,7 @@ class TestReporter { useActionsSummary = core.getInput('use-actions-summary', { required: false }) === 'true'; badgeTitle = core.getInput('badge-title', { required: false }); reportTitle = core.getInput('report-title', { required: false }); + collapsed = core.getInput('collapsed', { required: false }); token = core.getInput('token', { required: true }); octokit; context = (0, github_utils_1.getCheckRunContext)(); @@ -322,6 +324,10 @@ class TestReporter { core.setFailed(`Input parameter 'list-tests' has invalid value`); return; } + if (this.collapsed !== 'auto' && this.collapsed !== 'always' && this.collapsed !== 'never') { + core.setFailed(`Input parameter 'collapsed' has invalid value`); + return; + } if (isNaN(this.maxAnnotations) || this.maxAnnotations < 0 || this.maxAnnotations > 50) { core.setFailed(`Input parameter 'max-annotations' has invalid value`); return; @@ -401,7 +407,11 @@ class TestReporter { throw error; } } - const { listSuites, listTests, onlySummary, useActionsSummary, badgeTitle, reportTitle } = this; + const { listSuites, listTests, onlySummary, useActionsSummary, badgeTitle, reportTitle, collapsed } = this; + const passed = results.reduce((sum, tr) => sum + tr.passed, 0); + const failed = results.reduce((sum, tr) => sum + tr.failed, 0); + const skipped = results.reduce((sum, tr) => sum + tr.skipped, 0); + const shortSummary = `${passed} passed, ${failed} failed and ${skipped} skipped `; let baseUrl = ''; if (this.useActionsSummary) { const summary = (0, get_report_1.getReport)(results, { @@ -411,8 +421,9 @@ class TestReporter { onlySummary, useActionsSummary, badgeTitle, - reportTitle - }); + reportTitle, + collapsed + }, shortSummary); core.info('Summary content:'); core.info(summary); await core.summary.addRaw(summary).write(); @@ -438,16 +449,13 @@ class TestReporter { onlySummary, useActionsSummary, badgeTitle, - reportTitle + reportTitle, + collapsed }); core.info('Creating annotations'); const annotations = (0, get_annotations_1.getAnnotations)(results, this.maxAnnotations); const isFailed = this.failOnError && results.some(tr => tr.result === 'failed'); const conclusion = isFailed ? 'failure' : 'success'; - const passed = results.reduce((sum, tr) => sum + tr.passed, 0); - const failed = results.reduce((sum, tr) => sum + tr.failed, 0); - const skipped = results.reduce((sum, tr) => sum + tr.skipped, 0); - const shortSummary = `${passed} passed, ${failed} failed and ${skipped} skipped `; core.info(`Updating check run conclusion (${conclusion}) and output`); const resp = await this.octokit.rest.checks.update({ check_run_id: createResp.data.id, @@ -486,6 +494,8 @@ class TestReporter { return new jest_junit_parser_1.JestJunitParser(options); case 'mocha-json': return new mocha_json_parser_1.MochaJsonParser(options); + case 'python-xunit': + return new python_xunit_parser_1.PythonXunitParser(options); case 'rspec-json': return new rspec_json_parser_1.RspecJsonParser(options); case 'swift-xunit': @@ -836,12 +846,12 @@ class DotnetNunitParser { .map(suite => suite.$.name) .join('.'); const groupName = suitesWithoutTheories[suitesWithoutTheories.length - 1].$.name; - let existingSuite = result.find(existingSuite => existingSuite.name === suiteName); + let existingSuite = result.find(suite => suite.name === suiteName); if (existingSuite === undefined) { existingSuite = new test_results_1.TestSuiteResult(suiteName, []); result.push(existingSuite); } - let existingGroup = existingSuite.groups.find(existingGroup => existingGroup.name === groupName); + let existingGroup = existingSuite.groups.find(group => group.name === groupName); if (existingGroup === undefined) { existingGroup = new test_results_1.TestGroupResult(groupName, []); existingSuite.groups.push(existingGroup); @@ -973,7 +983,7 @@ class DotnetTrxParser { })); const testClasses = {}; for (const r of unitTestsResults) { - const className = r.test.TestMethod[0].$.className; + const className = r.test.TestMethod[0].$.className ?? "Unclassified"; let tc = testClasses[className]; if (tc === undefined) { tc = new TestClass(className); @@ -1024,8 +1034,8 @@ class DotnetTrxParser { error.StackTrace.length === 0) { return undefined; } - const message = test.error.Message[0]; const stackTrace = test.error.StackTrace[0]; + const message = `${test.error.Message[0]}\n${stackTrace}`; let path; let line; const src = this.exceptionThrowSource(stackTrace); @@ -1037,7 +1047,7 @@ class DotnetTrxParser { path, line, message, - details: `${message}\n${stackTrace}` + details: `${message}` }; } exceptionThrowSource(stackTrace) { @@ -1656,6 +1666,26 @@ class MochaJsonParser { exports.MochaJsonParser = MochaJsonParser; +/***/ }), + +/***/ 6578: +/***/ ((__unused_webpack_module, exports, __nccwpck_require__) => { + +"use strict"; + +Object.defineProperty(exports, "__esModule", ({ value: true })); +exports.PythonXunitParser = void 0; +const java_junit_parser_1 = __nccwpck_require__(8342); +class PythonXunitParser extends java_junit_parser_1.JavaJunitParser { + options; + constructor(options) { + super(options); + this.options = options; + } +} +exports.PythonXunitParser = PythonXunitParser; + + /***/ }), /***/ 9768: @@ -1908,6 +1938,7 @@ var __importStar = (this && this.__importStar) || (function () { Object.defineProperty(exports, "__esModule", ({ value: true })); exports.DEFAULT_OPTIONS = void 0; exports.getReport = getReport; +exports.getBadge = getBadge; const core = __importStar(__nccwpck_require__(7484)); const markdown_utils_1 = __nccwpck_require__(5129); const node_utils_1 = __nccwpck_require__(5384); @@ -1922,13 +1953,13 @@ exports.DEFAULT_OPTIONS = { onlySummary: false, useActionsSummary: true, badgeTitle: 'tests', - reportTitle: '' + reportTitle: '', + collapsed: 'auto' }; -function getReport(results, options = exports.DEFAULT_OPTIONS) { - core.info('Generating check run summary'); +function getReport(results, options = exports.DEFAULT_OPTIONS, shortSummary = '') { applySort(results); const opts = { ...options }; - let lines = renderReport(results, opts); + let lines = renderReport(results, opts, shortSummary); let report = lines.join('\n'); if (getByteLength(report) <= getMaxReportLength(options)) { return report; @@ -1936,7 +1967,7 @@ function getReport(results, options = exports.DEFAULT_OPTIONS) { if (opts.listTests === 'all') { core.info("Test report summary is too big - setting 'listTests' to 'failed'"); opts.listTests = 'failed'; - lines = renderReport(results, opts); + lines = renderReport(results, opts, shortSummary); report = lines.join('\n'); if (getByteLength(report) <= getMaxReportLength(options)) { return report; @@ -1983,12 +2014,15 @@ function applySort(results) { function getByteLength(text) { return Buffer.byteLength(text, 'utf8'); } -function renderReport(results, options) { +function renderReport(results, options, shortSummary) { const sections = []; const reportTitle = options.reportTitle.trim(); if (reportTitle) { sections.push(`# ${reportTitle}`); } + if (shortSummary) { + sections.push(`## ${shortSummary}`); + } const badge = getReportBadge(results, options); sections.push(badge); const runs = getTestRunsReport(results, options); @@ -2021,26 +2055,33 @@ function getBadge(passed, failed, skipped, options) { color = 'yellow'; } const hint = failed > 0 ? 'Tests failed' : 'Tests passed successfully'; - const uri = encodeURIComponent(`${options.badgeTitle}-${message}-${color}`); - return `![${hint}](https://img.shields.io/badge/${uri})`; + const encodedBadgeTitle = encodeImgShieldsURIComponent(options.badgeTitle); + const encodedMessage = encodeImgShieldsURIComponent(message); + const encodedColor = encodeImgShieldsURIComponent(color); + return `![${hint}](https://img.shields.io/badge/${encodedBadgeTitle}-${encodedMessage}-${encodedColor})`; } function getTestRunsReport(testRuns, options) { const sections = []; const totalFailed = testRuns.reduce((sum, tr) => sum + tr.failed, 0); - if (totalFailed === 0) { + // Determine if report should be collapsed based on collapsed option + const shouldCollapse = options.collapsed === 'always' || (options.collapsed === 'auto' && totalFailed === 0); + if (shouldCollapse) { sections.push(`
Expand for details`); sections.push(` `); } if (testRuns.length > 0 || options.onlySummary) { const tableData = testRuns - .filter(tr => tr.passed > 0 || tr.failed > 0 || tr.skipped > 0) - .map(tr => { + .map((tr, originalIndex) => ({ tr, originalIndex })) + .filter(({ tr }) => tr.passed > 0 || tr.failed > 0 || tr.skipped > 0) + .map(({ tr, originalIndex }) => { const time = (0, markdown_utils_1.formatTime)(tr.time); const name = tr.path; + const addr = options.baseUrl + makeRunSlug(originalIndex, options).link; + const nameLink = (0, markdown_utils_1.link)(name, addr); const passed = tr.passed > 0 ? `${tr.passed} ${markdown_utils_1.Icon.success}` : ''; const failed = tr.failed > 0 ? `${tr.failed} ${markdown_utils_1.Icon.fail}` : ''; const skipped = tr.skipped > 0 ? `${tr.skipped} ${markdown_utils_1.Icon.skip}` : ''; - return [name, passed, failed, skipped, time]; + return [nameLink, passed, failed, skipped, time]; }); const resultsTable = (0, markdown_utils_1.table)(['Report', 'Passed', 'Failed', 'Skipped', 'Time'], [markdown_utils_1.Align.Left, markdown_utils_1.Align.Right, markdown_utils_1.Align.Right, markdown_utils_1.Align.Right, markdown_utils_1.Align.Right], ...tableData); sections.push(resultsTable); @@ -2049,7 +2090,7 @@ function getTestRunsReport(testRuns, options) { const suitesReports = testRuns.map((tr, i) => getSuitesReport(tr, i, options)).flat(); sections.push(...suitesReports); } - if (totalFailed === 0) { + if (shouldCollapse) { sections.push(`
`); } return sections; @@ -2111,6 +2152,9 @@ function getTestsReport(ts, runIndex, suiteIndex, options) { } const space = grp.name ? ' ' : ''; for (const tc of grp.tests) { + if (options.listTests === 'failed' && tc.result !== 'failed') { + continue; + } const result = getResultIcon(tc.result); sections.push(`${space}${result} ${tc.name}`); if (tc.error) { @@ -2146,6 +2190,9 @@ function getResultIcon(result) { return ''; } } +function encodeImgShieldsURIComponent(component) { + return encodeURIComponent(component).replace(/-/g, '--').replace(/_/g, '__'); +} /***/ }), @@ -2435,11 +2482,11 @@ async function downloadArtifact(octokit, artifactId, fileName, token) { }; const downloadStream = got_1.default.stream(req.url, { headers }); const fileWriterStream = (0, fs_1.createWriteStream)(fileName); - downloadStream.on('redirect', response => { + downloadStream.on('redirect', (response) => { core.info(`Downloading ${response.headers.location}`); }); - downloadStream.on('downloadProgress', ({ transferred }) => { - core.info(`Progress: ${transferred} B`); + downloadStream.on('downloadProgress', (progress) => { + core.info(`Progress: ${progress.transferred} B`); }); await asyncStream(downloadStream, fileWriterStream); } @@ -27840,6 +27887,7 @@ module.exports = { // Replace globs with equivalent patterns to reduce parsing time. REPLACEMENTS: { + __proto__: null, '***': '*', '**/**': '**', '**/**/**': '**' @@ -30274,8 +30322,11 @@ function runParallel (tasks, cb) { /***/ 2560: /***/ ((__unused_webpack_module, exports, __nccwpck_require__) => { -;(function (sax) { // wrapper for non-node envs - sax.parser = function (strict, opt) { return new SAXParser(strict, opt) } +;(function (sax) { + // wrapper for non-node envs + sax.parser = function (strict, opt) { + return new SAXParser(strict, opt) + } sax.SAXParser = SAXParser sax.SAXStream = SAXStream sax.createStream = createStream @@ -30292,9 +30343,18 @@ function runParallel (tasks, cb) { sax.MAX_BUFFER_LENGTH = 64 * 1024 var buffers = [ - 'comment', 'sgmlDecl', 'textNode', 'tagName', 'doctype', - 'procInstName', 'procInstBody', 'entity', 'attribName', - 'attribValue', 'cdata', 'script' + 'comment', + 'sgmlDecl', + 'textNode', + 'tagName', + 'doctype', + 'procInstName', + 'procInstBody', + 'entity', + 'attribName', + 'attribValue', + 'cdata', + 'script', ] sax.EVENTS = [ @@ -30315,10 +30375,10 @@ function runParallel (tasks, cb) { 'ready', 'script', 'opennamespace', - 'closenamespace' + 'closenamespace', ] - function SAXParser (strict, opt) { + function SAXParser(strict, opt) { if (!(this instanceof SAXParser)) { return new SAXParser(strict, opt) } @@ -30337,7 +30397,10 @@ function runParallel (tasks, cb) { parser.noscript = !!(strict || parser.opt.noscript) parser.state = S.BEGIN parser.strictEntities = parser.opt.strictEntities - parser.ENTITIES = parser.strictEntities ? Object.create(sax.XML_ENTITIES) : Object.create(sax.ENTITIES) + parser.ENTITIES = + parser.strictEntities ? + Object.create(sax.XML_ENTITIES) + : Object.create(sax.ENTITIES) parser.attribList = [] // namespaces form a prototype chain. @@ -30347,6 +30410,12 @@ function runParallel (tasks, cb) { parser.ns = Object.create(rootNS) } + // disallow unquoted attribute values if not otherwise configured + // and strict mode is true + if (parser.opt.unquotedAttributeValues === undefined) { + parser.opt.unquotedAttributeValues = !strict + } + // mostly just for error reporting parser.trackPosition = parser.opt.position !== false if (parser.trackPosition) { @@ -30357,7 +30426,7 @@ function runParallel (tasks, cb) { if (!Object.create) { Object.create = function (o) { - function F () {} + function F() {} F.prototype = o var newf = new F() return newf @@ -30372,7 +30441,7 @@ function runParallel (tasks, cb) { } } - function checkBufferLength (parser) { + function checkBufferLength(parser) { var maxAllowed = Math.max(sax.MAX_BUFFER_LENGTH, 10) var maxActual = 0 for (var i = 0, l = buffers.length; i < l; i++) { @@ -30408,13 +30477,13 @@ function runParallel (tasks, cb) { parser.bufferCheckPosition = m + parser.position } - function clearBuffers (parser) { + function clearBuffers(parser) { for (var i = 0, l = buffers.length; i < l; i++) { parser[buffers[i]] = '' } } - function flushBuffers (parser) { + function flushBuffers(parser) { closeText(parser) if (parser.cdata !== '') { emitNode(parser, 'oncdata', parser.cdata) @@ -30427,11 +30496,20 @@ function runParallel (tasks, cb) { } SAXParser.prototype = { - end: function () { end(this) }, + end: function () { + end(this) + }, write: write, - resume: function () { this.error = null; return this }, - close: function () { return this.write(null) }, - flush: function () { flushBuffers(this) } + resume: function () { + this.error = null + return this + }, + close: function () { + return this.write(null) + }, + flush: function () { + flushBuffers(this) + }, } var Stream @@ -30440,16 +30518,17 @@ function runParallel (tasks, cb) { } catch (ex) { Stream = function () {} } + if (!Stream) Stream = function () {} var streamWraps = sax.EVENTS.filter(function (ev) { return ev !== 'error' && ev !== 'end' }) - function createStream (strict, opt) { + function createStream(strict, opt) { return new SAXStream(strict, opt) } - function SAXStream (strict, opt) { + function SAXStream(strict, opt) { if (!(this instanceof SAXStream)) { return new SAXStream(strict, opt) } @@ -30490,21 +30569,23 @@ function runParallel (tasks, cb) { me.on(ev, h) }, enumerable: true, - configurable: false + configurable: false, }) }) } SAXStream.prototype = Object.create(Stream.prototype, { constructor: { - value: SAXStream - } + value: SAXStream, + }, }) SAXStream.prototype.write = function (data) { - if (typeof Buffer === 'function' && + if ( + typeof Buffer === 'function' && typeof Buffer.isBuffer === 'function' && - Buffer.isBuffer(data)) { + Buffer.isBuffer(data) + ) { if (!this._decoder) { var SD = (__nccwpck_require__(3193).StringDecoder) this._decoder = new SD('utf8') @@ -30529,7 +30610,10 @@ function runParallel (tasks, cb) { var me = this if (!me._parser['on' + ev] && streamWraps.indexOf(ev) !== -1) { me._parser['on' + ev] = function () { - var args = arguments.length === 1 ? [arguments[0]] : Array.apply(null, arguments) + var args = + arguments.length === 1 ? + [arguments[0]] + : Array.apply(null, arguments) args.splice(0, 0, ev) me.emit.apply(me, args) } @@ -30552,30 +30636,34 @@ function runParallel (tasks, cb) { // without a significant breaking change to either this parser, or the // JavaScript language. Implementation of an emoji-capable xml parser // is left as an exercise for the reader. - var nameStart = /[:_A-Za-z\u00C0-\u00D6\u00D8-\u00F6\u00F8-\u02FF\u0370-\u037D\u037F-\u1FFF\u200C-\u200D\u2070-\u218F\u2C00-\u2FEF\u3001-\uD7FF\uF900-\uFDCF\uFDF0-\uFFFD]/ + var nameStart = + /[:_A-Za-z\u00C0-\u00D6\u00D8-\u00F6\u00F8-\u02FF\u0370-\u037D\u037F-\u1FFF\u200C-\u200D\u2070-\u218F\u2C00-\u2FEF\u3001-\uD7FF\uF900-\uFDCF\uFDF0-\uFFFD]/ - var nameBody = /[:_A-Za-z\u00C0-\u00D6\u00D8-\u00F6\u00F8-\u02FF\u0370-\u037D\u037F-\u1FFF\u200C-\u200D\u2070-\u218F\u2C00-\u2FEF\u3001-\uD7FF\uF900-\uFDCF\uFDF0-\uFFFD\u00B7\u0300-\u036F\u203F-\u2040.\d-]/ + var nameBody = + /[:_A-Za-z\u00C0-\u00D6\u00D8-\u00F6\u00F8-\u02FF\u0370-\u037D\u037F-\u1FFF\u200C-\u200D\u2070-\u218F\u2C00-\u2FEF\u3001-\uD7FF\uF900-\uFDCF\uFDF0-\uFFFD\u00B7\u0300-\u036F\u203F-\u2040.\d-]/ - var entityStart = /[#:_A-Za-z\u00C0-\u00D6\u00D8-\u00F6\u00F8-\u02FF\u0370-\u037D\u037F-\u1FFF\u200C-\u200D\u2070-\u218F\u2C00-\u2FEF\u3001-\uD7FF\uF900-\uFDCF\uFDF0-\uFFFD]/ - var entityBody = /[#:_A-Za-z\u00C0-\u00D6\u00D8-\u00F6\u00F8-\u02FF\u0370-\u037D\u037F-\u1FFF\u200C-\u200D\u2070-\u218F\u2C00-\u2FEF\u3001-\uD7FF\uF900-\uFDCF\uFDF0-\uFFFD\u00B7\u0300-\u036F\u203F-\u2040.\d-]/ + var entityStart = + /[#:_A-Za-z\u00C0-\u00D6\u00D8-\u00F6\u00F8-\u02FF\u0370-\u037D\u037F-\u1FFF\u200C-\u200D\u2070-\u218F\u2C00-\u2FEF\u3001-\uD7FF\uF900-\uFDCF\uFDF0-\uFFFD]/ + var entityBody = + /[#:_A-Za-z\u00C0-\u00D6\u00D8-\u00F6\u00F8-\u02FF\u0370-\u037D\u037F-\u1FFF\u200C-\u200D\u2070-\u218F\u2C00-\u2FEF\u3001-\uD7FF\uF900-\uFDCF\uFDF0-\uFFFD\u00B7\u0300-\u036F\u203F-\u2040.\d-]/ - function isWhitespace (c) { + function isWhitespace(c) { return c === ' ' || c === '\n' || c === '\r' || c === '\t' } - function isQuote (c) { - return c === '"' || c === '\'' + function isQuote(c) { + return c === '"' || c === "'" } - function isAttribEnd (c) { + function isAttribEnd(c) { return c === '>' || isWhitespace(c) } - function isMatch (regex, c) { + function isMatch(regex, c) { return regex.test(c) } - function notMatch (regex, c) { + function notMatch(regex, c) { return !isMatch(regex, c) } @@ -30616,271 +30704,271 @@ function runParallel (tasks, cb) { CLOSE_TAG: S++, // SCRIPT: S++, //