diff --git a/.github/workflows/check-dist.yml b/.github/workflows/check-dist.yml
index c994dfa..e571511 100644
--- a/.github/workflows/check-dist.yml
+++ b/.github/workflows/check-dist.yml
@@ -21,10 +21,10 @@ jobs:
runs-on: ubuntu-latest
steps:
- - uses: actions/checkout@v6
+ - uses: actions/checkout@v4
- name: Set Node.js
- uses: actions/setup-node@v6
+ uses: actions/setup-node@v4
with:
node-version-file: '.nvmrc'
@@ -46,7 +46,7 @@ jobs:
id: diff
# If index.js was different than expected, upload the expected version as an artifact
- - uses: actions/upload-artifact@v5
+ - uses: actions/upload-artifact@v4
if: ${{ failure() && steps.diff.conclusion == 'failure' }}
with:
name: dist
diff --git a/.github/workflows/ci.yml b/.github/workflows/ci.yml
index aeec96d..a8a86e4 100644
--- a/.github/workflows/ci.yml
+++ b/.github/workflows/ci.yml
@@ -13,8 +13,8 @@ jobs:
name: Build & Test
runs-on: ubuntu-latest
steps:
- - uses: actions/checkout@v6
- - uses: actions/setup-node@v6
+ - uses: actions/checkout@v4
+ - uses: actions/setup-node@v4
with:
node-version-file: '.nvmrc'
- run: npm ci
@@ -24,8 +24,8 @@ jobs:
- run: npm test
- name: Upload test results
- if: ${{ !cancelled() }}
- uses: actions/upload-artifact@v5
+ if: success() || failure()
+ uses: actions/upload-artifact@v4
with:
name: test-results
path: __tests__/__results__/*.xml
diff --git a/.github/workflows/manual-run.yml b/.github/workflows/manual-run.yml
index c1875a9..f8b1ee0 100644
--- a/.github/workflows/manual-run.yml
+++ b/.github/workflows/manual-run.yml
@@ -8,14 +8,14 @@ jobs:
runs-on: ubuntu-latest
steps:
- - uses: actions/checkout@v6
+ - uses: actions/checkout@v4
- run: npm ci
- run: npm run build
- run: npm test
- name: Create test report
uses: ./
- if: ${{ !cancelled() }}
+ if: success() || failure()
with:
name: JEST Tests
path: __tests__/__results__/*.xml
diff --git a/.github/workflows/test-report.yml b/.github/workflows/test-report.yml
index 11b266a..e3f9555 100644
--- a/.github/workflows/test-report.yml
+++ b/.github/workflows/test-report.yml
@@ -11,7 +11,7 @@ jobs:
name: Workflow test
runs-on: ubuntu-latest
steps:
- - uses: actions/checkout@v6
+ - uses: actions/checkout@v4
- uses: ./
with:
artifact: test-results
diff --git a/.markdownlint.json b/.markdownlint.json
deleted file mode 100644
index 3f71d87..0000000
--- a/.markdownlint.json
+++ /dev/null
@@ -1,13 +0,0 @@
-{
- "blanks-around-headings": false,
- "blanks-around-lists": false,
- "blanks-around-tables": false,
- "blanks-around-fences": false,
- "no-bare-urls": false,
- "line-length": false,
- "ul-style": false,
- "no-inline-html": false,
- "no-multiple-blanks": {
- "maximum": 3
- }
-}
diff --git a/CHANGELOG.md b/CHANGELOG.md
index b0badd5..d47e76e 100644
--- a/CHANGELOG.md
+++ b/CHANGELOG.md
@@ -1,30 +1,5 @@
# Changelog
-## 2.3.0
-* Feature: Add Python support with `python-xunit` reporter (pytest) https://github.com/dorny/test-reporter/pull/643
-* Feature: Add pytest traceback parsing and `directory-mapping` option https://github.com/dorny/test-reporter/pull/238
-* Performance: Update sax.js to fix large XML file parsing https://github.com/dorny/test-reporter/pull/681
-* Documentation: Complete documentation for all supported reporters https://github.com/dorny/test-reporter/pull/691
-* Security: Bump js-yaml and mocha in /reports/mocha (fixes prototype pollution) https://github.com/dorny/test-reporter/pull/682
-
-## 2.2.0
-* Feature: Add collapsed option to control report summary visibility https://github.com/dorny/test-reporter/pull/664
-* Fix badge encoding for values including underscore and hyphens https://github.com/dorny/test-reporter/pull/672
-* Fix missing `report-title` attribute in action definition https://github.com/dorny/test-reporter/pull/637
-* Refactor variable names to fix shadowing issues https://github.com/dorny/test-reporter/pull/630
-
-## 2.1.1
-* Fix error when a TestMethod element does not have a className attribute in a trx file https://github.com/dorny/test-reporter/pull/623
-* Add stack trace from trx to summary https://github.com/dorny/test-reporter/pull/615
-* List only failed tests https://github.com/dorny/test-reporter/pull/606
-* Add type definitions to `github-utils.ts` https://github.com/dorny/test-reporter/pull/604
-* Avoid split on undefined https://github.com/dorny/test-reporter/pull/258
-* Return links to summary report https://github.com/dorny/test-reporter/pull/588
-* Add step summary short summary https://github.com/dorny/test-reporter/pull/589
-* Fix for empty TRX TestDefinitions https://github.com/dorny/test-reporter/pull/582
-* Increase step summary limit to 1MiB https://github.com/dorny/test-reporter/pull/581
-* Fix input description for list options https://github.com/dorny/test-reporter/pull/572
-
## 2.1.0
* Feature: Add summary title https://github.com/dorny/test-reporter/pull/568
* Feature: Add Golang test parser https://github.com/dorny/test-reporter/pull/571
diff --git a/README.md b/README.md
index d900926..676d21e 100644
--- a/README.md
+++ b/README.md
@@ -9,7 +9,7 @@ This [Github Action](https://github.com/features/actions) displays test results
✔️ Provides final `conclusion` and counts of `passed`, `failed` and `skipped` tests as output parameters
**How it looks:**
-|||||
+|||||
|:--:|:--:|:--:|:--:|
**Supported languages / frameworks:**
@@ -19,8 +19,6 @@ This [Github Action](https://github.com/features/actions) displays test results
- Go / [go test](https://pkg.go.dev/testing)
- Java / [JUnit](https://junit.org/)
- JavaScript / [JEST](https://jestjs.io/) / [Mocha](https://mochajs.org/)
-- Python / [pytest](https://docs.pytest.org/en/stable/) / [unittest](https://docs.python.org/3/library/unittest.html)
-- Ruby / [RSpec](https://rspec.info/)
- Swift / xUnit
For more information see [Supported formats](#supported-formats) section.
@@ -52,7 +50,7 @@ jobs:
- name: Test Report
uses: dorny/test-reporter@v2
- if: ${{ !cancelled() }} # run this step even if previous step failed
+ if: success() || failure() # run this step even if previous step failed
with:
name: JEST Tests # Name of the check run which will be created
path: reports/jest-*.xml # Path to test results
@@ -81,7 +79,7 @@ jobs:
- run: npm ci # install packages
- run: npm test # run tests (configured to use jest-junit reporter)
- uses: actions/upload-artifact@v4 # upload test results
- if: ${{ !cancelled() }} # run this step even if previous step failed
+ if: success() || failure() # run this step even if previous step failed
with:
name: test-results
path: jest-junit.xml
@@ -147,9 +145,7 @@ jobs:
# java-junit
# jest-junit
# mocha-json
- # python-xunit
# rspec-json
- # swift-xunit
reporter: ''
# Allows you to generate only the summary.
@@ -257,20 +253,6 @@ Supported testing frameworks:
For more information see [dotnet test](https://docs.microsoft.com/en-us/dotnet/core/tools/dotnet-test#examples)
-
- dotnet-nunit
-
-Test execution must be configured to generate [NUnit3](https://docs.nunit.org/articles/nunit/technical-notes/usage/Test-Result-XML-Format.html) XML test results.
-Install the [NUnit3TestAdapter](https://www.nuget.org/packages/NUnit3TestAdapter) package (required; it registers the `nunit` logger for `dotnet test`), then run tests with:
-
-`dotnet test --logger "nunit;LogFileName=test-results.xml"`
-
-Supported testing frameworks:
-- [NUnit](https://nunit.org/)
-
-For more information see [dotnet test](https://docs.microsoft.com/en-us/dotnet/core/tools/dotnet-test#examples)
-
-
flutter-json
@@ -367,41 +349,6 @@ Before version [v9.1.0](https://github.com/mochajs/mocha/releases/tag/v9.1.0), M
Please update Mocha to version [v9.1.0](https://github.com/mochajs/mocha/releases/tag/v9.1.0) or above if you encounter this issue.
-
- python-xunit (Experimental)
-
-Support for Python test results in xUnit format is experimental - should work but it was not extensively tested.
-
-For **pytest** support, configure [JUnit XML output](https://docs.pytest.org/en/stable/how-to/output.html#creating-junitxml-format-files) and run with the `--junit-xml` option, which also lets you specify the output path for test results.
-
-```shell
-pytest --junit-xml=test-report.xml
-```
-
-For **unittest** support, use a test runner that outputs the JUnit report format, such as [unittest-xml-reporting](https://pypi.org/project/unittest-xml-reporting/).
-
-
-
- rspec-json
-
-[RSpec](https://rspec.info/) testing framework support requires the usage of JSON formatter.
-You can configure RSpec to output JSON format by using the `--format json` option and redirecting to a file:
-
-```shell
-rspec --format json --out rspec-results.json
-```
-
-Or configure it in `.rspec` file:
-```
---format json
---out rspec-results.json
-```
-
-For more information see:
-- [RSpec documentation](https://rspec.info/)
-- [RSpec Formatters](https://relishapp.com/rspec/rspec-core/docs/formatters)
-
-
swift-xunit (Experimental)
diff --git a/__tests__/__outputs__/dart-json.md b/__tests__/__outputs__/dart-json.md
index 2b7ec09..9805d20 100644
--- a/__tests__/__outputs__/dart-json.md
+++ b/__tests__/__outputs__/dart-json.md
@@ -1,7 +1,7 @@

|Report|Passed|Failed|Skipped|Time|
|:---|---:|---:|---:|---:|
-|[fixtures/dart-json.json](#user-content-r0)|1 ✅|4 ❌|1 ⚪|4s|
+|fixtures/dart-json.json|1 ✅|4 ❌|1 ⚪|4s|
## ❌ fixtures/dart-json.json
**6** tests were completed in **4s** with **1** passed, **4** failed and **1** skipped.
|Test suite|Passed|Failed|Skipped|Time|
diff --git a/__tests__/__outputs__/dotnet-nunit.md b/__tests__/__outputs__/dotnet-nunit.md
index a0985f5..e33c66d 100644
--- a/__tests__/__outputs__/dotnet-nunit.md
+++ b/__tests__/__outputs__/dotnet-nunit.md
@@ -1,7 +1,7 @@

|Report|Passed|Failed|Skipped|Time|
|:---|---:|---:|---:|---:|
-|[fixtures/dotnet-nunit.xml](#user-content-r0)|3 ✅|5 ❌|1 ⚪|230ms|
+|fixtures/dotnet-nunit.xml|3 ✅|5 ❌|1 ⚪|230ms|
## ❌ fixtures/dotnet-nunit.xml
**9** tests were completed in **230ms** with **3** passed, **5** failed and **1** skipped.
|Test suite|Passed|Failed|Skipped|Time|
diff --git a/__tests__/__outputs__/dotnet-trx-only-failed.md b/__tests__/__outputs__/dotnet-trx-only-failed.md
deleted file mode 100644
index 5eeaec5..0000000
--- a/__tests__/__outputs__/dotnet-trx-only-failed.md
+++ /dev/null
@@ -1,34 +0,0 @@
-
-|Report|Passed|Failed|Skipped|Time|
-|:---|---:|---:|---:|---:|
-|[fixtures/dotnet-trx.trx](#user-content-r0)|5 ✅|5 ❌|1 ⚪|1s|
-## ❌ fixtures/dotnet-trx.trx
-**11** tests were completed in **1s** with **5** passed, **5** failed and **1** skipped.
-|Test suite|Passed|Failed|Skipped|Time|
-|:---|---:|---:|---:|---:|
-|[DotnetTests.XUnitTests.CalculatorTests](#user-content-r0s0)|5 ✅|5 ❌|1 ⚪|118ms|
-### ❌ DotnetTests.XUnitTests.CalculatorTests
-```
-❌ Exception_In_TargetTest
- System.DivideByZeroException : Attempted to divide by zero.
- at DotnetTests.Unit.Calculator.Div(Int32 a, Int32 b) in C:\Users\Michal\Workspace\dorny\test-reporter\reports\dotnet\DotnetTests.Unit\Calculator.cs:line 9
- at DotnetTests.XUnitTests.CalculatorTests.Exception_In_TargetTest() in C:\Users\Michal\Workspace\dorny\test-reporter\reports\dotnet\DotnetTests.XUnitTests\CalculatorTests.cs:line 33
-❌ Exception_In_Test
- System.Exception : Test
- at DotnetTests.XUnitTests.CalculatorTests.Exception_In_Test() in C:\Users\Michal\Workspace\dorny\test-reporter\reports\dotnet\DotnetTests.XUnitTests\CalculatorTests.cs:line 39
-❌ Failing_Test
- Assert.Equal() Failure
- Expected: 3
- Actual: 2
- at DotnetTests.XUnitTests.CalculatorTests.Failing_Test() in C:\Users\Michal\Workspace\dorny\test-reporter\reports\dotnet\DotnetTests.XUnitTests\CalculatorTests.cs:line 27
-❌ Is_Even_Number(i: 3)
- Assert.True() Failure
- Expected: True
- Actual: False
- at DotnetTests.XUnitTests.CalculatorTests.Is_Even_Number(Int32 i) in C:\Users\Michal\Workspace\dorny\test-reporter\reports\dotnet\DotnetTests.XUnitTests\CalculatorTests.cs:line 59
-❌ Should be even number(i: 3)
- Assert.True() Failure
- Expected: True
- Actual: False
- at DotnetTests.XUnitTests.CalculatorTests.Theory_With_Custom_Name(Int32 i) in C:\Users\Michal\Workspace\dorny\test-reporter\reports\dotnet\DotnetTests.XUnitTests\CalculatorTests.cs:line 67
-```
\ No newline at end of file
diff --git a/__tests__/__outputs__/dotnet-trx.md b/__tests__/__outputs__/dotnet-trx.md
index 40bcaf2..eedab46 100644
--- a/__tests__/__outputs__/dotnet-trx.md
+++ b/__tests__/__outputs__/dotnet-trx.md
@@ -1,7 +1,7 @@

|Report|Passed|Failed|Skipped|Time|
|:---|---:|---:|---:|---:|
-|[fixtures/dotnet-trx.trx](#user-content-r0)|5 ✅|5 ❌|1 ⚪|1s|
+|fixtures/dotnet-trx.trx|5 ✅|5 ❌|1 ⚪|1s|
## ❌ fixtures/dotnet-trx.trx
**11** tests were completed in **1s** with **5** passed, **5** failed and **1** skipped.
|Test suite|Passed|Failed|Skipped|Time|
@@ -12,29 +12,23 @@
✅ Custom Name
❌ Exception_In_TargetTest
System.DivideByZeroException : Attempted to divide by zero.
- at DotnetTests.Unit.Calculator.Div(Int32 a, Int32 b) in C:\Users\Michal\Workspace\dorny\test-reporter\reports\dotnet\DotnetTests.Unit\Calculator.cs:line 9
- at DotnetTests.XUnitTests.CalculatorTests.Exception_In_TargetTest() in C:\Users\Michal\Workspace\dorny\test-reporter\reports\dotnet\DotnetTests.XUnitTests\CalculatorTests.cs:line 33
❌ Exception_In_Test
System.Exception : Test
- at DotnetTests.XUnitTests.CalculatorTests.Exception_In_Test() in C:\Users\Michal\Workspace\dorny\test-reporter\reports\dotnet\DotnetTests.XUnitTests\CalculatorTests.cs:line 39
❌ Failing_Test
Assert.Equal() Failure
Expected: 3
Actual: 2
- at DotnetTests.XUnitTests.CalculatorTests.Failing_Test() in C:\Users\Michal\Workspace\dorny\test-reporter\reports\dotnet\DotnetTests.XUnitTests\CalculatorTests.cs:line 27
✅ Is_Even_Number(i: 2)
❌ Is_Even_Number(i: 3)
Assert.True() Failure
Expected: True
Actual: False
- at DotnetTests.XUnitTests.CalculatorTests.Is_Even_Number(Int32 i) in C:\Users\Michal\Workspace\dorny\test-reporter\reports\dotnet\DotnetTests.XUnitTests\CalculatorTests.cs:line 59
✅ Passing_Test
✅ Should be even number(i: 2)
❌ Should be even number(i: 3)
Assert.True() Failure
Expected: True
Actual: False
- at DotnetTests.XUnitTests.CalculatorTests.Theory_With_Custom_Name(Int32 i) in C:\Users\Michal\Workspace\dorny\test-reporter\reports\dotnet\DotnetTests.XUnitTests\CalculatorTests.cs:line 67
⚪ Skipped_Test
✅ Timeout_Test
```
\ No newline at end of file
diff --git a/__tests__/__outputs__/dotnet-xunitv3.md b/__tests__/__outputs__/dotnet-xunitv3.md
deleted file mode 100644
index f27f9e2..0000000
--- a/__tests__/__outputs__/dotnet-xunitv3.md
+++ /dev/null
@@ -1,26 +0,0 @@
-
-|Report|Passed|Failed|Skipped|Time|
-|:---|---:|---:|---:|---:|
-|[fixtures/dotnet-xunitv3.trx](#user-content-r0)|1 ✅|3 ❌||267ms|
-## ❌ fixtures/dotnet-xunitv3.trx
-**4** tests were completed in **267ms** with **1** passed, **3** failed and **0** skipped.
-|Test suite|Passed|Failed|Skipped|Time|
-|:---|---:|---:|---:|---:|
-|[DotnetTests.XUnitV3Tests.FixtureTests](#user-content-r0s0)|1 ✅|1 ❌||18ms|
-|[Unclassified](#user-content-r0s1)||2 ❌||0ms|
-### ❌ DotnetTests.XUnitV3Tests.FixtureTests
-```
-❌ Failing_Test
- Assert.Null() Failure: Value is not null
- Expected: null
- Actual: Fixture { }
- at DotnetTests.XUnitV3Tests.FixtureTests.Failing_Test() in /_/reports/dotnet/DotnetTests.XUnitV3Tests/FixtureTests.cs:line 25
- at System.RuntimeMethodHandle.InvokeMethod(Object target, Void** arguments, Signature sig, Boolean isConstructor)
- at System.Reflection.MethodBaseInvoker.InvokeWithNoArgs(Object obj, BindingFlags invokeAttr)
-✅ Passing_Test
-```
-### ❌ Unclassified
-```
-❌ [Test Class Cleanup Failure (DotnetTests.XUnitV3Tests.FixtureTests.Failing_Test)]
-❌ [Test Class Cleanup Failure (DotnetTests.XUnitV3Tests.FixtureTests.Passing_Test)]
-```
\ No newline at end of file
diff --git a/__tests__/__outputs__/fluent-validation-test-results.md b/__tests__/__outputs__/fluent-validation-test-results.md
index 83d0503..9337a62 100644
--- a/__tests__/__outputs__/fluent-validation-test-results.md
+++ b/__tests__/__outputs__/fluent-validation-test-results.md
@@ -3,7 +3,7 @@
|Report|Passed|Failed|Skipped|Time|
|:---|---:|---:|---:|---:|
-|[fixtures/external/FluentValidation.Tests.trx](#user-content-r0)|803 ✅||1 ⚪|4s|
+|fixtures/external/FluentValidation.Tests.trx|803 ✅||1 ⚪|4s|
## ✅ fixtures/external/FluentValidation.Tests.trx
**804** tests were completed in **4s** with **803** passed, **0** failed and **1** skipped.
|Test suite|Passed|Failed|Skipped|Time|
diff --git a/__tests__/__outputs__/golang-json.md b/__tests__/__outputs__/golang-json.md
index 8b63704..b3640d8 100644
--- a/__tests__/__outputs__/golang-json.md
+++ b/__tests__/__outputs__/golang-json.md
@@ -1,7 +1,7 @@

|Report|Passed|Failed|Skipped|Time|
|:---|---:|---:|---:|---:|
-|[fixtures/golang-json.json](#user-content-r0)|5 ✅|6 ❌|1 ⚪|6s|
+|fixtures/golang-json.json|5 ✅|6 ❌|1 ⚪|6s|
## ❌ fixtures/golang-json.json
**12** tests were completed in **6s** with **5** passed, **6** failed and **1** skipped.
|Test suite|Passed|Failed|Skipped|Time|
diff --git a/__tests__/__outputs__/jest-junit-eslint.md b/__tests__/__outputs__/jest-junit-eslint.md
index 5ebb57e..d3ad9b9 100644
--- a/__tests__/__outputs__/jest-junit-eslint.md
+++ b/__tests__/__outputs__/jest-junit-eslint.md
@@ -3,7 +3,7 @@
|Report|Passed|Failed|Skipped|Time|
|:---|---:|---:|---:|---:|
-|[fixtures/jest-junit-eslint.xml](#user-content-r0)|1 ✅|||0ms|
+|fixtures/jest-junit-eslint.xml|1 ✅|||0ms|
## ✅ fixtures/jest-junit-eslint.xml
**1** tests were completed in **0ms** with **1** passed, **0** failed and **0** skipped.
|Test suite|Passed|Failed|Skipped|Time|
diff --git a/__tests__/__outputs__/jest-junit.md b/__tests__/__outputs__/jest-junit.md
index 951256f..ed5a174 100644
--- a/__tests__/__outputs__/jest-junit.md
+++ b/__tests__/__outputs__/jest-junit.md
@@ -1,7 +1,7 @@

|Report|Passed|Failed|Skipped|Time|
|:---|---:|---:|---:|---:|
-|[fixtures/jest-junit.xml](#user-content-r0)|1 ✅|4 ❌|1 ⚪|1s|
+|fixtures/jest-junit.xml|1 ✅|4 ❌|1 ⚪|1s|
## ❌ fixtures/jest-junit.xml
**6** tests were completed in **1s** with **1** passed, **4** failed and **1** skipped.
|Test suite|Passed|Failed|Skipped|Time|
diff --git a/__tests__/__outputs__/jest-react-component-test-results.md b/__tests__/__outputs__/jest-react-component-test-results.md
index 1365818..d71db4c 100644
--- a/__tests__/__outputs__/jest-react-component-test-results.md
+++ b/__tests__/__outputs__/jest-react-component-test-results.md
@@ -3,7 +3,7 @@
|Report|Passed|Failed|Skipped|Time|
|:---|---:|---:|---:|---:|
-|[fixtures/external/jest/jest-react-component-test-results.xml](#user-content-r0)|1 ✅|||1000ms|
+|fixtures/external/jest/jest-react-component-test-results.xml|1 ✅|||1000ms|
## ✅ fixtures/external/jest/jest-react-component-test-results.xml
**1** tests were completed in **1000ms** with **1** passed, **0** failed and **0** skipped.
|Test suite|Passed|Failed|Skipped|Time|
diff --git a/__tests__/__outputs__/jest-test-results.md b/__tests__/__outputs__/jest-test-results.md
index cfbc169..25dd567 100644
--- a/__tests__/__outputs__/jest-test-results.md
+++ b/__tests__/__outputs__/jest-test-results.md
@@ -1,7 +1,7 @@

|Report|Passed|Failed|Skipped|Time|
|:---|---:|---:|---:|---:|
-|[fixtures/external/jest/jest-test-results.xml](#user-content-r0)|4207 ✅|2 ❌|30 ⚪|166s|
+|fixtures/external/jest/jest-test-results.xml|4207 ✅|2 ❌|30 ⚪|166s|
## ❌ fixtures/external/jest/jest-test-results.xml
**4239** tests were completed in **166s** with **4207** passed, **2** failed and **30** skipped.
|Test suite|Passed|Failed|Skipped|Time|
diff --git a/__tests__/__outputs__/junit-with-message.md b/__tests__/__outputs__/junit-with-message.md
index 634a402..988d8fc 100644
--- a/__tests__/__outputs__/junit-with-message.md
+++ b/__tests__/__outputs__/junit-with-message.md
@@ -1,7 +1,7 @@

|Report|Passed|Failed|Skipped|Time|
|:---|---:|---:|---:|---:|
-|[fixtures/junit-with-message.xml](#user-content-r0)||1 ❌||1ms|
+|fixtures/junit-with-message.xml||1 ❌||1ms|
## ❌ fixtures/junit-with-message.xml
**1** tests were completed in **1ms** with **0** passed, **1** failed and **0** skipped.
|Test suite|Passed|Failed|Skipped|Time|
diff --git a/__tests__/__outputs__/mocha-json.md b/__tests__/__outputs__/mocha-json.md
index 875f881..50419e1 100644
--- a/__tests__/__outputs__/mocha-json.md
+++ b/__tests__/__outputs__/mocha-json.md
@@ -1,7 +1,7 @@

|Report|Passed|Failed|Skipped|Time|
|:---|---:|---:|---:|---:|
-|[fixtures/mocha-json.json](#user-content-r0)|1 ✅|4 ❌|1 ⚪|12ms|
+|fixtures/mocha-json.json|1 ✅|4 ❌|1 ⚪|12ms|
## ❌ fixtures/mocha-json.json
**6** tests were completed in **12ms** with **1** passed, **4** failed and **1** skipped.
|Test suite|Passed|Failed|Skipped|Time|
diff --git a/__tests__/__outputs__/mocha-test-results.md b/__tests__/__outputs__/mocha-test-results.md
index 4a6e2f6..8831d7b 100644
--- a/__tests__/__outputs__/mocha-test-results.md
+++ b/__tests__/__outputs__/mocha-test-results.md
@@ -3,7 +3,7 @@
|Report|Passed|Failed|Skipped|Time|
|:---|---:|---:|---:|---:|
-|[fixtures/external/mocha/mocha-test-results.json](#user-content-r0)|833 ✅||6 ⚪|6s|
+|fixtures/external/mocha/mocha-test-results.json|833 ✅||6 ⚪|6s|
## ✅ fixtures/external/mocha/mocha-test-results.json
**839** tests were completed in **6s** with **833** passed, **0** failed and **6** skipped.
|Test suite|Passed|Failed|Skipped|Time|
diff --git a/__tests__/__outputs__/provider-test-results.md b/__tests__/__outputs__/provider-test-results.md
index 172f070..b2216ad 100644
--- a/__tests__/__outputs__/provider-test-results.md
+++ b/__tests__/__outputs__/provider-test-results.md
@@ -1,7 +1,7 @@

|Report|Passed|Failed|Skipped|Time|
|:---|---:|---:|---:|---:|
-|[fixtures/external/flutter/provider-test-results.json](#user-content-r0)|268 ✅|1 ❌||0ms|
+|fixtures/external/flutter/provider-test-results.json|268 ✅|1 ❌||0ms|
## ❌ fixtures/external/flutter/provider-test-results.json
**269** tests were completed in **0ms** with **268** passed, **1** failed and **0** skipped.
|Test suite|Passed|Failed|Skipped|Time|
diff --git a/__tests__/__outputs__/pulsar-test-results-no-merge.md b/__tests__/__outputs__/pulsar-test-results-no-merge.md
index de5a9b6..b1738a4 100644
--- a/__tests__/__outputs__/pulsar-test-results-no-merge.md
+++ b/__tests__/__outputs__/pulsar-test-results-no-merge.md
@@ -1,7 +1,7 @@

|Report|Passed|Failed|Skipped|Time|
|:---|---:|---:|---:|---:|
-|[fixtures/external/java/TEST-org.apache.pulsar.AddMissingPatchVersionTest.xml](#user-content-r0)||1 ❌|1 ⚪|116ms|
+|fixtures/external/java/TEST-org.apache.pulsar.AddMissingPatchVersionTest.xml||1 ❌|1 ⚪|116ms|
## ❌ fixtures/external/java/TEST-org.apache.pulsar.AddMissingPatchVersionTest.xml
**2** tests were completed in **116ms** with **0** passed, **1** failed and **1** skipped.
|Test suite|Passed|Failed|Skipped|Time|
diff --git a/__tests__/__outputs__/pulsar-test-results.md b/__tests__/__outputs__/pulsar-test-results.md
index aaaa82e..a9a5290 100644
--- a/__tests__/__outputs__/pulsar-test-results.md
+++ b/__tests__/__outputs__/pulsar-test-results.md
@@ -1,7 +1,7 @@

|Report|Passed|Failed|Skipped|Time|
|:---|---:|---:|---:|---:|
-|[fixtures/external/java/pulsar-test-report.xml](#user-content-r0)|793 ✅|1 ❌|14 ⚪|2127s|
+|fixtures/external/java/pulsar-test-report.xml|793 ✅|1 ❌|14 ⚪|2127s|
## ❌ fixtures/external/java/pulsar-test-report.xml
**808** tests were completed in **2127s** with **793** passed, **1** failed and **14** skipped.
|Test suite|Passed|Failed|Skipped|Time|
diff --git a/__tests__/__outputs__/python-xunit-pytest.md b/__tests__/__outputs__/python-xunit-pytest.md
deleted file mode 100644
index 7b13e28..0000000
--- a/__tests__/__outputs__/python-xunit-pytest.md
+++ /dev/null
@@ -1,26 +0,0 @@
-
-|Report|Passed|Failed|Skipped|Time|
-|:---|---:|---:|---:|---:|
-|[fixtures/python-xunit-pytest.xml](#user-content-r0)|6 ✅|2 ❌|2 ⚪|19ms|
-## ❌ fixtures/python-xunit-pytest.xml
-**10** tests were completed in **19ms** with **6** passed, **2** failed and **2** skipped.
-|Test suite|Passed|Failed|Skipped|Time|
-|:---|---:|---:|---:|---:|
-|[pytest](#user-content-r0s0)|6 ✅|2 ❌|2 ⚪|19ms|
-### ❌ pytest
-```
-tests.test_lib
- ✅ test_always_pass
- ✅ test_with_subtests
- ✅ test_parameterized[param1]
- ✅ test_parameterized[param2]
- ⚪ test_always_skip
- ❌ test_always_fail
- assert False
- ⚪ test_expected_failure
- ❌ test_error
- Exception: error
- ✅ test_with_record_property
-custom_classname
- ✅ test_with_record_xml_attribute
-```
\ No newline at end of file
diff --git a/__tests__/__outputs__/python-xunit-unittest.md b/__tests__/__outputs__/python-xunit-unittest.md
deleted file mode 100644
index 230d186..0000000
--- a/__tests__/__outputs__/python-xunit-unittest.md
+++ /dev/null
@@ -1,23 +0,0 @@
-
-|Report|Passed|Failed|Skipped|Time|
-|:---|---:|---:|---:|---:|
-|[fixtures/python-xunit-unittest.xml](#user-content-r0)|4 ✅|2 ❌|2 ⚪|1ms|
-## ❌ fixtures/python-xunit-unittest.xml
-**8** tests were completed in **1ms** with **4** passed, **2** failed and **2** skipped.
-|Test suite|Passed|Failed|Skipped|Time|
-|:---|---:|---:|---:|---:|
-|[TestAcme-20251114214921](#user-content-r0s0)|4 ✅|2 ❌|2 ⚪|1ms|
-### ❌ TestAcme-20251114214921
-```
-TestAcme
- ✅ test_always_pass
- ✅ test_parameterized_0_param1
- ✅ test_parameterized_1_param2
- ✅ test_with_subtests
- ❌ test_always_fail
- AssertionError: failed
- ❌ test_error
- Exception: error
- ⚪ test_always_skip
- ⚪ test_expected_failure
-```
\ No newline at end of file
diff --git a/__tests__/__outputs__/rspec-json.md b/__tests__/__outputs__/rspec-json.md
index d64cf44..7444608 100644
--- a/__tests__/__outputs__/rspec-json.md
+++ b/__tests__/__outputs__/rspec-json.md
@@ -1,7 +1,7 @@

|Report|Passed|Failed|Skipped|Time|
|:---|---:|---:|---:|---:|
-|[fixtures/rspec-json.json](#user-content-r0)|1 ✅|1 ❌|1 ⚪|0ms|
+|fixtures/rspec-json.json|1 ✅|1 ❌|1 ⚪|0ms|
## ❌ fixtures/rspec-json.json
**3** tests were completed in **0ms** with **1** passed, **1** failed and **1** skipped.
|Test suite|Passed|Failed|Skipped|Time|
diff --git a/__tests__/__outputs__/silent-notes-test-results.md b/__tests__/__outputs__/silent-notes-test-results.md
index e3abc49..34f5bab 100644
--- a/__tests__/__outputs__/silent-notes-test-results.md
+++ b/__tests__/__outputs__/silent-notes-test-results.md
@@ -3,7 +3,7 @@
|Report|Passed|Failed|Skipped|Time|
|:---|---:|---:|---:|---:|
-|[fixtures/external/SilentNotes.trx](#user-content-r0)|67 ✅||12 ⚪|1s|
+|fixtures/external/SilentNotes.trx|67 ✅||12 ⚪|1s|
## ✅ fixtures/external/SilentNotes.trx
**79** tests were completed in **1s** with **67** passed, **0** failed and **12** skipped.
|Test suite|Passed|Failed|Skipped|Time|
diff --git a/__tests__/__outputs__/swift-xunit.md b/__tests__/__outputs__/swift-xunit.md
index b001151..6f9ed46 100644
--- a/__tests__/__outputs__/swift-xunit.md
+++ b/__tests__/__outputs__/swift-xunit.md
@@ -1,7 +1,7 @@

|Report|Passed|Failed|Skipped|Time|
|:---|---:|---:|---:|---:|
-|[fixtures/swift-xunit.xml](#user-content-r0)|2 ✅|1 ❌||220ms|
+|fixtures/swift-xunit.xml|2 ✅|1 ❌||220ms|
## ❌ fixtures/swift-xunit.xml
**3** tests were completed in **220ms** with **2** passed, **1** failed and **0** skipped.
|Test suite|Passed|Failed|Skipped|Time|
diff --git a/__tests__/__snapshots__/dart-json.test.ts.snap b/__tests__/__snapshots__/dart-json.test.ts.snap
index 88a7349..a499822 100644
--- a/__tests__/__snapshots__/dart-json.test.ts.snap
+++ b/__tests__/__snapshots__/dart-json.test.ts.snap
@@ -1,4 +1,4 @@
-// Jest Snapshot v1, https://jestjs.io/docs/snapshot-testing
+// Jest Snapshot v1, https://goo.gl/fbAQLP
exports[`dart-json tests matches report snapshot 1`] = `
TestRunResult {
diff --git a/__tests__/__snapshots__/dotnet-nunit.test.ts.snap b/__tests__/__snapshots__/dotnet-nunit.test.ts.snap
index 529f702..60d55f2 100644
--- a/__tests__/__snapshots__/dotnet-nunit.test.ts.snap
+++ b/__tests__/__snapshots__/dotnet-nunit.test.ts.snap
@@ -1,4 +1,4 @@
-// Jest Snapshot v1, https://jestjs.io/docs/snapshot-testing
+// Jest Snapshot v1, https://goo.gl/fbAQLP
exports[`dotnet-nunit tests report from ./reports/dotnet test results matches snapshot 1`] = `
TestRunResult {
diff --git a/__tests__/__snapshots__/dotnet-trx.test.ts.snap b/__tests__/__snapshots__/dotnet-trx.test.ts.snap
index b9d272d..1ca07eb 100644
--- a/__tests__/__snapshots__/dotnet-trx.test.ts.snap
+++ b/__tests__/__snapshots__/dotnet-trx.test.ts.snap
@@ -1,6 +1,6 @@
-// Jest Snapshot v1, https://jestjs.io/docs/snapshot-testing
+// Jest Snapshot v1, https://goo.gl/fbAQLP
-exports[`dotnet-trx tests matches dotnet-trx report snapshot 1`] = `
+exports[`dotnet-trx tests matches report snapshot 1`] = `
TestRunResult {
"path": "fixtures/dotnet-trx.trx",
"suites": [
@@ -21,9 +21,7 @@ TestRunResult {
at DotnetTests.Unit.Calculator.Div(Int32 a, Int32 b) in C:\\Users\\Michal\\Workspace\\dorny\\test-reporter\\reports\\dotnet\\DotnetTests.Unit\\Calculator.cs:line 9
at DotnetTests.XUnitTests.CalculatorTests.Exception_In_TargetTest() in C:\\Users\\Michal\\Workspace\\dorny\\test-reporter\\reports\\dotnet\\DotnetTests.XUnitTests\\CalculatorTests.cs:line 33",
"line": 9,
- "message": "System.DivideByZeroException : Attempted to divide by zero.
- at DotnetTests.Unit.Calculator.Div(Int32 a, Int32 b) in C:\\Users\\Michal\\Workspace\\dorny\\test-reporter\\reports\\dotnet\\DotnetTests.Unit\\Calculator.cs:line 9
- at DotnetTests.XUnitTests.CalculatorTests.Exception_In_TargetTest() in C:\\Users\\Michal\\Workspace\\dorny\\test-reporter\\reports\\dotnet\\DotnetTests.XUnitTests\\CalculatorTests.cs:line 33",
+ "message": "System.DivideByZeroException : Attempted to divide by zero.",
"path": "DotnetTests.Unit/Calculator.cs",
},
"name": "Exception_In_TargetTest",
@@ -35,8 +33,7 @@ TestRunResult {
"details": "System.Exception : Test
at DotnetTests.XUnitTests.CalculatorTests.Exception_In_Test() in C:\\Users\\Michal\\Workspace\\dorny\\test-reporter\\reports\\dotnet\\DotnetTests.XUnitTests\\CalculatorTests.cs:line 39",
"line": 39,
- "message": "System.Exception : Test
- at DotnetTests.XUnitTests.CalculatorTests.Exception_In_Test() in C:\\Users\\Michal\\Workspace\\dorny\\test-reporter\\reports\\dotnet\\DotnetTests.XUnitTests\\CalculatorTests.cs:line 39",
+ "message": "System.Exception : Test",
"path": "DotnetTests.XUnitTests/CalculatorTests.cs",
},
"name": "Exception_In_Test",
@@ -52,8 +49,7 @@ Actual: 2
"line": 27,
"message": "Assert.Equal() Failure
Expected: 3
-Actual: 2
- at DotnetTests.XUnitTests.CalculatorTests.Failing_Test() in C:\\Users\\Michal\\Workspace\\dorny\\test-reporter\\reports\\dotnet\\DotnetTests.XUnitTests\\CalculatorTests.cs:line 27",
+Actual: 2",
"path": "DotnetTests.XUnitTests/CalculatorTests.cs",
},
"name": "Failing_Test",
@@ -75,8 +71,7 @@ Actual: False
"line": 59,
"message": "Assert.True() Failure
Expected: True
-Actual: False
- at DotnetTests.XUnitTests.CalculatorTests.Is_Even_Number(Int32 i) in C:\\Users\\Michal\\Workspace\\dorny\\test-reporter\\reports\\dotnet\\DotnetTests.XUnitTests\\CalculatorTests.cs:line 59",
+Actual: False",
"path": "DotnetTests.XUnitTests/CalculatorTests.cs",
},
"name": "Is_Even_Number(i: 3)",
@@ -104,213 +99,7 @@ Actual: False
"line": 67,
"message": "Assert.True() Failure
Expected: True
-Actual: False
- at DotnetTests.XUnitTests.CalculatorTests.Theory_With_Custom_Name(Int32 i) in C:\\Users\\Michal\\Workspace\\dorny\\test-reporter\\reports\\dotnet\\DotnetTests.XUnitTests\\CalculatorTests.cs:line 67",
- "path": "DotnetTests.XUnitTests/CalculatorTests.cs",
- },
- "name": "Should be even number(i: 3)",
- "result": "failed",
- "time": 0.6537000000000001,
- },
- TestCaseResult {
- "error": undefined,
- "name": "Skipped_Test",
- "result": "skipped",
- "time": 1,
- },
- TestCaseResult {
- "error": undefined,
- "name": "Timeout_Test",
- "result": "success",
- "time": 108.42580000000001,
- },
- ],
- },
- ],
- "name": "DotnetTests.XUnitTests.CalculatorTests",
- "totalTime": undefined,
- },
- ],
- "totalTime": 1116,
-}
-`;
-
-exports[`dotnet-trx tests matches dotnet-xunitv3 report snapshot 1`] = `
-TestRunResult {
- "path": "fixtures/dotnet-xunitv3.trx",
- "suites": [
- TestSuiteResult {
- "groups": [
- TestGroupResult {
- "name": null,
- "tests": [
- TestCaseResult {
- "error": {
- "details": "Assert.Null() Failure: Value is not null
-Expected: null
-Actual: Fixture { }
- at DotnetTests.XUnitV3Tests.FixtureTests.Failing_Test() in /_/reports/dotnet/DotnetTests.XUnitV3Tests/FixtureTests.cs:line 25
- at System.RuntimeMethodHandle.InvokeMethod(Object target, Void** arguments, Signature sig, Boolean isConstructor)
- at System.Reflection.MethodBaseInvoker.InvokeWithNoArgs(Object obj, BindingFlags invokeAttr)",
- "line": 25,
- "message": "Assert.Null() Failure: Value is not null
-Expected: null
-Actual: Fixture { }
- at DotnetTests.XUnitV3Tests.FixtureTests.Failing_Test() in /_/reports/dotnet/DotnetTests.XUnitV3Tests/FixtureTests.cs:line 25
- at System.RuntimeMethodHandle.InvokeMethod(Object target, Void** arguments, Signature sig, Boolean isConstructor)
- at System.Reflection.MethodBaseInvoker.InvokeWithNoArgs(Object obj, BindingFlags invokeAttr)",
- "path": "DotnetTests.XUnitV3Tests/FixtureTests.cs",
- },
- "name": "Failing_Test",
- "result": "failed",
- "time": 17.0545,
- },
- TestCaseResult {
- "error": undefined,
- "name": "Passing_Test",
- "result": "success",
- "time": 0.8786,
- },
- ],
- },
- ],
- "name": "DotnetTests.XUnitV3Tests.FixtureTests",
- "totalTime": undefined,
- },
- TestSuiteResult {
- "groups": [
- TestGroupResult {
- "name": null,
- "tests": [
- TestCaseResult {
- "error": undefined,
- "name": "[Test Class Cleanup Failure (DotnetTests.XUnitV3Tests.FixtureTests.Failing_Test)]",
- "result": "failed",
- "time": 0,
- },
- TestCaseResult {
- "error": undefined,
- "name": "[Test Class Cleanup Failure (DotnetTests.XUnitV3Tests.FixtureTests.Passing_Test)]",
- "result": "failed",
- "time": 0,
- },
- ],
- },
- ],
- "name": "Unclassified",
- "totalTime": undefined,
- },
- ],
- "totalTime": 267,
-}
-`;
-
-exports[`dotnet-trx tests matches report snapshot (only failed tests) 1`] = `
-TestRunResult {
- "path": "fixtures/dotnet-trx.trx",
- "suites": [
- TestSuiteResult {
- "groups": [
- TestGroupResult {
- "name": null,
- "tests": [
- TestCaseResult {
- "error": undefined,
- "name": "Custom Name",
- "result": "success",
- "time": 0.1371,
- },
- TestCaseResult {
- "error": {
- "details": "System.DivideByZeroException : Attempted to divide by zero.
- at DotnetTests.Unit.Calculator.Div(Int32 a, Int32 b) in C:\\Users\\Michal\\Workspace\\dorny\\test-reporter\\reports\\dotnet\\DotnetTests.Unit\\Calculator.cs:line 9
- at DotnetTests.XUnitTests.CalculatorTests.Exception_In_TargetTest() in C:\\Users\\Michal\\Workspace\\dorny\\test-reporter\\reports\\dotnet\\DotnetTests.XUnitTests\\CalculatorTests.cs:line 33",
- "line": 9,
- "message": "System.DivideByZeroException : Attempted to divide by zero.
- at DotnetTests.Unit.Calculator.Div(Int32 a, Int32 b) in C:\\Users\\Michal\\Workspace\\dorny\\test-reporter\\reports\\dotnet\\DotnetTests.Unit\\Calculator.cs:line 9
- at DotnetTests.XUnitTests.CalculatorTests.Exception_In_TargetTest() in C:\\Users\\Michal\\Workspace\\dorny\\test-reporter\\reports\\dotnet\\DotnetTests.XUnitTests\\CalculatorTests.cs:line 33",
- "path": "DotnetTests.Unit/Calculator.cs",
- },
- "name": "Exception_In_TargetTest",
- "result": "failed",
- "time": 0.8377,
- },
- TestCaseResult {
- "error": {
- "details": "System.Exception : Test
- at DotnetTests.XUnitTests.CalculatorTests.Exception_In_Test() in C:\\Users\\Michal\\Workspace\\dorny\\test-reporter\\reports\\dotnet\\DotnetTests.XUnitTests\\CalculatorTests.cs:line 39",
- "line": 39,
- "message": "System.Exception : Test
- at DotnetTests.XUnitTests.CalculatorTests.Exception_In_Test() in C:\\Users\\Michal\\Workspace\\dorny\\test-reporter\\reports\\dotnet\\DotnetTests.XUnitTests\\CalculatorTests.cs:line 39",
- "path": "DotnetTests.XUnitTests/CalculatorTests.cs",
- },
- "name": "Exception_In_Test",
- "result": "failed",
- "time": 2.5175,
- },
- TestCaseResult {
- "error": {
- "details": "Assert.Equal() Failure
-Expected: 3
-Actual: 2
- at DotnetTests.XUnitTests.CalculatorTests.Failing_Test() in C:\\Users\\Michal\\Workspace\\dorny\\test-reporter\\reports\\dotnet\\DotnetTests.XUnitTests\\CalculatorTests.cs:line 27",
- "line": 27,
- "message": "Assert.Equal() Failure
-Expected: 3
-Actual: 2
- at DotnetTests.XUnitTests.CalculatorTests.Failing_Test() in C:\\Users\\Michal\\Workspace\\dorny\\test-reporter\\reports\\dotnet\\DotnetTests.XUnitTests\\CalculatorTests.cs:line 27",
- "path": "DotnetTests.XUnitTests/CalculatorTests.cs",
- },
- "name": "Failing_Test",
- "result": "failed",
- "time": 3.8697,
- },
- TestCaseResult {
- "error": undefined,
- "name": "Is_Even_Number(i: 2)",
- "result": "success",
- "time": 0.0078,
- },
- TestCaseResult {
- "error": {
- "details": "Assert.True() Failure
-Expected: True
-Actual: False
- at DotnetTests.XUnitTests.CalculatorTests.Is_Even_Number(Int32 i) in C:\\Users\\Michal\\Workspace\\dorny\\test-reporter\\reports\\dotnet\\DotnetTests.XUnitTests\\CalculatorTests.cs:line 59",
- "line": 59,
- "message": "Assert.True() Failure
-Expected: True
-Actual: False
- at DotnetTests.XUnitTests.CalculatorTests.Is_Even_Number(Int32 i) in C:\\Users\\Michal\\Workspace\\dorny\\test-reporter\\reports\\dotnet\\DotnetTests.XUnitTests\\CalculatorTests.cs:line 59",
- "path": "DotnetTests.XUnitTests/CalculatorTests.cs",
- },
- "name": "Is_Even_Number(i: 3)",
- "result": "failed",
- "time": 0.41409999999999997,
- },
- TestCaseResult {
- "error": undefined,
- "name": "Passing_Test",
- "result": "success",
- "time": 0.1365,
- },
- TestCaseResult {
- "error": undefined,
- "name": "Should be even number(i: 2)",
- "result": "success",
- "time": 0.0097,
- },
- TestCaseResult {
- "error": {
- "details": "Assert.True() Failure
-Expected: True
-Actual: False
- at DotnetTests.XUnitTests.CalculatorTests.Theory_With_Custom_Name(Int32 i) in C:\\Users\\Michal\\Workspace\\dorny\\test-reporter\\reports\\dotnet\\DotnetTests.XUnitTests\\CalculatorTests.cs:line 67",
- "line": 67,
- "message": "Assert.True() Failure
-Expected: True
-Actual: False
- at DotnetTests.XUnitTests.CalculatorTests.Theory_With_Custom_Name(Int32 i) in C:\\Users\\Michal\\Workspace\\dorny\\test-reporter\\reports\\dotnet\\DotnetTests.XUnitTests\\CalculatorTests.cs:line 67",
+Actual: False",
"path": "DotnetTests.XUnitTests/CalculatorTests.cs",
},
"name": "Should be even number(i: 3)",
diff --git a/__tests__/__snapshots__/golang-json.test.ts.snap b/__tests__/__snapshots__/golang-json.test.ts.snap
index bd28d4a..75c6de1 100644
--- a/__tests__/__snapshots__/golang-json.test.ts.snap
+++ b/__tests__/__snapshots__/golang-json.test.ts.snap
@@ -1,4 +1,4 @@
-// Jest Snapshot v1, https://jestjs.io/docs/snapshot-testing
+// Jest Snapshot v1, https://goo.gl/fbAQLP
exports[`golang-json tests report from ./reports/dotnet test results matches snapshot 1`] = `
TestRunResult {
diff --git a/__tests__/__snapshots__/java-junit.test.ts.snap b/__tests__/__snapshots__/java-junit.test.ts.snap
index 38daca9..341b092 100644
--- a/__tests__/__snapshots__/java-junit.test.ts.snap
+++ b/__tests__/__snapshots__/java-junit.test.ts.snap
@@ -1,4 +1,4 @@
-// Jest Snapshot v1, https://jestjs.io/docs/snapshot-testing
+// Jest Snapshot v1, https://goo.gl/fbAQLP
exports[`java-junit tests report from apache/pulsar single suite test results matches snapshot 1`] = `
TestRunResult {
diff --git a/__tests__/__snapshots__/jest-junit.test.ts.snap b/__tests__/__snapshots__/jest-junit.test.ts.snap
index eca0092..eb20dfe 100644
--- a/__tests__/__snapshots__/jest-junit.test.ts.snap
+++ b/__tests__/__snapshots__/jest-junit.test.ts.snap
@@ -1,4 +1,4 @@
-// Jest Snapshot v1, https://jestjs.io/docs/snapshot-testing
+// Jest Snapshot v1, https://goo.gl/fbAQLP
exports[`jest-junit tests parsing ESLint report without timing information works - PR #134 1`] = `
TestRunResult {
diff --git a/__tests__/__snapshots__/mocha-json.test.ts.snap b/__tests__/__snapshots__/mocha-json.test.ts.snap
index 4a1448c..7038239 100644
--- a/__tests__/__snapshots__/mocha-json.test.ts.snap
+++ b/__tests__/__snapshots__/mocha-json.test.ts.snap
@@ -1,4 +1,4 @@
-// Jest Snapshot v1, https://jestjs.io/docs/snapshot-testing
+// Jest Snapshot v1, https://goo.gl/fbAQLP
exports[`mocha-json tests report from ./reports/mocha-json test results matches snapshot 1`] = `
TestRunResult {
diff --git a/__tests__/__snapshots__/python-xunit.test.ts.snap b/__tests__/__snapshots__/python-xunit.test.ts.snap
deleted file mode 100644
index f325c84..0000000
--- a/__tests__/__snapshots__/python-xunit.test.ts.snap
+++ /dev/null
@@ -1,192 +0,0 @@
-// Jest Snapshot v1, https://jestjs.io/docs/snapshot-testing
-
-exports[`python-xunit pytest report report from python test results matches snapshot 1`] = `
-TestRunResult {
- "path": "fixtures/python-xunit-pytest.xml",
- "suites": [
- TestSuiteResult {
- "groups": [
- TestGroupResult {
- "name": "tests.test_lib",
- "tests": [
- TestCaseResult {
- "error": undefined,
- "name": "test_always_pass",
- "result": "success",
- "time": 2,
- },
- TestCaseResult {
- "error": undefined,
- "name": "test_with_subtests",
- "result": "success",
- "time": 5,
- },
- TestCaseResult {
- "error": undefined,
- "name": "test_parameterized[param1]",
- "result": "success",
- "time": 0,
- },
- TestCaseResult {
- "error": undefined,
- "name": "test_parameterized[param2]",
- "result": "success",
- "time": 0,
- },
- TestCaseResult {
- "error": undefined,
- "name": "test_always_skip",
- "result": "skipped",
- "time": 0,
- },
- TestCaseResult {
- "error": {
- "details": "def test_always_fail():
- > assert False
- E assert False
-
- tests/test_lib.py:25: AssertionError
- ",
- "line": undefined,
- "message": "assert False",
- "path": undefined,
- },
- "name": "test_always_fail",
- "result": "failed",
- "time": 0,
- },
- TestCaseResult {
- "error": undefined,
- "name": "test_expected_failure",
- "result": "skipped",
- "time": 0,
- },
- TestCaseResult {
- "error": {
- "details": "def test_error():
- > raise Exception("error")
- E Exception: error
-
- tests/test_lib.py:32: Exception
- ",
- "line": undefined,
- "message": "Exception: error",
- "path": undefined,
- },
- "name": "test_error",
- "result": "failed",
- "time": 0,
- },
- TestCaseResult {
- "error": undefined,
- "name": "test_with_record_property",
- "result": "success",
- "time": 0,
- },
- ],
- },
- TestGroupResult {
- "name": "custom_classname",
- "tests": [
- TestCaseResult {
- "error": undefined,
- "name": "test_with_record_xml_attribute",
- "result": "success",
- "time": 0,
- },
- ],
- },
- ],
- "name": "pytest",
- "totalTime": 19,
- },
- ],
- "totalTime": undefined,
-}
-`;
-
-exports[`python-xunit unittest report report from python test results matches snapshot 1`] = `
-TestRunResult {
- "path": "fixtures/python-xunit-unittest.xml",
- "suites": [
- TestSuiteResult {
- "groups": [
- TestGroupResult {
- "name": "TestAcme",
- "tests": [
- TestCaseResult {
- "error": undefined,
- "name": "test_always_pass",
- "result": "success",
- "time": 0,
- },
- TestCaseResult {
- "error": undefined,
- "name": "test_parameterized_0_param1",
- "result": "success",
- "time": 1,
- },
- TestCaseResult {
- "error": undefined,
- "name": "test_parameterized_1_param2",
- "result": "success",
- "time": 0,
- },
- TestCaseResult {
- "error": undefined,
- "name": "test_with_subtests",
- "result": "success",
- "time": 0,
- },
- TestCaseResult {
- "error": {
- "details": "Traceback (most recent call last):
- File "/Users/foo/Projects/python-test/tests/test_lib.py", line 24, in test_always_fail
- self.fail("failed")
-AssertionError: failed
-",
- "line": undefined,
- "message": "AssertionError: failed",
- "path": undefined,
- },
- "name": "test_always_fail",
- "result": "failed",
- "time": 0,
- },
- TestCaseResult {
- "error": {
- "details": "Traceback (most recent call last):
- File "/Users/foo/Projects/python-test/tests/test_lib.py", line 31, in test_error
- raise Exception("error")
-Exception: error
-",
- "line": undefined,
- "message": "Exception: error",
- "path": undefined,
- },
- "name": "test_error",
- "result": "failed",
- "time": 0,
- },
- TestCaseResult {
- "error": undefined,
- "name": "test_always_skip",
- "result": "skipped",
- "time": 0,
- },
- TestCaseResult {
- "error": undefined,
- "name": "test_expected_failure",
- "result": "skipped",
- "time": 0,
- },
- ],
- },
- ],
- "name": "TestAcme-20251114214921",
- "totalTime": 1,
- },
- ],
- "totalTime": 1,
-}
-`;
diff --git a/__tests__/__snapshots__/rspec-json.test.ts.snap b/__tests__/__snapshots__/rspec-json.test.ts.snap
index 51c1943..cc14bfb 100644
--- a/__tests__/__snapshots__/rspec-json.test.ts.snap
+++ b/__tests__/__snapshots__/rspec-json.test.ts.snap
@@ -1,4 +1,4 @@
-// Jest Snapshot v1, https://jestjs.io/docs/snapshot-testing
+// Jest Snapshot v1, https://goo.gl/fbAQLP
exports[`rspec-json tests report from ./reports/rspec-json test results matches snapshot 1`] = `
TestRunResult {
diff --git a/__tests__/__snapshots__/swift-xunit.test.ts.snap b/__tests__/__snapshots__/swift-xunit.test.ts.snap
index bddc6ea..ae34deb 100644
--- a/__tests__/__snapshots__/swift-xunit.test.ts.snap
+++ b/__tests__/__snapshots__/swift-xunit.test.ts.snap
@@ -1,4 +1,4 @@
-// Jest Snapshot v1, https://jestjs.io/docs/snapshot-testing
+// Jest Snapshot v1, https://goo.gl/fbAQLP
exports[`swift-xunit tests report from swift test results matches snapshot 1`] = `
TestRunResult {
diff --git a/__tests__/dotnet-trx.test.ts b/__tests__/dotnet-trx.test.ts
index a7aead6..e7f83ee 100644
--- a/__tests__/dotnet-trx.test.ts
+++ b/__tests__/dotnet-trx.test.ts
@@ -3,7 +3,7 @@ import * as path from 'path'
import {DotnetTrxParser} from '../src/parsers/dotnet-trx/dotnet-trx-parser'
import {ParseOptions} from '../src/test-parser'
-import {DEFAULT_OPTIONS, getReport, ReportOptions} from '../src/report/get-report'
+import {DEFAULT_OPTIONS, getReport} from '../src/report/get-report'
import {normalizeFilePath} from '../src/utils/path-utils'
describe('dotnet-trx tests', () => {
@@ -39,34 +39,9 @@ describe('dotnet-trx tests', () => {
expect(result.result).toBe('success')
})
- it.each([['dotnet-trx'], ['dotnet-xunitv3']])('matches %s report snapshot', async reportName => {
- const fixturePath = path.join(__dirname, 'fixtures', `${reportName}.trx`)
- const outputPath = path.join(__dirname, '__outputs__', `${reportName}.md`)
- const filePath = normalizeFilePath(path.relative(__dirname, fixturePath))
- const fileContent = fs.readFileSync(fixturePath, {encoding: 'utf8'})
-
- const opts: ParseOptions = {
- parseErrors: true,
- trackedFiles: [
- 'DotnetTests.Unit/Calculator.cs',
- 'DotnetTests.XUnitTests/CalculatorTests.cs',
- 'DotnetTests.XUnitV3Tests/FixtureTests.cs'
- ]
- //workDir: 'C:/Users/Michal/Workspace/dorny/test-check/reports/dotnet/'
- }
-
- const parser = new DotnetTrxParser(opts)
- const result = await parser.parse(filePath, fileContent)
- expect(result).toMatchSnapshot()
-
- const report = getReport([result])
- fs.mkdirSync(path.dirname(outputPath), {recursive: true})
- fs.writeFileSync(outputPath, report)
- })
-
- it('matches report snapshot (only failed tests)', async () => {
+ it('matches report snapshot', async () => {
const fixturePath = path.join(__dirname, 'fixtures', 'dotnet-trx.trx')
- const outputPath = path.join(__dirname, '__outputs__', 'dotnet-trx-only-failed.md')
+ const outputPath = path.join(__dirname, '__outputs__', 'dotnet-trx.md')
const filePath = normalizeFilePath(path.relative(__dirname, fixturePath))
const fileContent = fs.readFileSync(fixturePath, {encoding: 'utf8'})
@@ -80,12 +55,7 @@ describe('dotnet-trx tests', () => {
const result = await parser.parse(filePath, fileContent)
expect(result).toMatchSnapshot()
- const reportOptions: ReportOptions = {
- ...DEFAULT_OPTIONS,
- listSuites: 'all',
- listTests: 'failed'
- }
- const report = getReport([result], reportOptions)
+ const report = getReport([result])
fs.mkdirSync(path.dirname(outputPath), {recursive: true})
fs.writeFileSync(outputPath, report)
})
diff --git a/__tests__/fixtures/dotnet-xunitv3.trx b/__tests__/fixtures/dotnet-xunitv3.trx
deleted file mode 100644
index 6bd9c25..0000000
--- a/__tests__/fixtures/dotnet-xunitv3.trx
+++ /dev/null
@@ -1,60 +0,0 @@
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
- Exit code indicates failure: '2'. Please refer to https://aka.ms/testingplatform/exitcodes for more information.
-
-
-
-
\ No newline at end of file
diff --git a/__tests__/fixtures/python-xunit-pytest.xml b/__tests__/fixtures/python-xunit-pytest.xml
deleted file mode 100644
index fcb044a..0000000
--- a/__tests__/fixtures/python-xunit-pytest.xml
+++ /dev/null
@@ -1,42 +0,0 @@
-
-
-
-
-
-
-
-
-
-
-
- /Users/mike/Projects/python-test/tests/test_lib.py:20: skipped
-
-
-
- def test_always_fail():
- > assert False
- E assert False
-
- tests/test_lib.py:25: AssertionError
-
-
-
-
-
-
- def test_error():
- > raise Exception("error")
- E Exception: error
-
- tests/test_lib.py:32: Exception
-
-
-
-
-
-
-
-
-
-
diff --git a/__tests__/fixtures/python-xunit-unittest.xml b/__tests__/fixtures/python-xunit-unittest.xml
deleted file mode 100644
index ecc67d4..0000000
--- a/__tests__/fixtures/python-xunit-unittest.xml
+++ /dev/null
@@ -1,27 +0,0 @@
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
diff --git a/__tests__/jest-junit.test.ts b/__tests__/jest-junit.test.ts
index 912ebde..f4b8335 100644
--- a/__tests__/jest-junit.test.ts
+++ b/__tests__/jest-junit.test.ts
@@ -207,143 +207,4 @@ describe('jest-junit tests', () => {
// Report should have the title as the first line
expect(report).toMatch(/^# My Custom Title\n/)
})
-
- it('report can be collapsed when configured', async () => {
- const fixturePath = path.join(__dirname, 'fixtures', 'jest-junit.xml')
- const filePath = normalizeFilePath(path.relative(__dirname, fixturePath))
- const fileContent = fs.readFileSync(fixturePath, {encoding: 'utf8'})
-
- const opts: ParseOptions = {
- parseErrors: true,
- trackedFiles: []
- }
-
- const parser = new JestJunitParser(opts)
- const result = await parser.parse(filePath, fileContent)
- const report = getReport([result], {
- ...DEFAULT_OPTIONS,
- collapsed: 'always'
- })
- // Report should include collapsible details
- expect(report).toContain('Expand for details
')
- expect(report).toContain(' ')
- })
-
- it('report is not collapsed when configured to never', async () => {
- const fixturePath = path.join(__dirname, 'fixtures', 'jest-junit.xml')
- const filePath = normalizeFilePath(path.relative(__dirname, fixturePath))
- const fileContent = fs.readFileSync(fixturePath, {encoding: 'utf8'})
-
- const opts: ParseOptions = {
- parseErrors: true,
- trackedFiles: []
- }
-
- const parser = new JestJunitParser(opts)
- const result = await parser.parse(filePath, fileContent)
- const report = getReport([result], {
- ...DEFAULT_OPTIONS,
- collapsed: 'never'
- })
- // Report should not include collapsible details
- expect(report).not.toContain('Expand for details
')
- expect(report).not.toContain(' ')
- })
-
- it('report auto-collapses when all tests pass', async () => {
- // Test with a fixture that has all passing tests (no failures)
- const fixturePath = path.join(__dirname, 'fixtures', 'jest-junit-eslint.xml')
- const filePath = normalizeFilePath(path.relative(__dirname, fixturePath))
- const fileContent = fs.readFileSync(fixturePath, {encoding: 'utf8'})
-
- const opts: ParseOptions = {
- parseErrors: true,
- trackedFiles: []
- }
-
- const parser = new JestJunitParser(opts)
- const result = await parser.parse(filePath, fileContent)
-
- // Verify this fixture has no failures
- expect(result.failed).toBe(0)
-
- const report = getReport([result], {
- ...DEFAULT_OPTIONS,
- collapsed: 'auto'
- })
-
- // Should collapse when all tests pass
- expect(report).toContain('Expand for details
')
- expect(report).toContain(' ')
- })
-
- it('report does not auto-collapse when tests fail', async () => {
- // Test with a fixture that has failing tests
- const fixturePath = path.join(__dirname, 'fixtures', 'jest-junit.xml')
- const filePath = normalizeFilePath(path.relative(__dirname, fixturePath))
- const fileContent = fs.readFileSync(fixturePath, {encoding: 'utf8'})
-
- const opts: ParseOptions = {
- parseErrors: true,
- trackedFiles: []
- }
-
- const parser = new JestJunitParser(opts)
- const result = await parser.parse(filePath, fileContent)
-
- // Verify this fixture has failures
- expect(result.failed).toBeGreaterThan(0)
-
- const report = getReport([result], {
- ...DEFAULT_OPTIONS,
- collapsed: 'auto'
- })
-
- // Should not collapse when there are failures
- expect(report).not.toContain('Expand for details
')
- expect(report).not.toContain(' ')
- })
-
- it('report includes the short summary', async () => {
- const fixturePath = path.join(__dirname, 'fixtures', 'jest-junit.xml')
- const filePath = normalizeFilePath(path.relative(__dirname, fixturePath))
- const fileContent = fs.readFileSync(fixturePath, {encoding: 'utf8'})
-
- const opts: ParseOptions = {
- parseErrors: true,
- trackedFiles: []
- }
-
- const parser = new JestJunitParser(opts)
- const result = await parser.parse(filePath, fileContent)
- const shortSummary = '1 passed, 4 failed and 1 skipped'
- const report = getReport([result], DEFAULT_OPTIONS, shortSummary)
- // Report should have the title as the first line
- expect(report).toMatch(/^## 1 passed, 4 failed and 1 skipped\n/)
- })
-
- it('report includes a custom report title and short summary', async () => {
- const fixturePath = path.join(__dirname, 'fixtures', 'jest-junit.xml')
- const filePath = normalizeFilePath(path.relative(__dirname, fixturePath))
- const fileContent = fs.readFileSync(fixturePath, {encoding: 'utf8'})
-
- const opts: ParseOptions = {
- parseErrors: true,
- trackedFiles: []
- }
-
- const parser = new JestJunitParser(opts)
- const result = await parser.parse(filePath, fileContent)
- const shortSummary = '1 passed, 4 failed and 1 skipped'
- const report = getReport(
- [result],
- {
- ...DEFAULT_OPTIONS,
- reportTitle: 'My Custom Title'
- },
- shortSummary
- )
- // Report should have the title as the first line
- expect(report).toMatch(/^# My Custom Title\n## 1 passed, 4 failed and 1 skipped\n/)
- })
})
diff --git a/__tests__/python-xunit.test.ts b/__tests__/python-xunit.test.ts
deleted file mode 100644
index c1550a4..0000000
--- a/__tests__/python-xunit.test.ts
+++ /dev/null
@@ -1,93 +0,0 @@
-import * as fs from 'fs'
-import * as path from 'path'
-
-import {PythonXunitParser} from '../src/parsers/python-xunit/python-xunit-parser'
-import {ParseOptions} from '../src/test-parser'
-import {DEFAULT_OPTIONS, getReport} from '../src/report/get-report'
-import {normalizeFilePath} from '../src/utils/path-utils'
-
-const defaultOpts: ParseOptions = {
- parseErrors: true,
- trackedFiles: []
-}
-
-describe('python-xunit unittest report', () => {
- const fixturePath = path.join(__dirname, 'fixtures', 'python-xunit-unittest.xml')
- const filePath = normalizeFilePath(path.relative(__dirname, fixturePath))
- const fileContent = fs.readFileSync(fixturePath, {encoding: 'utf8'})
- const outputPath = path.join(__dirname, '__outputs__', 'python-xunit-unittest.md')
-
- it('report from python test results matches snapshot', async () => {
- const trackedFiles = ['tests/test_lib.py']
- const opts: ParseOptions = {
- ...defaultOpts,
- trackedFiles
- }
-
- const parser = new PythonXunitParser(opts)
- const result = await parser.parse(filePath, fileContent)
- expect(result).toMatchSnapshot()
-
- const report = getReport([result])
- fs.mkdirSync(path.dirname(outputPath), {recursive: true})
- fs.writeFileSync(outputPath, report)
- })
-
- it('report does not include a title by default', async () => {
- const parser = new PythonXunitParser(defaultOpts)
- const result = await parser.parse(filePath, fileContent)
- const report = getReport([result])
- // Report should have the badge as the first line
- expect(report).toMatch(/^!\[Tests failed]/)
- })
-
- it.each([
- ['empty string', ''],
- ['space', ' '],
- ['tab', '\t'],
- ['newline', '\n']
- ])('report does not include a title when configured value is %s', async (_, reportTitle) => {
- const parser = new PythonXunitParser(defaultOpts)
- const result = await parser.parse(filePath, fileContent)
- const report = getReport([result], {
- ...DEFAULT_OPTIONS,
- reportTitle
- })
- // Report should have the badge as the first line
- expect(report).toMatch(/^!\[Tests failed]/)
- })
-
- it('report includes a custom report title', async () => {
- const parser = new PythonXunitParser(defaultOpts)
- const result = await parser.parse(filePath, fileContent)
- const report = getReport([result], {
- ...DEFAULT_OPTIONS,
- reportTitle: 'My Custom Title'
- })
- // Report should have the title as the first line
- expect(report).toMatch(/^# My Custom Title\n/)
- })
-})
-
-describe('python-xunit pytest report', () => {
- const fixturePath = path.join(__dirname, 'fixtures', 'python-xunit-pytest.xml')
- const filePath = normalizeFilePath(path.relative(__dirname, fixturePath))
- const fileContent = fs.readFileSync(fixturePath, {encoding: 'utf8'})
- const outputPath = path.join(__dirname, '__outputs__', 'python-xunit-pytest.md')
-
- it('report from python test results matches snapshot', async () => {
- const trackedFiles = ['tests/test_lib.py']
- const opts: ParseOptions = {
- ...defaultOpts,
- trackedFiles
- }
-
- const parser = new PythonXunitParser(opts)
- const result = await parser.parse(filePath, fileContent)
- expect(result).toMatchSnapshot()
-
- const report = getReport([result])
- fs.mkdirSync(path.dirname(outputPath), {recursive: true})
- fs.writeFileSync(outputPath, report)
- })
-})
diff --git a/__tests__/report/get-report.test.ts b/__tests__/report/get-report.test.ts
deleted file mode 100644
index 670b0ad..0000000
--- a/__tests__/report/get-report.test.ts
+++ /dev/null
@@ -1,120 +0,0 @@
-import {getBadge, DEFAULT_OPTIONS, ReportOptions} from '../../src/report/get-report'
-
-describe('getBadge', () => {
- describe('URI encoding with special characters', () => {
- it('generates correct URI with simple badge title', () => {
- const options: ReportOptions = {
- ...DEFAULT_OPTIONS,
- badgeTitle: 'tests'
- }
- const badge = getBadge(5, 0, 1, options)
- expect(badge).toBe('')
- })
-
- it('handles badge title with single hyphen', () => {
- const options: ReportOptions = {
- ...DEFAULT_OPTIONS,
- badgeTitle: 'unit-tests'
- }
- const badge = getBadge(3, 0, 0, options)
- // The hyphen in the badge title should be encoded as --
- expect(badge).toBe('')
- })
-
- it('handles badge title with multiple hyphens', () => {
- const options: ReportOptions = {
- ...DEFAULT_OPTIONS,
- badgeTitle: 'integration-api-tests'
- }
- const badge = getBadge(10, 0, 0, options)
- // All hyphens in the title should be encoded as --
- expect(badge).toBe('')
- })
-
- it('handles badge title with multiple underscores', () => {
- const options: ReportOptions = {
- ...DEFAULT_OPTIONS,
- badgeTitle: 'my_integration_test'
- }
- const badge = getBadge(10, 0, 0, options)
- // All underscores in the title should be encoded as __
- expect(badge).toBe('')
- })
-
- it('handles badge title with version format containing hyphen', () => {
- const options: ReportOptions = {
- ...DEFAULT_OPTIONS,
- badgeTitle: 'MariaDb 12.0-ubi database tests'
- }
- const badge = getBadge(1, 0, 0, options)
- // The hyphen in "12.0-ubi" should be encoded as --
- expect(badge).toBe('')
- })
-
- it('handles badge title with dots and hyphens', () => {
- const options: ReportOptions = {
- ...DEFAULT_OPTIONS,
- badgeTitle: 'v1.2.3-beta-test'
- }
- const badge = getBadge(4, 1, 0, options)
- expect(badge).toBe('')
- })
-
- it('preserves structural hyphens between label and message', () => {
- const options: ReportOptions = {
- ...DEFAULT_OPTIONS,
- badgeTitle: 'test-suite'
- }
- const badge = getBadge(2, 3, 1, options)
- // The URI should have literal hyphens separating title-message-color
- expect(badge).toBe('')
- })
- })
-
- describe('generates test outcome as color name for imgshields', () => {
- it('uses success color when all tests pass', () => {
- const options: ReportOptions = {...DEFAULT_OPTIONS}
- const badge = getBadge(5, 0, 0, options)
- expect(badge).toContain('-success)')
- })
-
- it('uses critical color when tests fail', () => {
- const options: ReportOptions = {...DEFAULT_OPTIONS}
- const badge = getBadge(5, 2, 0, options)
- expect(badge).toContain('-critical)')
- })
-
- it('uses yellow color when no tests found', () => {
- const options: ReportOptions = {...DEFAULT_OPTIONS}
- const badge = getBadge(0, 0, 0, options)
- expect(badge).toContain('-yellow)')
- })
- })
-
- describe('badge message composition', () => {
- it('includes only passed count when no failures or skips', () => {
- const options: ReportOptions = {...DEFAULT_OPTIONS}
- const badge = getBadge(5, 0, 0, options)
- expect(badge).toBe('')
- })
-
- it('includes passed and failed counts', () => {
- const options: ReportOptions = {...DEFAULT_OPTIONS}
- const badge = getBadge(5, 2, 0, options)
- expect(badge).toBe('')
- })
-
- it('includes passed, failed and skipped counts', () => {
- const options: ReportOptions = {...DEFAULT_OPTIONS}
- const badge = getBadge(5, 2, 1, options)
- expect(badge).toBe('')
- })
-
- it('uses "none" message when no tests', () => {
- const options: ReportOptions = {...DEFAULT_OPTIONS}
- const badge = getBadge(0, 0, 0, options)
- expect(badge).toBe('')
- })
- })
-})
-
diff --git a/__tests__/utils/parse-utils.test.ts b/__tests__/utils/parse-utils.test.ts
index 0f02867..83689ef 100644
--- a/__tests__/utils/parse-utils.test.ts
+++ b/__tests__/utils/parse-utils.test.ts
@@ -32,6 +32,6 @@ describe('parseNetDuration', () => {
})
it('throws when string has invalid format', () => {
- expect(() => parseNetDuration('12:34:56 not a duration')).toThrow(/^Invalid format/)
+ expect(() => parseNetDuration('12:34:56 not a duration')).toThrowError(/^Invalid format/)
})
})
diff --git a/action.yml b/action.yml
index 530435c..ec4772f 100644
--- a/action.yml
+++ b/action.yml
@@ -1,5 +1,6 @@
name: Test Reporter
-description: Displays test results from popular testing frameworks directly in GitHub
+description: |
+ Shows test results in GitHub UI: .NET (xUnit, NUnit, MSTest), Dart, Flutter, Java (JUnit), JavaScript (JEST, Mocha)
author: Michal Dorner
inputs:
artifact:
@@ -28,11 +29,9 @@ inputs:
- dotnet-nunit
- dotnet-trx
- flutter-json
- - golang-json
- java-junit
- jest-junit
- mocha-json
- - python-xunit
- rspec-json
- swift-xunit
required: true
@@ -69,10 +68,6 @@ inputs:
working-directory:
description: Relative path under $GITHUB_WORKSPACE where the repository was checked out
required: false
- report-title:
- description: Title for the test report summary
- required: false
- default: ''
only-summary:
description: |
Allows you to generate only the summary.
@@ -90,14 +85,6 @@ inputs:
description: Customize badge title
required: false
default: 'tests'
- collapsed:
- description: |
- Controls whether test report details are collapsed or expanded. Supported options:
- - auto: Collapse only if all tests pass (default behavior)
- - always: Always collapse the report details
- - never: Always expand the report details
- required: false
- default: 'auto'
token:
description: GitHub Access Token
required: false
diff --git a/dist/index.js b/dist/index.js
index 42cb52e..9cb9528 100644
--- a/dist/index.js
+++ b/dist/index.js
@@ -277,7 +277,6 @@ const golang_json_parser_1 = __nccwpck_require__(5162);
const java_junit_parser_1 = __nccwpck_require__(8342);
const jest_junit_parser_1 = __nccwpck_require__(1042);
const mocha_json_parser_1 = __nccwpck_require__(5402);
-const python_xunit_parser_1 = __nccwpck_require__(6578);
const rspec_json_parser_1 = __nccwpck_require__(9768);
const swift_xunit_parser_1 = __nccwpck_require__(7330);
const path_utils_1 = __nccwpck_require__(9132);
@@ -310,7 +309,6 @@ class TestReporter {
useActionsSummary = core.getInput('use-actions-summary', { required: false }) === 'true';
badgeTitle = core.getInput('badge-title', { required: false });
reportTitle = core.getInput('report-title', { required: false });
- collapsed = core.getInput('collapsed', { required: false });
token = core.getInput('token', { required: true });
octokit;
context = (0, github_utils_1.getCheckRunContext)();
@@ -324,10 +322,6 @@ class TestReporter {
core.setFailed(`Input parameter 'list-tests' has invalid value`);
return;
}
- if (this.collapsed !== 'auto' && this.collapsed !== 'always' && this.collapsed !== 'never') {
- core.setFailed(`Input parameter 'collapsed' has invalid value`);
- return;
- }
if (isNaN(this.maxAnnotations) || this.maxAnnotations < 0 || this.maxAnnotations > 50) {
core.setFailed(`Input parameter 'max-annotations' has invalid value`);
return;
@@ -407,11 +401,7 @@ class TestReporter {
throw error;
}
}
- const { listSuites, listTests, onlySummary, useActionsSummary, badgeTitle, reportTitle, collapsed } = this;
- const passed = results.reduce((sum, tr) => sum + tr.passed, 0);
- const failed = results.reduce((sum, tr) => sum + tr.failed, 0);
- const skipped = results.reduce((sum, tr) => sum + tr.skipped, 0);
- const shortSummary = `${passed} passed, ${failed} failed and ${skipped} skipped `;
+ const { listSuites, listTests, onlySummary, useActionsSummary, badgeTitle, reportTitle } = this;
let baseUrl = '';
if (this.useActionsSummary) {
const summary = (0, get_report_1.getReport)(results, {
@@ -421,9 +411,8 @@ class TestReporter {
onlySummary,
useActionsSummary,
badgeTitle,
- reportTitle,
- collapsed
- }, shortSummary);
+ reportTitle
+ });
core.info('Summary content:');
core.info(summary);
await core.summary.addRaw(summary).write();
@@ -449,13 +438,16 @@ class TestReporter {
onlySummary,
useActionsSummary,
badgeTitle,
- reportTitle,
- collapsed
+ reportTitle
});
core.info('Creating annotations');
const annotations = (0, get_annotations_1.getAnnotations)(results, this.maxAnnotations);
const isFailed = this.failOnError && results.some(tr => tr.result === 'failed');
const conclusion = isFailed ? 'failure' : 'success';
+ const passed = results.reduce((sum, tr) => sum + tr.passed, 0);
+ const failed = results.reduce((sum, tr) => sum + tr.failed, 0);
+ const skipped = results.reduce((sum, tr) => sum + tr.skipped, 0);
+ const shortSummary = `${passed} passed, ${failed} failed and ${skipped} skipped `;
core.info(`Updating check run conclusion (${conclusion}) and output`);
const resp = await this.octokit.rest.checks.update({
check_run_id: createResp.data.id,
@@ -494,8 +486,6 @@ class TestReporter {
return new jest_junit_parser_1.JestJunitParser(options);
case 'mocha-json':
return new mocha_json_parser_1.MochaJsonParser(options);
- case 'python-xunit':
- return new python_xunit_parser_1.PythonXunitParser(options);
case 'rspec-json':
return new rspec_json_parser_1.RspecJsonParser(options);
case 'swift-xunit':
@@ -846,12 +836,12 @@ class DotnetNunitParser {
.map(suite => suite.$.name)
.join('.');
const groupName = suitesWithoutTheories[suitesWithoutTheories.length - 1].$.name;
- let existingSuite = result.find(suite => suite.name === suiteName);
+ let existingSuite = result.find(existingSuite => existingSuite.name === suiteName);
if (existingSuite === undefined) {
existingSuite = new test_results_1.TestSuiteResult(suiteName, []);
result.push(existingSuite);
}
- let existingGroup = existingSuite.groups.find(group => group.name === groupName);
+ let existingGroup = existingSuite.groups.find(existingGroup => existingGroup.name === groupName);
if (existingGroup === undefined) {
existingGroup = new test_results_1.TestGroupResult(groupName, []);
existingSuite.groups.push(existingGroup);
@@ -983,7 +973,7 @@ class DotnetTrxParser {
}));
const testClasses = {};
for (const r of unitTestsResults) {
- const className = r.test.TestMethod[0].$.className ?? "Unclassified";
+ const className = r.test.TestMethod[0].$.className;
let tc = testClasses[className];
if (tc === undefined) {
tc = new TestClass(className);
@@ -1034,8 +1024,8 @@ class DotnetTrxParser {
error.StackTrace.length === 0) {
return undefined;
}
+ const message = test.error.Message[0];
const stackTrace = test.error.StackTrace[0];
- const message = `${test.error.Message[0]}\n${stackTrace}`;
let path;
let line;
const src = this.exceptionThrowSource(stackTrace);
@@ -1047,7 +1037,7 @@ class DotnetTrxParser {
path,
line,
message,
- details: `${message}`
+ details: `${message}\n${stackTrace}`
};
}
exceptionThrowSource(stackTrace) {
@@ -1666,26 +1656,6 @@ class MochaJsonParser {
exports.MochaJsonParser = MochaJsonParser;
-/***/ }),
-
-/***/ 6578:
-/***/ ((__unused_webpack_module, exports, __nccwpck_require__) => {
-
-"use strict";
-
-Object.defineProperty(exports, "__esModule", ({ value: true }));
-exports.PythonXunitParser = void 0;
-const java_junit_parser_1 = __nccwpck_require__(8342);
-class PythonXunitParser extends java_junit_parser_1.JavaJunitParser {
- options;
- constructor(options) {
- super(options);
- this.options = options;
- }
-}
-exports.PythonXunitParser = PythonXunitParser;
-
-
/***/ }),
/***/ 9768:
@@ -1938,7 +1908,6 @@ var __importStar = (this && this.__importStar) || (function () {
Object.defineProperty(exports, "__esModule", ({ value: true }));
exports.DEFAULT_OPTIONS = void 0;
exports.getReport = getReport;
-exports.getBadge = getBadge;
const core = __importStar(__nccwpck_require__(7484));
const markdown_utils_1 = __nccwpck_require__(5129);
const node_utils_1 = __nccwpck_require__(5384);
@@ -1953,13 +1922,13 @@ exports.DEFAULT_OPTIONS = {
onlySummary: false,
useActionsSummary: true,
badgeTitle: 'tests',
- reportTitle: '',
- collapsed: 'auto'
+ reportTitle: ''
};
-function getReport(results, options = exports.DEFAULT_OPTIONS, shortSummary = '') {
+function getReport(results, options = exports.DEFAULT_OPTIONS) {
+ core.info('Generating check run summary');
applySort(results);
const opts = { ...options };
- let lines = renderReport(results, opts, shortSummary);
+ let lines = renderReport(results, opts);
let report = lines.join('\n');
if (getByteLength(report) <= getMaxReportLength(options)) {
return report;
@@ -1967,7 +1936,7 @@ function getReport(results, options = exports.DEFAULT_OPTIONS, shortSummary = ''
if (opts.listTests === 'all') {
core.info("Test report summary is too big - setting 'listTests' to 'failed'");
opts.listTests = 'failed';
- lines = renderReport(results, opts, shortSummary);
+ lines = renderReport(results, opts);
report = lines.join('\n');
if (getByteLength(report) <= getMaxReportLength(options)) {
return report;
@@ -2014,15 +1983,12 @@ function applySort(results) {
function getByteLength(text) {
return Buffer.byteLength(text, 'utf8');
}
-function renderReport(results, options, shortSummary) {
+function renderReport(results, options) {
const sections = [];
const reportTitle = options.reportTitle.trim();
if (reportTitle) {
sections.push(`# ${reportTitle}`);
}
- if (shortSummary) {
- sections.push(`## ${shortSummary}`);
- }
const badge = getReportBadge(results, options);
sections.push(badge);
const runs = getTestRunsReport(results, options);
@@ -2055,33 +2021,26 @@ function getBadge(passed, failed, skipped, options) {
color = 'yellow';
}
const hint = failed > 0 ? 'Tests failed' : 'Tests passed successfully';
- const encodedBadgeTitle = encodeImgShieldsURIComponent(options.badgeTitle);
- const encodedMessage = encodeImgShieldsURIComponent(message);
- const encodedColor = encodeImgShieldsURIComponent(color);
- return ``;
+ const uri = encodeURIComponent(`${options.badgeTitle}-${message}-${color}`);
+ return ``;
}
function getTestRunsReport(testRuns, options) {
const sections = [];
const totalFailed = testRuns.reduce((sum, tr) => sum + tr.failed, 0);
- // Determine if report should be collapsed based on collapsed option
- const shouldCollapse = options.collapsed === 'always' || (options.collapsed === 'auto' && totalFailed === 0);
- if (shouldCollapse) {
+ if (totalFailed === 0) {
sections.push(`Expand for details
`);
sections.push(` `);
}
if (testRuns.length > 0 || options.onlySummary) {
const tableData = testRuns
- .map((tr, originalIndex) => ({ tr, originalIndex }))
- .filter(({ tr }) => tr.passed > 0 || tr.failed > 0 || tr.skipped > 0)
- .map(({ tr, originalIndex }) => {
+ .filter(tr => tr.passed > 0 || tr.failed > 0 || tr.skipped > 0)
+ .map(tr => {
const time = (0, markdown_utils_1.formatTime)(tr.time);
const name = tr.path;
- const addr = options.baseUrl + makeRunSlug(originalIndex, options).link;
- const nameLink = (0, markdown_utils_1.link)(name, addr);
const passed = tr.passed > 0 ? `${tr.passed} ${markdown_utils_1.Icon.success}` : '';
const failed = tr.failed > 0 ? `${tr.failed} ${markdown_utils_1.Icon.fail}` : '';
const skipped = tr.skipped > 0 ? `${tr.skipped} ${markdown_utils_1.Icon.skip}` : '';
- return [nameLink, passed, failed, skipped, time];
+ return [name, passed, failed, skipped, time];
});
const resultsTable = (0, markdown_utils_1.table)(['Report', 'Passed', 'Failed', 'Skipped', 'Time'], [markdown_utils_1.Align.Left, markdown_utils_1.Align.Right, markdown_utils_1.Align.Right, markdown_utils_1.Align.Right, markdown_utils_1.Align.Right], ...tableData);
sections.push(resultsTable);
@@ -2090,7 +2049,7 @@ function getTestRunsReport(testRuns, options) {
const suitesReports = testRuns.map((tr, i) => getSuitesReport(tr, i, options)).flat();
sections.push(...suitesReports);
}
- if (shouldCollapse) {
+ if (totalFailed === 0) {
sections.push(` `);
}
return sections;
@@ -2152,9 +2111,6 @@ function getTestsReport(ts, runIndex, suiteIndex, options) {
}
const space = grp.name ? ' ' : '';
for (const tc of grp.tests) {
- if (options.listTests === 'failed' && tc.result !== 'failed') {
- continue;
- }
const result = getResultIcon(tc.result);
sections.push(`${space}${result} ${tc.name}`);
if (tc.error) {
@@ -2190,9 +2146,6 @@ function getResultIcon(result) {
return '';
}
}
-function encodeImgShieldsURIComponent(component) {
- return encodeURIComponent(component).replace(/-/g, '--').replace(/_/g, '__');
-}
/***/ }),
@@ -2482,11 +2435,11 @@ async function downloadArtifact(octokit, artifactId, fileName, token) {
};
const downloadStream = got_1.default.stream(req.url, { headers });
const fileWriterStream = (0, fs_1.createWriteStream)(fileName);
- downloadStream.on('redirect', (response) => {
+ downloadStream.on('redirect', response => {
core.info(`Downloading ${response.headers.location}`);
});
- downloadStream.on('downloadProgress', (progress) => {
- core.info(`Progress: ${progress.transferred} B`);
+ downloadStream.on('downloadProgress', ({ transferred }) => {
+ core.info(`Progress: ${transferred} B`);
});
await asyncStream(downloadStream, fileWriterStream);
}
@@ -27887,7 +27840,6 @@ module.exports = {
// Replace globs with equivalent patterns to reduce parsing time.
REPLACEMENTS: {
- __proto__: null,
'***': '*',
'**/**': '**',
'**/**/**': '**'
@@ -30322,11 +30274,8 @@ function runParallel (tasks, cb) {
/***/ 2560:
/***/ ((__unused_webpack_module, exports, __nccwpck_require__) => {
-;(function (sax) {
- // wrapper for non-node envs
- sax.parser = function (strict, opt) {
- return new SAXParser(strict, opt)
- }
+;(function (sax) { // wrapper for non-node envs
+ sax.parser = function (strict, opt) { return new SAXParser(strict, opt) }
sax.SAXParser = SAXParser
sax.SAXStream = SAXStream
sax.createStream = createStream
@@ -30343,18 +30292,9 @@ function runParallel (tasks, cb) {
sax.MAX_BUFFER_LENGTH = 64 * 1024
var buffers = [
- 'comment',
- 'sgmlDecl',
- 'textNode',
- 'tagName',
- 'doctype',
- 'procInstName',
- 'procInstBody',
- 'entity',
- 'attribName',
- 'attribValue',
- 'cdata',
- 'script',
+ 'comment', 'sgmlDecl', 'textNode', 'tagName', 'doctype',
+ 'procInstName', 'procInstBody', 'entity', 'attribName',
+ 'attribValue', 'cdata', 'script'
]
sax.EVENTS = [
@@ -30375,10 +30315,10 @@ function runParallel (tasks, cb) {
'ready',
'script',
'opennamespace',
- 'closenamespace',
+ 'closenamespace'
]
- function SAXParser(strict, opt) {
+ function SAXParser (strict, opt) {
if (!(this instanceof SAXParser)) {
return new SAXParser(strict, opt)
}
@@ -30397,10 +30337,7 @@ function runParallel (tasks, cb) {
parser.noscript = !!(strict || parser.opt.noscript)
parser.state = S.BEGIN
parser.strictEntities = parser.opt.strictEntities
- parser.ENTITIES =
- parser.strictEntities ?
- Object.create(sax.XML_ENTITIES)
- : Object.create(sax.ENTITIES)
+ parser.ENTITIES = parser.strictEntities ? Object.create(sax.XML_ENTITIES) : Object.create(sax.ENTITIES)
parser.attribList = []
// namespaces form a prototype chain.
@@ -30410,12 +30347,6 @@ function runParallel (tasks, cb) {
parser.ns = Object.create(rootNS)
}
- // disallow unquoted attribute values if not otherwise configured
- // and strict mode is true
- if (parser.opt.unquotedAttributeValues === undefined) {
- parser.opt.unquotedAttributeValues = !strict
- }
-
// mostly just for error reporting
parser.trackPosition = parser.opt.position !== false
if (parser.trackPosition) {
@@ -30426,7 +30357,7 @@ function runParallel (tasks, cb) {
if (!Object.create) {
Object.create = function (o) {
- function F() {}
+ function F () {}
F.prototype = o
var newf = new F()
return newf
@@ -30441,7 +30372,7 @@ function runParallel (tasks, cb) {
}
}
- function checkBufferLength(parser) {
+ function checkBufferLength (parser) {
var maxAllowed = Math.max(sax.MAX_BUFFER_LENGTH, 10)
var maxActual = 0
for (var i = 0, l = buffers.length; i < l; i++) {
@@ -30477,13 +30408,13 @@ function runParallel (tasks, cb) {
parser.bufferCheckPosition = m + parser.position
}
- function clearBuffers(parser) {
+ function clearBuffers (parser) {
for (var i = 0, l = buffers.length; i < l; i++) {
parser[buffers[i]] = ''
}
}
- function flushBuffers(parser) {
+ function flushBuffers (parser) {
closeText(parser)
if (parser.cdata !== '') {
emitNode(parser, 'oncdata', parser.cdata)
@@ -30496,20 +30427,11 @@ function runParallel (tasks, cb) {
}
SAXParser.prototype = {
- end: function () {
- end(this)
- },
+ end: function () { end(this) },
write: write,
- resume: function () {
- this.error = null
- return this
- },
- close: function () {
- return this.write(null)
- },
- flush: function () {
- flushBuffers(this)
- },
+ resume: function () { this.error = null; return this },
+ close: function () { return this.write(null) },
+ flush: function () { flushBuffers(this) }
}
var Stream
@@ -30518,17 +30440,16 @@ function runParallel (tasks, cb) {
} catch (ex) {
Stream = function () {}
}
- if (!Stream) Stream = function () {}
var streamWraps = sax.EVENTS.filter(function (ev) {
return ev !== 'error' && ev !== 'end'
})
- function createStream(strict, opt) {
+ function createStream (strict, opt) {
return new SAXStream(strict, opt)
}
- function SAXStream(strict, opt) {
+ function SAXStream (strict, opt) {
if (!(this instanceof SAXStream)) {
return new SAXStream(strict, opt)
}
@@ -30569,23 +30490,21 @@ function runParallel (tasks, cb) {
me.on(ev, h)
},
enumerable: true,
- configurable: false,
+ configurable: false
})
})
}
SAXStream.prototype = Object.create(Stream.prototype, {
constructor: {
- value: SAXStream,
- },
+ value: SAXStream
+ }
})
SAXStream.prototype.write = function (data) {
- if (
- typeof Buffer === 'function' &&
+ if (typeof Buffer === 'function' &&
typeof Buffer.isBuffer === 'function' &&
- Buffer.isBuffer(data)
- ) {
+ Buffer.isBuffer(data)) {
if (!this._decoder) {
var SD = (__nccwpck_require__(3193).StringDecoder)
this._decoder = new SD('utf8')
@@ -30610,10 +30529,7 @@ function runParallel (tasks, cb) {
var me = this
if (!me._parser['on' + ev] && streamWraps.indexOf(ev) !== -1) {
me._parser['on' + ev] = function () {
- var args =
- arguments.length === 1 ?
- [arguments[0]]
- : Array.apply(null, arguments)
+ var args = arguments.length === 1 ? [arguments[0]] : Array.apply(null, arguments)
args.splice(0, 0, ev)
me.emit.apply(me, args)
}
@@ -30636,34 +30552,30 @@ function runParallel (tasks, cb) {
// without a significant breaking change to either this parser, or the
// JavaScript language. Implementation of an emoji-capable xml parser
// is left as an exercise for the reader.
- var nameStart =
- /[:_A-Za-z\u00C0-\u00D6\u00D8-\u00F6\u00F8-\u02FF\u0370-\u037D\u037F-\u1FFF\u200C-\u200D\u2070-\u218F\u2C00-\u2FEF\u3001-\uD7FF\uF900-\uFDCF\uFDF0-\uFFFD]/
+ var nameStart = /[:_A-Za-z\u00C0-\u00D6\u00D8-\u00F6\u00F8-\u02FF\u0370-\u037D\u037F-\u1FFF\u200C-\u200D\u2070-\u218F\u2C00-\u2FEF\u3001-\uD7FF\uF900-\uFDCF\uFDF0-\uFFFD]/
- var nameBody =
- /[:_A-Za-z\u00C0-\u00D6\u00D8-\u00F6\u00F8-\u02FF\u0370-\u037D\u037F-\u1FFF\u200C-\u200D\u2070-\u218F\u2C00-\u2FEF\u3001-\uD7FF\uF900-\uFDCF\uFDF0-\uFFFD\u00B7\u0300-\u036F\u203F-\u2040.\d-]/
+ var nameBody = /[:_A-Za-z\u00C0-\u00D6\u00D8-\u00F6\u00F8-\u02FF\u0370-\u037D\u037F-\u1FFF\u200C-\u200D\u2070-\u218F\u2C00-\u2FEF\u3001-\uD7FF\uF900-\uFDCF\uFDF0-\uFFFD\u00B7\u0300-\u036F\u203F-\u2040.\d-]/
- var entityStart =
- /[#:_A-Za-z\u00C0-\u00D6\u00D8-\u00F6\u00F8-\u02FF\u0370-\u037D\u037F-\u1FFF\u200C-\u200D\u2070-\u218F\u2C00-\u2FEF\u3001-\uD7FF\uF900-\uFDCF\uFDF0-\uFFFD]/
- var entityBody =
- /[#:_A-Za-z\u00C0-\u00D6\u00D8-\u00F6\u00F8-\u02FF\u0370-\u037D\u037F-\u1FFF\u200C-\u200D\u2070-\u218F\u2C00-\u2FEF\u3001-\uD7FF\uF900-\uFDCF\uFDF0-\uFFFD\u00B7\u0300-\u036F\u203F-\u2040.\d-]/
+ var entityStart = /[#:_A-Za-z\u00C0-\u00D6\u00D8-\u00F6\u00F8-\u02FF\u0370-\u037D\u037F-\u1FFF\u200C-\u200D\u2070-\u218F\u2C00-\u2FEF\u3001-\uD7FF\uF900-\uFDCF\uFDF0-\uFFFD]/
+ var entityBody = /[#:_A-Za-z\u00C0-\u00D6\u00D8-\u00F6\u00F8-\u02FF\u0370-\u037D\u037F-\u1FFF\u200C-\u200D\u2070-\u218F\u2C00-\u2FEF\u3001-\uD7FF\uF900-\uFDCF\uFDF0-\uFFFD\u00B7\u0300-\u036F\u203F-\u2040.\d-]/
- function isWhitespace(c) {
+ function isWhitespace (c) {
return c === ' ' || c === '\n' || c === '\r' || c === '\t'
}
- function isQuote(c) {
- return c === '"' || c === "'"
+ function isQuote (c) {
+ return c === '"' || c === '\''
}
- function isAttribEnd(c) {
+ function isAttribEnd (c) {
return c === '>' || isWhitespace(c)
}
- function isMatch(regex, c) {
+ function isMatch (regex, c) {
return regex.test(c)
}
- function notMatch(regex, c) {
+ function notMatch (regex, c) {
return !isMatch(regex, c)
}
@@ -30704,271 +30616,271 @@ function runParallel (tasks, cb) {
CLOSE_TAG: S++, //
SCRIPT: S++, //