Help us improve Gearset
We love getting feedback from our users on how we can make Gearset even better. Post your ideas for improvements, new features, and bug fixes alike, and vote for others – let us know what’s important to you.
23 results found
-
Ability to export results of historic test run
When successfully validating a pull request, there is a "view details" link under the "Merge and validation checks" section. When clicked, in the new page that opens there is a button "Export test results…" that allows me to download the test results in excel file with detailed precision, reporting each test method duration.
Unfortunately this feature is only accessible when PR are being validated, and disappears when checking historic PRs already closed and deployed. Would it be possible to provide such data?
1 vote -
Attach specific Tests to Feature Branch Manually
I would like to have the ability to force a features jobs to run specific test(s) from the CI screen.
So if GS fails to detect a test class (either by error or b/c its not named in the expected convention) OR would have failed to catch breaking changes elsewhere, we can avoid/prevent this attaching a set of test classes to that feature branches deployment configuration in GS.
3 votes -
Ability to use include/exclude patterns for Apex tests
Now we can only have patterns to match test classes to include.
Would love to have ability to use include/exclude patterns (like for comparisons).
We have many tests which we don't want to execute within CI jobs. These could be from (stable) 3rd party libraries like FFLib.2 votes -
Ability to export the unit test coverage in the json format and push it to the repository automatically
The sfdx CLI has the ability to generate a report of the unit test coverage in JSON format with the default naming convention test-result-codecoverage.
I would like to be able to retrieve this JSON file and push it automatically to the repository on every unit test execution.
As an example please find a possible sfdx command to achieve it: - sfdx force:apex:test:report --testrunid 7075J00001NZBNIQA5 -c -r json -d folderDirectory #testRunID is the value from the ApexTestResult obj AsyncApexJobId field
Please find more information on the official salesforce documentation about it https://developer.salesforce.com/docs/atlas.en-us.sfdx_dev.meta/sfdx_dev/sfdx_dev_testing.htm3 votes -
Add setting to skip Salesforce validation for Aura/LWC only changes
Since Salesforce tests don't cover LWC/Aura components anyhow, if the changes are all Aura/LWC the validation should be skipped.
1 vote -
Rerun faild tests on the unit tests screen
It would be nice to be able to rerun just the failed tests with a separate button, when debugging scheduled tests
1 vote -
Run Flow tests during test monitoring jobs
With the new flow tests, it would be nice if we could also have these run within test monitoring jobs.
5 votes -
8 votes
-
Exclude specific Classes from Scheduled Unit Tests
Some Testclasses only run successfully, when the Org is connected to a Backend Application. Usually sandboxes for Development are not connected to the Backend. These are the Orgs, that we need to watch. It would be useful to exclude specific Classes on specified Environments from being tested in scheduled Unit Tests.
10 votes -
Throw Error When Invalid Test is Entered
When entering test class names in the text field, invalid entries can be added without any indication that there is something wrong, which would lead to uncovered code and most likely a failed deployment. For example, entering in ThisIsAnInvalidTestName would not result in any error.
10 votes -
it would be great if deployments could be done automatically based on a event trigger
To facilitate Automated Testing, it would be great if a deployment could be triggered by an event (such as a commit by a dev), so that an org is created to allow functional automated tests (not unit tests) to be run automatically without requiring any manual process
1 vote -
Provide the option to skip code coverage when running tests on a schedule
skipping code coverage on apex test execution, helps tests finish faster.
1 vote -
Add 'run selected tests' in nightly scheduled Unit Tests Jobs.
We can select which tests run in Validation. It would be great if we could do the same in a nightly monitoring job.
2 votes -
Ability to have Tests to Run during Validation to default to Unchecked all
Ability to have Tests to Run during Validation to default to Unchecked all vs all checked (even before they finish loading into the component). We often validate with only a few tests or predefined group of test classes that we input into 'Specify Tests to run' and if a user inputs those test classes in the text area before all the Tests populate above in the list, it will include all of the ones available. It defaults to checked all and they need to remember to uncheck them or to wait for them all to load which can take awhile…
3 votes -
Allow run tests not only by specific name but by regex
We have a lot of tests and it takes about 2 hours to run them all. But we also have some small services almost completely isolated and which have a specific test class name convention. Currently, we're using ant to specify which exactly we want to run on these services build, so we don't have to wait for 2 hours to deploy a small fix to an isolated service.
2 votes -
3 votes
-
Give Permissions to Edit Automated Job Settings
Adding a feature to give users permissions to edit & delete Automated Test Jobs would be great. Today I had to hunt down a colleague to be have him edit his job when all I had to do was change the schedule. This feature would have saved a lot of time.
2 votes -
Automated User Test Suite Integration
Such as Selenium and/or Cypress.io
7 votes -
Display which apex triggers have low code coverage if a deployment fails
It would be great if Gearset can surface which Apex trigger has 0% code coverage when a deployment fails.
9 votes -
Provide a way to get the debug logs of failing tests
Sometimes my tests are failing when they run during the automated run at night, but, when I log in and run the tests manually, they pass. I can see the failure stack trace, but I was wondering if there is any way to keep/view the debug logs from the failed tests so I can try to understand why they fail at night, but work during the day?
6 votes
- Don't see your idea?