any code coverage tools(statement, branch, MCDC) for QT C++ codes? - qt

I need to find the code coverage metrics values for my QT based GUI code. Please suggest any tool which would allow to create a test case and generate the coverage values.
Thanks,
Nayan

GCOV can be used to get statement and branch coverage
ref :Qt and gcov, coverage files are not generated
other tool is TUG
ref:
http://pedromateo.github.io/tug_qt_unit_testing_fw/

Related

Exclude some design unit from code coverage on Questasim

I run a code coverage on questasim and I got ucdb file as output. But I need to exclude code coverages of some modules that connect to the top module.I don't need some of modules to be covered which this I can improve coverage report.
How can I do it without run simulation again?
Thanks.
I used a command as shown below.
coverage exclude -du <design_unit>
or
coverage exclude -srcfile <design_file>
Reference: QuestaSim User Manual

What is instrumentation in nyc istanbul?

What is instumentation used in nyc?
nyc's instrument command can be used to instrument source files outside of the context of your unit-tests:
I assume it will do coverage outside unit-testing. I tried it with
nyc instrument src coverage/instrument
then run the application and tries hitting an endpoint
npm start
but when I do above, it doesn't generate a file in nyc_output thus can't report anything.
Do I have to finish the nyc instrument command? how to do so?
nyc instrument is used to instrument your code. It produces output that, when ran, will gather coverage data. This is useless unless you actually do something with that data... like report it or use it somehow. When you run a file that has been instrumented, it will store coverage data in global.__coverage__ I believe. You can then do what you want with that data. So, you could create a reporter that will run the instrumented file, then take a look at global.__coverage__ to see what the coverage is like. Simply running an instrumented file won't generate any output
To see what the coverage is of a file that has been instrumented, you can either create your own reporter where you require the instrumented file, then take a look at global.__coverage__ or you could run the nyc command to generate coverage data like normal.
Here are a few examples:
Let's say you have a file file.js that you want to check the coverage of and you've run the command:
nyc instrument file.js > file_instrumented.js
Now, you'll have a file named file_instrumented.js that has all the code necessary to generate code coverage.
If I run that file with node file_instumented.js nothing happens... other than the file executes same as file.js
But, if I create a file named coverage.js with this code:
require("./file_instrumented.js");
console.log(global.__coverage__)
Then, I run node coverage.js you'll be able to see coverage data. You can then output whatever data you want. It's sort of a lower level access to the coverage data
If you want to generate a report in nyc_output you'll need to use the nyc command against the instrumented file. For example, something like this:
nyc --reporter=text --report-dir=./nyc_output node file_instrumented.js
A command like this would work too if you made the file_instrumented.js file executable:
nyc --reporter=text --report-dir=./nyc_output file_instrumented.js
However, if we try to run that same command against the original file.js like this:
nyc --reporter=text --report-dir=./nyc_output node file.js
You'll see that we get a report that shows no coverage. And this is because the file.js file isn't instrumented and therefore doesn't give the nyc reporter any data to report
You are correct that using nyc instrument will do coverage outside unit-testing frameworks as I've demonstrated above. It's a bit confusing because the docs aren't as clear as they should be. There are no good examples that I could find on how to get coverage of files outside of test frameworks so I figured this all out by looking at the source code for nyc as well as some of the testing frameworks.
The thing is that the testing frameworks instrument the file for you so when you run a command like this using the Mocha testing framework for example:
nyc --reporter=text mocha --ui bdd test.js
What's happening is:
nyc is executing mocha...
then mocha is instrumenting your code for you behind the scenes (correction - as per #Dining Philosopher - it looks like it’s not actuallymocha that is instrumenting your code. It’s actually nyc that instruments the code because it knows about mocha. Note: I haven’t verified this in code)
then mocha is running that instrumented code
which runs tests while collecting coverage data
which gives nyc the global.__coverage__ that it needs to generate a report
finally, nyc uses that data to output a report in your nyc_output folder
Hope this all makes sense...
In the nyc v15.1.0, we can create the report without instrumented file, just run
nyc --reporter=text --report-dir=./nyc_output node file.js
is worked !!
But i think we if we want to get the running time report, we still need the instrumented file to run
Instrumentation in this context means adding some code in between the original code. You can see this with a simple nyc instrument example.js. The output looks but confusing but is just valid javascript that still does the same things as the original program.
The important thing to know is that the added code has some side-effects. In the case of nyc the added code modifies the global.__coverage__ object (thanks Ray Perea).
Confusingly, nyc only instruments code that is tested (it knows about some testing frameworks such as mocha). To override this behavior, pass the --all flag to nyc.
A simple example running instrumented javascript:
nyc --all node example.js
Where example.js contains some uninstrumented javascript (e.g. console.log("Hello world"). This will instrument all javascript files (nested) in the current directory and report their coverage when running node example.js.
P.S. I would have edited Ray Perea's answer but their edit queue is full.

Where are the source-codes contained in the IDL-directory?

can someone tell me, where i can find the source-codes contained in the IDL-Directory.
In IDL one can run "demo". Under "Math and Statistics" -> "Math and Statistics Demo", there is as an example a polynomial fit. Where is the source-code for this polynomial fit contained within the idl directory?
best regards
So your answer has several parts. You can find the "demo" source code in your IDL installation, inside examples/demo/demosrc. If you want the source code to the actual polynomial fit routines, those will either be in source code inside the "lib" directory, or they might be written in C code for speed. The documentation will tell you if it is written in pro code.
For pro code routines, once you've compiled the routine, you can always use the "routine_filepath" command to find the source code location. Hope this helps!

R: How to enable JIT compiler via script, not environment?

I work on several R projects at the same time. One of them involves a simulation with a for-loop, which I hope to speed up by using a JIT-compiler. To do so, I added to the file Rcmd_environ in my R-directory/etc the lines following this recommendation.
R_COMPILE_PKGS=TRUE
R_ENABLE_JIT=3
Now I wonder, whether it is possible, to turn this on and off via a script. This way, I wouldn't have JIT-compilation in my other projects. Any ideas?
You can load the compiler library and then set the JIT level via calling the 'enableJIT` function.
For e.g. you can do
require(compiler)
enableJIT(3)
to get full JIT compilation.

Graphical representation of test results of phpunit

I am using phpunit to do functional tests. I use the log-junit option to generate results in JUnit-XML format. I then use phing to read this XML and generate a HTML report. The report is fine and neat. However, I have two questions:--
Can I also show the results in graphical format in the same JUnit HTML report file(generated by Phing)? A pie chart or any other chart for that matter? (In terms of passed to failed tests)
The JUnit summary which is generated by using the option -log-junit when running PHPUnit tests shows the test times in seconds. It is not easily readable when the number is big. Can I convert this to minutes by setting some option in command line? Is there any other way to do this?
I am trying to do this without the use of jenkins.
Please share if you know something about this.
Thanks.
What you can do is to use PHPUnit and generate HTML report for test coverage. It's similar to junit log:
phpunit --log-junit results/phpunit/junit.xml --coverage-html=results/phpunit/covegare -c tests/phpunit.xml
Then you can use Jenkins - Post-build Actions
Publish Junit test result report
Publish HTML reports
This way you will have tests results but also very useful code coverage report.

Resources