[lttng-dev] Pull request: lttng-ust tests clean-up
Christian Babeux
christian.babeux at efficios.com
Tue Mar 26 11:08:51 EDT 2013
Hi Jérémie,
> I'm wondering what would be the best way to get tests
> to run conditionally depending on the "configure" options now that we
> are using Christian's "prove + test lists" scheme.
What I have in mind is something along this way:
In the lttng-tools/tests/Makefile.am:
if USE_PYTHON
check-am:
./run.sh unit_tests
./run.sh fast_regression
./run.sh with_bindings_regression
else
check-am:
./run.sh unit_tests
./run.sh fast_regression
endif
The testlist with_bindings_regression would contain the appropriate
tests that have a dependency on the lttng-tools Python bindings.
> While this would certainly be good
> enough for now, manually maintaining test lists based on configuration
> dependancies may grow tedious as we add configuration options and
> tests.
Agreed. This would only apply of course if we add those tests to the
default "make check" target.
> Perhaps we could dynamically generate test lists depending on the
> configuration options and each tests' requirements... but that
> certainly sounds like overkill right now.
Let's keep it simple for now :).
Christian
On Tue, Mar 26, 2013 at 10:48 AM, Jérémie Galarneau
<jeremie.galarneau at efficios.com> wrote:
> On Tue, Mar 26, 2013 at 9:48 AM, Mathieu Desnoyers
> <mathieu.desnoyers at efficios.com> wrote:
>> * Jérémie Galarneau (jeremie.galarneau at efficios.com) wrote:
>>> Hi all,
>>>
>>> I'd like to propose a set of patches to clean-up the lttng-ust tests.
>>> Since the clean-up now weighs in at a fairly hefty 21 patches, I think
>>> a link to my personal repository might be more appropriate than
>>> posting the patch-set on this list.
>>>
>>> The relevant commits are 8ae5122...0060eb2
>>>
>>> I will also be posting a set of patches that fixes and moves some of
>>> the lttng-ust tests to lttng-tools since they depend on it.
>>>
>>> Git access:
>>> git://github.com/jgalar/lttng-ust-tests-cleanup.git -b test-cleanup
>>>
>>> Web interface:
>>> https://github.com/jgalar/lttng-ust-tests-cleanup
>>
>> I tried running the various batch files under tests/ in your tree, and
>> they pretty much all complain about:
>>
>> compudj at thinkos:~/git/jgalar/lttng-ust-tests-cleanup/tests$
>> (git:test-cleanup)> ./runtests
>> ./runtests: line 31: .//snprintf/run: No such file or directory
>>
>> Is that expected ?
>>
>
> I meant to remove that file since, with Christian's recent lttng-tools
> patches, we moved away from using runner scripts to using prove + test
> lists. I'll correct the last commit (Tests: Use Perl prove as the
> testsuite runner).
>
>> Also, make check only runs a single test:
>>
>> snprintf/test_snprintf .. ok
>> All tests successful.
>> Files=1, Tests=1, 0 wallclock secs ( 0.01 usr + 0.00 sys = 0.01 CPU)
>> Result: PASS
>>
>> Is that expected too ?
>
> Yes, most of the tests were moved to lttng-tools since they depend on
> lttng-sessiond. Hopefully, now that the infrastructure is in place,
> It'll be easier to add self-contained UST tests.
>
> I'll take the opportunity to ask if you, or anyone else really, have
> suggestions for UST unit tests that don't depend on lttng-tools?
>
> I can think of a few, such as validating the header files outputted by
> lttng-gen-tp, testing the filter bytecode interpretation mechanisms,
> etc. But I wonder if unit testing the control interface, for instance,
> is realistic since it tends to change fairly often. Of course, lttng
> contributors are welcome to submit new tests!
>
> As for my lttng-tools patches, some of the new tests depend on the
> lttng-tools python bindings (not included in the default
> configuration). I'm wondering what woud be the best way to get tests
> to run conditionally depending on the "configure" options now that we
> are using Christian's "prove + test lists" scheme.
>
> I have discussed the issue with him privately and we both agreed that
> we could have separate test lists that would be used depending on the
> current project configuration. While this would certainly be good
> enough for now, manually maintaining test lists based on configuration
> dependancies may grow tedious as we add configuration options and
> tests.
>
> Perhaps we could dynamically generate test lists depending on the
> configuration options and each tests' requirements... but that
> certainly sounds like overkill right now. Thoughts?
>
> Jérémie
>
>>
>> Thanks,
>>
>> Mathieu
>>
>>>
>>> Comments are welcome, as always!
>>>
>>> Regards,
>>> Jérémie
>>>
>>> --
>>> Jérémie Galarneau
>>> EfficiOS Inc.
>>> http://www.efficios.com
>>>
>>> _______________________________________________
>>> lttng-dev mailing list
>>> lttng-dev at lists.lttng.org
>>> http://lists.lttng.org/cgi-bin/mailman/listinfo/lttng-dev
>>
>> --
>> Mathieu Desnoyers
>> EfficiOS Inc.
>> http://www.efficios.com
>
>
>
> --
> Jérémie Galarneau
> EfficiOS Inc.
> http://www.efficios.com
More information about the lttng-dev
mailing list