All management ops are synchronous, and execute serially. Maybe you are thinking of test
ordering issues?
Sent from my iPhone
On Jul 9, 2012, at 9:30 PM, Ondřej Žižka <ozizka(a)redhat.com> wrote:
I'd also suggest to add an information to the docs of DMR
operations whether they are sync or async.
Often I can see tests broken due to race condition caused by async operation, like
unfinished removal of something in one test while being added in next test.
my2c
Ondra
Jason T. Greene píše v Po 09. 07. 2012 v 13:16 -0500:
>
> We always have the problem of having a set of tests which fail one out
> of 10 runs, but we leave the test around hoping one day someone will fix
> it. The problem is no one does, and it makes regression catching hard.
> Right now people that submit pull requests have to scan through test
> results and ask around to figure out if they broke something or not.
>
> So I propose a new policy. Any test which intermittently fails will be
> ignored and a JIRA opened to the author for up to a month. If that test
> is not passing in one month time, it will be removed from the codebase.
>
> The biggest problem with this policy is that we might completely lose
> coverage. A number of the clustering tests for example fail
> intermittently, and if we removed them we would have no other coverage.
> So for special cases like clustering, I am thinking of relocating them
> to a different test run called "broken-clustering", or something like
> that. This run would only be monitored by those working on clustering,
> and would not be included in the main "all tests" run.
>
> Any other ideas?
>