[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]
[bug #42390] fopen-fail fails on Debian GNU/Linux unstable
From: |
Paul D. Smith |
Subject: |
[bug #42390] fopen-fail fails on Debian GNU/Linux unstable |
Date: |
Sun, 12 Jul 2015 22:21:07 +0000 |
User-agent: |
Mozilla/5.0 (X11; Ubuntu; Linux x86_64; rv:39.0) Gecko/20100101 Firefox/39.0 |
Follow-up Comment #5, bug #42390 (project make):
There is already a way to choose a longer timeout: the run_make_test() method
takes an optional extra argument which is the timeout value.
I increased this to 300; hopefully that will solve this problem (I cannot
reproduce the original problem on my current Ubuntu GNOME system, which is
Debian-based of course: even with the original timeout I get the expected
failure).
I don't understand YunQiang's comment; the point of the test is to verify that
we get a non-zero error if we open too many file descriptors. I don't know
why it should take so long to open more file descriptors than are valid for
the process. Maybe a difference in the number of open file descriptors
allowed via ulimit -n is at play somehow here? I've also added a step to the
check-regression rule (used by "make check") that sets ulimit -n 512 just in
case. Of course this only works if you invoke the test suite through make,
not separately. I couldn't find a portable way (without requiring extra
modules) of setting ulimit from within Perl itself.
_______________________________________________________
Reply to this item at:
<http://savannah.gnu.org/bugs/?42390>
_______________________________________________
Message sent via/by Savannah
http://savannah.gnu.org/
- [bug #42390] fopen-fail fails on Debian GNU/Linux unstable,
Paul D. Smith <=