glib2.0/glib/tests/timeout.c

221 lines
6.0 KiB
C

#include <glib.h>
#ifdef G_OS_UNIX
#include <unistd.h>
#endif
static GMainLoop *loop;
static void
stop_waiting (gpointer data)
{
g_main_loop_quit (loop);
}
static gboolean
unreachable_callback (gpointer data)
{
g_assert_not_reached ();
return G_SOURCE_REMOVE;
}
static void
unreachable_void_callback (gpointer data)
{
g_assert_not_reached ();
}
static void
test_seconds (void)
{
guint id;
/* Bug 642052 mentions that g_timeout_add_seconds(21475) schedules a
* job that runs once per second.
*
* Test that that isn't true anymore by scheduling two jobs:
* - one, as above
* - another that runs in 2100ms
*
* If everything is working properly, the 2100ms one should run first
* (and exit the mainloop). If we ever see the 21475 second job run
* then we have trouble (since it ran in less than 2 seconds).
*
* We need a timeout of at least 2 seconds because
* g_timeout_add_seconds() can add as much as an additional second of
* latency.
*/
g_test_bug ("https://bugzilla.gnome.org/show_bug.cgi?id=642052");
loop = g_main_loop_new (NULL, FALSE);
g_timeout_add_once (2100, stop_waiting, NULL);
id = g_timeout_add_seconds (21475, unreachable_callback, NULL);
g_main_loop_run (loop);
g_main_loop_unref (loop);
g_source_remove (id);
}
static void
test_seconds_once (void)
{
/* Use the same principle as in test_seconds() */
loop = g_main_loop_new (NULL, FALSE);
g_timeout_add_once (2100, stop_waiting, NULL);
g_timeout_add_seconds_once (21475, unreachable_void_callback, NULL);
g_main_loop_run (loop);
g_main_loop_unref (loop);
}
static void
test_weeks_overflow (void)
{
guint id;
guint interval_seconds;
/* Internally, the guint interval (in seconds) was converted to milliseconds
* then stored in a guint variable. This meant that any interval larger than
* G_MAXUINT / 1000 would overflow.
*
* On a system with 32-bit guint, the interval (G_MAXUINT / 1000) + 1 seconds
* (49.7 days) would end wrapping to 704 milliseconds.
*
* Test that that isn't true anymore by scheduling two jobs:
* - one, as above
* - another that runs in 2100ms
*
* If everything is working properly, the 2100ms one should run first
* (and exit the mainloop). If we ever see the other job run
* then we have trouble (since it ran in less than 2 seconds).
*
* We need a timeout of at least 2 seconds because
* g_timeout_add_seconds() can add as much as an additional second of
* latency.
*/
g_test_bug ("https://gitlab.gnome.org/GNOME/glib/issues/1600");
loop = g_main_loop_new (NULL, FALSE);
g_timeout_add_once (2100, stop_waiting, NULL);
interval_seconds = 1 + G_MAXUINT / 1000;
id = g_timeout_add_seconds (interval_seconds, unreachable_callback, NULL);
g_main_loop_run (loop);
g_main_loop_unref (loop);
g_source_remove (id);
}
/* The ready_time for a GSource is stored as a gint64, as an absolute monotonic
* time in microseconds. To call poll(), this must be converted to a relative
* timeout, in milliseconds, as a gint. If the ready_time is sufficiently far
* in the future, the timeout will not fit. Previously, it would be narrowed in
* an implementation-defined way; if this gave a negative result, poll() would
* block forever.
*
* This test creates a GSource with the largest possible ready_time (a little
* over 292 millennia, assuming g_get_monotonic_time() starts from near 0 when
* the system boots), adds it to a GMainContext, queries it for the parameters
* to pass to poll() -- essentially the first half of
* g_main_context_iteration() -- and checks that the timeout is a large
* positive number.
*/
static void
test_far_future_ready_time (void)
{
GSourceFuncs source_funcs = { 0 };
GMainContext *context = g_main_context_new ();
GSource *source = g_source_new (&source_funcs, sizeof (GSource));
gboolean acquired, ready;
gint priority, timeout_, n_fds;
g_source_set_ready_time (source, G_MAXINT64);
g_source_attach (source, context);
acquired = g_main_context_acquire (context);
g_assert_true (acquired);
ready = g_main_context_prepare (context, &priority);
g_assert_false (ready);
n_fds = 0;
n_fds = g_main_context_query (context, priority, &timeout_, NULL, n_fds);
g_assert_cmpint (n_fds, >=, 0);
/* The true timeout in milliseconds doesn't fit into a gint. We definitely
* don't want poll() to block forever:
*/
g_assert_cmpint (timeout_, >=, 0);
/* Instead, we want it to block for as long as possible: */
g_assert_cmpint (timeout_, ==, G_MAXINT);
g_main_context_release (context);
g_main_context_unref (context);
g_source_unref (source);
}
static gint64 last_time;
static gint count;
static gboolean
test_func (gpointer data)
{
gint64 current_time;
current_time = g_get_monotonic_time ();
/* We accept 2 on the first iteration because _add_seconds() can
* have an initial latency of 1 second, see its documentation.
*
* Allow up to 500ms leeway for rounding and scheduling.
*/
if (count == 0)
g_assert_cmpint (current_time / 1000 - last_time / 1000, <=, 2500);
else
g_assert_cmpint (current_time / 1000 - last_time / 1000, <=, 1500);
last_time = current_time;
count++;
/* Make the timeout take up to 0.1 seconds.
* We should still get scheduled for the next second.
*/
g_usleep (count * 10000);
if (count < 10)
return TRUE;
g_main_loop_quit (loop);
return FALSE;
}
static void
test_rounding (void)
{
loop = g_main_loop_new (NULL, FALSE);
last_time = g_get_monotonic_time ();
g_timeout_add_seconds (1, test_func, NULL);
g_main_loop_run (loop);
g_main_loop_unref (loop);
}
int
main (int argc, char *argv[])
{
g_test_init (&argc, &argv, NULL);
g_test_add_func ("/timeout/seconds", test_seconds);
g_test_add_func ("/timeout/seconds-once", test_seconds_once);
g_test_add_func ("/timeout/weeks-overflow", test_weeks_overflow);
g_test_add_func ("/timeout/far-future-ready-time", test_far_future_ready_time);
g_test_add_func ("/timeout/rounding", test_rounding);
return g_test_run ();
}