[libqb] Fix testsuite ipc test artifacts

David Vossel dvossel at fedoraproject.org
Wed Feb 26 05:48:34 UTC 2014


commit 240ba43e0eb3583691428f8e392c7514b583e15f
Author: David Vossel <dvossel at redhat.com>
Date:   Thu Feb 20 21:26:39 2014 -0600

    Fix testsuite ipc test artifacts

 libqb.spec             |    9 +-
 test-timeout-fix.patch |  590 ++++++++++++++++++++++++++++++++++++++++++++++++
 2 files changed, 597 insertions(+), 2 deletions(-)
---
diff --git a/libqb.spec b/libqb.spec
index 2631570..01c01d2 100644
--- a/libqb.spec
+++ b/libqb.spec
@@ -1,12 +1,13 @@
 Name:           libqb
 Version:        0.17.0
-Release:        1%{?dist}
+Release:        2%{?dist}
 Summary:        An IPC library for high performance servers
 
 Group:          System Environment/Libraries
 License:        LGPLv2+
 URL:            http://www.libqb.org
 Source0:        https://fedorahosted.org/releases/q/u/quarterback/%{name}-%{version}.tar.xz
+Patch0:			test-timeout-fix.patch	
 BuildRoot:      %{_tmppath}/%{name}-%{version}-%{release}-root-%(%{__id_u} -n)
 
 BuildRequires:  libtool doxygen procps check-devel automake
@@ -19,6 +20,7 @@ Initially these are IPC and poll.
 
 %prep
 %setup -q
+%patch0 -p1
 
 # Make sure the timestamps are correct
 find . -exec touch \{\} \;
@@ -28,7 +30,7 @@ find . -exec touch \{\} \;
 make %{?_smp_mflags}
 
 %check
-make check
+make -j1 check
 
 %install
 rm -rf $RPM_BUILD_ROOT
@@ -69,6 +71,9 @@ developing applications that use %{name}.
 
 %changelog
 
+* Thu Feb 20 2014 David Vossel <dvossel at redhat.com> - 0.17.0-2
+Fix testsuite timing values
+
 * Wed Feb 19 2014 David Vossel <dvossel at redhat.com> - 0.17.0-1
 Fix: build: Allow 'make rpm' to work with lightweight tags for release candidates
 Fix: spec: reference correct url in spec file
diff --git a/test-timeout-fix.patch b/test-timeout-fix.patch
new file mode 100644
index 0000000..01d6b18
--- /dev/null
+++ b/test-timeout-fix.patch
@@ -0,0 +1,590 @@
+diff --git a/lib/Makefile.am b/lib/Makefile.am
+index 91c6bd1..5bcdc22 100644
+--- a/lib/Makefile.am
++++ b/lib/Makefile.am
+@@ -42,6 +42,8 @@ source_to_lint		= util.c hdb.c ringbuffer.c ringbuffer_helper.c \
+ libqb_la_SOURCES	= $(source_to_lint) unix.c
+ libqb_la_LIBADD	        = @LTLIBOBJS@
+ 
++LDFLAGS 	= $(LDFLAGS_COPY:-Bsymbolic-functions=)
++
+ if HAVE_SEM_TIMEDWAIT
+ else
+   libqb_la_SOURCES+=rpl_sem.c
+diff --git a/tests/check_ipc.c b/tests/check_ipc.c
+index 6a80fec..e0df9e7 100644
+--- a/tests/check_ipc.c
++++ b/tests/check_ipc.c
+@@ -139,6 +139,7 @@ s1_msg_process_fn(qb_ipcs_connection_t *c,
+ 		int32_t m;
+ 		int32_t num;
+ 		struct qb_ipcs_connection_stats_2 *stats;
++		uint32_t max_size = MAX_MSG_SIZE;
+ 
+ 		response.size = sizeof(struct qb_ipc_response_header);
+ 		response.error = 0;
+@@ -148,8 +149,7 @@ s1_msg_process_fn(qb_ipcs_connection_t *c,
+ 		free(stats);
+ 
+ 		/* crazy large message */
+-		res = qb_ipcs_event_send(c, &response,
+-					 MAX_MSG_SIZE*10);
++		res = qb_ipcs_event_send(c, &response, max_size*10);
+ 		ck_assert_int_eq(res, -EMSGSIZE);
+ 
+ 		/* send one event before responding */
+@@ -157,6 +157,11 @@ s1_msg_process_fn(qb_ipcs_connection_t *c,
+ 		ck_assert_int_eq(res, sizeof(response));
+ 		response.id++;
+ 
++		/* There should be one more item in the event queue now. */
++		stats = qb_ipcs_connection_stats_get_2(c, QB_FALSE);
++		ck_assert_int_eq(stats->event_q_length - num, 1);
++		free(stats);
++
+ 		/* send response */
+ 		response.id = IPC_MSG_RES_BULK_EVENTS;
+ 		res = qb_ipcs_response_send(c, &response, response.size);
+@@ -175,9 +180,6 @@ s1_msg_process_fn(qb_ipcs_connection_t *c,
+ 			ck_assert_int_eq(res, sizeof(response));
+ 			response.id++;
+ 		}
+-		stats = qb_ipcs_connection_stats_get_2(c, QB_FALSE);
+-		ck_assert_int_eq(stats->event_q_length - num, num_bulk_events);
+-		free(stats);
+ 
+ 	} else if (req_pt->id == IPC_MSG_REQ_STRESS_EVENT) {
+ 		struct {
+@@ -313,7 +315,7 @@ s1_connection_destroyed(qb_ipcs_connection_t *c)
+ static void
+ s1_connection_created(qb_ipcs_connection_t *c)
+ {
+-	int32_t max = MAX_MSG_SIZE;
++	uint32_t max = MAX_MSG_SIZE;
+ 
+ 	if (send_event_on_created) {
+ 		struct qb_ipc_response_header response;
+@@ -362,6 +364,7 @@ run_ipc_server(void)
+ 		.dispatch_mod = my_dispatch_mod,
+ 		.dispatch_del = my_dispatch_del,
+ 	};
++	uint32_t max_size = MAX_MSG_SIZE;
+ 
+ 	qb_loop_signal_add(my_loop, QB_LOOP_HIGH, SIGSTOP,
+ 			   NULL, exit_handler, &handle);
+@@ -374,7 +377,7 @@ run_ipc_server(void)
+ 	fail_if(s1 == 0);
+ 
+ 	if (enforce_server_buffer) {
+-		qb_ipcs_enforce_buffer_size(s1, MAX_MSG_SIZE);
++		qb_ipcs_enforce_buffer_size(s1, max_size);
+ 	}
+ 	qb_ipcs_poll_handlers_set(s1, &ph);
+ 
+@@ -402,13 +405,70 @@ run_function_in_new_process(void (*run_ipc_server_fn)(void))
+ 	return pid;
+ }
+ 
+-static int32_t
+-stop_process(pid_t pid)
++static void
++request_server_exit(void)
++{
++	struct qb_ipc_request_header req_header;
++	struct qb_ipc_response_header res_header;
++	struct iovec iov[1];
++	int32_t res;
++
++	/*
++	 * tell the server to exit
++	 */
++	req_header.id = IPC_MSG_REQ_SERVER_FAIL;
++	req_header.size = sizeof(struct qb_ipc_request_header);
++
++	iov[0].iov_len = req_header.size;
++	iov[0].iov_base = &req_header;
++
++	ck_assert_int_eq(QB_TRUE, qb_ipcc_is_connected(conn));
++
++	res = qb_ipcc_sendv_recv(conn, iov, 1,
++				 &res_header,
++				 sizeof(struct qb_ipc_response_header), -1);
++	/*
++	 * confirm we get -ENOTCONN or ECONNRESET
++	 */
++	if (res != -ECONNRESET && res != -ENOTCONN) {
++		qb_log(LOG_ERR, "id:%d size:%d", res_header.id, res_header.size);
++		ck_assert_int_eq(res, -ENOTCONN);
++	}
++}
++
++static void
++kill_server(pid_t pid)
+ {
+-	/* wait a bit for the server to shutdown by it's self */
+-	usleep(100000);
+ 	kill(pid, SIGTERM);
+ 	waitpid(pid, NULL, 0);
++}
++
++static int32_t
++verify_graceful_stop(pid_t pid)
++{
++	int wait_rc = 0;
++	int status = 0;
++	int rc = 0;
++	int tries;
++
++	/* We need the server to be able to exit by itself */
++	for (tries = 10;  tries >= 0; tries--) {
++		sleep(1);
++		wait_rc = waitpid(pid, &status, WNOHANG);
++		if (wait_rc > 0) {
++			break;
++		}
++	}
++
++	ck_assert_int_eq(wait_rc, pid);
++	rc = WIFEXITED(status);
++	if (rc) {
++		rc = WEXITSTATUS(status);
++		ck_assert_int_eq(rc, 0);
++	} else {
++		fail_if(rc == 0);
++	}
++	
+ 	return 0;
+ }
+ 
+@@ -425,6 +485,7 @@ send_and_check(int32_t req_id, uint32_t size,
+ 	struct qb_ipc_response_header res_header;
+ 	int32_t res;
+ 	int32_t try_times = 0;
++	uint32_t max_size = MAX_MSG_SIZE;
+ 
+ 	request.hdr.id = req_id;
+ 	request.hdr.size = sizeof(struct qb_ipc_request_header) + size;
+@@ -432,7 +493,7 @@ send_and_check(int32_t req_id, uint32_t size,
+ 	/* check that we can't send a message that is too big
+ 	 * and we get the right return code.
+ 	 */
+-	res = qb_ipcc_send(conn, &request, MAX_MSG_SIZE*2);
++	res = qb_ipcc_send(conn, &request, max_size*2);
+ 	ck_assert_int_eq(res, -EMSGSIZE);
+ 
+ repeat_send:
+@@ -484,13 +545,14 @@ test_ipc_txrx(void)
+ 	int32_t c = 0;
+ 	size_t size;
+ 	pid_t pid;
++	uint32_t max_size = MAX_MSG_SIZE;
+ 
+ 	pid = run_function_in_new_process(run_ipc_server);
+ 	fail_if(pid == -1);
+ 	sleep(1);
+ 
+ 	do {
+-		conn = qb_ipcc_connect(ipc_name, MAX_MSG_SIZE);
++		conn = qb_ipcc_connect(ipc_name, max_size);
+ 		if (conn == NULL) {
+ 			j = waitpid(pid, NULL, WNOHANG);
+ 			ck_assert_int_eq(j, 0);
+@@ -503,7 +565,7 @@ test_ipc_txrx(void)
+ 	size = QB_MIN(sizeof(struct qb_ipc_request_header), 64);
+ 	for (j = 1; j < 19; j++) {
+ 		size *= 2;
+-		if (size >= MAX_MSG_SIZE)
++		if (size >= max_size)
+ 			break;
+ 		if (send_and_check(IPC_MSG_REQ_TX_RX, size,
+ 				   recv_timeout, QB_TRUE) < 0) {
+@@ -511,10 +573,17 @@ test_ipc_txrx(void)
+ 		}
+ 	}
+ 	if (turn_on_fc) {
++		/* can't signal server to shutdown if flow control is on */
+ 		ck_assert_int_eq(fc_enabled, QB_TRUE);
++		qb_ipcc_disconnect(conn);
++		/* TODO - figure out why this sleep is necessary */
++		sleep(1);
++		kill_server(pid);
++	} else {
++		request_server_exit();
++		qb_ipcc_disconnect(conn);
++		verify_graceful_stop(pid);
+ 	}
+-	qb_ipcc_disconnect(conn);
+-	stop_process(pid);
+ }
+ 
+ static void
+@@ -527,13 +596,14 @@ test_ipc_exit(void)
+ 	int32_t c = 0;
+ 	int32_t j = 0;
+ 	pid_t pid;
++	uint32_t max_size = MAX_MSG_SIZE;
+ 
+ 	pid = run_function_in_new_process(run_ipc_server);
+ 	fail_if(pid == -1);
+ 	sleep(1);
+ 
+ 	do {
+-		conn = qb_ipcc_connect(ipc_name, MAX_MSG_SIZE);
++		conn = qb_ipcc_connect(ipc_name, max_size);
+ 		if (conn == NULL) {
+ 			j = waitpid(pid, NULL, WNOHANG);
+ 			ck_assert_int_eq(j, 0);
+@@ -554,8 +624,8 @@ test_ipc_exit(void)
+ 				 sizeof(struct qb_ipc_response_header), -1);
+ 	ck_assert_int_eq(res, sizeof(struct qb_ipc_response_header));
+ 
+-	/* kill the server */
+-	stop_process(pid);
++	request_server_exit();
++	verify_graceful_stop(pid);
+ 
+ 	/*
+ 	 * wait a bit for the server to die.
+@@ -674,13 +744,14 @@ test_ipc_dispatch(void)
+ 	int32_t c = 0;
+ 	pid_t pid;
+ 	int32_t size;
++	uint32_t max_size = MAX_MSG_SIZE;
+ 
+ 	pid = run_function_in_new_process(run_ipc_server);
+ 	fail_if(pid == -1);
+ 	sleep(1);
+ 
+ 	do {
+-		conn = qb_ipcc_connect(ipc_name, MAX_MSG_SIZE);
++		conn = qb_ipcc_connect(ipc_name, max_size);
+ 		if (conn == NULL) {
+ 			j = waitpid(pid, NULL, WNOHANG);
+ 			ck_assert_int_eq(j, 0);
+@@ -693,7 +764,7 @@ test_ipc_dispatch(void)
+ 	size = QB_MIN(sizeof(struct qb_ipc_request_header), 64);
+ 	for (j = 1; j < 19; j++) {
+ 		size *= 2;
+-		if (size >= MAX_MSG_SIZE)
++		if (size >= max_size)
+ 			break;
+ 		if (send_and_check(IPC_MSG_REQ_DISPATCH, size,
+ 				   recv_timeout, QB_TRUE) < 0) {
+@@ -701,8 +772,9 @@ test_ipc_dispatch(void)
+ 		}
+ 	}
+ 
++	request_server_exit();
+ 	qb_ipcc_disconnect(conn);
+-	stop_process(pid);
++	verify_graceful_stop(pid);
+ }
+ 
+ START_TEST(test_ipc_disp_us)
+@@ -781,22 +853,20 @@ count_bulk_events(int32_t fd, int32_t revents, void *data)
+ static void
+ test_ipc_bulk_events(void)
+ {
+-	struct qb_ipc_request_header req_header;
+-	struct qb_ipc_response_header res_header;
+-	struct iovec iov[1];
+ 	int32_t c = 0;
+ 	int32_t j = 0;
+ 	pid_t pid;
+ 	int32_t res;
+ 	qb_loop_t *cl;
+ 	int32_t fd;
++	uint32_t max_size = MAX_MSG_SIZE;
+ 
+ 	pid = run_function_in_new_process(run_ipc_server);
+ 	fail_if(pid == -1);
+ 	sleep(1);
+ 
+ 	do {
+-		conn = qb_ipcc_connect(ipc_name, MAX_MSG_SIZE);
++		conn = qb_ipcc_connect(ipc_name, max_size);
+ 		if (conn == NULL) {
+ 			j = waitpid(pid, NULL, WNOHANG);
+ 			ck_assert_int_eq(j, 0);
+@@ -823,21 +893,9 @@ test_ipc_bulk_events(void)
+ 	qb_loop_run(cl);
+ 	ck_assert_int_eq(events_received, num_bulk_events);
+ 
+-	req_header.id = IPC_MSG_REQ_SERVER_FAIL;
+-	req_header.size = sizeof(struct qb_ipc_request_header);
+-
+-	iov[0].iov_len = req_header.size;
+-	iov[0].iov_base = &req_header;
+-	res = qb_ipcc_sendv_recv(conn, iov, 1,
+-				 &res_header,
+-				 sizeof(struct qb_ipc_response_header), -1);
+-	if (res != -ECONNRESET && res != -ENOTCONN) {
+-		qb_log(LOG_ERR, "id:%d size:%d", res_header.id, res_header.size);
+-		ck_assert_int_eq(res, -ENOTCONN);
+-	}
+-
++	request_server_exit();
+ 	qb_ipcc_disconnect(conn);
+-	stop_process(pid);
++	verify_graceful_stop(pid);
+ }
+ 
+ static void
+@@ -857,13 +915,14 @@ test_ipc_stress_test(void)
+ 	int32_t res;
+ 	qb_loop_t *cl;
+ 	int32_t fd;
++	uint32_t max_size = MAX_MSG_SIZE;
+ 	/* This looks strange, but it serves an important purpose.
+ 	 * This test forces the server to enforce the MAX_MSG_SIZE
+ 	 * limit from the server side, which overrides the client's
+ 	 * buffer limit.  To verify this functionality is working
+ 	 * we set the client limit lower than what the server
+ 	 * is enforcing. */
+-	int32_t client_buf_size = MAX_MSG_SIZE - 1024;
++	int32_t client_buf_size = max_size - 1024;
+ 	int32_t real_buf_size;
+ 
+ 	enforce_server_buffer = 1;
+@@ -884,7 +943,7 @@ test_ipc_stress_test(void)
+ 	fail_if(conn == NULL);
+ 
+ 	real_buf_size = qb_ipcc_get_buffer_size(conn);
+-	ck_assert_int_eq(real_buf_size, MAX_MSG_SIZE);
++	ck_assert_int_eq(real_buf_size, max_size);
+ 
+ 	qb_log(LOG_DEBUG, "Testing %d iterations of EVENT msg passing.", num_stress_events);
+ 
+@@ -920,7 +979,7 @@ test_ipc_stress_test(void)
+ 	}
+ 
+ 	qb_ipcc_disconnect(conn);
+-	stop_process(pid);
++	verify_graceful_stop(pid);
+ }
+ 
+ START_TEST(test_ipc_stress_test_us)
+@@ -954,6 +1013,7 @@ test_ipc_event_on_created(void)
+ 	int32_t res;
+ 	qb_loop_t *cl;
+ 	int32_t fd;
++	uint32_t max_size = MAX_MSG_SIZE;
+ 
+ 	num_bulk_events = 1;
+ 
+@@ -962,7 +1022,7 @@ test_ipc_event_on_created(void)
+ 	sleep(1);
+ 
+ 	do {
+-		conn = qb_ipcc_connect(ipc_name, MAX_MSG_SIZE);
++		conn = qb_ipcc_connect(ipc_name, max_size);
+ 		if (conn == NULL) {
+ 			j = waitpid(pid, NULL, WNOHANG);
+ 			ck_assert_int_eq(j, 0);
+@@ -984,8 +1044,9 @@ test_ipc_event_on_created(void)
+ 	qb_loop_run(cl);
+ 	ck_assert_int_eq(events_received, num_bulk_events);
+ 
++	request_server_exit();
+ 	qb_ipcc_disconnect(conn);
+-	stop_process(pid);
++	verify_graceful_stop(pid);
+ }
+ 
+ START_TEST(test_ipc_event_on_created_us)
+@@ -1009,13 +1070,14 @@ test_ipc_disconnect_after_created(void)
+ 	int32_t j = 0;
+ 	pid_t pid;
+ 	int32_t res;
++	uint32_t max_size = MAX_MSG_SIZE;
+ 
+ 	pid = run_function_in_new_process(run_ipc_server);
+ 	fail_if(pid == -1);
+ 	sleep(1);
+ 
+ 	do {
+-		conn = qb_ipcc_connect(ipc_name, MAX_MSG_SIZE);
++		conn = qb_ipcc_connect(ipc_name, max_size);
+ 		if (conn == NULL) {
+ 			j = waitpid(pid, NULL, WNOHANG);
+ 			ck_assert_int_eq(j, 0);
+@@ -1046,7 +1108,7 @@ test_ipc_disconnect_after_created(void)
+ 	ck_assert_int_eq(QB_FALSE, qb_ipcc_is_connected(conn));
+ 
+ 	qb_ipcc_disconnect(conn);
+-	stop_process(pid);
++	kill_server(pid);
+ }
+ 
+ START_TEST(test_ipc_disconnect_after_created_us)
+@@ -1063,20 +1125,17 @@ END_TEST
+ static void
+ test_ipc_server_fail(void)
+ {
+-	struct qb_ipc_request_header req_header;
+-	struct qb_ipc_response_header res_header;
+-	struct iovec iov[1];
+-	int32_t res;
+ 	int32_t j;
+ 	int32_t c = 0;
+ 	pid_t pid;
++	uint32_t max_size = MAX_MSG_SIZE;
+ 
+ 	pid = run_function_in_new_process(run_ipc_server);
+ 	fail_if(pid == -1);
+ 	sleep(1);
+ 
+ 	do {
+-		conn = qb_ipcc_connect(ipc_name, MAX_MSG_SIZE);
++		conn = qb_ipcc_connect(ipc_name, max_size);
+ 		if (conn == NULL) {
+ 			j = waitpid(pid, NULL, WNOHANG);
+ 			ck_assert_int_eq(j, 0);
+@@ -1086,31 +1145,10 @@ test_ipc_server_fail(void)
+ 	} while (conn == NULL && c < 5);
+ 	fail_if(conn == NULL);
+ 
+-	/*
+-	 * tell the server to exit
+-	 */
+-	req_header.id = IPC_MSG_REQ_SERVER_FAIL;
+-	req_header.size = sizeof(struct qb_ipc_request_header);
+-
+-	iov[0].iov_len = req_header.size;
+-	iov[0].iov_base = &req_header;
+-
+-	ck_assert_int_eq(QB_TRUE, qb_ipcc_is_connected(conn));
+-
+-	res = qb_ipcc_sendv_recv(conn, iov, 1,
+-				 &res_header,
+-				 sizeof(struct qb_ipc_response_header), -1);
+-	/*
+-	 * confirm we get -ENOTCONN or ECONNRESET
+-	 */
+-	if (res != -ECONNRESET && res != -ENOTCONN) {
+-		qb_log(LOG_ERR, "id:%d size:%d", res_header.id, res_header.size);
+-		ck_assert_int_eq(res, -ENOTCONN);
+-	}
++	request_server_exit();
+ 	ck_assert_int_eq(QB_FALSE, qb_ipcc_is_connected(conn));
+-
+ 	qb_ipcc_disconnect(conn);
+-	stop_process(pid);
++	verify_graceful_stop(pid);
+ }
+ 
+ START_TEST(test_ipc_server_fail_soc)
+@@ -1181,6 +1219,7 @@ test_ipc_service_ref_count(void)
+ 	int32_t c = 0;
+ 	int32_t j = 0;
+ 	pid_t pid;
++	uint32_t max_size = MAX_MSG_SIZE;
+ 
+ 	reference_count_test = QB_TRUE;
+ 
+@@ -1189,7 +1228,7 @@ test_ipc_service_ref_count(void)
+ 	sleep(1);
+ 
+ 	do {
+-		conn = qb_ipcc_connect(ipc_name, MAX_MSG_SIZE);
++		conn = qb_ipcc_connect(ipc_name, max_size);
+ 		if (conn == NULL) {
+ 			j = waitpid(pid, NULL, WNOHANG);
+ 			ck_assert_int_eq(j, 0);
+@@ -1201,8 +1240,7 @@ test_ipc_service_ref_count(void)
+ 
+ 	sleep(5);
+ 
+-	qb_ipcc_disconnect(conn);
+-	stop_process(pid);
++	kill_server(pid);
+ }
+ 
+ 
+@@ -1265,22 +1303,22 @@ make_shm_suite(void)
+ 
+ 	tc = tcase_create("ipc_server_fail_shm");
+ 	tcase_add_test(tc, test_ipc_server_fail_shm);
+-	tcase_set_timeout(tc, 6);
++	tcase_set_timeout(tc, 8);
+ 	suite_add_tcase(s, tc);
+ 
+ 	tc = tcase_create("ipc_txrx_shm_block");
+ 	tcase_add_test(tc, test_ipc_txrx_shm_block);
+-	tcase_set_timeout(tc, 6);
++	tcase_set_timeout(tc, 8);
+ 	suite_add_tcase(s, tc);
+ 
+ 	tc = tcase_create("ipc_txrx_shm_tmo");
+ 	tcase_add_test(tc, test_ipc_txrx_shm_tmo);
+-	tcase_set_timeout(tc, 6);
++	tcase_set_timeout(tc, 8);
+ 	suite_add_tcase(s, tc);
+ 
+ 	tc = tcase_create("ipc_fc_shm");
+ 	tcase_add_test(tc, test_ipc_fc_shm);
+-	tcase_set_timeout(tc, 6);
++	tcase_set_timeout(tc, 8);
+ 	suite_add_tcase(s, tc);
+ 
+ 	tc = tcase_create("ipc_dispatch_shm");
+@@ -1300,11 +1338,12 @@ make_shm_suite(void)
+ 
+ 	tc = tcase_create("ipc_exit_shm");
+ 	tcase_add_test(tc, test_ipc_exit_shm);
+-	tcase_set_timeout(tc, 3);
++	tcase_set_timeout(tc, 8);
+ 	suite_add_tcase(s, tc);
+ 
+ 	tc = tcase_create("ipc_event_on_created_shm");
+ 	tcase_add_test(tc, test_ipc_event_on_created_shm);
++	tcase_set_timeout(tc, 10);
+ 	suite_add_tcase(s, tc);
+ 
+ 	tc = tcase_create("ipc_service_ref_count_shm");
+@@ -1328,27 +1367,27 @@ make_soc_suite(void)
+ 
+ 	tc = tcase_create("ipc_server_fail_soc");
+ 	tcase_add_test(tc, test_ipc_server_fail_soc);
+-	tcase_set_timeout(tc, 6);
++	tcase_set_timeout(tc, 8);
+ 	suite_add_tcase(s, tc);
+ 
+ 	tc = tcase_create("ipc_txrx_us_block");
+ 	tcase_add_test(tc, test_ipc_txrx_us_block);
+-	tcase_set_timeout(tc, 6);
++	tcase_set_timeout(tc, 8);
+ 	suite_add_tcase(s, tc);
+ 
+ 	tc = tcase_create("ipc_txrx_us_tmo");
+ 	tcase_add_test(tc, test_ipc_txrx_us_tmo);
+-	tcase_set_timeout(tc, 6);
++	tcase_set_timeout(tc, 8);
+ 	suite_add_tcase(s, tc);
+ 
+ 	tc = tcase_create("ipc_fc_us");
+ 	tcase_add_test(tc, test_ipc_fc_us);
+-	tcase_set_timeout(tc, 6);
++	tcase_set_timeout(tc, 8);
+ 	suite_add_tcase(s, tc);
+ 
+ 	tc = tcase_create("ipc_exit_us");
+ 	tcase_add_test(tc, test_ipc_exit_us);
+-	tcase_set_timeout(tc, 6);
++	tcase_set_timeout(tc, 8);
+ 	suite_add_tcase(s, tc);
+ 
+ 	tc = tcase_create("ipc_dispatch_us");
+@@ -1368,10 +1407,12 @@ make_soc_suite(void)
+ 
+ 	tc = tcase_create("ipc_event_on_created_us");
+ 	tcase_add_test(tc, test_ipc_event_on_created_us);
++	tcase_set_timeout(tc, 10);
+ 	suite_add_tcase(s, tc);
+ 
+ 	tc = tcase_create("ipc_disconnect_after_created_us");
+ 	tcase_add_test(tc, test_ipc_disconnect_after_created_us);
++	tcase_set_timeout(tc, 10);
+ 	suite_add_tcase(s, tc);
+ 
+ 	tc = tcase_create("ipc_service_ref_count_us");


More information about the scm-commits mailing list