[Pacemaker] The problem with which queue between cib and stonith-ng overflows

Yusuke Iida yusk.iida at gmail.com
Mon Jun 2 07:05:27 CEST 2014


Hi, Andrew

I use the newest of 1.1 brunches and am testing by eight sets of nodes.

Although the problem was settled once,
Now, the problem with which queue overflows between cib and stonithd
has recurred.

As an example, I paste the log of the DC node.
The problem is occurring on all nodes.

Jun  2 11:34:02 vm04 cib[3940]:    error: crm_ipcs_flush_events:
Evicting slow client 0x250afe0[3941]: event queue reached 638 entries
Jun  2 11:34:02 vm04 stonith-ng[3941]:    error: crm_ipc_read:
Connection to cib_rw failed
Jun  2 11:34:02 vm04 stonith-ng[3941]:    error:
mainloop_gio_callback: Connection to cib_rw[0x662510] closed (I/O
condition=17)
Jun  2 11:34:02 vm04 stonith-ng[3941]:   notice:
cib_connection_destroy: Connection to the CIB terminated. Shutting
down.
Jun  2 11:34:02 vm04 stonith-ng[3941]:     info: stonith_shutdown:
Terminating with  2 clients
Jun  2 11:34:02 vm04 stonith-ng[3941]:     info: qb_ipcs_us_withdraw:
withdrawing server sockets

After loading a resource setup, time for stonithd to build device
information is long.
It has taken the time for about about 15 seconds.
It seems that the diff message of cib accumulates between them.

Are there any plans to improve on this issue?

I attach a report when a problem occurs.
https://drive.google.com/file/d/0BwMFJItoO-fVUEFEN1NlelNWRjg/edit?usp=sharing

Regards,
Yusuke
-- 
----------------------------------------
METRO SYSTEMS CO., LTD

Yusuke Iida
Mail: yusk.iida at gmail.com
----------------------------------------



More information about the Pacemaker mailing list