2007-04-08 14:39:58 +00:00
|
|
|
/*
|
|
|
|
* FD polling functions for generic poll()
|
|
|
|
*
|
MAJOR: polling: rework the whole polling system
This commit heavily changes the polling system in order to definitely
fix the frequent breakage of SSL which needs to remember the last
EAGAIN before deciding whether to poll or not. Now we have a state per
direction for each FD, as opposed to a previous and current state
previously. An FD can have up to 8 different states for each direction,
each of which being the result of a 3-bit combination. These 3 bits
indicate a wish to access the FD, the readiness of the FD and the
subscription of the FD to the polling system.
This means that it will now be possible to remember the state of a
file descriptor across disable/enable sequences that generally happen
during forwarding, where enabling reading on a previously disabled FD
would result in forgetting the EAGAIN flag it met last time.
Several new state manipulation functions have been introduced or
adapted :
- fd_want_{recv,send} : enable receiving/sending on the FD regardless
of its state (sets the ACTIVE flag) ;
- fd_stop_{recv,send} : stop receiving/sending on the FD regardless
of its state (clears the ACTIVE flag) ;
- fd_cant_{recv,send} : report a failure to receive/send on the FD
corresponding to EAGAIN (clears the READY flag) ;
- fd_may_{recv,send} : report the ability to receive/send on the FD
as reported by poll() (sets the READY flag) ;
Some functions are used to report the current FD status :
- fd_{recv,send}_active
- fd_{recv,send}_ready
- fd_{recv,send}_polled
Some functions were removed :
- fd_ev_clr(), fd_ev_set(), fd_ev_rem(), fd_ev_wai()
The POLLHUP/POLLERR flags are now reported as ready so that the I/O layers
knows it can try to access the file descriptor to get this information.
In order to simplify the conditions to add/remove cache entries, a new
function fd_alloc_or_release_cache_entry() was created to be used from
pollers while scanning for updates.
The following pollers have been updated :
ev_select() : done, built, tested on Linux 3.10
ev_poll() : done, built, tested on Linux 3.10
ev_epoll() : done, built, tested on Linux 3.10 & 3.13
ev_kqueue() : done, built, tested on OpenBSD 5.2
2014-01-10 15:58:45 +00:00
|
|
|
* Copyright 2000-2014 Willy Tarreau <w@1wt.eu>
|
2007-04-08 14:39:58 +00:00
|
|
|
*
|
|
|
|
* This program is free software; you can redistribute it and/or
|
|
|
|
* modify it under the terms of the GNU General Public License
|
|
|
|
* as published by the Free Software Foundation; either version
|
|
|
|
* 2 of the License, or (at your option) any later version.
|
|
|
|
*
|
|
|
|
*/
|
|
|
|
|
2017-03-13 16:14:51 +00:00
|
|
|
#define _GNU_SOURCE // for POLLRDHUP on Linux
|
|
|
|
|
2007-04-08 14:39:58 +00:00
|
|
|
#include <unistd.h>
|
2017-03-13 16:14:51 +00:00
|
|
|
#include <poll.h>
|
2007-04-08 14:39:58 +00:00
|
|
|
#include <sys/time.h>
|
|
|
|
#include <sys/types.h>
|
|
|
|
|
2020-06-09 07:07:15 +00:00
|
|
|
#include <haproxy/activity.h>
|
2020-05-27 10:58:42 +00:00
|
|
|
#include <haproxy/api.h>
|
2021-10-08 07:33:24 +00:00
|
|
|
#include <haproxy/clock.h>
|
2020-06-09 07:07:15 +00:00
|
|
|
#include <haproxy/fd.h>
|
|
|
|
#include <haproxy/global.h>
|
2022-09-08 15:46:31 +00:00
|
|
|
#include <haproxy/signal.h>
|
2021-09-30 15:53:22 +00:00
|
|
|
#include <haproxy/task.h>
|
2020-06-02 16:15:32 +00:00
|
|
|
#include <haproxy/ticks.h>
|
2007-04-08 14:39:58 +00:00
|
|
|
|
|
|
|
|
2017-03-13 16:14:51 +00:00
|
|
|
#ifndef POLLRDHUP
|
|
|
|
/* POLLRDHUP was defined late in libc, and it appeared in kernel 2.6.17 */
|
|
|
|
#define POLLRDHUP 0
|
|
|
|
#endif
|
|
|
|
|
2018-01-26 20:48:23 +00:00
|
|
|
static int maxfd; /* # of the highest fd + 1 */
|
2013-03-31 12:06:57 +00:00
|
|
|
static unsigned int *fd_evts[2];
|
2007-04-08 14:39:58 +00:00
|
|
|
|
|
|
|
/* private data */
|
MAJOR: threads/fd: Make fd stuffs thread-safe
Many changes have been made to do so. First, the fd_updt array, where all
pending FDs for polling are stored, is now a thread-local array. Then 3 locks
have been added to protect, respectively, the fdtab array, the fd_cache array
and poll information. In addition, a lock for each entry in the fdtab array has
been added to protect all accesses to a specific FD or its information.
For pollers, according to the poller, the way to manage the concurrency is
different. There is a poller loop on each thread. So the set of monitored FDs
may need to be protected. epoll and kqueue are thread-safe per-se, so there few
things to do to protect these pollers. This is not possible with select and
poll, so there is no sharing between the threads. The poller on each thread is
independant from others.
Finally, per-thread init/deinit functions are used for each pollers and for FD
part for manage thread-local ressources.
Now, you must be carefull when a FD is created during the HAProxy startup. All
update on the FD state must be made in the threads context and never before
their creation. This is mandatory because fd_updt array is thread-local and
initialized only for threads. Because there is no pollers for the main one, this
array remains uninitialized in this context. For this reason, listeners are now
enabled in run_thread_poll_loop function, just like the worker pipe.
2017-05-29 08:40:41 +00:00
|
|
|
static THREAD_LOCAL int nbfd = 0;
|
|
|
|
static THREAD_LOCAL struct pollfd *poll_events = NULL;
|
2007-04-08 14:39:58 +00:00
|
|
|
|
2020-02-25 06:38:05 +00:00
|
|
|
static void __fd_clo(int fd)
|
2013-03-31 12:06:57 +00:00
|
|
|
{
|
|
|
|
hap_fd_clr(fd, fd_evts[DIR_RD]);
|
|
|
|
hap_fd_clr(fd, fd_evts[DIR_WR]);
|
2007-04-08 14:39:58 +00:00
|
|
|
}
|
|
|
|
|
2018-04-25 14:58:25 +00:00
|
|
|
static void _update_fd(int fd, int *max_add_fd)
|
|
|
|
{
|
|
|
|
int en;
|
2022-07-06 08:37:31 +00:00
|
|
|
ulong pr, ps;
|
2018-04-25 14:58:25 +00:00
|
|
|
|
|
|
|
en = fdtab[fd].state;
|
2022-07-06 08:37:31 +00:00
|
|
|
pr = _HA_ATOMIC_LOAD(&polled_mask[fd].poll_recv);
|
|
|
|
ps = _HA_ATOMIC_LOAD(&polled_mask[fd].poll_send);
|
2018-04-25 14:58:25 +00:00
|
|
|
|
|
|
|
/* we have a single state for all threads, which is why we
|
|
|
|
* don't check the tid_bit. First thread to see the update
|
|
|
|
* takes it for every other one.
|
|
|
|
*/
|
2019-09-04 07:52:57 +00:00
|
|
|
if (!(en & FD_EV_ACTIVE_RW)) {
|
2022-07-06 08:37:31 +00:00
|
|
|
if (!(pr | ps)) {
|
2018-04-25 14:58:25 +00:00
|
|
|
/* fd was not watched, it's still not */
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
/* fd totally removed from poll list */
|
|
|
|
hap_fd_clr(fd, fd_evts[DIR_RD]);
|
|
|
|
hap_fd_clr(fd, fd_evts[DIR_WR]);
|
2019-07-25 14:00:18 +00:00
|
|
|
_HA_ATOMIC_AND(&polled_mask[fd].poll_recv, 0);
|
|
|
|
_HA_ATOMIC_AND(&polled_mask[fd].poll_send, 0);
|
2018-04-25 14:58:25 +00:00
|
|
|
}
|
|
|
|
else {
|
|
|
|
/* OK fd has to be monitored, it was either added or changed */
|
2019-09-04 07:52:57 +00:00
|
|
|
if (!(en & FD_EV_ACTIVE_R)) {
|
2018-04-25 14:58:25 +00:00
|
|
|
hap_fd_clr(fd, fd_evts[DIR_RD]);
|
2022-07-06 08:37:31 +00:00
|
|
|
if (pr & ti->ltid_bit)
|
|
|
|
_HA_ATOMIC_AND(&polled_mask[fd].poll_recv, ~ti->ltid_bit);
|
2019-07-25 14:00:18 +00:00
|
|
|
} else {
|
2018-04-25 14:58:25 +00:00
|
|
|
hap_fd_set(fd, fd_evts[DIR_RD]);
|
2022-07-06 08:37:31 +00:00
|
|
|
if (!(pr & ti->ltid_bit))
|
|
|
|
_HA_ATOMIC_OR(&polled_mask[fd].poll_recv, ti->ltid_bit);
|
2019-07-25 14:00:18 +00:00
|
|
|
}
|
2018-04-25 14:58:25 +00:00
|
|
|
|
2019-09-04 07:52:57 +00:00
|
|
|
if (!(en & FD_EV_ACTIVE_W)) {
|
2018-04-25 14:58:25 +00:00
|
|
|
hap_fd_clr(fd, fd_evts[DIR_WR]);
|
2022-07-06 08:37:31 +00:00
|
|
|
if (ps & ti->ltid_bit)
|
|
|
|
_HA_ATOMIC_AND(&polled_mask[fd].poll_send, ~ti->ltid_bit);
|
|
|
|
} else {
|
2018-04-25 14:58:25 +00:00
|
|
|
hap_fd_set(fd, fd_evts[DIR_WR]);
|
2022-07-06 08:37:31 +00:00
|
|
|
if (!(ps & ti->ltid_bit))
|
|
|
|
_HA_ATOMIC_OR(&polled_mask[fd].poll_send, ti->ltid_bit);
|
2019-07-25 14:00:18 +00:00
|
|
|
}
|
2018-04-25 14:58:25 +00:00
|
|
|
|
|
|
|
if (fd > *max_add_fd)
|
|
|
|
*max_add_fd = fd;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2007-04-08 14:39:58 +00:00
|
|
|
/*
|
|
|
|
* Poll() poller
|
|
|
|
*/
|
2020-02-25 06:38:05 +00:00
|
|
|
static void _do_poll(struct poller *p, int exp, int wake)
|
2007-04-08 14:39:58 +00:00
|
|
|
{
|
|
|
|
int status;
|
MAJOR: threads/fd: Make fd stuffs thread-safe
Many changes have been made to do so. First, the fd_updt array, where all
pending FDs for polling are stored, is now a thread-local array. Then 3 locks
have been added to protect, respectively, the fdtab array, the fd_cache array
and poll information. In addition, a lock for each entry in the fdtab array has
been added to protect all accesses to a specific FD or its information.
For pollers, according to the poller, the way to manage the concurrency is
different. There is a poller loop on each thread. So the set of monitored FDs
may need to be protected. epoll and kqueue are thread-safe per-se, so there few
things to do to protect these pollers. This is not possible with select and
poll, so there is no sharing between the threads. The poller on each thread is
independant from others.
Finally, per-thread init/deinit functions are used for each pollers and for FD
part for manage thread-local ressources.
Now, you must be carefull when a FD is created during the HAProxy startup. All
update on the FD state must be made in the threads context and never before
their creation. This is mandatory because fd_updt array is thread-local and
initialized only for threads. Because there is no pollers for the main one, this
array remains uninitialized in this context. For this reason, listeners are now
enabled in run_thread_poll_loop function, just like the worker pipe.
2017-05-29 08:40:41 +00:00
|
|
|
int fd;
|
2007-05-12 20:35:00 +00:00
|
|
|
int wait_time;
|
2018-04-25 14:58:25 +00:00
|
|
|
int updt_idx;
|
2007-04-08 14:39:58 +00:00
|
|
|
int fds, count;
|
|
|
|
int sr, sw;
|
2018-01-26 20:48:23 +00:00
|
|
|
int old_maxfd, new_maxfd, max_add_fd;
|
2007-04-08 14:39:58 +00:00
|
|
|
unsigned rn, wn; /* read new, write new */
|
2018-04-25 14:58:25 +00:00
|
|
|
int old_fd;
|
2007-04-08 14:39:58 +00:00
|
|
|
|
2018-01-26 20:48:23 +00:00
|
|
|
max_add_fd = -1;
|
|
|
|
|
2012-11-11 16:25:15 +00:00
|
|
|
/* first, scan the update list to find changes */
|
|
|
|
for (updt_idx = 0; updt_idx < fd_nbupdt; updt_idx++) {
|
|
|
|
fd = fd_updt[updt_idx];
|
MAJOR: polling: rework the whole polling system
This commit heavily changes the polling system in order to definitely
fix the frequent breakage of SSL which needs to remember the last
EAGAIN before deciding whether to poll or not. Now we have a state per
direction for each FD, as opposed to a previous and current state
previously. An FD can have up to 8 different states for each direction,
each of which being the result of a 3-bit combination. These 3 bits
indicate a wish to access the FD, the readiness of the FD and the
subscription of the FD to the polling system.
This means that it will now be possible to remember the state of a
file descriptor across disable/enable sequences that generally happen
during forwarding, where enabling reading on a previously disabled FD
would result in forgetting the EAGAIN flag it met last time.
Several new state manipulation functions have been introduced or
adapted :
- fd_want_{recv,send} : enable receiving/sending on the FD regardless
of its state (sets the ACTIVE flag) ;
- fd_stop_{recv,send} : stop receiving/sending on the FD regardless
of its state (clears the ACTIVE flag) ;
- fd_cant_{recv,send} : report a failure to receive/send on the FD
corresponding to EAGAIN (clears the READY flag) ;
- fd_may_{recv,send} : report the ability to receive/send on the FD
as reported by poll() (sets the READY flag) ;
Some functions are used to report the current FD status :
- fd_{recv,send}_active
- fd_{recv,send}_ready
- fd_{recv,send}_polled
Some functions were removed :
- fd_ev_clr(), fd_ev_set(), fd_ev_rem(), fd_ev_wai()
The POLLHUP/POLLERR flags are now reported as ready so that the I/O layers
knows it can try to access the file descriptor to get this information.
In order to simplify the conditions to add/remove cache entries, a new
function fd_alloc_or_release_cache_entry() was created to be used from
pollers while scanning for updates.
The following pollers have been updated :
ev_select() : done, built, tested on Linux 3.10
ev_poll() : done, built, tested on Linux 3.10
ev_epoll() : done, built, tested on Linux 3.10 & 3.13
ev_kqueue() : done, built, tested on OpenBSD 5.2
2014-01-10 15:58:45 +00:00
|
|
|
|
2022-07-05 17:21:06 +00:00
|
|
|
_HA_ATOMIC_AND(&fdtab[fd].update_mask, ~ti->ltid_bit);
|
2018-01-20 18:30:13 +00:00
|
|
|
if (!fdtab[fd].owner) {
|
2020-06-17 18:35:33 +00:00
|
|
|
activity[tid].poll_drop_fd++;
|
MAJOR: polling: rework the whole polling system
This commit heavily changes the polling system in order to definitely
fix the frequent breakage of SSL which needs to remember the last
EAGAIN before deciding whether to poll or not. Now we have a state per
direction for each FD, as opposed to a previous and current state
previously. An FD can have up to 8 different states for each direction,
each of which being the result of a 3-bit combination. These 3 bits
indicate a wish to access the FD, the readiness of the FD and the
subscription of the FD to the polling system.
This means that it will now be possible to remember the state of a
file descriptor across disable/enable sequences that generally happen
during forwarding, where enabling reading on a previously disabled FD
would result in forgetting the EAGAIN flag it met last time.
Several new state manipulation functions have been introduced or
adapted :
- fd_want_{recv,send} : enable receiving/sending on the FD regardless
of its state (sets the ACTIVE flag) ;
- fd_stop_{recv,send} : stop receiving/sending on the FD regardless
of its state (clears the ACTIVE flag) ;
- fd_cant_{recv,send} : report a failure to receive/send on the FD
corresponding to EAGAIN (clears the READY flag) ;
- fd_may_{recv,send} : report the ability to receive/send on the FD
as reported by poll() (sets the READY flag) ;
Some functions are used to report the current FD status :
- fd_{recv,send}_active
- fd_{recv,send}_ready
- fd_{recv,send}_polled
Some functions were removed :
- fd_ev_clr(), fd_ev_set(), fd_ev_rem(), fd_ev_wai()
The POLLHUP/POLLERR flags are now reported as ready so that the I/O layers
knows it can try to access the file descriptor to get this information.
In order to simplify the conditions to add/remove cache entries, a new
function fd_alloc_or_release_cache_entry() was created to be used from
pollers while scanning for updates.
The following pollers have been updated :
ev_select() : done, built, tested on Linux 3.10
ev_poll() : done, built, tested on Linux 3.10
ev_epoll() : done, built, tested on Linux 3.10 & 3.13
ev_kqueue() : done, built, tested on OpenBSD 5.2
2014-01-10 15:58:45 +00:00
|
|
|
continue;
|
2018-01-20 18:30:13 +00:00
|
|
|
}
|
2018-04-25 14:58:25 +00:00
|
|
|
_update_fd(fd, &max_add_fd);
|
|
|
|
}
|
MAJOR: polling: rework the whole polling system
This commit heavily changes the polling system in order to definitely
fix the frequent breakage of SSL which needs to remember the last
EAGAIN before deciding whether to poll or not. Now we have a state per
direction for each FD, as opposed to a previous and current state
previously. An FD can have up to 8 different states for each direction,
each of which being the result of a 3-bit combination. These 3 bits
indicate a wish to access the FD, the readiness of the FD and the
subscription of the FD to the polling system.
This means that it will now be possible to remember the state of a
file descriptor across disable/enable sequences that generally happen
during forwarding, where enabling reading on a previously disabled FD
would result in forgetting the EAGAIN flag it met last time.
Several new state manipulation functions have been introduced or
adapted :
- fd_want_{recv,send} : enable receiving/sending on the FD regardless
of its state (sets the ACTIVE flag) ;
- fd_stop_{recv,send} : stop receiving/sending on the FD regardless
of its state (clears the ACTIVE flag) ;
- fd_cant_{recv,send} : report a failure to receive/send on the FD
corresponding to EAGAIN (clears the READY flag) ;
- fd_may_{recv,send} : report the ability to receive/send on the FD
as reported by poll() (sets the READY flag) ;
Some functions are used to report the current FD status :
- fd_{recv,send}_active
- fd_{recv,send}_ready
- fd_{recv,send}_polled
Some functions were removed :
- fd_ev_clr(), fd_ev_set(), fd_ev_rem(), fd_ev_wai()
The POLLHUP/POLLERR flags are now reported as ready so that the I/O layers
knows it can try to access the file descriptor to get this information.
In order to simplify the conditions to add/remove cache entries, a new
function fd_alloc_or_release_cache_entry() was created to be used from
pollers while scanning for updates.
The following pollers have been updated :
ev_select() : done, built, tested on Linux 3.10
ev_poll() : done, built, tested on Linux 3.10
ev_epoll() : done, built, tested on Linux 3.10 & 3.13
ev_kqueue() : done, built, tested on OpenBSD 5.2
2014-01-10 15:58:45 +00:00
|
|
|
|
2018-04-25 14:58:25 +00:00
|
|
|
/* Now scan the global update list */
|
2022-07-08 09:33:43 +00:00
|
|
|
for (old_fd = fd = update_list[tgid - 1].first; fd != -1; fd = fdtab[fd].update.next) {
|
2018-04-25 14:58:25 +00:00
|
|
|
if (fd == -2) {
|
|
|
|
fd = old_fd;
|
|
|
|
continue;
|
MAJOR: polling: rework the whole polling system
This commit heavily changes the polling system in order to definitely
fix the frequent breakage of SSL which needs to remember the last
EAGAIN before deciding whether to poll or not. Now we have a state per
direction for each FD, as opposed to a previous and current state
previously. An FD can have up to 8 different states for each direction,
each of which being the result of a 3-bit combination. These 3 bits
indicate a wish to access the FD, the readiness of the FD and the
subscription of the FD to the polling system.
This means that it will now be possible to remember the state of a
file descriptor across disable/enable sequences that generally happen
during forwarding, where enabling reading on a previously disabled FD
would result in forgetting the EAGAIN flag it met last time.
Several new state manipulation functions have been introduced or
adapted :
- fd_want_{recv,send} : enable receiving/sending on the FD regardless
of its state (sets the ACTIVE flag) ;
- fd_stop_{recv,send} : stop receiving/sending on the FD regardless
of its state (clears the ACTIVE flag) ;
- fd_cant_{recv,send} : report a failure to receive/send on the FD
corresponding to EAGAIN (clears the READY flag) ;
- fd_may_{recv,send} : report the ability to receive/send on the FD
as reported by poll() (sets the READY flag) ;
Some functions are used to report the current FD status :
- fd_{recv,send}_active
- fd_{recv,send}_ready
- fd_{recv,send}_polled
Some functions were removed :
- fd_ev_clr(), fd_ev_set(), fd_ev_rem(), fd_ev_wai()
The POLLHUP/POLLERR flags are now reported as ready so that the I/O layers
knows it can try to access the file descriptor to get this information.
In order to simplify the conditions to add/remove cache entries, a new
function fd_alloc_or_release_cache_entry() was created to be used from
pollers while scanning for updates.
The following pollers have been updated :
ev_select() : done, built, tested on Linux 3.10
ev_poll() : done, built, tested on Linux 3.10
ev_epoll() : done, built, tested on Linux 3.10 & 3.13
ev_kqueue() : done, built, tested on OpenBSD 5.2
2014-01-10 15:58:45 +00:00
|
|
|
}
|
2018-04-25 14:58:25 +00:00
|
|
|
else if (fd <= -3)
|
|
|
|
fd = -fd -4;
|
|
|
|
if (fd == -1)
|
|
|
|
break;
|
2022-07-05 17:21:06 +00:00
|
|
|
if (fdtab[fd].update_mask & ti->ltid_bit) {
|
2018-04-25 14:58:25 +00:00
|
|
|
/* Cheat a bit, as the state is global to all pollers
|
2021-01-06 16:35:12 +00:00
|
|
|
* we don't need every thread to take care of the
|
2018-04-25 14:58:25 +00:00
|
|
|
* update.
|
|
|
|
*/
|
2022-07-05 17:21:06 +00:00
|
|
|
_HA_ATOMIC_AND(&fdtab[fd].update_mask, ~tg->threads_enabled);
|
2018-04-25 14:58:25 +00:00
|
|
|
done_update_polling(fd);
|
|
|
|
} else
|
|
|
|
continue;
|
|
|
|
if (!fdtab[fd].owner)
|
|
|
|
continue;
|
|
|
|
_update_fd(fd, &max_add_fd);
|
2012-11-11 16:25:15 +00:00
|
|
|
}
|
2018-01-26 20:48:23 +00:00
|
|
|
|
|
|
|
/* maybe we added at least one fd larger than maxfd */
|
|
|
|
for (old_maxfd = maxfd; old_maxfd <= max_add_fd; ) {
|
2019-03-08 17:49:54 +00:00
|
|
|
if (_HA_ATOMIC_CAS(&maxfd, &old_maxfd, max_add_fd + 1))
|
2018-01-26 20:48:23 +00:00
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* maxfd doesn't need to be precise but it needs to cover *all* active
|
|
|
|
* FDs. Thus we only shrink it if we have such an opportunity. The algo
|
|
|
|
* is simple : look for the previous used place, try to update maxfd to
|
|
|
|
* point to it, abort if maxfd changed in the mean time.
|
|
|
|
*/
|
|
|
|
old_maxfd = maxfd;
|
|
|
|
do {
|
|
|
|
new_maxfd = old_maxfd;
|
|
|
|
while (new_maxfd - 1 >= 0 && !fdtab[new_maxfd - 1].owner)
|
|
|
|
new_maxfd--;
|
|
|
|
if (new_maxfd >= old_maxfd)
|
|
|
|
break;
|
2019-03-08 17:49:54 +00:00
|
|
|
} while (!_HA_ATOMIC_CAS(&maxfd, &old_maxfd, new_maxfd));
|
2018-01-26 20:48:23 +00:00
|
|
|
|
MEDIUM: threads: add a stronger thread_isolate_full() call
The current principle of running under isolation was made to access
sensitive data while being certain that no other thread was using them
in parallel, without necessarily having to place locks everywhere. The
main use case are "show sess" and "show fd" which run over long chains
of pointers.
The thread_isolate() call relies on the "harmless" bit that indicates
for a given thread that it's not currently doing such sensitive things,
which is advertised using thread_harmless_now() and which ends usings
thread_harmless_end(), which also waits for possibly concurrent threads
to complete their work if they took this opportunity for starting
something tricky.
As some system calls were notoriously slow (e.g. mmap()), a bunch of
thread_harmless_now() / thread_harmless_end() were placed around them
to let waiting threads do their work while such other threads were not
able to modify memory contents.
But this is not sufficient for performing memory modifications. One such
example is the server deletion code. By modifying memory, it not only
requires that other threads are not playing with it, but are not either
in the process of touching it. The fact that a pool_alloc() or pool_free()
on some structure may call thread_harmless_now() and let another thread
start to release the same object's memory is not acceptable.
This patch introduces the concept of "idle threads". Threads entering
the polling loop are idle, as well as those that are waiting for all
others to become idle via the new function thread_isolate_full(). Once
thread_isolate_full() is granted, the thread is not idle anymore, and
it is released using thread_release() just like regular isolation. Its
users have to keep in mind that across this call nothing is granted as
another thread might have performed shared memory modifications. But
such users are extremely rare and are actually expecting this from their
peers as well.
Note that that in case of backport, this patch depends on previous patch:
MINOR: threads: make thread_release() not wait for other ones to complete
2021-08-04 09:44:17 +00:00
|
|
|
thread_idle_now();
|
2018-08-02 08:16:17 +00:00
|
|
|
thread_harmless_now();
|
|
|
|
|
2012-11-11 16:25:15 +00:00
|
|
|
fd_nbupdt = 0;
|
|
|
|
|
2007-04-08 14:39:58 +00:00
|
|
|
nbfd = 0;
|
2013-03-31 12:06:57 +00:00
|
|
|
for (fds = 0; (fds * 8*sizeof(**fd_evts)) < maxfd; fds++) {
|
|
|
|
rn = fd_evts[DIR_RD][fds];
|
|
|
|
wn = fd_evts[DIR_WR][fds];
|
MAJOR: threads/fd: Make fd stuffs thread-safe
Many changes have been made to do so. First, the fd_updt array, where all
pending FDs for polling are stored, is now a thread-local array. Then 3 locks
have been added to protect, respectively, the fdtab array, the fd_cache array
and poll information. In addition, a lock for each entry in the fdtab array has
been added to protect all accesses to a specific FD or its information.
For pollers, according to the poller, the way to manage the concurrency is
different. There is a poller loop on each thread. So the set of monitored FDs
may need to be protected. epoll and kqueue are thread-safe per-se, so there few
things to do to protect these pollers. This is not possible with select and
poll, so there is no sharing between the threads. The poller on each thread is
independant from others.
Finally, per-thread init/deinit functions are used for each pollers and for FD
part for manage thread-local ressources.
Now, you must be carefull when a FD is created during the HAProxy startup. All
update on the FD state must be made in the threads context and never before
their creation. This is mandatory because fd_updt array is thread-local and
initialized only for threads. Because there is no pollers for the main one, this
array remains uninitialized in this context. For this reason, listeners are now
enabled in run_thread_poll_loop function, just like the worker pipe.
2017-05-29 08:40:41 +00:00
|
|
|
|
2013-03-31 12:06:57 +00:00
|
|
|
if (!(rn|wn))
|
|
|
|
continue;
|
|
|
|
|
|
|
|
for (count = 0, fd = fds * 8*sizeof(**fd_evts); count < 8*sizeof(**fd_evts) && fd < maxfd; count++, fd++) {
|
|
|
|
sr = (rn >> count) & 1;
|
|
|
|
sw = (wn >> count) & 1;
|
|
|
|
if ((sr|sw)) {
|
2018-01-20 18:30:13 +00:00
|
|
|
if (!fdtab[fd].owner) {
|
|
|
|
/* should normally not happen here except
|
|
|
|
* due to rare thread concurrency
|
|
|
|
*/
|
|
|
|
continue;
|
|
|
|
}
|
|
|
|
|
2022-07-07 06:23:03 +00:00
|
|
|
if (!(fdtab[fd].thread_mask & ti->ltid_bit)) {
|
2018-01-20 18:30:13 +00:00
|
|
|
continue;
|
|
|
|
}
|
|
|
|
|
2013-03-31 12:06:57 +00:00
|
|
|
poll_events[nbfd].fd = fd;
|
2017-03-13 16:14:51 +00:00
|
|
|
poll_events[nbfd].events = (sr ? (POLLIN | POLLRDHUP) : 0) | (sw ? POLLOUT : 0);
|
2013-03-31 12:06:57 +00:00
|
|
|
nbfd++;
|
2007-04-08 14:39:58 +00:00
|
|
|
}
|
MAJOR: threads/fd: Make fd stuffs thread-safe
Many changes have been made to do so. First, the fd_updt array, where all
pending FDs for polling are stored, is now a thread-local array. Then 3 locks
have been added to protect, respectively, the fdtab array, the fd_cache array
and poll information. In addition, a lock for each entry in the fdtab array has
been added to protect all accesses to a specific FD or its information.
For pollers, according to the poller, the way to manage the concurrency is
different. There is a poller loop on each thread. So the set of monitored FDs
may need to be protected. epoll and kqueue are thread-safe per-se, so there few
things to do to protect these pollers. This is not possible with select and
poll, so there is no sharing between the threads. The poller on each thread is
independant from others.
Finally, per-thread init/deinit functions are used for each pollers and for FD
part for manage thread-local ressources.
Now, you must be carefull when a FD is created during the HAProxy startup. All
update on the FD state must be made in the threads context and never before
their creation. This is mandatory because fd_updt array is thread-local and
initialized only for threads. Because there is no pollers for the main one, this
array remains uninitialized in this context. For this reason, listeners are now
enabled in run_thread_poll_loop function, just like the worker pipe.
2017-05-29 08:40:41 +00:00
|
|
|
}
|
2007-04-08 14:39:58 +00:00
|
|
|
}
|
MAJOR: threads/fd: Make fd stuffs thread-safe
Many changes have been made to do so. First, the fd_updt array, where all
pending FDs for polling are stored, is now a thread-local array. Then 3 locks
have been added to protect, respectively, the fdtab array, the fd_cache array
and poll information. In addition, a lock for each entry in the fdtab array has
been added to protect all accesses to a specific FD or its information.
For pollers, according to the poller, the way to manage the concurrency is
different. There is a poller loop on each thread. So the set of monitored FDs
may need to be protected. epoll and kqueue are thread-safe per-se, so there few
things to do to protect these pollers. This is not possible with select and
poll, so there is no sharing between the threads. The poller on each thread is
independant from others.
Finally, per-thread init/deinit functions are used for each pollers and for FD
part for manage thread-local ressources.
Now, you must be carefull when a FD is created during the HAProxy startup. All
update on the FD state must be made in the threads context and never before
their creation. This is mandatory because fd_updt array is thread-local and
initialized only for threads. Because there is no pollers for the main one, this
array remains uninitialized in this context. For this reason, listeners are now
enabled in run_thread_poll_loop function, just like the worker pipe.
2017-05-29 08:40:41 +00:00
|
|
|
|
2022-09-09 08:21:00 +00:00
|
|
|
/* Now let's wait for polled events. */
|
|
|
|
wait_time = wake ? 0 : compute_poll_timeout(exp);
|
2021-10-08 08:43:59 +00:00
|
|
|
clock_entering_poll();
|
2007-04-08 14:39:58 +00:00
|
|
|
status = poll(poll_events, nbfd, wait_time);
|
2021-10-08 07:33:24 +00:00
|
|
|
clock_update_date(wait_time, status);
|
2007-04-08 14:39:58 +00:00
|
|
|
|
2022-06-22 13:21:34 +00:00
|
|
|
fd_leaving_poll(wait_time, status);
|
2018-08-02 08:16:17 +00:00
|
|
|
|
2020-06-17 18:25:18 +00:00
|
|
|
if (status > 0)
|
|
|
|
activity[tid].poll_io++;
|
|
|
|
|
2007-04-08 14:39:58 +00:00
|
|
|
for (count = 0; status > 0 && count < nbfd; count++) {
|
2017-08-30 08:34:36 +00:00
|
|
|
unsigned int n;
|
2012-07-06 09:16:01 +00:00
|
|
|
int e = poll_events[count].revents;
|
2022-07-09 16:55:37 +00:00
|
|
|
|
2007-04-08 14:39:58 +00:00
|
|
|
fd = poll_events[count].fd;
|
MAJOR: threads/fd: Make fd stuffs thread-safe
Many changes have been made to do so. First, the fd_updt array, where all
pending FDs for polling are stored, is now a thread-local array. Then 3 locks
have been added to protect, respectively, the fdtab array, the fd_cache array
and poll information. In addition, a lock for each entry in the fdtab array has
been added to protect all accesses to a specific FD or its information.
For pollers, according to the poller, the way to manage the concurrency is
different. There is a poller loop on each thread. So the set of monitored FDs
may need to be protected. epoll and kqueue are thread-safe per-se, so there few
things to do to protect these pollers. This is not possible with select and
poll, so there is no sharing between the threads. The poller on each thread is
independant from others.
Finally, per-thread init/deinit functions are used for each pollers and for FD
part for manage thread-local ressources.
Now, you must be carefull when a FD is created during the HAProxy startup. All
update on the FD state must be made in the threads context and never before
their creation. This is mandatory because fd_updt array is thread-local and
initialized only for threads. Because there is no pollers for the main one, this
array remains uninitialized in this context. For this reason, listeners are now
enabled in run_thread_poll_loop function, just like the worker pipe.
2017-05-29 08:40:41 +00:00
|
|
|
|
2021-07-29 14:19:24 +00:00
|
|
|
if ((e & POLLRDHUP) && !(cur_poller.flags & HAP_POLL_F_RDHUP))
|
|
|
|
_HA_ATOMIC_OR(&cur_poller.flags, HAP_POLL_F_RDHUP);
|
|
|
|
|
2020-06-23 08:04:54 +00:00
|
|
|
#ifdef DEBUG_FD
|
2021-04-06 11:53:36 +00:00
|
|
|
_HA_ATOMIC_INC(&fdtab[fd].event_count);
|
2020-06-23 08:04:54 +00:00
|
|
|
#endif
|
2017-03-13 16:14:51 +00:00
|
|
|
if (!(e & ( POLLOUT | POLLIN | POLLERR | POLLHUP | POLLRDHUP )))
|
2007-04-08 14:39:58 +00:00
|
|
|
continue;
|
|
|
|
|
2012-07-06 14:02:29 +00:00
|
|
|
/* ok, we found one active fd */
|
|
|
|
status--;
|
|
|
|
|
2019-09-06 17:05:50 +00:00
|
|
|
n = ((e & POLLIN) ? FD_EV_READY_R : 0) |
|
|
|
|
((e & POLLOUT) ? FD_EV_READY_W : 0) |
|
|
|
|
((e & POLLRDHUP) ? FD_EV_SHUT_R : 0) |
|
|
|
|
((e & POLLHUP) ? FD_EV_SHUT_RW : 0) |
|
|
|
|
((e & POLLERR) ? FD_EV_ERR_RW : 0);
|
2012-07-06 09:16:01 +00:00
|
|
|
|
2022-07-09 16:55:37 +00:00
|
|
|
fd_update_events(fd, n);
|
MEDIUM: fd: rely more on fd_update_events() to detect changes
This function already performs a number of checks prior to calling the
IOCB, and detects the change of thread (FD migration). Half of the
controls are still in each poller, and these pollers also maintain
activity counters for various cases.
Note that the unreliable test on thread_mask was removed so that only
the one performed by fd_set_running() is now used, since this one is
reliable.
Let's centralize all that fd-specific logic into the function and make
it return a status among:
FD_UPDT_DONE, // update done, nothing else to be done
FD_UPDT_DEAD, // FD was already dead, ignore it
FD_UPDT_CLOSED, // FD was closed
FD_UPDT_MIGRATED, // FD was migrated, ignore it now
Some pollers already used to call it last and have nothing to do after
it, regardless of the result. epoll has to delete the FD in case a
migration is detected. Overall this removes more code than it adds.
2021-07-29 14:57:19 +00:00
|
|
|
}
|
2007-04-08 14:39:58 +00:00
|
|
|
}
|
|
|
|
|
MAJOR: threads/fd: Make fd stuffs thread-safe
Many changes have been made to do so. First, the fd_updt array, where all
pending FDs for polling are stored, is now a thread-local array. Then 3 locks
have been added to protect, respectively, the fdtab array, the fd_cache array
and poll information. In addition, a lock for each entry in the fdtab array has
been added to protect all accesses to a specific FD or its information.
For pollers, according to the poller, the way to manage the concurrency is
different. There is a poller loop on each thread. So the set of monitored FDs
may need to be protected. epoll and kqueue are thread-safe per-se, so there few
things to do to protect these pollers. This is not possible with select and
poll, so there is no sharing between the threads. The poller on each thread is
independant from others.
Finally, per-thread init/deinit functions are used for each pollers and for FD
part for manage thread-local ressources.
Now, you must be carefull when a FD is created during the HAProxy startup. All
update on the FD state must be made in the threads context and never before
their creation. This is mandatory because fd_updt array is thread-local and
initialized only for threads. Because there is no pollers for the main one, this
array remains uninitialized in this context. For this reason, listeners are now
enabled in run_thread_poll_loop function, just like the worker pipe.
2017-05-29 08:40:41 +00:00
|
|
|
|
|
|
|
static int init_poll_per_thread()
|
|
|
|
{
|
|
|
|
poll_events = calloc(1, sizeof(struct pollfd) * global.maxsock);
|
|
|
|
if (poll_events == NULL)
|
|
|
|
return 0;
|
|
|
|
return 1;
|
|
|
|
}
|
|
|
|
|
|
|
|
static void deinit_poll_per_thread()
|
|
|
|
{
|
2021-02-20 09:46:51 +00:00
|
|
|
ha_free(&poll_events);
|
MAJOR: threads/fd: Make fd stuffs thread-safe
Many changes have been made to do so. First, the fd_updt array, where all
pending FDs for polling are stored, is now a thread-local array. Then 3 locks
have been added to protect, respectively, the fdtab array, the fd_cache array
and poll information. In addition, a lock for each entry in the fdtab array has
been added to protect all accesses to a specific FD or its information.
For pollers, according to the poller, the way to manage the concurrency is
different. There is a poller loop on each thread. So the set of monitored FDs
may need to be protected. epoll and kqueue are thread-safe per-se, so there few
things to do to protect these pollers. This is not possible with select and
poll, so there is no sharing between the threads. The poller on each thread is
independant from others.
Finally, per-thread init/deinit functions are used for each pollers and for FD
part for manage thread-local ressources.
Now, you must be carefull when a FD is created during the HAProxy startup. All
update on the FD state must be made in the threads context and never before
their creation. This is mandatory because fd_updt array is thread-local and
initialized only for threads. Because there is no pollers for the main one, this
array remains uninitialized in this context. For this reason, listeners are now
enabled in run_thread_poll_loop function, just like the worker pipe.
2017-05-29 08:40:41 +00:00
|
|
|
}
|
|
|
|
|
2007-04-09 07:23:31 +00:00
|
|
|
/*
|
|
|
|
* Initialization of the poll() poller.
|
|
|
|
* Returns 0 in case of failure, non-zero in case of success. If it fails, it
|
|
|
|
* disables the poller by setting its pref to 0.
|
|
|
|
*/
|
2020-02-25 06:38:05 +00:00
|
|
|
static int _do_init(struct poller *p)
|
2007-04-09 07:23:31 +00:00
|
|
|
{
|
2017-10-27 11:53:47 +00:00
|
|
|
__label__ fail_swevt, fail_srevt;
|
2013-03-31 12:06:57 +00:00
|
|
|
int fd_evts_bytes;
|
2007-04-09 07:23:31 +00:00
|
|
|
|
|
|
|
p->private = NULL;
|
2022-07-09 21:38:46 +00:00
|
|
|
|
|
|
|
/* this old poller uses a process-wide FD list that cannot work with
|
|
|
|
* groups.
|
|
|
|
*/
|
|
|
|
if (global.nbtgroups > 1)
|
|
|
|
goto fail_srevt;
|
|
|
|
|
2018-01-17 14:48:53 +00:00
|
|
|
fd_evts_bytes = (global.maxsock + sizeof(**fd_evts) * 8 - 1) / (sizeof(**fd_evts) * 8) * sizeof(**fd_evts);
|
2007-04-09 07:23:31 +00:00
|
|
|
|
2013-03-31 12:06:57 +00:00
|
|
|
if ((fd_evts[DIR_RD] = calloc(1, fd_evts_bytes)) == NULL)
|
2007-04-09 07:23:31 +00:00
|
|
|
goto fail_srevt;
|
2013-03-31 12:06:57 +00:00
|
|
|
if ((fd_evts[DIR_WR] = calloc(1, fd_evts_bytes)) == NULL)
|
2007-04-09 07:23:31 +00:00
|
|
|
goto fail_swevt;
|
|
|
|
|
2017-10-27 11:53:47 +00:00
|
|
|
hap_register_per_thread_init(init_poll_per_thread);
|
|
|
|
hap_register_per_thread_deinit(deinit_poll_per_thread);
|
|
|
|
|
2007-04-09 07:23:31 +00:00
|
|
|
return 1;
|
|
|
|
|
|
|
|
fail_swevt:
|
|
|
|
free(fd_evts[DIR_RD]);
|
|
|
|
fail_srevt:
|
|
|
|
p->pref = 0;
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Termination of the poll() poller.
|
|
|
|
* Memory is released and the poller is marked as unselectable.
|
|
|
|
*/
|
2020-02-25 06:38:05 +00:00
|
|
|
static void _do_term(struct poller *p)
|
2007-04-09 07:23:31 +00:00
|
|
|
{
|
2008-08-03 10:19:50 +00:00
|
|
|
free(fd_evts[DIR_WR]);
|
|
|
|
free(fd_evts[DIR_RD]);
|
2007-04-09 07:23:31 +00:00
|
|
|
p->private = NULL;
|
|
|
|
p->pref = 0;
|
|
|
|
}
|
|
|
|
|
2007-04-09 17:29:56 +00:00
|
|
|
/*
|
|
|
|
* Check that the poller works.
|
|
|
|
* Returns 1 if OK, otherwise 0.
|
|
|
|
*/
|
2020-02-25 06:38:05 +00:00
|
|
|
static int _do_test(struct poller *p)
|
2007-04-09 17:29:56 +00:00
|
|
|
{
|
|
|
|
return 1;
|
|
|
|
}
|
|
|
|
|
2007-04-08 14:39:58 +00:00
|
|
|
/*
|
2022-04-25 17:00:55 +00:00
|
|
|
* Registers the poller.
|
2007-04-08 14:39:58 +00:00
|
|
|
*/
|
2007-04-15 22:25:25 +00:00
|
|
|
static void _do_register(void)
|
2007-04-08 14:39:58 +00:00
|
|
|
{
|
2007-04-15 22:25:25 +00:00
|
|
|
struct poller *p;
|
|
|
|
|
|
|
|
if (nbpollers >= MAX_POLLERS)
|
|
|
|
return;
|
|
|
|
p = &pollers[nbpollers++];
|
|
|
|
|
2007-04-08 14:39:58 +00:00
|
|
|
p->name = "poll";
|
|
|
|
p->pref = 200;
|
2019-11-28 17:17:33 +00:00
|
|
|
p->flags = HAP_POLL_F_ERRHUP;
|
2007-04-08 14:39:58 +00:00
|
|
|
p->private = NULL;
|
|
|
|
|
2012-11-11 20:02:34 +00:00
|
|
|
p->clo = __fd_clo;
|
2007-04-15 22:25:25 +00:00
|
|
|
p->test = _do_test;
|
|
|
|
p->init = _do_init;
|
|
|
|
p->term = _do_term;
|
|
|
|
p->poll = _do_poll;
|
2007-04-08 14:39:58 +00:00
|
|
|
}
|
|
|
|
|
2022-04-25 17:00:55 +00:00
|
|
|
INITCALL0(STG_REGISTER, _do_register);
|
2007-04-08 14:39:58 +00:00
|
|
|
|
|
|
|
/*
|
|
|
|
* Local variables:
|
|
|
|
* c-indent-level: 8
|
|
|
|
* c-basic-offset: 8
|
|
|
|
* End:
|
|
|
|
*/
|