2006-06-26 00:48:02 +00:00
|
|
|
/*
|
2010-08-27 15:56:48 +00:00
|
|
|
* include/proto/task.h
|
|
|
|
* Functions for task management.
|
|
|
|
*
|
|
|
|
* Copyright (C) 2000-2010 Willy Tarreau - w@1wt.eu
|
|
|
|
*
|
|
|
|
* This library is free software; you can redistribute it and/or
|
|
|
|
* modify it under the terms of the GNU Lesser General Public
|
|
|
|
* License as published by the Free Software Foundation, version 2.1
|
|
|
|
* exclusively.
|
|
|
|
*
|
|
|
|
* This library is distributed in the hope that it will be useful,
|
|
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
|
|
* Lesser General Public License for more details.
|
|
|
|
*
|
|
|
|
* You should have received a copy of the GNU Lesser General Public
|
|
|
|
* License along with this library; if not, write to the Free Software
|
|
|
|
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
|
|
|
|
*/
|
2006-06-26 00:48:02 +00:00
|
|
|
|
|
|
|
#ifndef _PROTO_TASK_H
|
|
|
|
#define _PROTO_TASK_H
|
|
|
|
|
|
|
|
|
|
|
|
#include <sys/time.h>
|
2006-06-29 16:54:54 +00:00
|
|
|
|
|
|
|
#include <common/config.h>
|
2006-06-29 15:53:05 +00:00
|
|
|
#include <common/memory.h>
|
2007-04-29 08:41:56 +00:00
|
|
|
#include <common/mini-clist.h>
|
|
|
|
#include <common/standard.h>
|
2009-03-08 14:53:06 +00:00
|
|
|
#include <common/ticks.h>
|
2017-09-27 12:59:38 +00:00
|
|
|
#include <common/hathreads.h>
|
|
|
|
|
MAJOR: task: make use of the scope-aware ebtree functions
Currently the task scheduler suffers from an O(n) lookup when
skipping tasks that are not for the current thread. The reason
is that eb32_lookup_ge() has no information about the current
thread so it always revisits many tasks for other threads before
finding its own tasks.
This is particularly visible with HTTP/2 since the number of
concurrent streams created at once causes long series of tasks
for the same stream in the scheduler. With only 10 connections
and 100 streams each, by running on two threads, the performance
drops from 640kreq/s to 11.2kreq/s! Lookup metrics show that for
only 200000 task lookups, 430 million skips had to be performed,
which means that on average, each lookup leads to 2150 nodes to
be visited.
This commit backports the principle of scope lookups for ebtrees
from the ebtree_v7 development tree. The idea is that each node
contains a mask indicating the union of the scopes for the nodes
below it, which is fed during insertion, and used during lookups.
Then during lookups, branches that do not contain any leaf matching
the requested scope are simply ignored. This perfectly matches a
thread mask, allowing a thread to only extract the tasks it cares
about from the run queue, and to always find them in O(log(n))
instead of O(n). Thus the scheduler uses tid_bit and
task->thread_mask as the ebtree scope here.
Doing this has recovered most of the performance, as can be seen on
the test below with two threads, 10 connections, 100 streams each,
and 1 million requests total :
Before After Gain
test duration : 89.6s 4.73s x19
HTTP requests/s (DEBUG) : 11200 211300 x19
HTTP requests/s (PROD) : 15900 447000 x28
spin_lock time : 85.2s 0.46s /185
time per lookup : 13us 40ns /325
Even when going to 6 threads (on 3 hyperthreaded CPU cores), the
performance stays around 284000 req/s, showing that the contention
is much lower.
A test showed that there's no benefit in using this for the wait queue
though.
2017-11-05 12:34:20 +00:00
|
|
|
#include <eb32sctree.h>
|
2009-10-26 20:10:04 +00:00
|
|
|
#include <eb32tree.h>
|
2007-04-29 08:41:56 +00:00
|
|
|
|
2014-11-13 15:57:19 +00:00
|
|
|
#include <types/global.h>
|
2006-06-29 16:54:54 +00:00
|
|
|
#include <types/task.h>
|
2006-06-26 00:48:02 +00:00
|
|
|
|
2009-03-08 14:53:06 +00:00
|
|
|
/* Principle of the wait queue.
|
|
|
|
*
|
|
|
|
* We want to be able to tell whether an expiration date is before of after the
|
|
|
|
* current time <now>. We KNOW that expiration dates are never too far apart,
|
|
|
|
* because they are measured in ticks (milliseconds). We also know that almost
|
|
|
|
* all dates will be in the future, and that a very small part of them will be
|
|
|
|
* in the past, they are the ones which have expired since last time we checked
|
|
|
|
* them. Using ticks, we know if a date is in the future or in the past, but we
|
|
|
|
* cannot use that to store sorted information because that reference changes
|
|
|
|
* all the time.
|
|
|
|
*
|
2009-03-21 09:01:42 +00:00
|
|
|
* We'll use the fact that the time wraps to sort timers. Timers above <now>
|
|
|
|
* are in the future, timers below <now> are in the past. Here, "above" and
|
|
|
|
* "below" are to be considered modulo 2^31.
|
2009-03-08 14:53:06 +00:00
|
|
|
*
|
2009-03-21 09:01:42 +00:00
|
|
|
* Timers are stored sorted in an ebtree. We use the new ability for ebtrees to
|
|
|
|
* lookup values starting from X to only expire tasks between <now> - 2^31 and
|
|
|
|
* <now>. If the end of the tree is reached while walking over it, we simply
|
|
|
|
* loop back to the beginning. That way, we have no problem keeping sorted
|
|
|
|
* wrapping timers in a tree, between (now - 24 days) and (now + 24 days). The
|
|
|
|
* keys in the tree always reflect their real position, none can be infinite.
|
|
|
|
* This reduces the number of checks to be performed.
|
2009-03-08 14:53:06 +00:00
|
|
|
*
|
|
|
|
* Another nice optimisation is to allow a timer to stay at an old place in the
|
|
|
|
* queue as long as it's not further than the real expiration date. That way,
|
|
|
|
* we use the tree as a place holder for a minorant of the real expiration
|
|
|
|
* date. Since we have a very low chance of hitting a timeout anyway, we can
|
|
|
|
* bounce the nodes to their right place when we scan the tree if we encounter
|
|
|
|
* a misplaced node once in a while. This even allows us not to remove the
|
|
|
|
* infinite timers from the wait queue.
|
|
|
|
*
|
|
|
|
* So, to summarize, we have :
|
|
|
|
* - node->key always defines current position in the wait queue
|
|
|
|
* - timer is the real expiration date (possibly infinite)
|
2009-03-21 09:01:42 +00:00
|
|
|
* - node->key is always before or equal to timer
|
2009-03-08 14:53:06 +00:00
|
|
|
*
|
|
|
|
* The run queue works similarly to the wait queue except that the current date
|
|
|
|
* is replaced by an insertion counter which can also wrap without any problem.
|
|
|
|
*/
|
|
|
|
|
2009-03-21 09:01:42 +00:00
|
|
|
/* The farthest we can look back in a timer tree */
|
|
|
|
#define TIMER_LOOK_BACK (1U << 31)
|
2009-03-08 14:53:06 +00:00
|
|
|
|
|
|
|
/* a few exported variables */
|
2009-03-21 17:13:21 +00:00
|
|
|
extern unsigned int nb_tasks; /* total number of tasks */
|
2017-11-14 09:26:53 +00:00
|
|
|
extern unsigned long active_tasks_mask; /* Mask of threads with active tasks */
|
2016-12-06 08:15:30 +00:00
|
|
|
extern unsigned int tasks_run_queue; /* run queue size */
|
|
|
|
extern unsigned int tasks_run_queue_cur;
|
2009-03-21 17:33:52 +00:00
|
|
|
extern unsigned int nb_tasks_cur;
|
2008-06-30 05:51:00 +00:00
|
|
|
extern unsigned int niced_tasks; /* number of niced tasks in the run queue */
|
2017-11-24 16:34:44 +00:00
|
|
|
extern struct pool_head *pool_head_task;
|
2018-05-18 16:45:28 +00:00
|
|
|
extern struct pool_head *pool_head_tasklet;
|
2017-11-24 16:34:44 +00:00
|
|
|
extern struct pool_head *pool_head_notification;
|
2018-05-04 13:46:16 +00:00
|
|
|
extern THREAD_LOCAL struct task *curr_task; /* task currently running or NULL */
|
|
|
|
extern THREAD_LOCAL struct eb32sc_node *rq_next; /* Next task to be potentially run */
|
2018-06-06 12:22:03 +00:00
|
|
|
#ifdef USE_THREAD
|
2018-05-18 16:38:23 +00:00
|
|
|
extern struct eb_root rqueue; /* tree constituting the run queue */
|
2018-06-06 12:22:03 +00:00
|
|
|
#endif
|
2018-05-18 16:38:23 +00:00
|
|
|
extern struct eb_root rqueue_local[MAX_THREADS]; /* tree constituting the per-thread run queue */
|
2018-05-18 16:45:28 +00:00
|
|
|
extern struct list task_list[MAX_THREADS]; /* List of tasks to be run, mixing tasks and tasklets */
|
|
|
|
extern int task_list_size[MAX_THREADS]; /* Number of task sin the task_list */
|
2017-11-13 09:34:01 +00:00
|
|
|
|
|
|
|
__decl_hathreads(extern HA_SPINLOCK_T rq_lock); /* spin lock related to run queue */
|
|
|
|
__decl_hathreads(extern HA_SPINLOCK_T wq_lock); /* spin lock related to wait queue */
|
2007-05-13 17:43:47 +00:00
|
|
|
|
2009-03-07 16:25:21 +00:00
|
|
|
/* return 0 if task is in run queue, otherwise non-zero */
|
|
|
|
static inline int task_in_rq(struct task *t)
|
|
|
|
{
|
2018-05-18 16:45:28 +00:00
|
|
|
/* Check if leaf_p is NULL, in case he's not in the runqueue, and if
|
|
|
|
* it's not 0x1, which would mean it's in the tasklet list.
|
|
|
|
*/
|
|
|
|
return t->rq.node.leaf_p != NULL && t->rq.node.leaf_p != (void *)0x1;
|
2009-03-07 16:25:21 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
/* return 0 if task is in wait queue, otherwise non-zero */
|
|
|
|
static inline int task_in_wq(struct task *t)
|
|
|
|
{
|
|
|
|
return t->wq.node.leaf_p != NULL;
|
|
|
|
}
|
|
|
|
|
2008-08-29 16:19:04 +00:00
|
|
|
/* puts the task <t> in run queue with reason flags <f>, and returns <t> */
|
2018-05-18 16:38:23 +00:00
|
|
|
/* This will put the task in the local runqueue if the task is only runnable
|
|
|
|
* by the current thread, in the global runqueue otherwies.
|
|
|
|
*/
|
|
|
|
void __task_wakeup(struct task *t, struct eb_root *);
|
|
|
|
static inline void task_wakeup(struct task *t, unsigned int f)
|
2008-08-29 13:26:14 +00:00
|
|
|
{
|
2018-05-18 16:38:23 +00:00
|
|
|
unsigned short state;
|
|
|
|
|
|
|
|
#ifdef USE_THREAD
|
|
|
|
struct eb_root *root;
|
|
|
|
|
2018-05-18 16:45:28 +00:00
|
|
|
if (t->thread_mask == tid_bit || global.nbthread == 1)
|
2018-05-18 16:38:23 +00:00
|
|
|
root = &rqueue_local[tid];
|
|
|
|
else
|
|
|
|
root = &rqueue;
|
|
|
|
#else
|
2018-06-06 12:01:08 +00:00
|
|
|
struct eb_root *root = &rqueue_local[tid];
|
2018-05-18 16:38:23 +00:00
|
|
|
#endif
|
|
|
|
|
|
|
|
state = HA_ATOMIC_OR(&t->state, f);
|
|
|
|
if (!(state & TASK_RUNNING))
|
|
|
|
__task_wakeup(t, root);
|
2008-08-29 13:26:14 +00:00
|
|
|
}
|
2006-06-26 00:48:02 +00:00
|
|
|
|
2017-10-31 15:06:06 +00:00
|
|
|
/* change the thread affinity of a task to <thread_mask> */
|
2017-09-27 12:59:38 +00:00
|
|
|
static inline void task_set_affinity(struct task *t, unsigned long thread_mask)
|
|
|
|
{
|
2017-10-31 15:06:06 +00:00
|
|
|
t->thread_mask = thread_mask;
|
2017-09-27 12:59:38 +00:00
|
|
|
}
|
2017-10-31 15:06:06 +00:00
|
|
|
|
2009-03-07 16:25:21 +00:00
|
|
|
/*
|
|
|
|
* Unlink the task from the wait queue, and possibly update the last_timer
|
|
|
|
* pointer. A pointer to the task itself is returned. The task *must* already
|
|
|
|
* be in the wait queue before calling this function. If unsure, use the safer
|
|
|
|
* task_unlink_wq() function.
|
2006-06-26 00:48:02 +00:00
|
|
|
*/
|
2009-03-07 16:25:21 +00:00
|
|
|
static inline struct task *__task_unlink_wq(struct task *t)
|
|
|
|
{
|
|
|
|
eb32_delete(&t->wq);
|
|
|
|
return t;
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline struct task *task_unlink_wq(struct task *t)
|
2006-06-26 00:48:02 +00:00
|
|
|
{
|
2017-11-07 09:42:54 +00:00
|
|
|
HA_SPIN_LOCK(TASK_WQ_LOCK, &wq_lock);
|
2009-03-07 16:25:21 +00:00
|
|
|
if (likely(task_in_wq(t)))
|
|
|
|
__task_unlink_wq(t);
|
2017-11-07 09:42:54 +00:00
|
|
|
HA_SPIN_UNLOCK(TASK_WQ_LOCK, &wq_lock);
|
2007-04-29 08:41:56 +00:00
|
|
|
return t;
|
2006-06-26 00:48:02 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
2016-12-06 08:15:30 +00:00
|
|
|
* Unlink the task from the run queue. The tasks_run_queue size and number of
|
|
|
|
* niced tasks are updated too. A pointer to the task itself is returned. The
|
|
|
|
* task *must* already be in the run queue before calling this function. If
|
|
|
|
* unsure, use the safer task_unlink_rq() function. Note that the pointer to the
|
|
|
|
* next run queue entry is neither checked nor updated.
|
2006-06-26 00:48:02 +00:00
|
|
|
*/
|
2009-03-07 16:25:21 +00:00
|
|
|
static inline struct task *__task_unlink_rq(struct task *t)
|
|
|
|
{
|
2018-05-28 11:51:06 +00:00
|
|
|
HA_ATOMIC_SUB(&tasks_run_queue, 1);
|
MAJOR: task: make use of the scope-aware ebtree functions
Currently the task scheduler suffers from an O(n) lookup when
skipping tasks that are not for the current thread. The reason
is that eb32_lookup_ge() has no information about the current
thread so it always revisits many tasks for other threads before
finding its own tasks.
This is particularly visible with HTTP/2 since the number of
concurrent streams created at once causes long series of tasks
for the same stream in the scheduler. With only 10 connections
and 100 streams each, by running on two threads, the performance
drops from 640kreq/s to 11.2kreq/s! Lookup metrics show that for
only 200000 task lookups, 430 million skips had to be performed,
which means that on average, each lookup leads to 2150 nodes to
be visited.
This commit backports the principle of scope lookups for ebtrees
from the ebtree_v7 development tree. The idea is that each node
contains a mask indicating the union of the scopes for the nodes
below it, which is fed during insertion, and used during lookups.
Then during lookups, branches that do not contain any leaf matching
the requested scope are simply ignored. This perfectly matches a
thread mask, allowing a thread to only extract the tasks it cares
about from the run queue, and to always find them in O(log(n))
instead of O(n). Thus the scheduler uses tid_bit and
task->thread_mask as the ebtree scope here.
Doing this has recovered most of the performance, as can be seen on
the test below with two threads, 10 connections, 100 streams each,
and 1 million requests total :
Before After Gain
test duration : 89.6s 4.73s x19
HTTP requests/s (DEBUG) : 11200 211300 x19
HTTP requests/s (PROD) : 15900 447000 x28
spin_lock time : 85.2s 0.46s /185
time per lookup : 13us 40ns /325
Even when going to 6 threads (on 3 hyperthreaded CPU cores), the
performance stays around 284000 req/s, showing that the contention
is much lower.
A test showed that there's no benefit in using this for the wait queue
though.
2017-11-05 12:34:20 +00:00
|
|
|
eb32sc_delete(&t->rq);
|
2018-07-26 14:19:58 +00:00
|
|
|
if (t->state & TASK_GLOBAL)
|
|
|
|
HA_ATOMIC_AND(&t->state, ~TASK_GLOBAL);
|
2009-03-07 16:25:21 +00:00
|
|
|
if (likely(t->nice))
|
2018-05-18 16:38:23 +00:00
|
|
|
HA_ATOMIC_SUB(&niced_tasks, 1);
|
2009-03-07 16:25:21 +00:00
|
|
|
return t;
|
|
|
|
}
|
|
|
|
|
2015-02-23 15:07:01 +00:00
|
|
|
/* This function unlinks task <t> from the run queue if it is in it. It also
|
|
|
|
* takes care of updating the next run queue task if it was this task.
|
|
|
|
*/
|
2009-03-07 16:25:21 +00:00
|
|
|
static inline struct task *task_unlink_rq(struct task *t)
|
2006-06-26 00:48:02 +00:00
|
|
|
{
|
2018-05-18 16:38:23 +00:00
|
|
|
if (t->thread_mask != tid_bit)
|
|
|
|
HA_SPIN_LOCK(TASK_RQ_LOCK, &rq_lock);
|
2018-05-04 13:46:16 +00:00
|
|
|
if (likely(task_in_rq(t))) {
|
|
|
|
if (&t->rq == rq_next)
|
|
|
|
rq_next = eb32sc_next(rq_next, tid_bit);
|
2009-03-07 16:25:21 +00:00
|
|
|
__task_unlink_rq(t);
|
2018-05-04 13:46:16 +00:00
|
|
|
}
|
2018-05-18 16:38:23 +00:00
|
|
|
if (t->thread_mask != tid_bit)
|
|
|
|
HA_SPIN_UNLOCK(TASK_RQ_LOCK, &rq_lock);
|
2008-07-05 16:16:19 +00:00
|
|
|
return t;
|
|
|
|
}
|
2008-06-24 06:17:16 +00:00
|
|
|
|
2018-05-18 16:45:28 +00:00
|
|
|
static inline void tasklet_wakeup(struct tasklet *tl)
|
|
|
|
{
|
2018-07-17 16:29:22 +00:00
|
|
|
if (!LIST_ISEMPTY(&tl->list))
|
|
|
|
return;
|
2018-05-18 16:45:28 +00:00
|
|
|
LIST_ADDQ(&task_list[tid], &tl->list);
|
|
|
|
task_list_size[tid]++;
|
|
|
|
HA_ATOMIC_ADD(&tasks_run_queue, 1);
|
|
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline void task_insert_into_tasklet_list(struct task *t)
|
|
|
|
{
|
|
|
|
struct tasklet *tl;
|
|
|
|
void *expected = NULL;
|
|
|
|
|
|
|
|
/* Protect ourself against anybody trying to insert the task into
|
|
|
|
* another runqueue. We set leaf_p to 0x1 to indicate that the node is
|
|
|
|
* not in a tree but that it's in the tasklet list. See task_in_rq().
|
|
|
|
*/
|
2018-06-01 12:32:39 +00:00
|
|
|
if (unlikely(!HA_ATOMIC_CAS(&t->rq.node.leaf_p, &expected, (void *)0x1)))
|
2018-05-18 16:45:28 +00:00
|
|
|
return;
|
2018-05-28 11:51:06 +00:00
|
|
|
HA_ATOMIC_ADD(&tasks_run_queue, 1);
|
2018-05-18 16:45:28 +00:00
|
|
|
task_list_size[tid]++;
|
|
|
|
tl = (struct tasklet *)t;
|
|
|
|
LIST_ADDQ(&task_list[tid], &tl->list);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline void task_remove_from_task_list(struct task *t)
|
|
|
|
{
|
|
|
|
LIST_DEL(&((struct tasklet *)t)->list);
|
2018-06-08 15:08:19 +00:00
|
|
|
LIST_INIT(&((struct tasklet *)t)->list);
|
2018-05-18 16:45:28 +00:00
|
|
|
task_list_size[tid]--;
|
|
|
|
HA_ATOMIC_SUB(&tasks_run_queue, 1);
|
|
|
|
if (!TASK_IS_TASKLET(t)) {
|
|
|
|
t->rq.node.leaf_p = NULL; // was 0x1
|
|
|
|
__ha_barrier_store();
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2008-07-05 16:16:19 +00:00
|
|
|
/*
|
|
|
|
* Unlinks the task and adjusts run queue stats.
|
|
|
|
* A pointer to the task itself is returned.
|
|
|
|
*/
|
|
|
|
static inline struct task *task_delete(struct task *t)
|
|
|
|
{
|
2009-03-07 16:25:21 +00:00
|
|
|
task_unlink_wq(t);
|
|
|
|
task_unlink_rq(t);
|
2008-06-24 06:17:16 +00:00
|
|
|
return t;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
2009-03-21 17:13:21 +00:00
|
|
|
* Initialize a new task. The bare minimum is performed (queue pointers and
|
|
|
|
* state). The task is returned. This function should not be used outside of
|
|
|
|
* task_new().
|
2008-06-24 06:17:16 +00:00
|
|
|
*/
|
2017-09-27 12:59:38 +00:00
|
|
|
static inline struct task *task_init(struct task *t, unsigned long thread_mask)
|
2008-06-24 06:17:16 +00:00
|
|
|
{
|
2009-03-07 16:25:21 +00:00
|
|
|
t->wq.node.leaf_p = NULL;
|
|
|
|
t->rq.node.leaf_p = NULL;
|
2018-05-18 16:38:23 +00:00
|
|
|
t->state = TASK_SLEEPING;
|
2017-10-31 15:06:06 +00:00
|
|
|
t->thread_mask = thread_mask;
|
2008-06-30 05:51:00 +00:00
|
|
|
t->nice = 0;
|
2009-03-28 16:54:35 +00:00
|
|
|
t->calls = 0;
|
2017-07-24 15:52:58 +00:00
|
|
|
t->expire = TICK_ETERNITY;
|
2006-06-26 00:48:02 +00:00
|
|
|
return t;
|
|
|
|
}
|
|
|
|
|
2018-05-18 16:45:28 +00:00
|
|
|
static inline void tasklet_init(struct tasklet *t)
|
|
|
|
{
|
|
|
|
t->nice = -32768;
|
|
|
|
t->calls = 0;
|
|
|
|
t->state = 0;
|
2018-07-19 14:02:16 +00:00
|
|
|
t->process = NULL;
|
2018-06-08 15:08:19 +00:00
|
|
|
LIST_INIT(&t->list);
|
2018-05-18 16:45:28 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
static inline struct tasklet *tasklet_new(void)
|
|
|
|
{
|
|
|
|
struct tasklet *t = pool_alloc(pool_head_tasklet);
|
|
|
|
|
|
|
|
if (t) {
|
|
|
|
tasklet_init(t);
|
|
|
|
}
|
|
|
|
return t;
|
|
|
|
}
|
|
|
|
|
2006-06-26 00:48:02 +00:00
|
|
|
/*
|
2009-03-21 17:13:21 +00:00
|
|
|
* Allocate and initialise a new task. The new task is returned, or NULL in
|
|
|
|
* case of lack of memory. The task count is incremented. Tasks should only
|
|
|
|
* be allocated this way, and must be freed using task_free().
|
|
|
|
*/
|
2017-09-27 12:59:38 +00:00
|
|
|
static inline struct task *task_new(unsigned long thread_mask)
|
2009-03-21 17:13:21 +00:00
|
|
|
{
|
2017-11-24 16:34:44 +00:00
|
|
|
struct task *t = pool_alloc(pool_head_task);
|
2009-03-21 17:13:21 +00:00
|
|
|
if (t) {
|
2017-09-27 12:59:38 +00:00
|
|
|
HA_ATOMIC_ADD(&nb_tasks, 1);
|
|
|
|
task_init(t, thread_mask);
|
2009-03-21 17:13:21 +00:00
|
|
|
}
|
|
|
|
return t;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Free a task. Its context must have been freed since it will be lost.
|
|
|
|
* The task count is decremented.
|
2006-06-26 00:48:02 +00:00
|
|
|
*/
|
2018-05-04 13:46:16 +00:00
|
|
|
static inline void __task_free(struct task *t)
|
2006-06-26 00:48:02 +00:00
|
|
|
{
|
2017-11-24 16:34:44 +00:00
|
|
|
pool_free(pool_head_task, t);
|
2014-11-13 15:57:19 +00:00
|
|
|
if (unlikely(stopping))
|
2017-11-24 16:34:44 +00:00
|
|
|
pool_flush(pool_head_task);
|
2017-09-27 12:59:38 +00:00
|
|
|
HA_ATOMIC_SUB(&nb_tasks, 1);
|
2006-06-26 00:48:02 +00:00
|
|
|
}
|
|
|
|
|
2018-05-04 13:46:16 +00:00
|
|
|
static inline void task_free(struct task *t)
|
|
|
|
{
|
|
|
|
/* There's no need to protect t->state with a lock, as the task
|
|
|
|
* has to run on the current thread.
|
|
|
|
*/
|
|
|
|
if (t == curr_task || !(t->state & TASK_RUNNING))
|
|
|
|
__task_free(t);
|
|
|
|
else
|
|
|
|
t->process = NULL;
|
|
|
|
}
|
|
|
|
|
2018-05-18 16:45:28 +00:00
|
|
|
static inline void tasklet_free(struct tasklet *tl)
|
|
|
|
{
|
2018-06-08 15:08:19 +00:00
|
|
|
LIST_DEL(&tl->list);
|
|
|
|
|
2018-05-18 16:45:28 +00:00
|
|
|
pool_free(pool_head_tasklet, tl);
|
|
|
|
if (unlikely(stopping))
|
|
|
|
pool_flush(pool_head_tasklet);
|
|
|
|
}
|
|
|
|
|
2009-03-07 16:25:21 +00:00
|
|
|
/* Place <task> into the wait queue, where it may already be. If the expiration
|
2009-03-08 15:35:27 +00:00
|
|
|
* timer is infinite, do nothing and rely on wake_expired_task to clean up.
|
2006-06-26 00:48:02 +00:00
|
|
|
*/
|
2009-03-08 15:35:27 +00:00
|
|
|
void __task_queue(struct task *task);
|
|
|
|
static inline void task_queue(struct task *task)
|
|
|
|
{
|
|
|
|
/* If we already have a place in the wait queue no later than the
|
|
|
|
* timeout we're trying to set, we'll stay there, because it is very
|
|
|
|
* unlikely that we will reach the timeout anyway. If the timeout
|
|
|
|
* has been disabled, it's useless to leave the queue as well. We'll
|
|
|
|
* rely on wake_expired_tasks() to catch the node and move it to the
|
|
|
|
* proper place should it ever happen. Finally we only add the task
|
|
|
|
* to the queue if it was not there or if it was further than what
|
|
|
|
* we want.
|
|
|
|
*/
|
|
|
|
if (!tick_isset(task->expire))
|
|
|
|
return;
|
|
|
|
|
2017-11-07 09:42:54 +00:00
|
|
|
HA_SPIN_LOCK(TASK_WQ_LOCK, &wq_lock);
|
2009-03-21 09:01:42 +00:00
|
|
|
if (!task_in_wq(task) || tick_is_lt(task->expire, task->wq.key))
|
2009-03-08 15:35:27 +00:00
|
|
|
__task_queue(task);
|
2017-11-07 09:42:54 +00:00
|
|
|
HA_SPIN_UNLOCK(TASK_WQ_LOCK, &wq_lock);
|
2009-03-08 15:35:27 +00:00
|
|
|
}
|
2006-06-26 00:48:02 +00:00
|
|
|
|
2011-07-25 12:30:42 +00:00
|
|
|
/* Ensure <task> will be woken up at most at <when>. If the task is already in
|
|
|
|
* the run queue (but not running), nothing is done. It may be used that way
|
|
|
|
* with a delay : task_schedule(task, tick_add(now_ms, delay));
|
|
|
|
*/
|
|
|
|
static inline void task_schedule(struct task *task, int when)
|
|
|
|
{
|
2017-09-27 12:59:38 +00:00
|
|
|
/* TODO: mthread, check if there is no tisk with this test */
|
2011-07-25 12:30:42 +00:00
|
|
|
if (task_in_rq(task))
|
|
|
|
return;
|
|
|
|
|
2017-11-07 09:42:54 +00:00
|
|
|
HA_SPIN_LOCK(TASK_WQ_LOCK, &wq_lock);
|
2011-07-25 12:30:42 +00:00
|
|
|
if (task_in_wq(task))
|
|
|
|
when = tick_first(when, task->expire);
|
|
|
|
|
|
|
|
task->expire = when;
|
|
|
|
if (!task_in_wq(task) || tick_is_lt(task->expire, task->wq.key))
|
|
|
|
__task_queue(task);
|
2017-11-07 09:42:54 +00:00
|
|
|
HA_SPIN_UNLOCK(TASK_WQ_LOCK, &wq_lock);
|
2011-07-25 12:30:42 +00:00
|
|
|
}
|
|
|
|
|
2017-07-12 12:31:10 +00:00
|
|
|
/* This function register a new signal. "lua" is the current lua
|
|
|
|
* execution context. It contains a pointer to the associated task.
|
|
|
|
* "link" is a list head attached to an other task that must be wake
|
|
|
|
* the lua task if an event occurs. This is useful with external
|
|
|
|
* events like TCP I/O or sleep functions. This funcion allocate
|
|
|
|
* memory for the signal.
|
|
|
|
*/
|
|
|
|
static inline struct notification *notification_new(struct list *purge, struct list *event, struct task *wakeup)
|
|
|
|
{
|
2017-11-24 16:34:44 +00:00
|
|
|
struct notification *com = pool_alloc(pool_head_notification);
|
2017-07-12 12:31:10 +00:00
|
|
|
if (!com)
|
|
|
|
return NULL;
|
|
|
|
LIST_ADDQ(purge, &com->purge_me);
|
|
|
|
LIST_ADDQ(event, &com->wake_me);
|
2017-11-07 09:42:54 +00:00
|
|
|
HA_SPIN_INIT(&com->lock);
|
2017-07-12 12:31:10 +00:00
|
|
|
com->task = wakeup;
|
|
|
|
return com;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* This function purge all the pending signals when the LUA execution
|
|
|
|
* is finished. This prevent than a coprocess try to wake a deleted
|
|
|
|
* task. This function remove the memory associated to the signal.
|
2017-12-10 16:14:07 +00:00
|
|
|
* The purge list is not locked because it is owned by only one
|
|
|
|
* process. before browsing this list, the caller must ensure to be
|
|
|
|
* the only one browser.
|
2017-07-12 12:31:10 +00:00
|
|
|
*/
|
|
|
|
static inline void notification_purge(struct list *purge)
|
|
|
|
{
|
|
|
|
struct notification *com, *back;
|
|
|
|
|
|
|
|
/* Delete all pending communication signals. */
|
|
|
|
list_for_each_entry_safe(com, back, purge, purge_me) {
|
2017-11-07 09:42:54 +00:00
|
|
|
HA_SPIN_LOCK(NOTIF_LOCK, &com->lock);
|
2017-07-12 12:31:10 +00:00
|
|
|
LIST_DEL(&com->purge_me);
|
2017-07-16 22:14:07 +00:00
|
|
|
if (!com->task) {
|
2017-11-07 09:42:54 +00:00
|
|
|
HA_SPIN_UNLOCK(NOTIF_LOCK, &com->lock);
|
2017-11-24 16:34:44 +00:00
|
|
|
pool_free(pool_head_notification, com);
|
2017-07-16 22:14:07 +00:00
|
|
|
continue;
|
|
|
|
}
|
|
|
|
com->task = NULL;
|
2017-11-07 09:42:54 +00:00
|
|
|
HA_SPIN_UNLOCK(NOTIF_LOCK, &com->lock);
|
2017-07-12 12:31:10 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2017-12-10 16:10:57 +00:00
|
|
|
/* In some cases, the disconnected notifications must be cleared.
|
|
|
|
* This function just release memory blocs. The purge list is not
|
|
|
|
* locked because it is owned by only one process. Before browsing
|
|
|
|
* this list, the caller must ensure to be the only one browser.
|
|
|
|
* The "com" is not locked because when com->task is NULL, the
|
|
|
|
* notification is no longer used.
|
|
|
|
*/
|
|
|
|
static inline void notification_gc(struct list *purge)
|
|
|
|
{
|
|
|
|
struct notification *com, *back;
|
|
|
|
|
|
|
|
/* Delete all pending communication signals. */
|
|
|
|
list_for_each_entry_safe (com, back, purge, purge_me) {
|
|
|
|
if (com->task)
|
|
|
|
continue;
|
|
|
|
LIST_DEL(&com->purge_me);
|
|
|
|
pool_free(pool_head_notification, com);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2017-07-12 12:31:10 +00:00
|
|
|
/* This function sends signals. It wakes all the tasks attached
|
|
|
|
* to a list head, and remove the signal, and free the used
|
2017-12-10 16:14:07 +00:00
|
|
|
* memory. The wake list is not locked because it is owned by
|
|
|
|
* only one process. before browsing this list, the caller must
|
|
|
|
* ensure to be the only one browser.
|
2017-07-12 12:31:10 +00:00
|
|
|
*/
|
|
|
|
static inline void notification_wake(struct list *wake)
|
|
|
|
{
|
|
|
|
struct notification *com, *back;
|
|
|
|
|
|
|
|
/* Wake task and delete all pending communication signals. */
|
|
|
|
list_for_each_entry_safe(com, back, wake, wake_me) {
|
2017-11-07 09:42:54 +00:00
|
|
|
HA_SPIN_LOCK(NOTIF_LOCK, &com->lock);
|
2017-07-12 12:31:10 +00:00
|
|
|
LIST_DEL(&com->wake_me);
|
2017-07-16 22:14:07 +00:00
|
|
|
if (!com->task) {
|
2017-11-07 09:42:54 +00:00
|
|
|
HA_SPIN_UNLOCK(NOTIF_LOCK, &com->lock);
|
2017-11-24 16:34:44 +00:00
|
|
|
pool_free(pool_head_notification, com);
|
2017-07-16 22:14:07 +00:00
|
|
|
continue;
|
|
|
|
}
|
2017-07-12 12:31:10 +00:00
|
|
|
task_wakeup(com->task, TASK_WOKEN_MSG);
|
2017-07-16 22:14:07 +00:00
|
|
|
com->task = NULL;
|
2017-11-07 09:42:54 +00:00
|
|
|
HA_SPIN_UNLOCK(NOTIF_LOCK, &com->lock);
|
2017-07-12 12:31:10 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2018-05-30 09:40:08 +00:00
|
|
|
/* This function returns true is some notification are pending
|
|
|
|
*/
|
|
|
|
static inline int notification_registered(struct list *wake)
|
|
|
|
{
|
|
|
|
return !LIST_ISEMPTY(wake);
|
|
|
|
}
|
|
|
|
|
2006-06-26 00:48:02 +00:00
|
|
|
/*
|
2011-07-25 14:33:49 +00:00
|
|
|
* This does 3 things :
|
2006-06-26 00:48:02 +00:00
|
|
|
* - wake up all expired tasks
|
|
|
|
* - call all runnable tasks
|
2007-05-12 20:35:00 +00:00
|
|
|
* - return the date of next event in <next> or eternity.
|
2006-06-26 00:48:02 +00:00
|
|
|
*/
|
|
|
|
|
2014-12-15 12:26:01 +00:00
|
|
|
void process_runnable_tasks();
|
2006-06-26 00:48:02 +00:00
|
|
|
|
2008-06-29 20:40:23 +00:00
|
|
|
/*
|
|
|
|
* Extract all expired timers from the timer queue, and wakes up all
|
|
|
|
* associated tasks. Returns the date of next event (or eternity).
|
|
|
|
*/
|
2014-12-15 12:26:01 +00:00
|
|
|
int wake_expired_tasks();
|
2008-06-29 20:40:23 +00:00
|
|
|
|
2009-03-08 14:53:06 +00:00
|
|
|
/* Perform minimal initializations, report 0 in case of error, 1 if OK. */
|
|
|
|
int init_task();
|
2006-06-26 00:48:02 +00:00
|
|
|
|
|
|
|
#endif /* _PROTO_TASK_H */
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Local variables:
|
|
|
|
* c-indent-level: 8
|
|
|
|
* c-basic-offset: 8
|
|
|
|
* End:
|
|
|
|
*/
|