mirror of
http://git.haproxy.org/git/haproxy.git/
synced 2024-12-11 14:05:12 +00:00
3c23a85550
Since commit 3dd6a25
("MINOR: stream-int: retrieve session pointer from
stream-int"), we can get the session from the task, so let's get rid of
this less obvious function.
301 lines
9.4 KiB
C
301 lines
9.4 KiB
C
/*
|
|
* include/proto/session.h
|
|
* This file defines everything related to sessions.
|
|
*
|
|
* Copyright (C) 2000-2010 Willy Tarreau - w@1wt.eu
|
|
*
|
|
* This library is free software; you can redistribute it and/or
|
|
* modify it under the terms of the GNU Lesser General Public
|
|
* License as published by the Free Software Foundation, version 2.1
|
|
* exclusively.
|
|
*
|
|
* This library is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
* Lesser General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU Lesser General Public
|
|
* License along with this library; if not, write to the Free Software
|
|
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
|
|
*/
|
|
|
|
#ifndef _PROTO_SESSION_H
|
|
#define _PROTO_SESSION_H
|
|
|
|
#include <common/config.h>
|
|
#include <common/memory.h>
|
|
#include <types/session.h>
|
|
#include <proto/fd.h>
|
|
#include <proto/freq_ctr.h>
|
|
#include <proto/stick_table.h>
|
|
#include <proto/task.h>
|
|
|
|
extern struct pool_head *pool2_session;
|
|
extern struct list sessions;
|
|
extern struct list buffer_wq;
|
|
|
|
extern struct data_cb sess_conn_cb;
|
|
|
|
int session_accept(struct listener *l, int cfd, struct sockaddr_storage *addr);
|
|
|
|
/* perform minimal intializations, report 0 in case of error, 1 if OK. */
|
|
int init_session();
|
|
|
|
/* kill a session and set the termination flags to <why> (one of SN_ERR_*) */
|
|
void session_shutdown(struct session *session, int why);
|
|
|
|
void session_process_counters(struct session *s);
|
|
void sess_change_server(struct session *sess, struct server *newsrv);
|
|
struct task *process_session(struct task *t);
|
|
void default_srv_error(struct session *s, struct stream_interface *si);
|
|
struct stkctr *smp_fetch_sc_stkctr(struct session *l4, const struct arg *args, const char *kw);
|
|
int parse_track_counters(char **args, int *arg,
|
|
int section_type, struct proxy *curpx,
|
|
struct track_ctr_prm *prm,
|
|
struct proxy *defpx, char **err);
|
|
|
|
/* Update the session's backend and server time stats */
|
|
void session_update_time_stats(struct session *s);
|
|
void __session_offer_buffers(int rqlimit);
|
|
static inline void session_offer_buffers();
|
|
int session_alloc_work_buffer(struct session *s);
|
|
void session_release_buffers(struct session *s);
|
|
int session_alloc_recv_buffer(struct session *s, struct buffer **buf);
|
|
|
|
/* sets the stick counter's entry pointer */
|
|
static inline void stkctr_set_entry(struct stkctr *stkctr, struct stksess *entry)
|
|
{
|
|
stkctr->entry = caddr_from_ptr(entry, 0);
|
|
}
|
|
|
|
/* returns the entry pointer from a stick counter */
|
|
static inline struct stksess *stkctr_entry(struct stkctr *stkctr)
|
|
{
|
|
return caddr_to_ptr(stkctr->entry);
|
|
}
|
|
|
|
/* returns the two flags from a stick counter */
|
|
static inline unsigned int stkctr_flags(struct stkctr *stkctr)
|
|
{
|
|
return caddr_to_data(stkctr->entry);
|
|
}
|
|
|
|
/* sets up to two flags at a time on a composite address */
|
|
static inline void stkctr_set_flags(struct stkctr *stkctr, unsigned int flags)
|
|
{
|
|
stkctr->entry = caddr_set_flags(stkctr->entry, flags);
|
|
}
|
|
|
|
/* returns the two flags from a stick counter */
|
|
static inline void stkctr_clr_flags(struct stkctr *stkctr, unsigned int flags)
|
|
{
|
|
stkctr->entry = caddr_clr_flags(stkctr->entry, flags);
|
|
}
|
|
|
|
/* Remove the refcount from the session to the tracked counters, and clear the
|
|
* pointer to ensure this is only performed once. The caller is responsible for
|
|
* ensuring that the pointer is valid first.
|
|
*/
|
|
static inline void session_store_counters(struct session *s)
|
|
{
|
|
void *ptr;
|
|
int i;
|
|
|
|
for (i = 0; i < MAX_SESS_STKCTR; i++) {
|
|
if (!stkctr_entry(&s->stkctr[i]))
|
|
continue;
|
|
ptr = stktable_data_ptr(s->stkctr[i].table, stkctr_entry(&s->stkctr[i]), STKTABLE_DT_CONN_CUR);
|
|
if (ptr)
|
|
stktable_data_cast(ptr, conn_cur)--;
|
|
stkctr_entry(&s->stkctr[i])->ref_cnt--;
|
|
stksess_kill_if_expired(s->stkctr[i].table, stkctr_entry(&s->stkctr[i]));
|
|
stkctr_set_entry(&s->stkctr[i], NULL);
|
|
}
|
|
}
|
|
|
|
/* Remove the refcount from the session counters tracked at the content level if
|
|
* any, and clear the pointer to ensure this is only performed once. The caller
|
|
* is responsible for ensuring that the pointer is valid first.
|
|
*/
|
|
static inline void session_stop_content_counters(struct session *s)
|
|
{
|
|
void *ptr;
|
|
int i;
|
|
|
|
for (i = 0; i < MAX_SESS_STKCTR; i++) {
|
|
if (!stkctr_entry(&s->stkctr[i]))
|
|
continue;
|
|
|
|
if (!(stkctr_flags(&s->stkctr[i]) & STKCTR_TRACK_CONTENT))
|
|
continue;
|
|
|
|
ptr = stktable_data_ptr(s->stkctr[i].table, stkctr_entry(&s->stkctr[i]), STKTABLE_DT_CONN_CUR);
|
|
if (ptr)
|
|
stktable_data_cast(ptr, conn_cur)--;
|
|
stkctr_entry(&s->stkctr[i])->ref_cnt--;
|
|
stksess_kill_if_expired(s->stkctr[i].table, stkctr_entry(&s->stkctr[i]));
|
|
stkctr_set_entry(&s->stkctr[i], NULL);
|
|
}
|
|
}
|
|
|
|
/* Increase total and concurrent connection count for stick entry <ts> of table
|
|
* <t>. The caller is responsible for ensuring that <t> and <ts> are valid
|
|
* pointers, and for calling this only once per connection.
|
|
*/
|
|
static inline void session_start_counters(struct stktable *t, struct stksess *ts)
|
|
{
|
|
void *ptr;
|
|
|
|
ptr = stktable_data_ptr(t, ts, STKTABLE_DT_CONN_CUR);
|
|
if (ptr)
|
|
stktable_data_cast(ptr, conn_cur)++;
|
|
|
|
ptr = stktable_data_ptr(t, ts, STKTABLE_DT_CONN_CNT);
|
|
if (ptr)
|
|
stktable_data_cast(ptr, conn_cnt)++;
|
|
|
|
ptr = stktable_data_ptr(t, ts, STKTABLE_DT_CONN_RATE);
|
|
if (ptr)
|
|
update_freq_ctr_period(&stktable_data_cast(ptr, conn_rate),
|
|
t->data_arg[STKTABLE_DT_CONN_RATE].u, 1);
|
|
if (tick_isset(t->expire))
|
|
ts->expire = tick_add(now_ms, MS_TO_TICKS(t->expire));
|
|
}
|
|
|
|
/* Enable tracking of session counters as <stkctr> on stksess <ts>. The caller is
|
|
* responsible for ensuring that <t> and <ts> are valid pointers. Some controls
|
|
* are performed to ensure the state can still change.
|
|
*/
|
|
static inline void session_track_stkctr(struct stkctr *ctr, struct stktable *t, struct stksess *ts)
|
|
{
|
|
if (stkctr_entry(ctr))
|
|
return;
|
|
|
|
ts->ref_cnt++;
|
|
ctr->table = t;
|
|
stkctr_set_entry(ctr, ts);
|
|
session_start_counters(t, ts);
|
|
}
|
|
|
|
/* Increase the number of cumulated HTTP requests in the tracked counters */
|
|
static void inline session_inc_http_req_ctr(struct session *s)
|
|
{
|
|
void *ptr;
|
|
int i;
|
|
|
|
for (i = 0; i < MAX_SESS_STKCTR; i++) {
|
|
if (!stkctr_entry(&s->stkctr[i]))
|
|
continue;
|
|
|
|
ptr = stktable_data_ptr(s->stkctr[i].table, stkctr_entry(&s->stkctr[i]), STKTABLE_DT_HTTP_REQ_CNT);
|
|
if (ptr)
|
|
stktable_data_cast(ptr, http_req_cnt)++;
|
|
|
|
ptr = stktable_data_ptr(s->stkctr[i].table, stkctr_entry(&s->stkctr[i]), STKTABLE_DT_HTTP_REQ_RATE);
|
|
if (ptr)
|
|
update_freq_ctr_period(&stktable_data_cast(ptr, http_req_rate),
|
|
s->stkctr[i].table->data_arg[STKTABLE_DT_HTTP_REQ_RATE].u, 1);
|
|
}
|
|
}
|
|
|
|
/* Increase the number of cumulated HTTP requests in the backend's tracked counters */
|
|
static void inline session_inc_be_http_req_ctr(struct session *s)
|
|
{
|
|
void *ptr;
|
|
int i;
|
|
|
|
for (i = 0; i < MAX_SESS_STKCTR; i++) {
|
|
if (!stkctr_entry(&s->stkctr[i]))
|
|
continue;
|
|
|
|
if (!(stkctr_flags(&s->stkctr[i]) & STKCTR_TRACK_BACKEND))
|
|
continue;
|
|
|
|
ptr = stktable_data_ptr(s->stkctr[i].table, stkctr_entry(&s->stkctr[i]), STKTABLE_DT_HTTP_REQ_CNT);
|
|
if (ptr)
|
|
stktable_data_cast(ptr, http_req_cnt)++;
|
|
|
|
ptr = stktable_data_ptr(s->stkctr[i].table, stkctr_entry(&s->stkctr[i]), STKTABLE_DT_HTTP_REQ_RATE);
|
|
if (ptr)
|
|
update_freq_ctr_period(&stktable_data_cast(ptr, http_req_rate),
|
|
s->stkctr[i].table->data_arg[STKTABLE_DT_HTTP_REQ_RATE].u, 1);
|
|
}
|
|
}
|
|
|
|
/* Increase the number of cumulated failed HTTP requests in the tracked
|
|
* counters. Only 4xx requests should be counted here so that we can
|
|
* distinguish between errors caused by client behaviour and other ones.
|
|
* Note that even 404 are interesting because they're generally caused by
|
|
* vulnerability scans.
|
|
*/
|
|
static void inline session_inc_http_err_ctr(struct session *s)
|
|
{
|
|
void *ptr;
|
|
int i;
|
|
|
|
for (i = 0; i < MAX_SESS_STKCTR; i++) {
|
|
if (!stkctr_entry(&s->stkctr[i]))
|
|
continue;
|
|
|
|
ptr = stktable_data_ptr(s->stkctr[i].table, stkctr_entry(&s->stkctr[i]), STKTABLE_DT_HTTP_ERR_CNT);
|
|
if (ptr)
|
|
stktable_data_cast(ptr, http_err_cnt)++;
|
|
|
|
ptr = stktable_data_ptr(s->stkctr[i].table, stkctr_entry(&s->stkctr[i]), STKTABLE_DT_HTTP_ERR_RATE);
|
|
if (ptr)
|
|
update_freq_ctr_period(&stktable_data_cast(ptr, http_err_rate),
|
|
s->stkctr[i].table->data_arg[STKTABLE_DT_HTTP_ERR_RATE].u, 1);
|
|
}
|
|
}
|
|
|
|
static void inline session_add_srv_conn(struct session *sess, struct server *srv)
|
|
{
|
|
sess->srv_conn = srv;
|
|
LIST_ADD(&srv->actconns, &sess->by_srv);
|
|
}
|
|
|
|
static void inline session_del_srv_conn(struct session *sess)
|
|
{
|
|
if (!sess->srv_conn)
|
|
return;
|
|
|
|
sess->srv_conn = NULL;
|
|
LIST_DEL(&sess->by_srv);
|
|
}
|
|
|
|
static void inline session_init_srv_conn(struct session *sess)
|
|
{
|
|
sess->srv_conn = NULL;
|
|
LIST_INIT(&sess->by_srv);
|
|
}
|
|
|
|
static inline void session_offer_buffers()
|
|
{
|
|
int avail;
|
|
|
|
if (LIST_ISEMPTY(&buffer_wq))
|
|
return;
|
|
|
|
/* all sessions will need 1 buffer, so we can stop waking up sessions
|
|
* once we have enough of them to eat all the buffers. Note that we
|
|
* don't really know if they are sessions or just other tasks, but
|
|
* that's a rough estimate. Similarly, for each cached event we'll need
|
|
* 1 buffer. If no buffer is currently used, always wake up the number
|
|
* of tasks we can offer a buffer based on what is allocated, and in
|
|
* any case at least one task per two reserved buffers.
|
|
*/
|
|
avail = pool2_buffer->allocated - pool2_buffer->used - global.tune.reserved_bufs / 2;
|
|
|
|
if (avail > (int)run_queue)
|
|
__session_offer_buffers(avail);
|
|
}
|
|
|
|
#endif /* _PROTO_SESSION_H */
|
|
|
|
/*
|
|
* Local variables:
|
|
* c-indent-level: 8
|
|
* c-basic-offset: 8
|
|
* End:
|
|
*/
|