/* * include/proto/session.h * This file defines everything related to sessions. * * Copyright (C) 2000-2010 Willy Tarreau - w@1wt.eu * * This library is free software; you can redistribute it and/or * modify it under the terms of the GNU Lesser General Public * License as published by the Free Software Foundation, version 2.1 * exclusively. * * This library is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU * Lesser General Public License for more details. * * You should have received a copy of the GNU Lesser General Public * License along with this library; if not, write to the Free Software * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA */ #ifndef _PROTO_SESSION_H #define _PROTO_SESSION_H #include #include #include #include #include #include #include extern struct pool_head *pool2_session; extern struct list sessions; extern struct list buffer_wq; extern struct data_cb sess_conn_cb; int session_accept(struct listener *l, int cfd, struct sockaddr_storage *addr); /* perform minimal intializations, report 0 in case of error, 1 if OK. */ int init_session(); /* kill a session and set the termination flags to (one of SN_ERR_*) */ void session_shutdown(struct session *session, int why); void session_process_counters(struct session *s); void sess_change_server(struct session *sess, struct server *newsrv); struct task *process_session(struct task *t); void default_srv_error(struct session *s, struct stream_interface *si); struct stkctr *smp_fetch_sc_stkctr(struct session *l4, const struct arg *args, const char *kw); int parse_track_counters(char **args, int *arg, int section_type, struct proxy *curpx, struct track_ctr_prm *prm, struct proxy *defpx, char **err); /* Update the session's backend and server time stats */ void session_update_time_stats(struct session *s); void __session_offer_buffers(int rqlimit); static inline void session_offer_buffers(); int session_alloc_work_buffer(struct session *s); void session_release_buffers(struct session *s); int session_alloc_recv_buffer(struct session *s, struct buffer **buf); /* sets the stick counter's entry pointer */ static inline void stkctr_set_entry(struct stkctr *stkctr, struct stksess *entry) { stkctr->entry = caddr_from_ptr(entry, 0); } /* returns the entry pointer from a stick counter */ static inline struct stksess *stkctr_entry(struct stkctr *stkctr) { return caddr_to_ptr(stkctr->entry); } /* returns the two flags from a stick counter */ static inline unsigned int stkctr_flags(struct stkctr *stkctr) { return caddr_to_data(stkctr->entry); } /* sets up to two flags at a time on a composite address */ static inline void stkctr_set_flags(struct stkctr *stkctr, unsigned int flags) { stkctr->entry = caddr_set_flags(stkctr->entry, flags); } /* returns the two flags from a stick counter */ static inline void stkctr_clr_flags(struct stkctr *stkctr, unsigned int flags) { stkctr->entry = caddr_clr_flags(stkctr->entry, flags); } /* Remove the refcount from the session to the tracked counters, and clear the * pointer to ensure this is only performed once. The caller is responsible for * ensuring that the pointer is valid first. */ static inline void session_store_counters(struct session *s) { void *ptr; int i; for (i = 0; i < MAX_SESS_STKCTR; i++) { if (!stkctr_entry(&s->stkctr[i])) continue; ptr = stktable_data_ptr(s->stkctr[i].table, stkctr_entry(&s->stkctr[i]), STKTABLE_DT_CONN_CUR); if (ptr) stktable_data_cast(ptr, conn_cur)--; stkctr_entry(&s->stkctr[i])->ref_cnt--; stksess_kill_if_expired(s->stkctr[i].table, stkctr_entry(&s->stkctr[i])); stkctr_set_entry(&s->stkctr[i], NULL); } } /* Remove the refcount from the session counters tracked at the content level if * any, and clear the pointer to ensure this is only performed once. The caller * is responsible for ensuring that the pointer is valid first. */ static inline void session_stop_content_counters(struct session *s) { void *ptr; int i; for (i = 0; i < MAX_SESS_STKCTR; i++) { if (!stkctr_entry(&s->stkctr[i])) continue; if (!(stkctr_flags(&s->stkctr[i]) & STKCTR_TRACK_CONTENT)) continue; ptr = stktable_data_ptr(s->stkctr[i].table, stkctr_entry(&s->stkctr[i]), STKTABLE_DT_CONN_CUR); if (ptr) stktable_data_cast(ptr, conn_cur)--; stkctr_entry(&s->stkctr[i])->ref_cnt--; stksess_kill_if_expired(s->stkctr[i].table, stkctr_entry(&s->stkctr[i])); stkctr_set_entry(&s->stkctr[i], NULL); } } /* Increase total and concurrent connection count for stick entry of table * . The caller is responsible for ensuring that and are valid * pointers, and for calling this only once per connection. */ static inline void session_start_counters(struct stktable *t, struct stksess *ts) { void *ptr; ptr = stktable_data_ptr(t, ts, STKTABLE_DT_CONN_CUR); if (ptr) stktable_data_cast(ptr, conn_cur)++; ptr = stktable_data_ptr(t, ts, STKTABLE_DT_CONN_CNT); if (ptr) stktable_data_cast(ptr, conn_cnt)++; ptr = stktable_data_ptr(t, ts, STKTABLE_DT_CONN_RATE); if (ptr) update_freq_ctr_period(&stktable_data_cast(ptr, conn_rate), t->data_arg[STKTABLE_DT_CONN_RATE].u, 1); if (tick_isset(t->expire)) ts->expire = tick_add(now_ms, MS_TO_TICKS(t->expire)); } /* Enable tracking of session counters as on stksess . The caller is * responsible for ensuring that and are valid pointers. Some controls * are performed to ensure the state can still change. */ static inline void session_track_stkctr(struct stkctr *ctr, struct stktable *t, struct stksess *ts) { if (stkctr_entry(ctr)) return; ts->ref_cnt++; ctr->table = t; stkctr_set_entry(ctr, ts); session_start_counters(t, ts); } /* Increase the number of cumulated HTTP requests in the tracked counters */ static void inline session_inc_http_req_ctr(struct session *s) { void *ptr; int i; for (i = 0; i < MAX_SESS_STKCTR; i++) { if (!stkctr_entry(&s->stkctr[i])) continue; ptr = stktable_data_ptr(s->stkctr[i].table, stkctr_entry(&s->stkctr[i]), STKTABLE_DT_HTTP_REQ_CNT); if (ptr) stktable_data_cast(ptr, http_req_cnt)++; ptr = stktable_data_ptr(s->stkctr[i].table, stkctr_entry(&s->stkctr[i]), STKTABLE_DT_HTTP_REQ_RATE); if (ptr) update_freq_ctr_period(&stktable_data_cast(ptr, http_req_rate), s->stkctr[i].table->data_arg[STKTABLE_DT_HTTP_REQ_RATE].u, 1); } } /* Increase the number of cumulated HTTP requests in the backend's tracked counters */ static void inline session_inc_be_http_req_ctr(struct session *s) { void *ptr; int i; for (i = 0; i < MAX_SESS_STKCTR; i++) { if (!stkctr_entry(&s->stkctr[i])) continue; if (!(stkctr_flags(&s->stkctr[i]) & STKCTR_TRACK_BACKEND)) continue; ptr = stktable_data_ptr(s->stkctr[i].table, stkctr_entry(&s->stkctr[i]), STKTABLE_DT_HTTP_REQ_CNT); if (ptr) stktable_data_cast(ptr, http_req_cnt)++; ptr = stktable_data_ptr(s->stkctr[i].table, stkctr_entry(&s->stkctr[i]), STKTABLE_DT_HTTP_REQ_RATE); if (ptr) update_freq_ctr_period(&stktable_data_cast(ptr, http_req_rate), s->stkctr[i].table->data_arg[STKTABLE_DT_HTTP_REQ_RATE].u, 1); } } /* Increase the number of cumulated failed HTTP requests in the tracked * counters. Only 4xx requests should be counted here so that we can * distinguish between errors caused by client behaviour and other ones. * Note that even 404 are interesting because they're generally caused by * vulnerability scans. */ static void inline session_inc_http_err_ctr(struct session *s) { void *ptr; int i; for (i = 0; i < MAX_SESS_STKCTR; i++) { if (!stkctr_entry(&s->stkctr[i])) continue; ptr = stktable_data_ptr(s->stkctr[i].table, stkctr_entry(&s->stkctr[i]), STKTABLE_DT_HTTP_ERR_CNT); if (ptr) stktable_data_cast(ptr, http_err_cnt)++; ptr = stktable_data_ptr(s->stkctr[i].table, stkctr_entry(&s->stkctr[i]), STKTABLE_DT_HTTP_ERR_RATE); if (ptr) update_freq_ctr_period(&stktable_data_cast(ptr, http_err_rate), s->stkctr[i].table->data_arg[STKTABLE_DT_HTTP_ERR_RATE].u, 1); } } static void inline session_add_srv_conn(struct session *sess, struct server *srv) { sess->srv_conn = srv; LIST_ADD(&srv->actconns, &sess->by_srv); } static void inline session_del_srv_conn(struct session *sess) { if (!sess->srv_conn) return; sess->srv_conn = NULL; LIST_DEL(&sess->by_srv); } static void inline session_init_srv_conn(struct session *sess) { sess->srv_conn = NULL; LIST_INIT(&sess->by_srv); } static inline void session_offer_buffers() { int avail; if (LIST_ISEMPTY(&buffer_wq)) return; /* all sessions will need 1 buffer, so we can stop waking up sessions * once we have enough of them to eat all the buffers. Note that we * don't really know if they are sessions or just other tasks, but * that's a rough estimate. Similarly, for each cached event we'll need * 1 buffer. If no buffer is currently used, always wake up the number * of tasks we can offer a buffer based on what is allocated, and in * any case at least one task per two reserved buffers. */ avail = pool2_buffer->allocated - pool2_buffer->used - global.tune.reserved_bufs / 2; if (avail > (int)run_queue) __session_offer_buffers(avail); } #endif /* _PROTO_SESSION_H */ /* * Local variables: * c-indent-level: 8 * c-basic-offset: 8 * End: */