1
0
mirror of http://git.haproxy.org/git/haproxy.git/ synced 2025-04-01 22:48:25 +00:00

[MEDIUM] move the HTTP request body analyser out of process_request().

A new function http_process_request_body() has been created to process
the request body. Next step is now to clean up process_request().
This commit is contained in:
Willy Tarreau 2008-11-30 23:36:37 +01:00
parent 60b85b0694
commit d34af78a34
3 changed files with 90 additions and 76 deletions

View File

@ -64,6 +64,7 @@ int process_srv_data(struct session *t);
int process_srv_conn(struct session *t);
int process_request(struct session *t);
int http_process_tarpit(struct session *s, struct buffer *req);
int http_process_request_body(struct session *s, struct buffer *req);
int process_response(struct session *t);
int produce_content(struct session *s);

View File

@ -2359,82 +2359,6 @@ int process_request(struct session *t)
; // to keep gcc happy
}
if (req->analysers & AN_REQ_HTTP_BODY) {
/* We have to parse the HTTP request body to find any required data.
* "balance url_param check_post" should have been the only way to get
* into this. We were brought here after HTTP header analysis, so all
* related structures are ready.
*/
struct http_msg *msg = &t->txn.req;
unsigned long body = msg->sol[msg->eoh] == '\r' ? msg->eoh + 2 : msg->eoh + 1;
long long limit = t->be->url_param_post_limit;
struct hdr_ctx ctx;
ctx.idx = 0;
/* now if we have a length, we'll take the hint */
http_find_header2("Transfer-Encoding", 17, msg->sol, &t->txn.hdr_idx, &ctx);
if (ctx.idx && ctx.vlen >= 7 && strncasecmp(ctx.line+ctx.val, "chunked", 7) == 0) {
unsigned int chunk = 0;
while (body < req->l && !HTTP_IS_CRLF(msg->sol[body])) {
char c = msg->sol[body];
if (ishex(c)) {
unsigned int hex = toupper(c) - '0';
if (hex > 9)
hex -= 'A' - '9' - 1;
chunk = (chunk << 4) | hex;
} else
break;
body++;
}
if (body + 2 >= req->l) /* we want CRLF too */
goto http_body_end; /* end of buffer? data missing! */
if (memcmp(msg->sol+body, "\r\n", 2) != 0)
goto http_body_end; /* chunked encoding len ends with CRLF, and we don't have it yet */
body += 2; // skip CRLF
/* if we support more then one chunk here, we have to do it again when assigning server
* 1. how much entity data do we have? new var
* 2. should save entity_start, entity_cursor, elen & rlen in req; so we don't repeat scanning here
* 3. test if elen > limit, or set new limit to elen if 0 (end of entity found)
*/
if (chunk < limit)
limit = chunk; /* only reading one chunk */
} else {
if (msg->hdr_content_len < limit)
limit = msg->hdr_content_len;
}
http_body_end:
/* we leave once we know we have nothing left to do. This means that we have
* enough bytes, or that we know we'll not get any more (buffer full, read
* buffer closed).
*/
if (req->l - body >= limit || /* enough bytes! */
req->flags & (BF_FULL | BF_READ_ERROR | BF_SHUTR | BF_READ_TIMEOUT) ||
tick_is_expired(req->analyse_exp, now_ms)) {
/* The situation will not evolve, so let's give up on the analysis. */
t->logs.tv_request = now; /* update the request timer to reflect full request */
req->analysers &= ~AN_REQ_HTTP_BODY;
req->analyse_exp = TICK_ETERNITY;
}
else {
/* Not enough data. We'll re-use the http-request
* timeout here. Ideally, we should set the timeout
* relative to the accept() date. We just set the
* request timeout once at the beginning of the
* request.
*/
buffer_write_dis(req);
if (!tick_isset(req->analyse_exp))
req->analyse_exp = tick_add_ifset(now_ms, t->fe->timeout.httpreq);
return 0;
}
}
/* Note: eventhough nobody should set an unknown flag, clearing them right now will
* probably reduce one day's debugging session.
*/
@ -2489,6 +2413,91 @@ int http_process_tarpit(struct session *s, struct buffer *req)
return 0;
}
/* This function is an analyser which processes the HTTP request body. It looks
* for parameters to be used for the load balancing algorithm (url_param). It
* must only be called after the standard HTTP request processing has occurred,
* because it expects the request to be parsed. It returns zero if it needs to
* read more data, or 1 once it has completed its analysis.
*/
int http_process_request_body(struct session *s, struct buffer *req)
{
struct http_msg *msg = &s->txn.req;
unsigned long body = msg->sol[msg->eoh] == '\r' ? msg->eoh + 2 : msg->eoh + 1;
long long limit = s->be->url_param_post_limit;
struct hdr_ctx ctx;
/* We have to parse the HTTP request body to find any required data.
* "balance url_param check_post" should have been the only way to get
* into this. We were brought here after HTTP header analysis, so all
* related structures are ready.
*/
ctx.idx = 0;
/* now if we have a length, we'll take the hint */
http_find_header2("Transfer-Encoding", 17, msg->sol, &s->txn.hdr_idx, &ctx);
if (ctx.idx && ctx.vlen >= 7 && strncasecmp(ctx.line+ctx.val, "chunked", 7) == 0) {
unsigned int chunk = 0;
while (body < req->l && !HTTP_IS_CRLF(msg->sol[body])) {
char c = msg->sol[body];
if (ishex(c)) {
unsigned int hex = toupper(c) - '0';
if (hex > 9)
hex -= 'A' - '9' - 1;
chunk = (chunk << 4) | hex;
} else
break;
body++;
}
if (body + 2 >= req->l) /* we want CRLF too */
goto http_body_end; /* end of buffer? data missing! */
if (memcmp(msg->sol+body, "\r\n", 2) != 0)
goto http_body_end; /* chunked encoding len ends with CRLF, and we don't have it yet */
body += 2; // skip CRLF
/* if we support more then one chunk here, we have to do it again when assigning server
* 1. how much entity data do we have? new var
* 2. should save entity_start, entity_cursor, elen & rlen in req; so we don't repeat scanning here
* 3. test if elen > limit, or set new limit to elen if 0 (end of entity found)
*/
if (chunk < limit)
limit = chunk; /* only reading one chunk */
} else {
if (msg->hdr_content_len < limit)
limit = msg->hdr_content_len;
}
http_body_end:
/* we leave once we know we have nothing left to do. This means that we have
* enough bytes, or that we know we'll not get any more (buffer full, read
* buffer closed).
*/
if (req->l - body >= limit || /* enough bytes! */
req->flags & (BF_FULL | BF_READ_ERROR | BF_SHUTR | BF_READ_TIMEOUT) ||
tick_is_expired(req->analyse_exp, now_ms)) {
/* The situation will not evolve, so let's give up on the analysis. */
s->logs.tv_request = now; /* update the request timer to reflect full request */
req->analysers &= ~AN_REQ_HTTP_BODY;
req->analyse_exp = TICK_ETERNITY;
return 1;
}
else {
/* Not enough data. We'll re-use the http-request
* timeout here. Ideally, we should set the timeout
* relative to the accept() date. We just set the
* request timeout once at the beginning of the
* request.
*/
buffer_write_dis(req);
if (!tick_isset(req->analyse_exp))
req->analyse_exp = tick_add_ifset(now_ms, s->fe->timeout.httpreq);
return 0;
}
}
/* This function performs all the processing enabled for the current response.
* It normally returns zero, but may return 1 if it absolutely needs to be
* called again after other functions. It relies on buffers flags, and updates

View File

@ -713,6 +713,10 @@ resync_stream_interface:
if (!http_process_tarpit(s, s->req))
break;
if (s->req->analysers & AN_REQ_HTTP_BODY)
if (!http_process_request_body(s, s->req))
break;
/* Just make sure that nobody set a wrong flag causing an endless loop */
s->req->analysers &= AN_REQ_INSPECT | AN_REQ_HTTP_HDR | AN_REQ_HTTP_TARPIT | AN_REQ_HTTP_BODY;