mirror of
http://git.haproxy.org/git/haproxy.git/
synced 2024-12-25 14:12:13 +00:00
b88d231773
The new functions are : __b_put_varint() : inserts a varint when it's known that it fits b_put_varint() : tries to insert a varint at the tail b_get_varint() : tries to get a varint from the head b_peek_varint() : tries to peek a varint at a specific offset Wrapping is supported so that they are expected to be safe to use to manipulate varints with buffers anywhere.
1018 lines
28 KiB
C
1018 lines
28 KiB
C
/*
|
|
* include/common/buf.h
|
|
* Simple buffer handling.
|
|
*
|
|
* Copyright (C) 2000-2018 Willy Tarreau - w@1wt.eu
|
|
*
|
|
* Permission is hereby granted, free of charge, to any person obtaining
|
|
* a copy of this software and associated documentation files (the
|
|
* "Software"), to deal in the Software without restriction, including
|
|
* without limitation the rights to use, copy, modify, merge, publish,
|
|
* distribute, sublicense, and/or sell copies of the Software, and to
|
|
* permit persons to whom the Software is furnished to do so, subject to
|
|
* the following conditions:
|
|
*
|
|
* The above copyright notice and this permission notice shall be
|
|
* included in all copies or substantial portions of the Software.
|
|
*
|
|
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
|
|
* EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES
|
|
* OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
|
|
* NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT
|
|
* HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY,
|
|
* WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
|
|
* FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR
|
|
* OTHER DEALINGS IN THE SOFTWARE.
|
|
*/
|
|
|
|
#ifndef _COMMON_BUF_H
|
|
#define _COMMON_BUF_H
|
|
|
|
#include <inttypes.h>
|
|
#include <string.h>
|
|
#include <unistd.h>
|
|
|
|
#include <common/debug.h>
|
|
|
|
/* Structure defining a buffer's head */
|
|
struct buffer {
|
|
size_t size; /* buffer size in bytes */
|
|
char *area; /* points to <size> bytes */
|
|
size_t data; /* amount of data after head including wrapping */
|
|
size_t head; /* start offset of remaining data relative to area */
|
|
};
|
|
|
|
/* A buffer may be in 3 different states :
|
|
* - unallocated : size == 0, area == 0 (b_is_null() is true)
|
|
* - waiting : size == 0, area != 0 (b_is_null() is true)
|
|
* - allocated : size > 0, area > 0 (b_is_null() is false)
|
|
*/
|
|
|
|
/* initializers for certain buffer states. It is important that the NULL buffer
|
|
* remains the one with all fields initialized to zero so that a calloc() or a
|
|
* memset() on a struct automatically sets a NULL buffer.
|
|
*/
|
|
#define BUF_NULL ((struct buffer){ })
|
|
#define BUF_WANTED ((struct buffer){ .area = (char *)1 })
|
|
#define BUF_RING ((struct buffer){ .area = (char *)2 })
|
|
|
|
|
|
/***************************************************************************/
|
|
/* Functions used to compute offsets and pointers. Most of them exist in */
|
|
/* both wrapping-safe and unchecked ("__" prefix) variants. Some returning */
|
|
/* a pointer are also provided with an "_ofs" suffix when they return an */
|
|
/* offset relative to the storage area. */
|
|
/***************************************************************************/
|
|
|
|
/* b_is_null() : returns true if (and only if) the buffer is not yet allocated
|
|
* and thus has an empty size. Its pointer may then be anything, including NULL
|
|
* (unallocated) or an invalid pointer such as (char*)1 (allocation pending).
|
|
*/
|
|
static inline int b_is_null(const struct buffer *buf)
|
|
{
|
|
return buf->size == 0;
|
|
}
|
|
|
|
/* b_orig() : returns the pointer to the origin of the storage, which is the
|
|
* location of byte at offset zero. This is mostly used by functions which
|
|
* handle the wrapping by themselves.
|
|
*/
|
|
static inline char *b_orig(const struct buffer *b)
|
|
{
|
|
return b->area;
|
|
}
|
|
|
|
/* b_size() : returns the size of the buffer. */
|
|
static inline size_t b_size(const struct buffer *b)
|
|
{
|
|
return b->size;
|
|
}
|
|
|
|
/* b_wrap() : returns the pointer to the wrapping position of the buffer area,
|
|
* which is by definition the first byte not part of the buffer.
|
|
*/
|
|
static inline char *b_wrap(const struct buffer *b)
|
|
{
|
|
return b->area + b->size;
|
|
}
|
|
|
|
/* b_data() : returns the number of bytes present in the buffer. */
|
|
static inline size_t b_data(const struct buffer *b)
|
|
{
|
|
return b->data;
|
|
}
|
|
|
|
/* b_room() : returns the amount of room left in the buffer */
|
|
static inline size_t b_room(const struct buffer *b)
|
|
{
|
|
return b->size - b_data(b);
|
|
}
|
|
|
|
/* b_full() : returns true if the buffer is full. */
|
|
static inline size_t b_full(const struct buffer *b)
|
|
{
|
|
return !b_room(b);
|
|
}
|
|
|
|
|
|
/* b_stop() : returns the pointer to the byte following the end of the buffer,
|
|
* which may be out of the buffer if the buffer ends on the last byte of the
|
|
* area.
|
|
*/
|
|
static inline size_t __b_stop_ofs(const struct buffer *b)
|
|
{
|
|
return b->head + b->data;
|
|
}
|
|
|
|
static inline const char *__b_stop(const struct buffer *b)
|
|
{
|
|
return b_orig(b) + __b_stop_ofs(b);
|
|
}
|
|
|
|
static inline size_t b_stop_ofs(const struct buffer *b)
|
|
{
|
|
size_t stop = __b_stop_ofs(b);
|
|
|
|
if (stop > b->size)
|
|
stop -= b->size;
|
|
return stop;
|
|
}
|
|
|
|
static inline const char *b_stop(const struct buffer *b)
|
|
{
|
|
return b_orig(b) + b_stop_ofs(b);
|
|
}
|
|
|
|
|
|
/* b_peek() : returns a pointer to the data at position <ofs> relative to the
|
|
* head of the buffer. Will typically point to input data if called with the
|
|
* amount of output data. The wrapped versions will only support wrapping once
|
|
* before the beginning or after the end.
|
|
*/
|
|
static inline size_t __b_peek_ofs(const struct buffer *b, size_t ofs)
|
|
{
|
|
return b->head + ofs;
|
|
}
|
|
|
|
static inline char *__b_peek(const struct buffer *b, size_t ofs)
|
|
{
|
|
return b_orig(b) + __b_peek_ofs(b, ofs);
|
|
}
|
|
|
|
static inline size_t b_peek_ofs(const struct buffer *b, size_t ofs)
|
|
{
|
|
size_t ret = __b_peek_ofs(b, ofs);
|
|
|
|
if (ret >= b->size)
|
|
ret -= b->size;
|
|
|
|
return ret;
|
|
}
|
|
|
|
static inline char *b_peek(const struct buffer *b, size_t ofs)
|
|
{
|
|
return b_orig(b) + b_peek_ofs(b, ofs);
|
|
}
|
|
|
|
|
|
/* b_head() : returns the pointer to the buffer's head, which is the location
|
|
* of the next byte to be dequeued. Note that for buffers of size zero, the
|
|
* returned pointer may be outside of the buffer or even invalid.
|
|
*/
|
|
static inline size_t __b_head_ofs(const struct buffer *b)
|
|
{
|
|
return b->head;
|
|
}
|
|
|
|
static inline char *__b_head(const struct buffer *b)
|
|
{
|
|
return b_orig(b) + __b_head_ofs(b);
|
|
}
|
|
|
|
static inline size_t b_head_ofs(const struct buffer *b)
|
|
{
|
|
return __b_head_ofs(b);
|
|
}
|
|
|
|
static inline char *b_head(const struct buffer *b)
|
|
{
|
|
return __b_head(b);
|
|
}
|
|
|
|
|
|
/* b_tail() : returns the pointer to the tail of the buffer, which is the
|
|
* location of the first byte where it is possible to enqueue new data. Note
|
|
* that for buffers of size zero, the returned pointer may be outside of the
|
|
* buffer or even invalid.
|
|
*/
|
|
static inline size_t __b_tail_ofs(const struct buffer *b)
|
|
{
|
|
return __b_peek_ofs(b, b_data(b));
|
|
}
|
|
|
|
static inline char *__b_tail(const struct buffer *b)
|
|
{
|
|
return __b_peek(b, b_data(b));
|
|
}
|
|
|
|
static inline size_t b_tail_ofs(const struct buffer *b)
|
|
{
|
|
return b_peek_ofs(b, b_data(b));
|
|
}
|
|
|
|
static inline char *b_tail(const struct buffer *b)
|
|
{
|
|
return b_peek(b, b_data(b));
|
|
}
|
|
|
|
|
|
/* b_next() : for an absolute pointer <p> or a relative offset <o> pointing to
|
|
* a valid location within buffer <b>, returns either the absolute pointer or
|
|
* the relative offset pointing to the next byte, which usually is at (p + 1)
|
|
* unless p reaches the wrapping point and wrapping is needed.
|
|
*/
|
|
static inline size_t b_next_ofs(const struct buffer *b, size_t o)
|
|
{
|
|
o++;
|
|
if (o == b->size)
|
|
o = 0;
|
|
return o;
|
|
}
|
|
|
|
static inline char *b_next(const struct buffer *b, const char *p)
|
|
{
|
|
p++;
|
|
if (p == b_wrap(b))
|
|
p = b_orig(b);
|
|
return (char *)p;
|
|
}
|
|
|
|
/* b_dist() : returns the distance between two pointers, taking into account
|
|
* the ability to wrap around the buffer's end. The operation is not defined if
|
|
* either of the pointers does not belong to the buffer or if their distance is
|
|
* greater than the buffer's size.
|
|
*/
|
|
static inline size_t b_dist(const struct buffer *b, const char *from, const char *to)
|
|
{
|
|
ssize_t dist = to - from;
|
|
|
|
dist += dist < 0 ? b_size(b) : 0;
|
|
return dist;
|
|
}
|
|
|
|
/* b_almost_full() : returns 1 if the buffer uses at least 3/4 of its capacity,
|
|
* otherwise zero. Buffers of size zero are considered full.
|
|
*/
|
|
static inline int b_almost_full(const struct buffer *b)
|
|
{
|
|
return b_data(b) >= b_size(b) * 3 / 4;
|
|
}
|
|
|
|
/* b_space_wraps() : returns non-zero only if the buffer's free space wraps :
|
|
* [ |xxxx| ] => yes
|
|
* [xxxx| ] => no
|
|
* [ |xxxx] => no
|
|
* [xxxx| |xxxx] => no
|
|
* [xxxxxxxxxx|xxxxxxxxxxx] => no
|
|
*
|
|
* So the only case where the buffer does not wrap is when there's data either
|
|
* at the beginning or at the end of the buffer. Thus we have this :
|
|
* - if (head <= 0) ==> doesn't wrap
|
|
* - if (tail >= size) ==> doesn't wrap
|
|
* - otherwise wraps
|
|
*/
|
|
static inline int b_space_wraps(const struct buffer *b)
|
|
{
|
|
if ((ssize_t)__b_head_ofs(b) <= 0)
|
|
return 0;
|
|
if (__b_tail_ofs(b) >= b_size(b))
|
|
return 0;
|
|
return 1;
|
|
}
|
|
|
|
/* b_contig_data() : returns the amount of data that can contiguously be read
|
|
* at once starting from a relative offset <start> (which allows to easily
|
|
* pre-compute blocks for memcpy). The start point will typically contain the
|
|
* amount of past data already returned by a previous call to this function.
|
|
*/
|
|
static inline size_t b_contig_data(const struct buffer *b, size_t start)
|
|
{
|
|
size_t data = b_wrap(b) - b_peek(b, start);
|
|
size_t limit = b_data(b) - start;
|
|
|
|
if (data > limit)
|
|
data = limit;
|
|
return data;
|
|
}
|
|
|
|
/* b_contig_space() : returns the amount of bytes that can be appended to the
|
|
* buffer at once. We have 8 possible cases :
|
|
*
|
|
* [____________________] return size
|
|
* [______|_____________] return size - tail_ofs
|
|
* [XXXXXX|_____________] return size - tail_ofs
|
|
* [___|XXXXXX|_________] return size - tail_ofs
|
|
* [______________XXXXXX] return head_ofs
|
|
* [XXXX|___________|XXX] return head_ofs - tail_ofs
|
|
* [XXXXXXXXXX|XXXXXXXXX] return 0
|
|
* [XXXXXXXXXXXXXXXXXXXX] return 0
|
|
*/
|
|
static inline size_t b_contig_space(const struct buffer *b)
|
|
{
|
|
size_t left, right;
|
|
|
|
right = b_head_ofs(b);
|
|
left = right + b_data(b);
|
|
|
|
left = b_size(b) - left;
|
|
if ((ssize_t)left <= 0)
|
|
left += right;
|
|
return left;
|
|
}
|
|
|
|
/* b_getblk() : gets one full block of data at once from a buffer, starting
|
|
* from offset <offset> after the buffer's head, and limited to no more than
|
|
* <len> bytes. The caller is responsible for ensuring that neither <offset>
|
|
* nor <offset>+<len> exceed the total number of bytes available in the buffer.
|
|
* Return values :
|
|
* >0 : number of bytes read, equal to requested size.
|
|
* =0 : not enough data available. <blk> is left undefined.
|
|
* The buffer is left unaffected.
|
|
*/
|
|
static inline size_t b_getblk(const struct buffer *buf, char *blk, size_t len, size_t offset)
|
|
{
|
|
size_t firstblock;
|
|
|
|
if (len + offset > b_data(buf))
|
|
return 0;
|
|
|
|
firstblock = b_wrap(buf) - b_head(buf);
|
|
if (firstblock > offset) {
|
|
if (firstblock >= len + offset) {
|
|
memcpy(blk, b_head(buf) + offset, len);
|
|
return len;
|
|
}
|
|
|
|
memcpy(blk, b_head(buf) + offset, firstblock - offset);
|
|
memcpy(blk + firstblock - offset, b_orig(buf), len - firstblock + offset);
|
|
return len;
|
|
}
|
|
|
|
memcpy(blk, b_orig(buf) + offset - firstblock, len);
|
|
return len;
|
|
}
|
|
|
|
/* b_getblk_nc() : gets one or two blocks of data at once from a buffer,
|
|
* starting from offset <ofs> after the beginning of its output, and limited to
|
|
* no more than <max> bytes. The caller is responsible for ensuring that
|
|
* neither <ofs> nor <ofs>+<max> exceed the total number of bytes available in
|
|
* the buffer. Return values :
|
|
* >0 : number of blocks filled (1 or 2). blk1 is always filled before blk2.
|
|
* =0 : not enough data available. <blk*> are left undefined.
|
|
* The buffer is left unaffected. Unused buffers are left in an undefined state.
|
|
*/
|
|
static inline size_t b_getblk_nc(const struct buffer *buf, const char **blk1, size_t *len1, const char **blk2, size_t *len2, size_t ofs, size_t max)
|
|
{
|
|
size_t l1;
|
|
|
|
if (!max)
|
|
return 0;
|
|
|
|
*blk1 = b_peek(buf, ofs);
|
|
l1 = b_wrap(buf) - *blk1;
|
|
if (l1 < max) {
|
|
*len1 = l1;
|
|
*len2 = max - l1;
|
|
*blk2 = b_orig(buf);
|
|
return 2;
|
|
}
|
|
*len1 = max;
|
|
return 1;
|
|
}
|
|
|
|
|
|
/*********************************************/
|
|
/* Functions used to modify the buffer state */
|
|
/*********************************************/
|
|
|
|
/* b_reset() : resets a buffer. The size is not touched. */
|
|
static inline void b_reset(struct buffer *b)
|
|
{
|
|
b->head = 0;
|
|
b->data = 0;
|
|
}
|
|
|
|
/* b_make() : make a buffer from all parameters */
|
|
static inline struct buffer b_make(char *area, size_t size, size_t head, size_t data)
|
|
{
|
|
struct buffer b;
|
|
|
|
b.area = area;
|
|
b.size = size;
|
|
b.head = head;
|
|
b.data = data;
|
|
return b;
|
|
}
|
|
|
|
/* b_sub() : decreases the buffer length by <count> */
|
|
static inline void b_sub(struct buffer *b, size_t count)
|
|
{
|
|
b->data -= count;
|
|
}
|
|
|
|
/* b_add() : increase the buffer length by <count> */
|
|
static inline void b_add(struct buffer *b, size_t count)
|
|
{
|
|
b->data += count;
|
|
}
|
|
|
|
/* b_set_data() : sets the buffer's length */
|
|
static inline void b_set_data(struct buffer *b, size_t len)
|
|
{
|
|
b->data = len;
|
|
}
|
|
|
|
/* b_del() : skips <del> bytes in a buffer <b>. Covers both the output and the
|
|
* input parts so it's up to the caller to know where it plays and that <del>
|
|
* is always smaller than the amount of data in the buffer.
|
|
*/
|
|
static inline void b_del(struct buffer *b, size_t del)
|
|
{
|
|
b->data -= del;
|
|
b->head += del;
|
|
if (b->head >= b->size)
|
|
b->head -= b->size;
|
|
}
|
|
|
|
/* b_realign_if_empty() : realigns a buffer if it's empty */
|
|
static inline void b_realign_if_empty(struct buffer *b)
|
|
{
|
|
if (!b_data(b))
|
|
b->head = 0;
|
|
}
|
|
|
|
/* b_slow_realign() : this function realigns a possibly wrapping buffer so that
|
|
* the part remaining to be parsed is contiguous and starts at the beginning of
|
|
* the buffer and the already parsed output part ends at the end of the buffer.
|
|
* This provides the best conditions since it allows the largest inputs to be
|
|
* processed at once and ensures that once the output data leaves, the whole
|
|
* buffer is available at once. The number of output bytes supposedly present
|
|
* at the beginning of the buffer and which need to be moved to the end must be
|
|
* passed in <output>. A temporary swap area at least as large as b->size must
|
|
* be provided in <swap>. It's up to the caller to ensure <output> is no larger
|
|
* than the difference between the whole buffer's length and its input.
|
|
*/
|
|
static inline void b_slow_realign(struct buffer *b, char *swap, size_t output)
|
|
{
|
|
size_t block1 = output;
|
|
size_t block2 = 0;
|
|
|
|
/* process output data in two steps to cover wrapping */
|
|
if (block1 > b_size(b) - b_head_ofs(b)) {
|
|
block2 = b_size(b) - b_head_ofs(b);
|
|
block1 -= block2;
|
|
}
|
|
memcpy(swap + b_size(b) - output, b_head(b), block1);
|
|
memcpy(swap + b_size(b) - block2, b_orig(b), block2);
|
|
|
|
/* process input data in two steps to cover wrapping */
|
|
block1 = b_data(b) - output;
|
|
block2 = 0;
|
|
|
|
if (block1 > b_tail_ofs(b)) {
|
|
block2 = b_tail_ofs(b);
|
|
block1 = block1 - block2;
|
|
}
|
|
memcpy(swap, b_peek(b, output), block1);
|
|
memcpy(swap + block1, b_orig(b), block2);
|
|
|
|
/* reinject changes into the buffer */
|
|
memcpy(b_orig(b), swap, b_data(b) - output);
|
|
memcpy(b_wrap(b) - output, swap + b_size(b) - output, output);
|
|
|
|
b->head = (output ? b_size(b) - output : 0);
|
|
}
|
|
|
|
/* b_putchar() : tries to append char <c> at the end of buffer <b>. Supports
|
|
* wrapping. Data are truncated if buffer is full.
|
|
*/
|
|
static inline void b_putchr(struct buffer *b, char c)
|
|
{
|
|
if (b_full(b))
|
|
return;
|
|
*b_tail(b) = c;
|
|
b->data++;
|
|
}
|
|
|
|
/* __b_putblk() : tries to append <len> bytes from block <blk> to the end of
|
|
* buffer <b> without checking for free space (it's up to the caller to do it).
|
|
* Supports wrapping. It must not be called with len == 0.
|
|
*/
|
|
static inline void __b_putblk(struct buffer *b, const char *blk, size_t len)
|
|
{
|
|
size_t half = b_contig_space(b);
|
|
|
|
if (half > len)
|
|
half = len;
|
|
|
|
memcpy(b_tail(b), blk, half);
|
|
|
|
if (len > half)
|
|
memcpy(b_peek(b, b_data(b) + half), blk + half, len - half);
|
|
b->data += len;
|
|
}
|
|
|
|
/* b_putblk() : tries to append block <blk> at the end of buffer <b>. Supports
|
|
* wrapping. Data are truncated if buffer is too short. It returns the number
|
|
* of bytes copied.
|
|
*/
|
|
static inline size_t b_putblk(struct buffer *b, const char *blk, size_t len)
|
|
{
|
|
if (len > b_room(b))
|
|
len = b_room(b);
|
|
if (len)
|
|
__b_putblk(b, blk, len);
|
|
return len;
|
|
}
|
|
|
|
/* b_xfer() : transfers at most <count> bytes from buffer <src> to buffer <dst>
|
|
* and returns the number of bytes copied. The bytes are removed from <src> and
|
|
* added to <dst>. The caller is responsible for ensuring that <count> is not
|
|
* larger than b_room(dst). Whenever possible (if the destination is empty and
|
|
* at least as much as the source was requested), the buffers are simply
|
|
* swapped instead of copied.
|
|
*/
|
|
static inline size_t b_xfer(struct buffer *dst, struct buffer *src, size_t count)
|
|
{
|
|
size_t ret, block1, block2;
|
|
|
|
ret = 0;
|
|
if (!count)
|
|
goto leave;
|
|
|
|
ret = b_data(src);
|
|
if (!ret)
|
|
goto leave;
|
|
|
|
if (ret > count)
|
|
ret = count;
|
|
else if (!b_data(dst)) {
|
|
/* zero copy is possible by just swapping buffers */
|
|
struct buffer tmp = *dst;
|
|
*dst = *src;
|
|
*src = tmp;
|
|
goto leave;
|
|
}
|
|
|
|
block1 = b_contig_data(src, 0);
|
|
if (block1 > ret)
|
|
block1 = ret;
|
|
block2 = ret - block1;
|
|
|
|
if (block1)
|
|
__b_putblk(dst, b_head(src), block1);
|
|
|
|
if (block2)
|
|
__b_putblk(dst, b_peek(src, block1), block2);
|
|
|
|
b_del(src, ret);
|
|
leave:
|
|
return ret;
|
|
}
|
|
|
|
/* Moves <len> bytes from absolute position <src> of buffer <b> by <shift>
|
|
* bytes, while supporting wrapping of both the source and the destination.
|
|
* The position is relative to the buffer's origin and may overlap with the
|
|
* target position. The <shift>'s absolute value must be strictly lower than
|
|
* the buffer's size. The main purpose is to aggregate data block during
|
|
* parsing while removing unused delimiters. The buffer's length is not
|
|
* modified, and the caller must take care of size adjustments and holes by
|
|
* itself.
|
|
*/
|
|
static inline void b_move(const struct buffer *b, size_t src, size_t len, ssize_t shift)
|
|
{
|
|
char *orig = b_orig(b);
|
|
size_t size = b_size(b);
|
|
size_t dst = src + size + shift;
|
|
size_t cnt;
|
|
|
|
if (dst >= size)
|
|
dst -= size;
|
|
|
|
if (shift < 0) {
|
|
/* copy from left to right */
|
|
for (; (cnt = len); len -= cnt) {
|
|
if (cnt > size - src)
|
|
cnt = size - src;
|
|
if (cnt > size - dst)
|
|
cnt = size - dst;
|
|
|
|
memmove(orig + dst, orig + src, cnt);
|
|
dst += cnt;
|
|
src += cnt;
|
|
if (dst >= size)
|
|
dst -= size;
|
|
if (src >= size)
|
|
src -= size;
|
|
}
|
|
}
|
|
else if (shift > 0) {
|
|
/* copy from right to left */
|
|
for (; (cnt = len); len -= cnt) {
|
|
size_t src_end = src + len;
|
|
size_t dst_end = dst + len;
|
|
|
|
if (dst_end > size)
|
|
dst_end -= size;
|
|
if (src_end > size)
|
|
src_end -= size;
|
|
|
|
if (cnt > dst_end)
|
|
cnt = dst_end;
|
|
if (cnt > src_end)
|
|
cnt = src_end;
|
|
|
|
memmove(orig + dst_end - cnt, orig + src_end - cnt, cnt);
|
|
}
|
|
}
|
|
}
|
|
|
|
/* b_rep_blk() : writes the block <blk> at position <pos> which must be in
|
|
* buffer <b>, and moves the part between <end> and the buffer's tail just
|
|
* after the end of the copy of <blk>. This effectively replaces the part
|
|
* located between <pos> and <end> with a copy of <blk> of length <len>. The
|
|
* buffer's length is automatically updated. This is used to replace a block
|
|
* with another one inside a buffer. The shift value (positive or negative) is
|
|
* returned. If there's no space left, the move is not done. If <len> is null,
|
|
* the <blk> pointer is allowed to be null, in order to erase a block.
|
|
*/
|
|
static inline int b_rep_blk(struct buffer *b, char *pos, char *end, const char *blk, size_t len)
|
|
{
|
|
int delta;
|
|
|
|
delta = len - (end - pos);
|
|
|
|
if (__b_tail(b) + delta > b_wrap(b))
|
|
return 0; /* no space left */
|
|
|
|
if (b_data(b) &&
|
|
b_tail(b) + delta > b_head(b) &&
|
|
b_head(b) >= b_tail(b))
|
|
return 0; /* no space left before wrapping data */
|
|
|
|
/* first, protect the end of the buffer */
|
|
memmove(end + delta, end, b_tail(b) - end);
|
|
|
|
/* now, copy blk over pos */
|
|
if (len)
|
|
memcpy(pos, blk, len);
|
|
|
|
b_add(b, delta);
|
|
b_realign_if_empty(b);
|
|
|
|
return delta;
|
|
}
|
|
|
|
|
|
/* __b_put_varint(): encode 64-bit value <v> as a varint into buffer <b>. The
|
|
* caller must have checked that the encoded value fits in the buffer so that
|
|
* there are no length checks. Wrapping is supported. You don't want to use
|
|
* this function but b_put_varint() instead.
|
|
*/
|
|
static inline void __b_put_varint(struct buffer *b, uint64_t v)
|
|
{
|
|
size_t data = b->data;
|
|
size_t size = b_size(b);
|
|
char *wrap = b_wrap(b);
|
|
char *tail = b_tail(b);
|
|
|
|
if (v >= 0xF0) {
|
|
/* more than one byte, first write the 4 least significant
|
|
* bits, then follow with 7 bits per byte.
|
|
*/
|
|
*tail = v | 0xF0;
|
|
v = (v - 0xF0) >> 4;
|
|
|
|
while (1) {
|
|
if (tail++ == wrap)
|
|
tail -= size;
|
|
data++;
|
|
if (v < 0x80)
|
|
break;
|
|
*tail = v | 0x80;
|
|
v = (v - 0x80) >> 7;
|
|
}
|
|
}
|
|
|
|
/* last byte */
|
|
*tail = v;
|
|
data++;
|
|
b->data = data;
|
|
}
|
|
|
|
/* b_put_varint(): try to encode value <v> as a varint into buffer <b>. Returns
|
|
* the number of bytes written in case of success, or 0 if there is not enough
|
|
* room. Wrapping is supported. No partial writes will be performed.
|
|
*/
|
|
static inline int b_put_varint(struct buffer *b, uint64_t v)
|
|
{
|
|
size_t data = b->data;
|
|
size_t size = b_size(b);
|
|
char *wrap = b_wrap(b);
|
|
char *tail = b_tail(b);
|
|
|
|
if (data != size && v >= 0xF0) {
|
|
/* more than one byte, first write the 4 least significant
|
|
* bits, then follow with 7 bits per byte.
|
|
*/
|
|
*tail = v | 0xF0;
|
|
v = (v - 0xF0) >> 4;
|
|
|
|
while (1) {
|
|
if (tail++ == wrap)
|
|
tail -= size;
|
|
data++;
|
|
if (data == size || v < 0x80)
|
|
break;
|
|
*tail = v | 0x80;
|
|
v = (v - 0x80) >> 7;
|
|
}
|
|
}
|
|
|
|
/* last byte */
|
|
if (data == size)
|
|
return 0;
|
|
|
|
*tail = v;
|
|
data++;
|
|
|
|
size = data - b->data;
|
|
b->data = data;
|
|
return size;
|
|
}
|
|
|
|
/* b_get_varint(): try to decode a varint from buffer <b> into value <vptr>.
|
|
* Returns the number of bytes read in case of success, or 0 if there were not
|
|
* enough bytes. Wrapping is supported. No partial reads will be performed.
|
|
*/
|
|
static inline int b_get_varint(struct buffer *b, uint64_t *vptr)
|
|
{
|
|
const uint8_t *head = (const uint8_t *)b_head(b);
|
|
const uint8_t *wrap = (const uint8_t *)b_wrap(b);
|
|
size_t data = b->data;
|
|
size_t size = b_size(b);
|
|
uint64_t v = 0;
|
|
int bits = 0;
|
|
|
|
if (data != 0 && (*head >= 0xF0)) {
|
|
v = *head;
|
|
bits += 4;
|
|
while (1) {
|
|
if (head++ == wrap)
|
|
head -= size;
|
|
data--;
|
|
if (!data || !(*head & 0x80))
|
|
break;
|
|
v += (uint64_t)*head << bits;
|
|
bits += 7;
|
|
}
|
|
}
|
|
|
|
/* last byte */
|
|
if (!data)
|
|
return 0;
|
|
|
|
v += (uint64_t)*head << bits;
|
|
*vptr = v;
|
|
data--;
|
|
size = b->data - data;
|
|
b_del(b, size);
|
|
return size;
|
|
}
|
|
|
|
/* b_peek_varint(): try to decode a varint from buffer <b> at offset <ofs>
|
|
* relative to head, into value <vptr>. Returns the number of bytes parsed in
|
|
* case of success, or 0 if there were not enough bytes, in which case the
|
|
* contents of <vptr> are not updated. Wrapping is supported. The buffer's head
|
|
* will NOT be updated. It is illegal to call this function with <ofs> greater
|
|
* than b->data.
|
|
*/
|
|
static inline int b_peek_varint(struct buffer *b, size_t ofs, uint64_t *vptr)
|
|
{
|
|
const uint8_t *head = (const uint8_t *)b_peek(b, ofs);
|
|
const uint8_t *wrap = (const uint8_t *)b_wrap(b);
|
|
size_t data = b_data(b) - ofs;
|
|
size_t size = b_size(b);
|
|
uint64_t v = 0;
|
|
int bits = 0;
|
|
|
|
if (data != 0 && (*head >= 0xF0)) {
|
|
v = *head;
|
|
bits += 4;
|
|
while (1) {
|
|
if (head++ == wrap)
|
|
head -= size;
|
|
data--;
|
|
if (!data || !(*head & 0x80))
|
|
break;
|
|
v += (uint64_t)*head << bits;
|
|
bits += 7;
|
|
}
|
|
}
|
|
|
|
/* last byte */
|
|
if (!data)
|
|
return 0;
|
|
|
|
v += (uint64_t)*head << bits;
|
|
*vptr = v;
|
|
data--;
|
|
size = b->data - ofs - data;
|
|
return size;
|
|
}
|
|
|
|
|
|
/*
|
|
* Buffer ring management.
|
|
*
|
|
* A buffer ring is a circular list of buffers, with a head buffer (the oldest,
|
|
* being read from) and a tail (the newest, being written to). Such a ring is
|
|
* declared as an array of buffers. The first element in the array is the root
|
|
* and is used differently. It stores the following elements :
|
|
* - size : number of allocated elements in the array, including the root
|
|
* - area : magic value BUF_RING (just to help debugging)
|
|
* - head : position of the head in the array (starts at one)
|
|
* - data : position of the tail in the array (starts at one).
|
|
*
|
|
* Note that contrary to a linear buffer, head and tail may be equal with room
|
|
* available, since the producer is expected to fill the tail. Also, the tail
|
|
* might pretty much be equal to BUF_WANTED if an allocation is pending, in
|
|
* which case it's illegal to try to allocate past this point (only one entry
|
|
* may be subscribed for allocation). It is illegal to allocate a buffer after
|
|
* an empty one, so that BUF_NULL is always the last buffer. It is also illegal
|
|
* to remove elements without freeing the buffers. Buffers between <tail> and
|
|
* <head> are in an undefined state, but <tail> and <head> are always valid.
|
|
* A ring may not contain less than 2 elements, since the root is mandatory,
|
|
* and at least one entry is required to always present a valid buffer.
|
|
*
|
|
* Given that buffers are 16- or 32- bytes long, it's convenient to set the
|
|
* size of the array to 2^N in order to keep (2^N)-1 elements, totalizing
|
|
* 2^N*16(or 32) bytes. For example on a 64-bit system, a ring of 31 usable
|
|
* buffers takes 1024 bytes.
|
|
*/
|
|
|
|
/* Initialization of a ring, the size argument contains the number of allocated
|
|
* elements, including the root. There must always be at least 2 elements, one
|
|
* for the root and one for storage.
|
|
*/
|
|
static inline void br_init(struct buffer *r, size_t size)
|
|
{
|
|
BUG_ON(size < 2);
|
|
|
|
r->size = size;
|
|
r->area = BUF_RING.area;
|
|
r->head = r->data = 1;
|
|
r[1] = BUF_NULL;
|
|
}
|
|
|
|
/* Returns number of elements in the ring, root included */
|
|
static inline unsigned int br_size(const struct buffer *r)
|
|
{
|
|
BUG_ON(r->area != BUF_RING.area);
|
|
|
|
return r->size;
|
|
}
|
|
|
|
/* Returns true if no more buffers may be added */
|
|
static inline unsigned int br_full(const struct buffer *r)
|
|
{
|
|
BUG_ON(r->area != BUF_RING.area);
|
|
|
|
return r->data + 1 == r->head || r->data + 1 == r->head - 1 + r->size;
|
|
}
|
|
|
|
/* Returns the index of the ring's head buffer */
|
|
static inline unsigned int br_head_idx(const struct buffer *r)
|
|
{
|
|
BUG_ON(r->area != BUF_RING.area);
|
|
|
|
return r->head;
|
|
}
|
|
|
|
/* Returns the index of the ring's tail buffer */
|
|
static inline unsigned int br_tail_idx(const struct buffer *r)
|
|
{
|
|
BUG_ON(r->area != BUF_RING.area);
|
|
|
|
return r->data;
|
|
}
|
|
|
|
/* Returns a pointer to the ring's head buffer */
|
|
static inline struct buffer *br_head(struct buffer *r)
|
|
{
|
|
BUG_ON(r->area != BUF_RING.area);
|
|
|
|
return r + br_head_idx(r);
|
|
}
|
|
|
|
/* Returns a pointer to the ring's tail buffer */
|
|
static inline struct buffer *br_tail(struct buffer *r)
|
|
{
|
|
BUG_ON(r->area != BUF_RING.area);
|
|
|
|
return r + br_tail_idx(r);
|
|
}
|
|
|
|
/* Returns the amount of data of the ring's HEAD buffer */
|
|
static inline unsigned int br_data(const struct buffer *r)
|
|
{
|
|
BUG_ON(r->area != BUF_RING.area);
|
|
|
|
return b_data(r + br_head_idx(r));
|
|
}
|
|
|
|
/* Returns non-zero if the ring is non-full or its tail has some room */
|
|
static inline unsigned int br_has_room(const struct buffer *r)
|
|
{
|
|
BUG_ON(r->area != BUF_RING.area);
|
|
|
|
if (!br_full(r))
|
|
return 1;
|
|
return b_room(r + br_tail_idx(r));
|
|
}
|
|
|
|
/* Advances the ring's tail if it points to a non-empty buffer, and returns the
|
|
* buffer, or NULL if the ring is full or the tail buffer is already empty. A
|
|
* new buffer is initialized to BUF_NULL before being returned. This is to be
|
|
* used after failing to append data, in order to decide to retry or not.
|
|
*/
|
|
static inline struct buffer *br_tail_add(struct buffer *r)
|
|
{
|
|
struct buffer *b;
|
|
|
|
BUG_ON(r->area != BUF_RING.area);
|
|
|
|
b = br_tail(r);
|
|
if (!b_size(b))
|
|
return NULL;
|
|
|
|
if (br_full(r))
|
|
return NULL;
|
|
|
|
r->data++;
|
|
if (r->data >= r->size)
|
|
r->data = 1;
|
|
|
|
b = br_tail(r);
|
|
*b = BUF_NULL;
|
|
return b;
|
|
}
|
|
|
|
/* Extracts the ring's head buffer and returns it. The last buffer (tail) is
|
|
* never removed but it is returned. This guarantees that we stop on BUF_WANTED
|
|
* or BUF_EMPTY and that at the end a valid buffer remains present. This is
|
|
* used for pre-extraction during a free() loop for example. The caller is
|
|
* expected to detect the end (e.g. using bsize() since b_free() voids the
|
|
* buffer).
|
|
*/
|
|
static inline struct buffer *br_head_pick(struct buffer *r)
|
|
{
|
|
struct buffer *b;
|
|
|
|
BUG_ON(r->area != BUF_RING.area);
|
|
|
|
b = br_head(r);
|
|
if (r->head != r->data) {
|
|
r->head++;
|
|
if (r->head >= r->size)
|
|
r->head = 1;
|
|
}
|
|
return b;
|
|
}
|
|
|
|
/* Advances the ring's head and returns the next buffer, unless it's already
|
|
* the tail, in which case the tail itself is returned. This is used for post-
|
|
* parsing deletion. The caller is expected to detect the end (e.g. a parser
|
|
* will typically purge the head before proceeding).
|
|
*/
|
|
static inline struct buffer *br_del_head(struct buffer *r)
|
|
{
|
|
BUG_ON(r->area != BUF_RING.area);
|
|
|
|
if (r->head != r->data) {
|
|
r->head++;
|
|
if (r->head >= r->size)
|
|
r->head = 1;
|
|
}
|
|
return br_head(r);
|
|
}
|
|
|
|
#endif /* _COMMON_BUF_H */
|
|
|
|
/*
|
|
* Local variables:
|
|
* c-indent-level: 8
|
|
* c-basic-offset: 8
|
|
* End:
|
|
*/
|