haproxy/doc/lua-api
Thierry FOURNIER / OZON.IO 8a1027aa45 MINOR: lua: Add tokenize function.
For tokenizing a string, standard Lua recommends to use regexes.
The followinf example splits words:

   for i in string.gmatch(example, "%S+") do
      print(i)
   end

This is a little bit overkill for simply split words. This patch
adds a tokenize function which quick and do not use regexes.
2016-11-24 21:35:34 +01:00
..
_static MINOR: lua: add function which return true if the channel is full. 2016-11-12 10:42:25 +01:00
conf.py
index.rst MINOR: lua: Add tokenize function. 2016-11-24 21:35:34 +01:00
Makefile