mirror of https://git.ffmpeg.org/ffmpeg.git
607 lines
16 KiB
C
607 lines
16 KiB
C
/*
|
|
* PNM image format
|
|
* Copyright (c) 2002, 2003 Fabrice Bellard.
|
|
*
|
|
* This file is part of FFmpeg.
|
|
*
|
|
* FFmpeg is free software; you can redistribute it and/or
|
|
* modify it under the terms of the GNU Lesser General Public
|
|
* License as published by the Free Software Foundation; either
|
|
* version 2.1 of the License, or (at your option) any later version.
|
|
*
|
|
* FFmpeg is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
* Lesser General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU Lesser General Public
|
|
* License along with FFmpeg; if not, write to the Free Software
|
|
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
|
|
*/
|
|
#include "avcodec.h"
|
|
#include "parser.h" //for ParseContext
|
|
|
|
typedef struct PNMContext {
|
|
uint8_t *bytestream;
|
|
uint8_t *bytestream_start;
|
|
uint8_t *bytestream_end;
|
|
AVFrame picture;
|
|
} PNMContext;
|
|
|
|
static inline int pnm_space(int c)
|
|
{
|
|
return (c == ' ' || c == '\n' || c == '\r' || c == '\t');
|
|
}
|
|
|
|
static void pnm_get(PNMContext *sc, char *str, int buf_size)
|
|
{
|
|
char *s;
|
|
int c;
|
|
|
|
/* skip spaces and comments */
|
|
for(;;) {
|
|
c = *sc->bytestream++;
|
|
if (c == '#') {
|
|
do {
|
|
c = *sc->bytestream++;
|
|
} while (c != '\n' && sc->bytestream < sc->bytestream_end);
|
|
} else if (!pnm_space(c)) {
|
|
break;
|
|
}
|
|
}
|
|
|
|
s = str;
|
|
while (sc->bytestream < sc->bytestream_end && !pnm_space(c)) {
|
|
if ((s - str) < buf_size - 1)
|
|
*s++ = c;
|
|
c = *sc->bytestream++;
|
|
}
|
|
*s = '\0';
|
|
}
|
|
|
|
static int common_init(AVCodecContext *avctx){
|
|
PNMContext *s = avctx->priv_data;
|
|
|
|
avcodec_get_frame_defaults((AVFrame*)&s->picture);
|
|
avctx->coded_frame= (AVFrame*)&s->picture;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int pnm_decode_header(AVCodecContext *avctx, PNMContext * const s){
|
|
char buf1[32], tuple_type[32];
|
|
int h, w, depth, maxval;
|
|
|
|
pnm_get(s, buf1, sizeof(buf1));
|
|
if (!strcmp(buf1, "P4")) {
|
|
avctx->pix_fmt = PIX_FMT_MONOWHITE;
|
|
} else if (!strcmp(buf1, "P5")) {
|
|
if (avctx->codec_id == CODEC_ID_PGMYUV)
|
|
avctx->pix_fmt = PIX_FMT_YUV420P;
|
|
else
|
|
avctx->pix_fmt = PIX_FMT_GRAY8;
|
|
} else if (!strcmp(buf1, "P6")) {
|
|
avctx->pix_fmt = PIX_FMT_RGB24;
|
|
} else if (!strcmp(buf1, "P7")) {
|
|
w = -1;
|
|
h = -1;
|
|
maxval = -1;
|
|
depth = -1;
|
|
tuple_type[0] = '\0';
|
|
for(;;) {
|
|
pnm_get(s, buf1, sizeof(buf1));
|
|
if (!strcmp(buf1, "WIDTH")) {
|
|
pnm_get(s, buf1, sizeof(buf1));
|
|
w = strtol(buf1, NULL, 10);
|
|
} else if (!strcmp(buf1, "HEIGHT")) {
|
|
pnm_get(s, buf1, sizeof(buf1));
|
|
h = strtol(buf1, NULL, 10);
|
|
} else if (!strcmp(buf1, "DEPTH")) {
|
|
pnm_get(s, buf1, sizeof(buf1));
|
|
depth = strtol(buf1, NULL, 10);
|
|
} else if (!strcmp(buf1, "MAXVAL")) {
|
|
pnm_get(s, buf1, sizeof(buf1));
|
|
maxval = strtol(buf1, NULL, 10);
|
|
} else if (!strcmp(buf1, "TUPLETYPE")) {
|
|
pnm_get(s, tuple_type, sizeof(tuple_type));
|
|
} else if (!strcmp(buf1, "ENDHDR")) {
|
|
break;
|
|
} else {
|
|
return -1;
|
|
}
|
|
}
|
|
/* check that all tags are present */
|
|
if (w <= 0 || h <= 0 || maxval <= 0 || depth <= 0 || tuple_type[0] == '\0' || avcodec_check_dimensions(avctx, w, h))
|
|
return -1;
|
|
|
|
avctx->width = w;
|
|
avctx->height = h;
|
|
if (depth == 1) {
|
|
if (maxval == 1)
|
|
avctx->pix_fmt = PIX_FMT_MONOWHITE;
|
|
else
|
|
avctx->pix_fmt = PIX_FMT_GRAY8;
|
|
} else if (depth == 3) {
|
|
avctx->pix_fmt = PIX_FMT_RGB24;
|
|
} else if (depth == 4) {
|
|
avctx->pix_fmt = PIX_FMT_RGB32;
|
|
} else {
|
|
return -1;
|
|
}
|
|
return 0;
|
|
} else {
|
|
return -1;
|
|
}
|
|
pnm_get(s, buf1, sizeof(buf1));
|
|
avctx->width = atoi(buf1);
|
|
if (avctx->width <= 0)
|
|
return -1;
|
|
pnm_get(s, buf1, sizeof(buf1));
|
|
avctx->height = atoi(buf1);
|
|
if(avcodec_check_dimensions(avctx, avctx->width, avctx->height))
|
|
return -1;
|
|
if (avctx->pix_fmt != PIX_FMT_MONOWHITE) {
|
|
pnm_get(s, buf1, sizeof(buf1));
|
|
if(atoi(buf1) == 65535 && avctx->pix_fmt == PIX_FMT_GRAY8)
|
|
avctx->pix_fmt = PIX_FMT_GRAY16BE;
|
|
}
|
|
/* more check if YUV420 */
|
|
if (avctx->pix_fmt == PIX_FMT_YUV420P) {
|
|
if ((avctx->width & 1) != 0)
|
|
return -1;
|
|
h = (avctx->height * 2);
|
|
if ((h % 3) != 0)
|
|
return -1;
|
|
h /= 3;
|
|
avctx->height = h;
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
static int pnm_decode_frame(AVCodecContext *avctx,
|
|
void *data, int *data_size,
|
|
uint8_t *buf, int buf_size)
|
|
{
|
|
PNMContext * const s = avctx->priv_data;
|
|
AVFrame *picture = data;
|
|
AVFrame * const p= (AVFrame*)&s->picture;
|
|
int i, n, linesize, h;
|
|
unsigned char *ptr;
|
|
|
|
s->bytestream_start=
|
|
s->bytestream= buf;
|
|
s->bytestream_end= buf + buf_size;
|
|
|
|
if(pnm_decode_header(avctx, s) < 0)
|
|
return -1;
|
|
|
|
if(p->data[0])
|
|
avctx->release_buffer(avctx, p);
|
|
|
|
p->reference= 0;
|
|
if(avctx->get_buffer(avctx, p) < 0){
|
|
av_log(avctx, AV_LOG_ERROR, "get_buffer() failed\n");
|
|
return -1;
|
|
}
|
|
p->pict_type= FF_I_TYPE;
|
|
p->key_frame= 1;
|
|
|
|
switch(avctx->pix_fmt) {
|
|
default:
|
|
return -1;
|
|
case PIX_FMT_RGB24:
|
|
n = avctx->width * 3;
|
|
goto do_read;
|
|
case PIX_FMT_GRAY8:
|
|
n = avctx->width;
|
|
goto do_read;
|
|
case PIX_FMT_GRAY16BE:
|
|
n = avctx->width * 2;
|
|
goto do_read;
|
|
case PIX_FMT_MONOWHITE:
|
|
case PIX_FMT_MONOBLACK:
|
|
n = (avctx->width + 7) >> 3;
|
|
do_read:
|
|
ptr = p->data[0];
|
|
linesize = p->linesize[0];
|
|
if(s->bytestream + n*avctx->height > s->bytestream_end)
|
|
return -1;
|
|
for(i = 0; i < avctx->height; i++) {
|
|
memcpy(ptr, s->bytestream, n);
|
|
s->bytestream += n;
|
|
ptr += linesize;
|
|
}
|
|
break;
|
|
case PIX_FMT_YUV420P:
|
|
{
|
|
unsigned char *ptr1, *ptr2;
|
|
|
|
n = avctx->width;
|
|
ptr = p->data[0];
|
|
linesize = p->linesize[0];
|
|
if(s->bytestream + n*avctx->height*3/2 > s->bytestream_end)
|
|
return -1;
|
|
for(i = 0; i < avctx->height; i++) {
|
|
memcpy(ptr, s->bytestream, n);
|
|
s->bytestream += n;
|
|
ptr += linesize;
|
|
}
|
|
ptr1 = p->data[1];
|
|
ptr2 = p->data[2];
|
|
n >>= 1;
|
|
h = avctx->height >> 1;
|
|
for(i = 0; i < h; i++) {
|
|
memcpy(ptr1, s->bytestream, n);
|
|
s->bytestream += n;
|
|
memcpy(ptr2, s->bytestream, n);
|
|
s->bytestream += n;
|
|
ptr1 += p->linesize[1];
|
|
ptr2 += p->linesize[2];
|
|
}
|
|
}
|
|
break;
|
|
case PIX_FMT_RGB32:
|
|
ptr = p->data[0];
|
|
linesize = p->linesize[0];
|
|
if(s->bytestream + avctx->width*avctx->height*4 > s->bytestream_end)
|
|
return -1;
|
|
for(i = 0; i < avctx->height; i++) {
|
|
int j, r, g, b, a;
|
|
|
|
for(j = 0;j < avctx->width; j++) {
|
|
r = *s->bytestream++;
|
|
g = *s->bytestream++;
|
|
b = *s->bytestream++;
|
|
a = *s->bytestream++;
|
|
((uint32_t *)ptr)[j] = (a << 24) | (r << 16) | (g << 8) | b;
|
|
}
|
|
ptr += linesize;
|
|
}
|
|
break;
|
|
}
|
|
*picture= *(AVFrame*)&s->picture;
|
|
*data_size = sizeof(AVPicture);
|
|
|
|
return s->bytestream - s->bytestream_start;
|
|
}
|
|
|
|
static int pnm_encode_frame(AVCodecContext *avctx, unsigned char *outbuf, int buf_size, void *data){
|
|
PNMContext *s = avctx->priv_data;
|
|
AVFrame *pict = data;
|
|
AVFrame * const p= (AVFrame*)&s->picture;
|
|
int i, h, h1, c, n, linesize;
|
|
uint8_t *ptr, *ptr1, *ptr2;
|
|
|
|
if(buf_size < avpicture_get_size(avctx->pix_fmt, avctx->width, avctx->height) + 200){
|
|
av_log(avctx, AV_LOG_ERROR, "encoded frame too large\n");
|
|
return -1;
|
|
}
|
|
|
|
*p = *pict;
|
|
p->pict_type= FF_I_TYPE;
|
|
p->key_frame= 1;
|
|
|
|
s->bytestream_start=
|
|
s->bytestream= outbuf;
|
|
s->bytestream_end= outbuf+buf_size;
|
|
|
|
h = avctx->height;
|
|
h1 = h;
|
|
switch(avctx->pix_fmt) {
|
|
case PIX_FMT_MONOWHITE:
|
|
c = '4';
|
|
n = (avctx->width + 7) >> 3;
|
|
break;
|
|
case PIX_FMT_GRAY8:
|
|
c = '5';
|
|
n = avctx->width;
|
|
break;
|
|
case PIX_FMT_GRAY16BE:
|
|
c = '5';
|
|
n = avctx->width * 2;
|
|
break;
|
|
case PIX_FMT_RGB24:
|
|
c = '6';
|
|
n = avctx->width * 3;
|
|
break;
|
|
case PIX_FMT_YUV420P:
|
|
c = '5';
|
|
n = avctx->width;
|
|
h1 = (h * 3) / 2;
|
|
break;
|
|
default:
|
|
return -1;
|
|
}
|
|
snprintf(s->bytestream, s->bytestream_end - s->bytestream,
|
|
"P%c\n%d %d\n",
|
|
c, avctx->width, h1);
|
|
s->bytestream += strlen(s->bytestream);
|
|
if (avctx->pix_fmt != PIX_FMT_MONOWHITE) {
|
|
snprintf(s->bytestream, s->bytestream_end - s->bytestream,
|
|
"%d\n", (avctx->pix_fmt != PIX_FMT_GRAY16BE) ? 255 : 65535);
|
|
s->bytestream += strlen(s->bytestream);
|
|
}
|
|
|
|
ptr = p->data[0];
|
|
linesize = p->linesize[0];
|
|
for(i=0;i<h;i++) {
|
|
memcpy(s->bytestream, ptr, n);
|
|
s->bytestream += n;
|
|
ptr += linesize;
|
|
}
|
|
|
|
if (avctx->pix_fmt == PIX_FMT_YUV420P) {
|
|
h >>= 1;
|
|
n >>= 1;
|
|
ptr1 = p->data[1];
|
|
ptr2 = p->data[2];
|
|
for(i=0;i<h;i++) {
|
|
memcpy(s->bytestream, ptr1, n);
|
|
s->bytestream += n;
|
|
memcpy(s->bytestream, ptr2, n);
|
|
s->bytestream += n;
|
|
ptr1 += p->linesize[1];
|
|
ptr2 += p->linesize[2];
|
|
}
|
|
}
|
|
return s->bytestream - s->bytestream_start;
|
|
}
|
|
|
|
static int pam_encode_frame(AVCodecContext *avctx, unsigned char *outbuf, int buf_size, void *data){
|
|
PNMContext *s = avctx->priv_data;
|
|
AVFrame *pict = data;
|
|
AVFrame * const p= (AVFrame*)&s->picture;
|
|
int i, h, w, n, linesize, depth, maxval;
|
|
const char *tuple_type;
|
|
uint8_t *ptr;
|
|
|
|
if(buf_size < avpicture_get_size(avctx->pix_fmt, avctx->width, avctx->height) + 200){
|
|
av_log(avctx, AV_LOG_ERROR, "encoded frame too large\n");
|
|
return -1;
|
|
}
|
|
|
|
*p = *pict;
|
|
p->pict_type= FF_I_TYPE;
|
|
p->key_frame= 1;
|
|
|
|
s->bytestream_start=
|
|
s->bytestream= outbuf;
|
|
s->bytestream_end= outbuf+buf_size;
|
|
|
|
h = avctx->height;
|
|
w = avctx->width;
|
|
switch(avctx->pix_fmt) {
|
|
case PIX_FMT_MONOWHITE:
|
|
n = (w + 7) >> 3;
|
|
depth = 1;
|
|
maxval = 1;
|
|
tuple_type = "BLACKANDWHITE";
|
|
break;
|
|
case PIX_FMT_GRAY8:
|
|
n = w;
|
|
depth = 1;
|
|
maxval = 255;
|
|
tuple_type = "GRAYSCALE";
|
|
break;
|
|
case PIX_FMT_RGB24:
|
|
n = w * 3;
|
|
depth = 3;
|
|
maxval = 255;
|
|
tuple_type = "RGB";
|
|
break;
|
|
case PIX_FMT_RGB32:
|
|
n = w * 4;
|
|
depth = 4;
|
|
maxval = 255;
|
|
tuple_type = "RGB_ALPHA";
|
|
break;
|
|
default:
|
|
return -1;
|
|
}
|
|
snprintf(s->bytestream, s->bytestream_end - s->bytestream,
|
|
"P7\nWIDTH %d\nHEIGHT %d\nDEPTH %d\nMAXVAL %d\nTUPLETYPE %s\nENDHDR\n",
|
|
w, h, depth, maxval, tuple_type);
|
|
s->bytestream += strlen(s->bytestream);
|
|
|
|
ptr = p->data[0];
|
|
linesize = p->linesize[0];
|
|
|
|
if (avctx->pix_fmt == PIX_FMT_RGB32) {
|
|
int j;
|
|
unsigned int v;
|
|
|
|
for(i=0;i<h;i++) {
|
|
for(j=0;j<w;j++) {
|
|
v = ((uint32_t *)ptr)[j];
|
|
*s->bytestream++ = v >> 16;
|
|
*s->bytestream++ = v >> 8;
|
|
*s->bytestream++ = v;
|
|
*s->bytestream++ = v >> 24;
|
|
}
|
|
ptr += linesize;
|
|
}
|
|
} else {
|
|
for(i=0;i<h;i++) {
|
|
memcpy(s->bytestream, ptr, n);
|
|
s->bytestream += n;
|
|
ptr += linesize;
|
|
}
|
|
}
|
|
return s->bytestream - s->bytestream_start;
|
|
}
|
|
|
|
#if 0
|
|
static int pnm_probe(AVProbeData *pd)
|
|
{
|
|
const char *p = pd->buf;
|
|
if (pd->buf_size >= 8 &&
|
|
p[0] == 'P' &&
|
|
p[1] >= '4' && p[1] <= '6' &&
|
|
pnm_space(p[2]) )
|
|
return AVPROBE_SCORE_MAX - 1; /* to permit pgmyuv probe */
|
|
else
|
|
return 0;
|
|
}
|
|
|
|
static int pgmyuv_probe(AVProbeData *pd)
|
|
{
|
|
if (match_ext(pd->filename, "pgmyuv"))
|
|
return AVPROBE_SCORE_MAX;
|
|
else
|
|
return 0;
|
|
}
|
|
|
|
static int pam_probe(AVProbeData *pd)
|
|
{
|
|
const char *p = pd->buf;
|
|
if (pd->buf_size >= 8 &&
|
|
p[0] == 'P' &&
|
|
p[1] == '7' &&
|
|
p[2] == '\n')
|
|
return AVPROBE_SCORE_MAX;
|
|
else
|
|
return 0;
|
|
}
|
|
#endif
|
|
|
|
#ifdef CONFIG_PNM_PARSER
|
|
static int pnm_parse(AVCodecParserContext *s,
|
|
AVCodecContext *avctx,
|
|
uint8_t **poutbuf, int *poutbuf_size,
|
|
const uint8_t *buf, int buf_size)
|
|
{
|
|
ParseContext *pc = s->priv_data;
|
|
PNMContext pnmctx;
|
|
int next;
|
|
|
|
for(; pc->overread>0; pc->overread--){
|
|
pc->buffer[pc->index++]= pc->buffer[pc->overread_index++];
|
|
}
|
|
retry:
|
|
if(pc->index){
|
|
pnmctx.bytestream_start=
|
|
pnmctx.bytestream= pc->buffer;
|
|
pnmctx.bytestream_end= pc->buffer + pc->index;
|
|
}else{
|
|
pnmctx.bytestream_start=
|
|
pnmctx.bytestream= (uint8_t *) buf; /* casts avoid warnings */
|
|
pnmctx.bytestream_end= (uint8_t *) buf + buf_size;
|
|
}
|
|
if(pnm_decode_header(avctx, &pnmctx) < 0){
|
|
if(pnmctx.bytestream < pnmctx.bytestream_end){
|
|
if(pc->index){
|
|
pc->index=0;
|
|
}else{
|
|
buf++;
|
|
buf_size--;
|
|
}
|
|
goto retry;
|
|
}
|
|
#if 0
|
|
if(pc->index && pc->index*2 + FF_INPUT_BUFFER_PADDING_SIZE < pc->buffer_size && buf_size > pc->index){
|
|
memcpy(pc->buffer + pc->index, buf, pc->index);
|
|
pc->index += pc->index;
|
|
buf += pc->index;
|
|
buf_size -= pc->index;
|
|
goto retry;
|
|
}
|
|
#endif
|
|
next= END_NOT_FOUND;
|
|
}else{
|
|
next= pnmctx.bytestream - pnmctx.bytestream_start
|
|
+ avpicture_get_size(avctx->pix_fmt, avctx->width, avctx->height);
|
|
if(pnmctx.bytestream_start!=buf)
|
|
next-= pc->index;
|
|
if(next > buf_size)
|
|
next= END_NOT_FOUND;
|
|
}
|
|
|
|
if(ff_combine_frame(pc, next, (uint8_t **)&buf, &buf_size)<0){
|
|
*poutbuf = NULL;
|
|
*poutbuf_size = 0;
|
|
return buf_size;
|
|
}
|
|
*poutbuf = (uint8_t *)buf;
|
|
*poutbuf_size = buf_size;
|
|
return next;
|
|
}
|
|
|
|
AVCodecParser pnm_parser = {
|
|
{ CODEC_ID_PGM, CODEC_ID_PGMYUV, CODEC_ID_PPM, CODEC_ID_PBM, CODEC_ID_PAM},
|
|
sizeof(ParseContext),
|
|
NULL,
|
|
pnm_parse,
|
|
ff_parse_close,
|
|
};
|
|
#endif /* CONFIG_PNM_PARSER */
|
|
|
|
#ifdef CONFIG_PGM_ENCODER
|
|
AVCodec pgm_encoder = {
|
|
"pgm",
|
|
CODEC_TYPE_VIDEO,
|
|
CODEC_ID_PGM,
|
|
sizeof(PNMContext),
|
|
common_init,
|
|
pnm_encode_frame,
|
|
NULL, //encode_end,
|
|
pnm_decode_frame,
|
|
.pix_fmts= (enum PixelFormat[]){PIX_FMT_GRAY8, PIX_FMT_GRAY16BE, -1},
|
|
};
|
|
#endif // CONFIG_PGM_ENCODER
|
|
|
|
#ifdef CONFIG_PGMYUV_ENCODER
|
|
AVCodec pgmyuv_encoder = {
|
|
"pgmyuv",
|
|
CODEC_TYPE_VIDEO,
|
|
CODEC_ID_PGMYUV,
|
|
sizeof(PNMContext),
|
|
common_init,
|
|
pnm_encode_frame,
|
|
NULL, //encode_end,
|
|
pnm_decode_frame,
|
|
.pix_fmts= (enum PixelFormat[]){PIX_FMT_YUV420P, -1},
|
|
};
|
|
#endif // CONFIG_PGMYUV_ENCODER
|
|
|
|
#ifdef CONFIG_PPM_ENCODER
|
|
AVCodec ppm_encoder = {
|
|
"ppm",
|
|
CODEC_TYPE_VIDEO,
|
|
CODEC_ID_PPM,
|
|
sizeof(PNMContext),
|
|
common_init,
|
|
pnm_encode_frame,
|
|
NULL, //encode_end,
|
|
pnm_decode_frame,
|
|
.pix_fmts= (enum PixelFormat[]){PIX_FMT_RGB24, -1},
|
|
};
|
|
#endif // CONFIG_PPM_ENCODER
|
|
|
|
#ifdef CONFIG_PBM_ENCODER
|
|
AVCodec pbm_encoder = {
|
|
"pbm",
|
|
CODEC_TYPE_VIDEO,
|
|
CODEC_ID_PBM,
|
|
sizeof(PNMContext),
|
|
common_init,
|
|
pnm_encode_frame,
|
|
NULL, //encode_end,
|
|
pnm_decode_frame,
|
|
.pix_fmts= (enum PixelFormat[]){PIX_FMT_MONOWHITE, -1},
|
|
};
|
|
#endif // CONFIG_PBM_ENCODER
|
|
|
|
#ifdef CONFIG_PAM_ENCODER
|
|
AVCodec pam_encoder = {
|
|
"pam",
|
|
CODEC_TYPE_VIDEO,
|
|
CODEC_ID_PAM,
|
|
sizeof(PNMContext),
|
|
common_init,
|
|
pam_encode_frame,
|
|
NULL, //encode_end,
|
|
pnm_decode_frame,
|
|
.pix_fmts= (enum PixelFormat[]){PIX_FMT_RGB24, PIX_FMT_RGB32, PIX_FMT_GRAY8, PIX_FMT_MONOWHITE, -1},
|
|
};
|
|
#endif // CONFIG_PAM_ENCODER
|