package main import ( "fmt" "image" "unsafe" ) // #cgo pkg-config: libavcodec libavutil libswscale // #include // #include // #include import "C" func frameData(frame *C.AVFrame) **C.uint8_t { return (**C.uint8_t)(unsafe.Pointer(&frame.data[0])) } func frameLineSize(frame *C.AVFrame) *C.int { return (*C.int)(unsafe.Pointer(&frame.linesize[0])) } // h264Decoder is a wrapper around ffmpeg's H264 decoder. type h264Decoder struct { codecCtx *C.AVCodecContext srcFrame *C.AVFrame swsCtx *C.struct_SwsContext dstFrame *C.AVFrame dstFramePtr []uint8 } // newH264Decoder allocates a new h264Decoder. func newH264Decoder() (*h264Decoder, error) { codec := C.avcodec_find_decoder(C.AV_CODEC_ID_H264) if codec == nil { return nil, fmt.Errorf("avcodec_find_decoder() failed") } codecCtx := C.avcodec_alloc_context3(codec) if codecCtx == nil { return nil, fmt.Errorf("avcodec_alloc_context3() failed") } res := C.avcodec_open2(codecCtx, codec, nil) if res < 0 { C.avcodec_close(codecCtx) return nil, fmt.Errorf("avcodec_open2() failed") } srcFrame := C.av_frame_alloc() if srcFrame == nil { C.avcodec_close(codecCtx) return nil, fmt.Errorf("av_frame_alloc() failed") } return &h264Decoder{ codecCtx: codecCtx, srcFrame: srcFrame, }, nil } // close closes the decoder. func (d *h264Decoder) close() { if d.dstFrame != nil { C.av_frame_free(&d.dstFrame) } if d.swsCtx != nil { C.sws_freeContext(d.swsCtx) } C.av_frame_free(&d.srcFrame) C.avcodec_close(d.codecCtx) } func (d *h264Decoder) decode(nalu []byte) (image.Image, error) { nalu = append([]uint8{0x00, 0x00, 0x00, 0x01}, []uint8(nalu)...) // send frame to decoder var avPacket C.AVPacket avPacket.data = (*C.uint8_t)(C.CBytes(nalu)) defer C.free(unsafe.Pointer(avPacket.data)) avPacket.size = C.int(len(nalu)) res := C.avcodec_send_packet(d.codecCtx, &avPacket) if res < 0 { return nil, nil } // receive frame if available res = C.avcodec_receive_frame(d.codecCtx, d.srcFrame) if res < 0 { return nil, nil } // if frame size has changed, allocate needed objects if d.dstFrame == nil || d.dstFrame.width != d.srcFrame.width || d.dstFrame.height != d.srcFrame.height { if d.dstFrame != nil { C.av_frame_free(&d.dstFrame) } if d.swsCtx != nil { C.sws_freeContext(d.swsCtx) } d.dstFrame = C.av_frame_alloc() d.dstFrame.format = C.AV_PIX_FMT_RGBA d.dstFrame.width = d.srcFrame.width d.dstFrame.height = d.srcFrame.height d.dstFrame.color_range = C.AVCOL_RANGE_JPEG res = C.av_frame_get_buffer(d.dstFrame, 1) if res < 0 { return nil, fmt.Errorf("av_frame_get_buffer() err") } d.swsCtx = C.sws_getContext(d.srcFrame.width, d.srcFrame.height, C.AV_PIX_FMT_YUV420P, d.dstFrame.width, d.dstFrame.height, (int32)(d.dstFrame.format), C.SWS_BILINEAR, nil, nil, nil) if d.swsCtx == nil { return nil, fmt.Errorf("sws_getContext() err") } dstFrameSize := C.av_image_get_buffer_size((int32)(d.dstFrame.format), d.dstFrame.width, d.dstFrame.height, 1) d.dstFramePtr = (*[1 << 30]uint8)(unsafe.Pointer(d.dstFrame.data[0]))[:dstFrameSize:dstFrameSize] } // convert frame from YUV420 to RGB res = C.sws_scale(d.swsCtx, frameData(d.srcFrame), frameLineSize(d.srcFrame), 0, d.srcFrame.height, frameData(d.dstFrame), frameLineSize(d.dstFrame)) if res < 0 { return nil, fmt.Errorf("sws_scale() err") } // embed frame into an image.Image return &image.RGBA{ Pix: d.dstFramePtr, Stride: 4 * (int)(d.dstFrame.width), Rect: image.Rectangle{ Max: image.Point{(int)(d.dstFrame.width), (int)(d.dstFrame.height)}, }, }, nil }