diff options
author | dec05eba <dec05eba@protonmail.com> | 2020-04-01 19:25:16 +0200 |
---|---|---|
committer | dec05eba <dec05eba@protonmail.com> | 2020-06-21 14:41:11 +0200 |
commit | 2fcd3ee3e5dbf65841d5e457aa1a558fee471433 (patch) | |
tree | 3858678f2954b91f092ecbfe6b60c0a5f941585b /src/main.cpp | |
parent | 5e60805c3681a705240d49a0c5c589ff65d7aa83 (diff) |
Add audio support
Diffstat (limited to 'src/main.cpp')
-rw-r--r-- | src/main.cpp | 262 |
1 files changed, 211 insertions, 51 deletions
diff --git a/src/main.cpp b/src/main.cpp index eb32c8a..27dd225 100644 --- a/src/main.cpp +++ b/src/main.cpp @@ -3,8 +3,13 @@ #include <stdlib.h> #include <string> #include <vector> +#include <thread> +#include <mutex> + #include <unistd.h> +#include "../include/sound.hpp" + #define GLX_GLXEXT_PROTOTYPES #include <GL/glew.h> #include <GL/glx.h> @@ -19,6 +24,8 @@ extern "C" { #include <libavformat/avformat.h> #include <libavutil/hwcontext.h> #include <libavutil/hwcontext_cuda.h> +#include <libavutil/opt.h> +#include <libswresample/swresample.h> } #include <cudaGL.h> @@ -28,6 +35,14 @@ extern "C" { //#include <CL/cl.h> +static char av_error_buffer[AV_ERROR_MAX_STRING_SIZE]; + +static char* av_error_to_string(int err) { + if(av_strerror(err, av_error_buffer, sizeof(av_error_buffer) < 0)) + strcpy(av_error_buffer, "Unknown error"); + return av_error_buffer; +} + struct ScopedGLXFBConfig { ~ScopedGLXFBConfig() { if (configs) @@ -236,7 +251,8 @@ std::vector<std::string> get_hardware_acceleration_device_names() { } static void receive_frames(AVCodecContext *av_codec_context, AVStream *stream, - AVFormatContext *av_format_context) { + AVFormatContext *av_format_context, + std::mutex &write_output_mutex) { AVPacket av_packet; av_init_packet(&av_packet); for (;;) { @@ -244,14 +260,17 @@ static void receive_frames(AVCodecContext *av_codec_context, AVStream *stream, av_packet.size = 0; int res = avcodec_receive_packet(av_codec_context, &av_packet); if (res == 0) { // we have a packet, send the packet to the muxer + assert(av_packet.stream_index == stream->id); av_packet_rescale_ts(&av_packet, av_codec_context->time_base, stream->time_base); av_packet.stream_index = stream->index; // Write the encoded video frame to disk // av_write_frame(av_format_context, &av_packet) // write(STDOUT_FILENO, av_packet.data, av_packet.size) - if (av_write_frame(av_format_context, &av_packet) < 0) { - fprintf(stderr, "Error: Failed to write frame to muxer\n"); + std::lock_guard<std::mutex> lock(write_output_mutex); + int ret = av_write_frame(av_format_context, &av_packet); + if(ret < 0) { + fprintf(stderr, "Error: Failed to write video frame to muxer, reason: %s (%d)\n", av_error_to_string(ret), ret); } av_packet_unref(&av_packet); } else if (res == AVERROR(EAGAIN)) { // we have no packet @@ -268,7 +287,46 @@ static void receive_frames(AVCodecContext *av_codec_context, AVStream *stream, //av_packet_unref(&av_packet); } -static AVStream *add_stream(AVFormatContext *av_format_context, AVCodec **codec, +static AVStream *add_audio_stream(AVFormatContext *av_format_context, AVCodec **codec, + enum AVCodecID codec_id) { + *codec = avcodec_find_encoder(AV_CODEC_ID_AAC); + if (!*codec) { + fprintf( + stderr, + "Error: Could not find aac encoder\n"); + exit(1); + } + + AVStream *stream = avformat_new_stream(av_format_context, *codec); + if (!stream) { + fprintf(stderr, "Error: Could not allocate stream\n"); + exit(1); + } + stream->id = av_format_context->nb_streams - 1; + fprintf(stderr, "audio stream id: %d\n", stream->id); + AVCodecContext *codec_context = stream->codec; + + assert((*codec)->type == AVMEDIA_TYPE_AUDIO); + /* + codec_context->sample_fmt = (*codec)->sample_fmts + ? (*codec)->sample_fmts[0] + : AV_SAMPLE_FMT_FLTP; + */ + codec_context->codec_id = AV_CODEC_ID_AAC; + codec_context->sample_fmt = AV_SAMPLE_FMT_FLTP; + //codec_context->bit_rate = 64000; + codec_context->sample_rate = 48000; + codec_context->channel_layout = AV_CH_LAYOUT_STEREO; + codec_context->channels = 2; + + // Some formats want stream headers to be seperate + //if (av_format_context->oformat->flags & AVFMT_GLOBALHEADER) + // av_format_context->flags |= AV_CODEC_FLAG_GLOBAL_HEADER; + + return stream; +} + +static AVStream *add_video_stream(AVFormatContext *av_format_context, AVCodec **codec, enum AVCodecID codec_id, const WindowPixmap &window_pixmap, int fps) { @@ -280,8 +338,7 @@ static AVStream *add_stream(AVFormatContext *av_format_context, AVCodec **codec, if (!*codec) { fprintf( stderr, - "Error: Could not find h264_nvenc or nvenc_h264 encoder for %s\n", - avcodec_get_name(codec_id)); + "Error: Could not find h264_nvenc or nvenc_h264 encoder\n"); exit(1); } @@ -291,48 +348,33 @@ static AVStream *add_stream(AVFormatContext *av_format_context, AVCodec **codec, exit(1); } stream->id = av_format_context->nb_streams - 1; + fprintf(stderr, "video stream id: %d\n", stream->id); AVCodecContext *codec_context = stream->codec; - switch ((*codec)->type) { - case AVMEDIA_TYPE_AUDIO: { - codec_context->sample_fmt = (*codec)->sample_fmts - ? (*codec)->sample_fmts[0] - : AV_SAMPLE_FMT_FLTP; - codec_context->bit_rate = 64000; - codec_context->sample_rate = 44100; - codec_context->channels = 2; - break; - } - case AVMEDIA_TYPE_VIDEO: { - codec_context->codec_id = codec_id; - // TODO: Scale bitrate by resolution. For 4k, 8000000 is a better value - codec_context->bit_rate = 5000000; - // Resolution must be a multiple of two - codec_context->width = window_pixmap.texture_width & ~1; - codec_context->height = window_pixmap.texture_height & ~1; - // Timebase: This is the fundamental unit of time (in seconds) in terms - // of which frame timestamps are represented. For fixed-fps content, - // timebase should be 1/framerate and timestamp increments should be - // identical to 1 - codec_context->time_base.num = 1; - codec_context->time_base.den = fps; - // codec_context->framerate.num = 60; - // codec_context->framerate.den = 1; - codec_context->sample_aspect_ratio.num = 1; - codec_context->sample_aspect_ratio.den = 1; - codec_context->gop_size = - 32; // Emit one intra frame every 32 frames at most - codec_context->pix_fmt = AV_PIX_FMT_CUDA; - if (codec_context->codec_id == AV_CODEC_ID_MPEG1VIDEO) - codec_context->mb_decision = 2; - - // stream->time_base = codec_context->time_base; - // codec_context->ticks_per_frame = 30; - break; - } - default: - break; - } + assert((*codec)->type == AVMEDIA_TYPE_VIDEO); + codec_context->codec_id = (*codec)->id; + fprintf(stderr, "codec id: %d\n", (*codec)->id); + codec_context->width = window_pixmap.texture_width & ~1; + codec_context->height = window_pixmap.texture_height & ~1; + codec_context->bit_rate = codec_context->width * codec_context->height; //5000000; + // Timebase: This is the fundamental unit of time (in seconds) in terms + // of which frame timestamps are represented. For fixed-fps content, + // timebase should be 1/framerate and timestamp increments should be + // identical to 1 + codec_context->time_base.num = 1; + codec_context->time_base.den = fps; + // codec_context->framerate.num = 60; + // codec_context->framerate.den = 1; + codec_context->sample_aspect_ratio.num = 1; + codec_context->sample_aspect_ratio.den = 1; + codec_context->gop_size = + 32; // Emit one intra frame every 32 frames at most + codec_context->pix_fmt = AV_PIX_FMT_CUDA; + if (codec_context->codec_id == AV_CODEC_ID_MPEG1VIDEO) + codec_context->mb_decision = 2; + + // stream->time_base = codec_context->time_base; + // codec_context->ticks_per_frame = 30; // Some formats want stream headers to be seperate if (av_format_context->oformat->flags & AVFMT_GLOBALHEADER) @@ -341,6 +383,36 @@ static AVStream *add_stream(AVFormatContext *av_format_context, AVCodec **codec, return stream; } +static AVFrame* open_audio(AVCodec *codec, AVStream *stream) { + int ret; + AVCodecContext *codec_context = stream->codec; + + ret = avcodec_open2(codec_context, codec, nullptr); + if(ret < 0) { + fprintf(stderr, "failed to open codec, reason: %s\n", av_error_to_string(ret)); + exit(1); + } + + AVFrame *frame = av_frame_alloc(); + if(!frame) { + fprintf(stderr, "failed to allocate audio frame\n"); + exit(1); + } + + frame->nb_samples = codec_context->frame_size; + frame->format = codec_context->sample_fmt; + frame->channels = codec_context->channels; + frame->channel_layout = codec_context->channel_layout; + + ret = av_frame_get_buffer(frame, 0); + if(ret < 0) { + fprintf(stderr, "failed to allocate audio data buffers, reason: %s\n", av_error_to_string(ret)); + exit(1); + } + + return frame; +} + static void open_video(AVCodec *codec, AVStream *stream, WindowPixmap &window_pixmap, AVBufferRef **device_ctx, CUgraphicsResource *cuda_graphics_resource) { @@ -528,15 +600,24 @@ int main(int argc, char **argv) { } AVOutputFormat *output_format = av_format_context->oformat; + AVCodec *video_codec; AVStream *video_stream = - add_stream(av_format_context, &video_codec, output_format->video_codec, + add_video_stream(av_format_context, &video_codec, output_format->video_codec, window_pixmap, fps); if (!video_stream) { fprintf(stderr, "Error: Failed to create video stream\n"); return 1; } + AVCodec *audio_codec; + AVStream *audio_stream = + add_audio_stream(av_format_context, &audio_codec, output_format->audio_codec); + if (!audio_stream) { + fprintf(stderr, "Error: Failed to create audio stream\n"); + return 1; + } + if (cuInit(0) < 0) { fprintf(stderr, "Error: cuInit failed\n"); return {}; @@ -547,7 +628,9 @@ int main(int argc, char **argv) { open_video(video_codec, video_stream, window_pixmap, &device_ctx, &cuda_graphics_resource); - av_dump_format(av_format_context, 0, filename, 1); + AVFrame *audio_frame = open_audio(audio_codec, audio_stream); + + //av_dump_format(av_format_context, 0, filename, 1); if (!(output_format->flags & AVFMT_NOFILE)) { int ret = avio_open(&av_format_context->pb, filename, AVIO_FLAG_WRITE); @@ -635,6 +718,69 @@ int main(int argc, char **argv) { int window_width = xwa.width; int window_height = xwa.height; + SoundDevice sound_device; + if(sound_device_get_by_name(&sound_device, "pulse", audio_stream->codec->channels, audio_stream->codec->frame_size) != 0) { + fprintf(stderr, "failed to get 'pulse' sound device\n"); + exit(1); + } + + int audio_buffer_size = av_samples_get_buffer_size(NULL, audio_stream->codec->channels, audio_stream->codec->frame_size, audio_stream->codec->sample_fmt, 1); + uint8_t *audio_frame_buf = (uint8_t *)av_malloc(audio_buffer_size); + avcodec_fill_audio_frame(audio_frame, audio_stream->codec->channels, audio_stream->codec->sample_fmt, (const uint8_t*)audio_frame_buf, audio_buffer_size, 1); + + AVPacket audio_packet; + av_new_packet(&audio_packet, audio_buffer_size); + + std::mutex write_output_mutex; + + bool running = true; + std::thread audio_thread([&running](AVFormatContext *av_format_context, AVStream *audio_stream, AVPacket *audio_packet, uint8_t *audio_frame_buf, SoundDevice *sound_device, AVFrame *audio_frame, std::mutex *write_output_mutex) { + SwrContext *swr = swr_alloc(); + if(!swr) { + fprintf(stderr, "Failed to create SwrContext\n"); + exit(1); + } + av_opt_set_int(swr, "in_channel_layout", audio_stream->codec->channel_layout, 0); + av_opt_set_int(swr, "out_channel_layout", audio_stream->codec->channel_layout, 0); + av_opt_set_int(swr, "in_sample_rate", audio_stream->codec->sample_rate, 0); + av_opt_set_int(swr, "out_sample_rate", audio_stream->codec->sample_rate, 0); + av_opt_set_sample_fmt(swr, "in_sample_fmt", AV_SAMPLE_FMT_S16, 0); + av_opt_set_sample_fmt(swr, "out_sample_fmt", AV_SAMPLE_FMT_FLTP, 0); + swr_init(swr); + + while(running) { + void *sound_buffer; + int sound_buffer_size = sound_device_read_next_chunk(sound_device, &sound_buffer); + if(sound_buffer_size >= 0) { + // TODO: Instead of converting audio, get float audio from alsa. Or does alsa do conversion internally to get this format? + swr_convert(swr, &audio_frame_buf, audio_frame->nb_samples, (const uint8_t**)&sound_buffer, sound_buffer_size); + audio_frame->extended_data = &audio_frame_buf; + // TODO: Fix this. Warning from ffmpeg: + // Timestamps are unset in a packet for stream 1. This is deprecated and will stop working in the future. Fix your code to set the timestamps properly + //audio_frame->pts=audio_frame_index*100; + //++audio_frame_index; + + int got_frame = 0; + int ret = avcodec_encode_audio2(audio_stream->codec, audio_packet, audio_frame, &got_frame); + if(ret < 0){ + printf("Failed to encode!\n"); + break; + } + if (got_frame==1){ + //printf("Succeed to encode 1 frame! \tsize:%5d\n",pkt.size); + audio_packet->stream_index = audio_stream->index; + std::lock_guard<std::mutex> lock(*write_output_mutex); + ret = av_write_frame(av_format_context, audio_packet); + av_free_packet(audio_packet); + } + } else { + fprintf(stderr, "failed to read sound from device, error: %d\n", sound_buffer_size); + } + } + + swr_free(&swr); + }, av_format_context, audio_stream, &audio_packet, audio_frame_buf, &sound_device, audio_frame, &write_output_mutex); + XEvent e; while (!glfwWindowShouldClose(window)) { glClear(GL_COLOR_BUFFER_BIT); @@ -719,7 +865,7 @@ int main(int argc, char **argv) { "Error: cuGraphicsGLRegisterImage failed, error %s, texture " "id: %u\n", err_str, window_pixmap.target_texture_id); - exit(1); + break; } res = cuGraphicsResourceSetMapFlags( @@ -730,7 +876,7 @@ int main(int argc, char **argv) { av_frame_unref(frame); if (av_hwframe_get_buffer(video_stream->codec->hw_frames_ctx, frame, 0) < 0) { fprintf(stderr, "Error: av_hwframe_get_buffer failed\n"); - exit(1); + break; } } @@ -741,7 +887,7 @@ int main(int argc, char **argv) { frame_count += 1; if (avcodec_send_frame(video_stream->codec, frame) >= 0) { receive_frames(video_stream->codec, video_stream, - av_format_context); + av_format_context, write_output_mutex); } else { fprintf(stderr, "Error: avcodec_send_frame failed\n"); } @@ -752,6 +898,20 @@ int main(int argc, char **argv) { usleep(5000); } + running = false; + audio_thread.join(); + + sound_device_close(&sound_device); + + //Flush Encoder + #if 0 + ret = flush_encoder(pFormatCtx,0); + if (ret < 0) { + printf("Flushing encoder failed\n"); + return -1; + } + #endif + if (av_write_trailer(av_format_context) != 0) { fprintf(stderr, "Failed to write trailer\n"); } |