+JNIEXPORT jint JNICALL Java_ai_suanzi_rtmpclient_Ffmpeg_process (JNIEnv *env, jobject obj, jbyteArray yuv){
+ int ret;
+ int enc_got_frame=0;
+ int i=0;
+
+ LOGE(" process data - ffmpeg");
+ pFrameYUV = av_frame_alloc();
+ uint8_t *out_buffer = (uint8_t *)av_malloc(avpicture_get_size(AV_PIX_FMT_YUV420P, pCodecCtx->width, pCodecCtx->height));
+ avpicture_fill((AVPicture *)pFrameYUV, out_buffer, AV_PIX_FMT_YUV420P, pCodecCtx->width, pCodecCtx->height);
+
+ //安卓摄像头数据为NV21格式,此处将其转换为YUV420P格式
+ jbyte* in= (jbyte*)env->GetByteArrayElements(yuv,0);
+ memcpy(pFrameYUV->data[0],in,y_length);
+ for(i=0;i<uv_length;i++)
+ {
+ *(pFrameYUV->data[2]+i)=*(in+y_length+i*2);
+ *(pFrameYUV->data[1]+i)=*(in+y_length+i*2+1);
+ }
+
+ pFrameYUV->format = AV_PIX_FMT_YUV420P;
+ pFrameYUV->width = yuv_width;
+ pFrameYUV->height = yuv_height;
+
+ enc_pkt.data = NULL;
+ enc_pkt.size = 0;
+ av_init_packet(&enc_pkt);
+ ret = avcodec_encode_video2(pCodecCtx, &enc_pkt, pFrameYUV, &enc_got_frame);
+ av_frame_free(&pFrameYUV);
+
+ if (enc_got_frame == 1){
+ //LOGE("Succeed to encode frame: %5d\tsize:%5d\n", framecnt, enc_pkt.size);
+ framecnt++;
+ enc_pkt.stream_index = video_st->index;
+
+ //Write PTS
+ AVRational time_base = ofmt_ctx->streams[0]->time_base;//{ 1, 1000 };
+ AVRational r_framerate1 = {60, 2 };//{ 50, 2 };
+ AVRational time_base_q = { 1, AV_TIME_BASE };
+ //Duration between 2 frames (us)
+ int64_t calc_duration = (double)(AV_TIME_BASE)*(1 / av_q2d(r_framerate1)); //内部时间戳
+ //Parameters
+ //enc_pkt.pts = (double)(framecnt*calc_duration)*(double)(av_q2d(time_base_q)) / (double)(av_q2d(time_base));
+ enc_pkt.pts = av_rescale_q(framecnt*calc_duration, time_base_q, time_base);
+ enc_pkt.dts = enc_pkt.pts;
+ enc_pkt.duration = av_rescale_q(calc_duration, time_base_q, time_base); //(double)(calc_duration)*(double)(av_q2d(time_base_q)) / (double)(av_q2d(time_base));
+ enc_pkt.pos = -1;
+
+ //Delay
+ int64_t pts_time = av_rescale_q(enc_pkt.dts, time_base, time_base_q);
+ int64_t now_time = av_gettime() - start_time;
+ if (pts_time > now_time)
+ av_usleep(pts_time - now_time);
+
+ ret = av_interleaved_write_frame(ofmt_ctx, &enc_pkt);
+ av_free_packet(&enc_pkt);
+ }
+ return 0;
+}
+
+JNIEXPORT jint JNICALL Java_ai_suanzi_rtmpclient_Ffmpeg_play (JNIEnv *env, jobject obj, jobject surface, jstring fname){
+
+
+
+
+
+ LOGE("###### video play #####");
+ // char * file_name = "/storage/emulated/0/Movies/big_buck_bunny_720p_10mb.mp4";
+ const char * file_name = env->GetStringUTFChars(fname, 0);
+
+ av_register_all();
+ avdevice_register_all();
+
+
+ AVFormatContext * pFormatCtx = avformat_alloc_context();
+
+
+//////////
+ av_log_set_callback(custom_log);
+
+ AVInputFormat *ifmt=av_find_input_format("video4linux2");
+ LOGE("===%s===", ifmt->name);
+ if(avformat_open_input(&pFormatCtx,"/dev/video0",ifmt,NULL)!=0){
+ LOGE("Couldn't open file:\n");
+ return -1; // Couldn't open file
+ }
+
+
+///////////
+
+/*
+ // Open video file
+ if(avformat_open_input(&pFormatCtx, file_name, NULL, NULL)!=0) {
+
+ LOGE("Couldn't open file:%s\n", file_name);
+ return -1; // Couldn't open file
+ }
+*/
+ // Retrieve stream information
+ if(avformat_find_stream_info(pFormatCtx, NULL)<0) {
+ LOGE("Couldn't find stream information.");
+ return -1;
+ }
+
+ // Find the first video stream
+ int videoStream = -1, i;
+ for (i = 0; i < pFormatCtx->nb_streams; i++) {
+ if (pFormatCtx->streams[i]->codec->codec_type == AVMEDIA_TYPE_VIDEO
+ && videoStream < 0) {
+ videoStream = i;
+ }
+ }
+ if(videoStream==-1) {
+ LOGE("Didn't find a video stream.");
+ return -1; // Didn't find a video stream
+ }
+
+ // Get a pointer to the codec context for the video stream
+ AVCodecContext * pCodecCtx = pFormatCtx->streams[videoStream]->codec;
+ LOGE("============= %d ========",__LINE__);
+ // Find the decoder for the video stream
+ AVCodec * pCodec = avcodec_find_decoder(pCodecCtx->codec_id);
+ if(pCodec==NULL) {
+ LOGE("Codec not found.");
+ return -1; // Codec not found
+ }
+
+ if(avcodec_open2(pCodecCtx, pCodec, NULL) < 0) {
+ LOGE("Could not open codec.");
+ return -1; // Could not open codec
+ }
+
+ // 获取native window
+ ANativeWindow* nativeWindow = ANativeWindow_fromSurface(env, surface);
+
+ // 获取视频宽高
+ int videoWidth = pCodecCtx->width;
+ int videoHeight = pCodecCtx->height;
+
+ // 设置native window的buffer大小,可自动拉伸
+ ANativeWindow_setBuffersGeometry(nativeWindow, videoWidth, videoHeight, WINDOW_FORMAT_RGBA_8888);
+ ANativeWindow_Buffer windowBuffer;
+
+ if(avcodec_open2(pCodecCtx, pCodec, NULL)<0) {
+ LOGE("Could not open codec.");
+ return -1; // Could not open codec
+ }
+
+ LOGE("stream format:%s", pFormatCtx->iformat->name);
+ LOGE("duration :%lld", (pFormatCtx->duration) / 1000000);
+ LOGE("Width, Height:%d x %d", pCodecCtx->width, pCodecCtx->height);
+ LOGE("Decoder name:%s", pCodec->name);
+
+ // Allocate video frame
+ AVFrame * pFrame = av_frame_alloc();
+
+ // 用于渲染
+ AVFrame * pFrameRGBA = av_frame_alloc();
+ if(pFrameRGBA == NULL || pFrame == NULL) {
+ LOGE("Could not allocate video frame.");
+ return -1;
+ }
+
+ // Determine required buffer size and allocate buffer
+ int numBytes=av_image_get_buffer_size(AV_PIX_FMT_RGBA, pCodecCtx->width, pCodecCtx->height, 1);
+ uint8_t * buffer=(uint8_t *)av_malloc(numBytes*sizeof(uint8_t));
+ av_image_fill_arrays(pFrameRGBA->data, pFrameRGBA->linesize, buffer, AV_PIX_FMT_RGBA,
+ pCodecCtx->width, pCodecCtx->height, 1);
+
+ // 由于解码出来的帧格式不是RGBA的,在渲染之前需要进行格式转换
+ struct SwsContext *sws_ctx = sws_getContext(pCodecCtx->width,
+ pCodecCtx->height,
+ pCodecCtx->pix_fmt,
+ pCodecCtx->width,
+ pCodecCtx->height,
+ AV_PIX_FMT_RGBA,
+ SWS_BILINEAR,
+ NULL,
+ NULL,
+ NULL);
+
+ int frameFinished;
+ AVPacket packet;
+ while(av_read_frame(pFormatCtx, &packet)>=0) {
+ // Is this a packet from the video stream?
+ if(packet.stream_index==videoStream) {
+
+ // Decode video frame
+ avcodec_decode_video2(pCodecCtx, pFrame, &frameFinished, &packet);
+
+ // 并不是decode一次就可解码出一帧
+ if (frameFinished) {
+
+ // lock native window buffer
+ ANativeWindow_lock(nativeWindow, &windowBuffer, 0);
+
+ // 格式转换
+ sws_scale(sws_ctx, (uint8_t const * const *)pFrame->data,
+ pFrame->linesize, 0, pCodecCtx->height,
+ pFrameRGBA->data, pFrameRGBA->linesize);
+
+ // 获取stride
+ uint8_t * dst = (uint8_t*) windowBuffer.bits;
+ int dstStride = windowBuffer.stride * 4;
+ uint8_t * src = (uint8_t*) (pFrameRGBA->data[0]);
+ int srcStride = pFrameRGBA->linesize[0];
+
+ // 由于window的stride和帧的stride不同,因此需要逐行复制
+ int h;
+ for (h = 0; h < videoHeight; h++) {
+ memcpy(dst + h * dstStride, src + h * srcStride, srcStride);
+ }
+
+ ANativeWindow_unlockAndPost(nativeWindow);
+ }
+
+ }
+ av_packet_unref(&packet);
+ }
+
+ av_free(buffer);
+ av_free(pFrameRGBA);
+
+ // Free the YUV frame
+ av_free(pFrame);
+
+ // Close the codecs
+ avcodec_close(pCodecCtx);
+
+ // Close the video file
+ avformat_close_input(&pFormatCtx);
+
+ env->ReleaseStringUTFChars(fname, file_name);