312 lines
11 KiB
Objective-C
312 lines
11 KiB
Objective-C
#include <kinc/video.h>
|
|
|
|
#import <AVFoundation/AVFoundation.h>
|
|
#include <kinc/audio1/audio.h>
|
|
#include <kinc/graphics4/texture.h>
|
|
#include <kinc/io/filereader.h>
|
|
#include <kinc/log.h>
|
|
#include <kinc/system.h>
|
|
#include <stdio.h>
|
|
#include <stdlib.h>
|
|
#include <string.h>
|
|
|
|
extern const char *iphonegetresourcepath(void);
|
|
extern const char *macgetresourcepath(void);
|
|
|
|
void kinc_internal_video_sound_stream_init(kinc_internal_video_sound_stream_t *stream, int channel_count, int frequency) {
|
|
stream->bufferSize = 1024 * 100;
|
|
stream->bufferReadPosition = 0;
|
|
stream->bufferWritePosition = 0;
|
|
stream->read = 0;
|
|
stream->written = 0;
|
|
stream->buffer = (float *)malloc(stream->bufferSize * sizeof(float));
|
|
}
|
|
|
|
void kinc_internal_video_sound_stream_destroy(kinc_internal_video_sound_stream_t *stream) {
|
|
free(stream->buffer);
|
|
}
|
|
|
|
void kinc_internal_video_sound_stream_insert_data(kinc_internal_video_sound_stream_t *stream, float *data, int sample_count) {
|
|
for (int i = 0; i < sample_count; ++i) {
|
|
float value = data[i]; // / 32767.0;
|
|
stream->buffer[stream->bufferWritePosition++] = value;
|
|
++stream->written;
|
|
if (stream->bufferWritePosition >= stream->bufferSize) {
|
|
stream->bufferWritePosition = 0;
|
|
}
|
|
}
|
|
}
|
|
|
|
static float samples[2] = {0};
|
|
|
|
float *kinc_internal_video_sound_stream_next_frame(kinc_internal_video_sound_stream_t *stream) {
|
|
++stream->read;
|
|
if (stream->written <= stream->read) {
|
|
kinc_log(KINC_LOG_LEVEL_WARNING, "Out of audio\n");
|
|
return 0;
|
|
}
|
|
|
|
if (stream->bufferReadPosition >= stream->bufferSize) {
|
|
stream->bufferReadPosition = 0;
|
|
kinc_log(KINC_LOG_LEVEL_INFO, "buffer read back - %i\n", (int)(stream->written - stream->read));
|
|
}
|
|
samples[0] = stream->buffer[stream->bufferReadPosition++];
|
|
|
|
if (stream->bufferReadPosition >= stream->bufferSize) {
|
|
stream->bufferReadPosition = 0;
|
|
kinc_log(KINC_LOG_LEVEL_INFO, "buffer read back - %i\n", (int)(stream->written - stream->read));
|
|
}
|
|
samples[1] = stream->buffer[stream->bufferReadPosition++];
|
|
|
|
return samples;
|
|
}
|
|
|
|
bool kinc_internal_video_sound_stream_ended(kinc_internal_video_sound_stream_t *stream) {
|
|
return false;
|
|
}
|
|
|
|
static void load(kinc_video_t *video, double startTime) {
|
|
video->impl.videoStart = startTime;
|
|
AVURLAsset *asset = [[AVURLAsset alloc] initWithURL:video->impl.url options:nil];
|
|
video->impl.videoAsset = asset;
|
|
|
|
video->impl.duration = [asset duration].value / [asset duration].timescale;
|
|
|
|
AVAssetTrack *videoTrack = [[asset tracksWithMediaType:AVMediaTypeVideo] objectAtIndex:0];
|
|
NSDictionary *videoOutputSettings =
|
|
[NSDictionary dictionaryWithObjectsAndKeys:[NSNumber numberWithInt:kCVPixelFormatType_32BGRA], kCVPixelBufferPixelFormatTypeKey, nil];
|
|
AVAssetReaderTrackOutput *videoOutput = [AVAssetReaderTrackOutput assetReaderTrackOutputWithTrack:videoTrack outputSettings:videoOutputSettings];
|
|
[videoOutput setSupportsRandomAccess:YES];
|
|
|
|
bool hasAudio = [[asset tracksWithMediaType:AVMediaTypeAudio] count] > 0;
|
|
AVAssetReaderAudioMixOutput *audioOutput = NULL;
|
|
if (hasAudio) {
|
|
AVAssetTrack *audioTrack = [[asset tracksWithMediaType:AVMediaTypeAudio] objectAtIndex:0];
|
|
NSDictionary *audioOutputSettings = [NSDictionary
|
|
dictionaryWithObjectsAndKeys:[NSNumber numberWithInt:kAudioFormatLinearPCM], AVFormatIDKey, [NSNumber numberWithFloat:44100.0], AVSampleRateKey,
|
|
[NSNumber numberWithInt:32], AVLinearPCMBitDepthKey, [NSNumber numberWithBool:NO], AVLinearPCMIsNonInterleaved,
|
|
[NSNumber numberWithBool:YES], AVLinearPCMIsFloatKey, [NSNumber numberWithBool:NO], AVLinearPCMIsBigEndianKey, nil];
|
|
audioOutput = [AVAssetReaderAudioMixOutput assetReaderAudioMixOutputWithAudioTracks:@[ audioTrack ] audioSettings:audioOutputSettings];
|
|
[audioOutput setSupportsRandomAccess:YES];
|
|
}
|
|
|
|
AVAssetReader *reader = [AVAssetReader assetReaderWithAsset:asset error:nil];
|
|
|
|
if (startTime > 0) {
|
|
CMTimeRange timeRange = CMTimeRangeMake(CMTimeMake(startTime * 1000, 1000), kCMTimePositiveInfinity);
|
|
reader.timeRange = timeRange;
|
|
}
|
|
|
|
[reader addOutput:videoOutput];
|
|
if (hasAudio) {
|
|
[reader addOutput:audioOutput];
|
|
}
|
|
|
|
video->impl.assetReader = reader;
|
|
video->impl.videoTrackOutput = videoOutput;
|
|
if (hasAudio) {
|
|
video->impl.audioTrackOutput = audioOutput;
|
|
}
|
|
else {
|
|
video->impl.audioTrackOutput = NULL;
|
|
}
|
|
|
|
if (video->impl.myWidth < 0)
|
|
video->impl.myWidth = [videoTrack naturalSize].width;
|
|
if (video->impl.myHeight < 0)
|
|
video->impl.myHeight = [videoTrack naturalSize].height;
|
|
int framerate = [videoTrack nominalFrameRate];
|
|
kinc_log(KINC_LOG_LEVEL_INFO, "Framerate: %i\n", framerate);
|
|
video->impl.next = video->impl.videoStart;
|
|
video->impl.audioTime = video->impl.videoStart * 44100;
|
|
}
|
|
|
|
void kinc_video_init(kinc_video_t *video, const char *filename) {
|
|
video->impl.playing = false;
|
|
video->impl.sound = NULL;
|
|
video->impl.image_initialized = false;
|
|
char name[2048];
|
|
#ifdef KINC_IOS
|
|
strcpy(name, iphonegetresourcepath());
|
|
#else
|
|
strcpy(name, macgetresourcepath());
|
|
#endif
|
|
strcat(name, "/");
|
|
strcat(name, KINC_DEBUGDIR);
|
|
strcat(name, "/");
|
|
strcat(name, filename);
|
|
video->impl.url = [NSURL fileURLWithPath:[NSString stringWithUTF8String:name]];
|
|
video->impl.myWidth = -1;
|
|
video->impl.myHeight = -1;
|
|
video->impl.finished = false;
|
|
video->impl.duration = 0;
|
|
load(video, 0);
|
|
}
|
|
|
|
void kinc_video_destroy(kinc_video_t *video) {
|
|
kinc_video_stop(video);
|
|
}
|
|
|
|
#ifdef KINC_IOS
|
|
void iosPlayVideoSoundStream(kinc_internal_video_sound_stream_t *video);
|
|
void iosStopVideoSoundStream(void);
|
|
#else
|
|
void macPlayVideoSoundStream(kinc_internal_video_sound_stream_t *video);
|
|
void macStopVideoSoundStream(void);
|
|
#endif
|
|
|
|
void kinc_video_play(kinc_video_t *video, bool loop) {
|
|
AVAssetReader *reader = video->impl.assetReader;
|
|
[reader startReading];
|
|
|
|
kinc_internal_video_sound_stream_t *stream = (kinc_internal_video_sound_stream_t *)malloc(sizeof(kinc_internal_video_sound_stream_t));
|
|
kinc_internal_video_sound_stream_init(stream, 2, 44100);
|
|
video->impl.sound = stream;
|
|
#ifdef KINC_IOS
|
|
iosPlayVideoSoundStream((kinc_internal_video_sound_stream_t *)video->impl.sound);
|
|
#else
|
|
macPlayVideoSoundStream((kinc_internal_video_sound_stream_t *)video->impl.sound);
|
|
#endif
|
|
|
|
video->impl.playing = true;
|
|
video->impl.start = kinc_time() - video->impl.videoStart;
|
|
video->impl.loop = loop;
|
|
}
|
|
|
|
void kinc_video_pause(kinc_video_t *video) {
|
|
video->impl.playing = false;
|
|
if (video->impl.sound != NULL) {
|
|
// Mixer::stop(sound);
|
|
#ifdef KINC_IOS
|
|
iosStopVideoSoundStream();
|
|
#else
|
|
macStopVideoSoundStream();
|
|
#endif
|
|
kinc_internal_video_sound_stream_destroy((kinc_internal_video_sound_stream_t *)video->impl.sound);
|
|
free(video->impl.sound);
|
|
video->impl.sound = NULL;
|
|
}
|
|
}
|
|
|
|
void kinc_video_stop(kinc_video_t *video) {
|
|
kinc_video_pause(video);
|
|
video->impl.finished = true;
|
|
}
|
|
|
|
static void updateImage(kinc_video_t *video) {
|
|
if (!video->impl.playing)
|
|
return;
|
|
|
|
{
|
|
AVAssetReaderTrackOutput *videoOutput = video->impl.videoTrackOutput;
|
|
CMSampleBufferRef buffer = [videoOutput copyNextSampleBuffer];
|
|
if (!buffer) {
|
|
if (video->impl.loop) {
|
|
CMTimeRange timeRange = CMTimeRangeMake(CMTimeMake(0, 1000), kCMTimePositiveInfinity);
|
|
[videoOutput resetForReadingTimeRanges:[NSArray arrayWithObject:[NSValue valueWithCMTimeRange:timeRange]]];
|
|
|
|
AVAssetReaderAudioMixOutput *audioOutput = video->impl.audioTrackOutput;
|
|
CMSampleBufferRef audio_buffer = [audioOutput copyNextSampleBuffer];
|
|
while (audio_buffer) {
|
|
audio_buffer = [audioOutput copyNextSampleBuffer];
|
|
}
|
|
[audioOutput resetForReadingTimeRanges:[NSArray arrayWithObject:[NSValue valueWithCMTimeRange:timeRange]]];
|
|
|
|
buffer = [videoOutput copyNextSampleBuffer];
|
|
|
|
video->impl.start = kinc_time() - video->impl.videoStart;
|
|
}
|
|
else {
|
|
kinc_video_stop(video);
|
|
return;
|
|
}
|
|
}
|
|
video->impl.next = CMTimeGetSeconds(CMSampleBufferGetOutputPresentationTimeStamp(buffer));
|
|
|
|
CVImageBufferRef pixelBuffer = CMSampleBufferGetImageBuffer(buffer);
|
|
|
|
if (!video->impl.image_initialized) {
|
|
CGSize size = CVImageBufferGetDisplaySize(pixelBuffer);
|
|
video->impl.myWidth = size.width;
|
|
video->impl.myHeight = size.height;
|
|
kinc_g4_texture_init(&video->impl.image, kinc_video_width(video), kinc_video_height(video), KINC_IMAGE_FORMAT_BGRA32);
|
|
video->impl.image_initialized = true;
|
|
}
|
|
|
|
if (pixelBuffer != NULL) {
|
|
CVPixelBufferLockBaseAddress(pixelBuffer, 0);
|
|
#ifdef KINC_OPENGL
|
|
kinc_g4_texture_upload(&video->impl.image, (uint8_t *)CVPixelBufferGetBaseAddress(pixelBuffer),
|
|
(int)(CVPixelBufferGetBytesPerRow(pixelBuffer) / 4));
|
|
#else
|
|
kinc_g4_texture_upload(&video->impl.image, (uint8_t *)CVPixelBufferGetBaseAddress(pixelBuffer), (int)(CVPixelBufferGetBytesPerRow(pixelBuffer)));
|
|
#endif
|
|
CVPixelBufferUnlockBaseAddress(pixelBuffer, 0);
|
|
}
|
|
CFRelease(buffer);
|
|
}
|
|
|
|
if (video->impl.audioTrackOutput != NULL) {
|
|
AVAssetReaderAudioMixOutput *audioOutput = video->impl.audioTrackOutput;
|
|
while (video->impl.audioTime / 44100.0 < video->impl.next + 0.1) {
|
|
CMSampleBufferRef buffer = [audioOutput copyNextSampleBuffer];
|
|
if (!buffer)
|
|
return;
|
|
CMItemCount numSamplesInBuffer = CMSampleBufferGetNumSamples(buffer);
|
|
AudioBufferList audioBufferList;
|
|
CMBlockBufferRef blockBufferOut = nil;
|
|
CMSampleBufferGetAudioBufferListWithRetainedBlockBuffer(buffer, NULL, &audioBufferList, sizeof(audioBufferList), NULL, NULL,
|
|
kCMSampleBufferFlag_AudioBufferList_Assure16ByteAlignment, &blockBufferOut);
|
|
for (int bufferCount = 0; bufferCount < audioBufferList.mNumberBuffers; ++bufferCount) {
|
|
float *samples = (float *)audioBufferList.mBuffers[bufferCount].mData;
|
|
kinc_internal_video_sound_stream_t *sound = (kinc_internal_video_sound_stream_t *)video->impl.sound;
|
|
if (video->impl.audioTime / 44100.0 > video->impl.next - 0.1) {
|
|
kinc_internal_video_sound_stream_insert_data(sound, samples, (int)numSamplesInBuffer * 2);
|
|
}
|
|
else {
|
|
// Send some data anyway because the buffers are huge
|
|
kinc_internal_video_sound_stream_insert_data(sound, samples, (int)numSamplesInBuffer);
|
|
}
|
|
video->impl.audioTime += numSamplesInBuffer;
|
|
}
|
|
CFRelease(blockBufferOut);
|
|
CFRelease(buffer);
|
|
}
|
|
}
|
|
}
|
|
|
|
void kinc_video_update(kinc_video_t *video, double time) {
|
|
if (video->impl.playing && time >= video->impl.start + video->impl.next) {
|
|
updateImage(video);
|
|
}
|
|
}
|
|
|
|
int kinc_video_width(kinc_video_t *video) {
|
|
return video->impl.myWidth;
|
|
}
|
|
|
|
int kinc_video_height(kinc_video_t *video) {
|
|
return video->impl.myHeight;
|
|
}
|
|
|
|
kinc_g4_texture_t *kinc_video_current_image(kinc_video_t *video) {
|
|
kinc_video_update(video, kinc_time());
|
|
return &video->impl.image;
|
|
}
|
|
|
|
double kinc_video_duration(kinc_video_t *video) {
|
|
return video->impl.duration;
|
|
}
|
|
|
|
bool kinc_video_finished(kinc_video_t *video) {
|
|
return video->impl.finished;
|
|
}
|
|
|
|
bool kinc_video_paused(kinc_video_t *video) {
|
|
return !video->impl.playing;
|
|
}
|
|
|
|
double kinc_video_position(kinc_video_t *video) {
|
|
return video->impl.next - video->impl.start;
|
|
}
|