ios - 保存高质量图像,进行实时处理——最好的方法是什么?
<p><p>我仍在学习 AVFoundation,所以我不确定我应该如何最好地解决需要捕获高质量静止图像但提供低质量预览视频流的问题。</p>
<p>我有一个应用程序需要拍摄高质量图像 (<code>AVCaptureSessionPresetPhoto</code>),但使用 OpenCV 处理预览视频流 - 可以接受低得多的分辨率。只需使用底座 <a href="https://github.com/Itseez/opencv/blob/master/modules/highgui/src/cap_ios_video_camera.mm" rel="noreferrer noopener nofollow">OpenCV Video Camera class</a>不好,因为将 <code>defaultAVCaptureSessionPreset</code> 设置为 <code>AVCaptureSessionPresetPhoto</code> 会导致将全分辨率帧传递给 <code>processImage</code> - 这确实很慢。</p>
<p>如何才能与可用于捕获静止图像的设备建立高质量连接,以及可处理和显示的低质量连接?对我需要如何设置 session /连接的描述将非常有帮助。是否有此类应用的开源示例?</p></p>
<br><hr><h1><strong>Best Answer-推荐答案</ strong></h1><br>
<p><p>我做了类似的事情——我在委托(delegate)方法中抓取了像素,为它们制作了一个 CGImageRef,然后将其发送到正常的优先级队列,在那里进行了修改。由于 AVFoundation 必须将 CADisplayLink 用于回调方法,因此它具有最高优先级。在我的特殊情况下,我没有抓取所有像素,因此它可以在 iPhone 4 上以 30fps 的速度运行。根据您要运行的设备,您需要在像素数、fps 等方面进行权衡。</p>
<p>另一个想法是获取 2 个像素子集的幂 - 例如,每行的每 4 个像素和每 4 行的像素。我再次在我的应用程序中以 20-30fps 的速度做了类似的事情。然后,您可以在分派(dispatch)的 block 中进一步操作这个较小的图像。</p>
<p>如果这看起来令人生畏,请悬赏工作代码。</p>
<p>代码:</p>
<pre><code>// Image is oriented with bottle neck to the left and the bottle bottom on the right
- (void)captureOutput:(AVCaptureVideoDataOutput *)captureOutput didOutputSampleBuffer:(CMSampleBufferRef)sampleBuffer fromConnection:(AVCaptureConnection *)connection
{
#if 1
AVCaptureDevice *camera = [(AVCaptureDeviceInput *) device];
if(camera.adjustingWhiteBalance || camera.adjustingExposure) NSLog(@"GOTCHA: %d %d", camera.adjustingWhiteBalance, camera.adjustingExposure);
printf("foo\n");
#endif
if(saveState != saveOne && saveState != saveAll) return;
@autoreleasepool {
CVImageBufferRef imageBuffer = CMSampleBufferGetImageBuffer(sampleBuffer);
//NSLog(@"PE: value=%lld timeScale=%d flags=%x", prStamp.value, prStamp.timescale, prStamp.flags);
/*Lock the image buffer*/
CVPixelBufferLockBaseAddress(imageBuffer,0);
NSRange captureRange;
if(saveState == saveOne) {
#if 0 // B G R A MODE !
NSLog(@"PIXEL_TYPE: 0x%lx", CVPixelBufferGetPixelFormatType(imageBuffer));
uint8_t *newPtr = (uint8_t *)CVPixelBufferGetBaseAddress(imageBuffer);
NSLog(@"ONE VAL %x %x %x %x", newPtr, newPtr, newPtr, newPtr);
}
exit(0);
#endif
;
BOOL success = ;
if(!success) {
dispatch_async(dispatch_get_main_queue(), ^{ edgeFinder = nil; ; });
saveState = saveNone;
} else
bottleRange = edgeFinder.sides;
xRange.location = edgeFinder.shoulder;
xRange.length = edgeFinder.bottom - xRange.location;
NSLog(@"bottleRange 1: %@ neck=%d bottom=%d", NSStringFromRange(bottleRange), edgeFinder.shoulder, edgeFinder.bottom );
//searchRows = ;
rowsPerSwath = lrintf((bottleRange.length*NUM_DEGREES_TO_GRAB)*(float)M_PI/360.0f);
NSLog(@"rowsPerSwath = %d", rowsPerSwath);
saveState = saveIdling;
captureRange = NSMakeRange(0, );
dispatch_async(dispatch_get_main_queue(), ^
{
;
edgeFinder = nil;
captureOutput.alwaysDiscardsLateVideoFrames = YES;
});
} else {
NSInteger rows = rowsPerSwath;
NSInteger newOffset = bottleRange.length - rows;
if(newOffset & 1) {
--newOffset;
++rows;
}
captureRange = NSMakeRange(bottleRange.location + newOffset/2, rows);
}
//NSLog(@"captureRange=%u %u", captureRange.location, captureRange.length);
/*Get information about the image*/
uint8_t *baseAddress = (uint8_t *)CVPixelBufferGetBaseAddress(imageBuffer);
size_t bytesPerRow = CVPixelBufferGetBytesPerRow(imageBuffer);
size_t width = CVPixelBufferGetWidth(imageBuffer);
// Note Apple sample code cheats big time - the phone is big endian so this reverses the "apparent" order of bytes
CGContextRef newContext = CGBitmapContextCreate(NULL, width, captureRange.length, 8, bytesPerRow, colorSpace, kCGImageAlphaNoneSkipFirst | kCGBitmapByteOrder32Little); // Video in ARGB format
assert(newContext);
uint8_t *newPtr = (uint8_t *)CGBitmapContextGetData(newContext);
size_t offset = captureRange.location * bytesPerRow;
memcpy(newPtr, baseAddress + offset, captureRange.length * bytesPerRow);
CVPixelBufferUnlockBaseAddress(imageBuffer, 0);
OSAtomicIncrement32(&totalImages);
int32_t curDepth = OSAtomicIncrement32(&queueDepth);
if(curDepth > maxDepth) maxDepth = curDepth;
#define kImageContext @"kImageContext"
#define kState @"kState"
#define kPresTime @"kPresTime"
CMTime prStamp = CMSampleBufferGetPresentationTimeStamp(sampleBuffer); // when it was taken?
//CMTime deStamp = CMSampleBufferGetDecodeTimeStamp(sampleBuffer); // now?
NSDictionary *dict = [NSDictionary dictionaryWithObjectsAndKeys:
, kState,
, kImageContext,
, kPresTime,
nil ];
dispatch_async(imageQueue, ^
{
// could be on any thread now
OSAtomicDecrement32(&queueDepth);
if(!isCancelled) {
saveImages state; [(NSValue *) getValue:&state];
CGContextRef context; [(NSValue *) getValue:&context];
CMTime stamp; [(NSValue *) getValue:&stamp];
CGImageRef newImageRef = CGBitmapContextCreateImage(context);
CGContextRelease(context);
UIImageOrientation orient = state == saveOne ? UIImageOrientationLeft : UIImageOrientationUp;
UIImage *image = ; // imageWithCGImage:UIImageOrientationUpUIImageOrientationLeft
CGImageRelease(newImageRef);
NSData *data = UIImagePNGRepresentation(image);
// NSLog(@"STATE:[%d]: value=%lld timeScale=%d flags=%x", state, stamp.value, stamp.timescale, stamp.flags);
{
NSString *name = ;
NSString *path = [ stringByAppendingPathComponent:name];
BOOL ret = ;
//NSLog(@"WROTE %d err=%d w/time %f path:%@", num, ret, (double)stamp.value/(double)stamp.timescale, path);
if(!ret) {
++errors;
} else {
dispatch_async(dispatch_get_main_queue(), ^
{
if(num) ;
} );
}
++num;
}
} else NSLog(@"CANCELLED");
} );
}
}
</code></pre></p>
<p style="font-size: 20px;">关于ios - 保存高质量图像,进行实时处理——最好的方法是什么?,我们在Stack Overflow上找到一个类似的问题:
<a href="https://stackoverflow.com/questions/16648394/" rel="noreferrer noopener nofollow" style="color: red;">
https://stackoverflow.com/questions/16648394/
</a>
</p>
页:
[1]