ad5e131f6a
* Add `onFrameProcessorPerformanceSuggestionAvailable` and make `frameProcessorFps` support `auto` * Implement performance suggestion and auto-adjusting * Fix FPS setting, evaluate correctly * Floor suggested FPS * Remove `console.log` for frame drop warnings. * Swift format * Use `30` magic number * only call if FPS is different * Update CameraView.swift * Implement Android 1/2 * Cleanup * Update `frameProcessorFps` if available * Optimize `FrameProcessorPerformanceDataCollector` initialization * Cache call * Set frameProcessorFps directly (Kotlin setter) * Don't suggest if same value * Call suggestion every second * reset time on set * Always store 15 last samples * reset counter too * Update FrameProcessorPerformanceDataCollector.swift * Update CameraView+RecordVideo.swift * Update CameraView.kt * iOS: Redesign evaluation * Update CameraView+RecordVideo.swift * Android: Redesign evaluation * Update CameraView.kt * Update REA to latest alpha and install RNScreens * Fix frameProcessorFps updating
166 lines
5.9 KiB
Swift
166 lines
5.9 KiB
Swift
//
|
|
// CameraViewManager.swift
|
|
// mrousavy
|
|
//
|
|
// Created by Marc Rousavy on 09.11.20.
|
|
// Copyright © 2020 mrousavy. All rights reserved.
|
|
//
|
|
|
|
import AVFoundation
|
|
import Foundation
|
|
|
|
@objc(CameraViewManager)
|
|
final class CameraViewManager: RCTViewManager {
|
|
// pragma MARK: Properties
|
|
|
|
private var runtimeManager: FrameProcessorRuntimeManager?
|
|
|
|
override var bridge: RCTBridge! {
|
|
didSet {
|
|
// Install Frame Processor bindings and setup Runtime
|
|
if VISION_CAMERA_ENABLE_FRAME_PROCESSORS {
|
|
CameraQueues.frameProcessorQueue.async {
|
|
self.runtimeManager = FrameProcessorRuntimeManager(bridge: self.bridge)
|
|
self.bridge.runOnJS {
|
|
self.runtimeManager!.installFrameProcessorBindings()
|
|
}
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
override var methodQueue: DispatchQueue! {
|
|
return DispatchQueue.main
|
|
}
|
|
|
|
override static func requiresMainQueueSetup() -> Bool {
|
|
return true
|
|
}
|
|
|
|
override final func view() -> UIView! {
|
|
return CameraView()
|
|
}
|
|
|
|
// pragma MARK: React Functions
|
|
|
|
@objc
|
|
final func startRecording(_ node: NSNumber, options: NSDictionary, onRecordCallback: @escaping RCTResponseSenderBlock) {
|
|
let component = getCameraView(withTag: node)
|
|
component.startRecording(options: options, callback: onRecordCallback)
|
|
}
|
|
|
|
@objc
|
|
final func stopRecording(_ node: NSNumber, resolve: @escaping RCTPromiseResolveBlock, reject: @escaping RCTPromiseRejectBlock) {
|
|
let component = getCameraView(withTag: node)
|
|
component.stopRecording(promise: Promise(resolver: resolve, rejecter: reject))
|
|
}
|
|
|
|
@objc
|
|
final func takePhoto(_ node: NSNumber, options: NSDictionary, resolve: @escaping RCTPromiseResolveBlock, reject: @escaping RCTPromiseRejectBlock) {
|
|
let component = getCameraView(withTag: node)
|
|
component.takePhoto(options: options, promise: Promise(resolver: resolve, rejecter: reject))
|
|
}
|
|
|
|
@objc
|
|
final func focus(_ node: NSNumber, point: NSDictionary, resolve: @escaping RCTPromiseResolveBlock, reject: @escaping RCTPromiseRejectBlock) {
|
|
let promise = Promise(resolver: resolve, rejecter: reject)
|
|
guard let x = point["x"] as? NSNumber, let y = point["y"] as? NSNumber else {
|
|
promise.reject(error: .parameter(.invalid(unionName: "point", receivedValue: point.description)))
|
|
return
|
|
}
|
|
let component = getCameraView(withTag: node)
|
|
component.focus(point: CGPoint(x: x.doubleValue, y: y.doubleValue), promise: promise)
|
|
}
|
|
|
|
@objc
|
|
final func getAvailableCameraDevices(_ resolve: @escaping RCTPromiseResolveBlock, reject: @escaping RCTPromiseRejectBlock) {
|
|
withPromise(resolve: resolve, reject: reject) {
|
|
let discoverySession = AVCaptureDevice.DiscoverySession(deviceTypes: getAllDeviceTypes(), mediaType: .video, position: .unspecified)
|
|
let devices = discoverySession.devices.filter {
|
|
if #available(iOS 11.1, *) {
|
|
// exclude the true-depth camera. The True-Depth camera has YUV and Infrared, can't take photos!
|
|
return $0.deviceType != .builtInTrueDepthCamera
|
|
}
|
|
return true
|
|
}
|
|
return devices.map {
|
|
return [
|
|
"id": $0.uniqueID,
|
|
"devices": $0.physicalDevices.map(\.deviceType.descriptor),
|
|
"position": $0.position.descriptor,
|
|
"name": $0.localizedName,
|
|
"hasFlash": $0.hasFlash,
|
|
"hasTorch": $0.hasTorch,
|
|
"minZoom": $0.minAvailableVideoZoomFactor,
|
|
"neutralZoom": $0.neutralZoomFactor,
|
|
"maxZoom": $0.maxAvailableVideoZoomFactor,
|
|
"isMultiCam": $0.isMultiCam,
|
|
"supportsParallelVideoProcessing": true,
|
|
"supportsDepthCapture": false, // TODO: supportsDepthCapture
|
|
"supportsRawCapture": false, // TODO: supportsRawCapture
|
|
"supportsLowLightBoost": $0.isLowLightBoostSupported,
|
|
"supportsFocus": $0.isFocusPointOfInterestSupported,
|
|
"formats": $0.formats.map { format -> [String: Any] in
|
|
format.toDictionary()
|
|
},
|
|
]
|
|
}
|
|
}
|
|
}
|
|
|
|
@objc
|
|
final func getCameraPermissionStatus(_ resolve: @escaping RCTPromiseResolveBlock, reject: @escaping RCTPromiseRejectBlock) {
|
|
withPromise(resolve: resolve, reject: reject) {
|
|
let status = AVCaptureDevice.authorizationStatus(for: .video)
|
|
return status.descriptor
|
|
}
|
|
}
|
|
|
|
@objc
|
|
final func getMicrophonePermissionStatus(_ resolve: @escaping RCTPromiseResolveBlock, reject: @escaping RCTPromiseRejectBlock) {
|
|
withPromise(resolve: resolve, reject: reject) {
|
|
let status = AVCaptureDevice.authorizationStatus(for: .audio)
|
|
return status.descriptor
|
|
}
|
|
}
|
|
|
|
@objc
|
|
final func requestCameraPermission(_ resolve: @escaping RCTPromiseResolveBlock, reject _: @escaping RCTPromiseRejectBlock) {
|
|
AVCaptureDevice.requestAccess(for: .video) { granted in
|
|
let result: AVAuthorizationStatus = granted ? .authorized : .denied
|
|
resolve(result.descriptor)
|
|
}
|
|
}
|
|
|
|
@objc
|
|
final func requestMicrophonePermission(_ resolve: @escaping RCTPromiseResolveBlock, reject _: @escaping RCTPromiseRejectBlock) {
|
|
AVCaptureDevice.requestAccess(for: .audio) { granted in
|
|
let result: AVAuthorizationStatus = granted ? .authorized : .denied
|
|
resolve(result.descriptor)
|
|
}
|
|
}
|
|
|
|
// MARK: Private
|
|
|
|
private func getCameraView(withTag tag: NSNumber) -> CameraView {
|
|
// swiftlint:disable force_cast
|
|
return bridge.uiManager.view(forReactTag: tag) as! CameraView
|
|
}
|
|
|
|
private final func getAllDeviceTypes() -> [AVCaptureDevice.DeviceType] {
|
|
var deviceTypes: [AVCaptureDevice.DeviceType] = []
|
|
if #available(iOS 13.0, *) {
|
|
deviceTypes.append(.builtInTripleCamera)
|
|
deviceTypes.append(.builtInDualWideCamera)
|
|
deviceTypes.append(.builtInUltraWideCamera)
|
|
}
|
|
if #available(iOS 11.1, *) {
|
|
deviceTypes.append(.builtInTrueDepthCamera)
|
|
}
|
|
deviceTypes.append(.builtInDualCamera)
|
|
deviceTypes.append(.builtInWideAngleCamera)
|
|
deviceTypes.append(.builtInTelephotoCamera)
|
|
return deviceTypes
|
|
}
|
|
}
|