Compare commits
No commits in common. "main" and "ivan/fix-last-segment-issue" have entirely different histories.
main
...
ivan/fix-l
@ -72,10 +72,6 @@ class CameraView(context: Context) :
|
||||
var zoom: Float = 1f // in "factor"
|
||||
var exposure: Double = 1.0
|
||||
var orientation: Orientation = Orientation.PORTRAIT
|
||||
set(value) {
|
||||
field = value
|
||||
previewView.orientation = value
|
||||
}
|
||||
var enableZoomGesture = false
|
||||
set(value) {
|
||||
field = value
|
||||
|
@ -10,7 +10,6 @@ import com.mrousavy.camera.types.CodeScannerOptions
|
||||
import com.mrousavy.camera.types.Orientation
|
||||
import com.mrousavy.camera.types.PixelFormat
|
||||
import com.mrousavy.camera.types.ResizeMode
|
||||
import android.util.Log
|
||||
import com.mrousavy.camera.types.Torch
|
||||
import com.mrousavy.camera.types.VideoStabilizationMode
|
||||
|
||||
@ -183,7 +182,6 @@ class CameraViewManager : ViewGroupManager<CameraView>() {
|
||||
fun setOrientation(view: CameraView, orientation: String?) {
|
||||
if (orientation != null) {
|
||||
val newMode = Orientation.fromUnionValue(orientation)
|
||||
Log.i(TAG, "Orientation set to: $newMode")
|
||||
view.orientation = newMode
|
||||
} else {
|
||||
view.orientation = Orientation.PORTRAIT
|
||||
|
@ -312,7 +312,7 @@ class CameraSession(private val context: Context, private val cameraManager: Cam
|
||||
enableHdr
|
||||
)
|
||||
outputs.add(output)
|
||||
|
||||
// Size is usually landscape, so we flip it here
|
||||
previewView?.setSurfaceSize(size.width, size.height, deviceDetails.sensorOrientation)
|
||||
}
|
||||
|
||||
|
@ -36,13 +36,6 @@ class PreviewView(context: Context, callback: SurfaceHolder.Callback) :
|
||||
updateLayout()
|
||||
}
|
||||
}
|
||||
var orientation: Orientation = Orientation.PORTRAIT
|
||||
set(value) {
|
||||
if (field != value) {
|
||||
Log.i(TAG, "View Orientation changed: $field -> $value")
|
||||
field = value
|
||||
}
|
||||
}
|
||||
private var inputOrientation: Orientation = Orientation.LANDSCAPE_LEFT
|
||||
set(value) {
|
||||
if (field != value) {
|
||||
@ -108,11 +101,6 @@ class PreviewView(context: Context, callback: SurfaceHolder.Callback) :
|
||||
}
|
||||
|
||||
private fun getSize(contentSize: Size, containerSize: Size, resizeMode: ResizeMode): Size {
|
||||
var contentSize = contentSize
|
||||
// Swap dimensions if orientation is landscape
|
||||
if (orientation.isLandscape()) {
|
||||
contentSize = Size(contentSize.height, contentSize.width)
|
||||
}
|
||||
val contentAspectRatio = contentSize.width.toDouble() / contentSize.height
|
||||
val containerAspectRatio = containerSize.width.toDouble() / containerSize.height
|
||||
if (!(contentAspectRatio > 0 && containerAspectRatio > 0)) {
|
||||
@ -140,11 +128,11 @@ class PreviewView(context: Context, callback: SurfaceHolder.Callback) :
|
||||
override fun onMeasure(widthMeasureSpec: Int, heightMeasureSpec: Int) {
|
||||
super.onMeasure(widthMeasureSpec, heightMeasureSpec)
|
||||
|
||||
val measuredViewSize = Size(MeasureSpec.getSize(widthMeasureSpec), MeasureSpec.getSize(heightMeasureSpec))
|
||||
val viewSize = Size(MeasureSpec.getSize(widthMeasureSpec), MeasureSpec.getSize(heightMeasureSpec))
|
||||
val surfaceSize = size.rotatedBy(inputOrientation)
|
||||
val fittedSize = getSize(surfaceSize, measuredViewSize, resizeMode)
|
||||
val fittedSize = getSize(surfaceSize, viewSize, resizeMode)
|
||||
|
||||
Log.i(TAG, "PreviewView is $measuredViewSize rendering $surfaceSize orientation ($orientation). Resizing to: $fittedSize ($resizeMode)")
|
||||
Log.i(TAG, "PreviewView is $viewSize, rendering $surfaceSize content ($inputOrientation). Resizing to: $fittedSize ($resizeMode)")
|
||||
setMeasuredDimension(fittedSize.width, fittedSize.height)
|
||||
}
|
||||
|
||||
|
@ -303,15 +303,6 @@ public final class CameraView: UIView, CameraSessionDelegate {
|
||||
}
|
||||
onInitialized([:])
|
||||
}
|
||||
|
||||
func onCameraConfigurationChanged(_ configuration: CameraConfiguration?, _ difference: CameraConfiguration.Difference?) {
|
||||
guard let configuration, let difference else { return }
|
||||
|
||||
if difference.orientationChanged, let connection = previewView.videoPreviewLayer.connection {
|
||||
let videoPreviewLayer = previewView.videoPreviewLayer
|
||||
connection.setOrientation(configuration.orientation)
|
||||
}
|
||||
}
|
||||
|
||||
func onCameraStarted() {
|
||||
ReactLogger.log(level: .info, message: "Camera started!")
|
||||
|
@ -195,7 +195,6 @@ class CameraSession: NSObject, AVCaptureVideoDataOutputSampleBufferDelegate, AVC
|
||||
self.delegate?.onSessionInitialized()
|
||||
}
|
||||
|
||||
self.delegate?.onCameraConfigurationChanged(config, difference)
|
||||
// After configuring, set this to the new configuration.
|
||||
self.configuration = config
|
||||
} catch {
|
||||
|
@ -21,8 +21,6 @@ protocol CameraSessionDelegate: AnyObject {
|
||||
Called when the [CameraSession] successfully initializes
|
||||
*/
|
||||
func onSessionInitialized()
|
||||
|
||||
func onCameraConfigurationChanged(_ configuration: CameraConfiguration?, _ difference: CameraConfiguration.Difference?)
|
||||
/**
|
||||
Called when the [CameraSession] starts streaming frames. (isActive=true)
|
||||
*/
|
||||
|
@ -11,7 +11,7 @@ import AVFoundation
|
||||
|
||||
|
||||
class ChunkedRecorder: NSObject {
|
||||
|
||||
|
||||
enum ChunkType {
|
||||
case initialization
|
||||
case data(index: UInt64, duration: CMTime?)
|
||||
@ -21,12 +21,12 @@ class ChunkedRecorder: NSObject {
|
||||
let url: URL
|
||||
let type: ChunkType
|
||||
}
|
||||
|
||||
|
||||
let outputURL: URL
|
||||
let onChunkReady: ((Chunk) -> Void)
|
||||
|
||||
|
||||
private var chunkIndex: UInt64 = 0
|
||||
|
||||
|
||||
init(outputURL: URL, onChunkReady: @escaping ((Chunk) -> Void)) throws {
|
||||
self.outputURL = outputURL
|
||||
self.onChunkReady = onChunkReady
|
||||
@ -34,16 +34,16 @@ class ChunkedRecorder: NSObject {
|
||||
throw CameraError.unknown(message: "output directory does not exist at: \(outputURL.path)", cause: nil)
|
||||
}
|
||||
}
|
||||
|
||||
|
||||
}
|
||||
|
||||
extension ChunkedRecorder: AVAssetWriterDelegate {
|
||||
|
||||
func assetWriter(_ writer: AVAssetWriter,
|
||||
|
||||
func assetWriter(_ writer: AVAssetWriter,
|
||||
didOutputSegmentData segmentData: Data,
|
||||
segmentType: AVAssetSegmentType,
|
||||
segmentReport: AVAssetSegmentReport?) {
|
||||
|
||||
|
||||
switch segmentType {
|
||||
case .initialization:
|
||||
saveInitSegment(segmentData)
|
||||
@ -53,13 +53,13 @@ extension ChunkedRecorder: AVAssetWriterDelegate {
|
||||
fatalError("Unknown AVAssetSegmentType!")
|
||||
}
|
||||
}
|
||||
|
||||
|
||||
private func saveInitSegment(_ data: Data) {
|
||||
let url = outputURL.appendingPathComponent("init.mp4")
|
||||
save(data: data, url: url)
|
||||
onChunkReady(url: url, type: .initialization)
|
||||
}
|
||||
|
||||
|
||||
private func saveSegment(_ data: Data, report: AVAssetSegmentReport?) {
|
||||
let name = "\(chunkIndex).mp4"
|
||||
let url = outputURL.appendingPathComponent(name)
|
||||
@ -72,7 +72,7 @@ extension ChunkedRecorder: AVAssetWriterDelegate {
|
||||
onChunkReady(url: url, type: .data(index: chunkIndex, duration: duration))
|
||||
chunkIndex += 1
|
||||
}
|
||||
|
||||
|
||||
private func save(data: Data, url: URL) {
|
||||
do {
|
||||
try data.write(to: url)
|
||||
@ -80,9 +80,9 @@ extension ChunkedRecorder: AVAssetWriterDelegate {
|
||||
ReactLogger.log(level: .error, message: "Unable to write \(url): \(error.localizedDescription)")
|
||||
}
|
||||
}
|
||||
|
||||
|
||||
private func onChunkReady(url: URL, type: ChunkType) {
|
||||
onChunkReady(Chunk(url: url, type: type))
|
||||
}
|
||||
|
||||
|
||||
}
|
||||
|
@ -84,7 +84,7 @@ class RecordingSession {
|
||||
assetWriter.shouldOptimizeForNetworkUse = false
|
||||
assetWriter.outputFileTypeProfile = .mpeg4AppleHLS
|
||||
assetWriter.preferredOutputSegmentInterval = CMTime(seconds: 6, preferredTimescale: 1)
|
||||
|
||||
|
||||
/*
|
||||
Apple HLS fMP4 does not have an Edit List Box ('elst') in an initialization segment to remove
|
||||
audio priming duration which advanced audio formats like AAC have, since the sample tables
|
||||
@ -95,7 +95,7 @@ class RecordingSession {
|
||||
*/
|
||||
let startTimeOffset = CMTime(value: 10, timescale: 1)
|
||||
assetWriter.initialSegmentStartTime = startTimeOffset
|
||||
|
||||
|
||||
assetWriter.delegate = recorder
|
||||
} catch let error as NSError {
|
||||
throw CameraError.capture(.createRecorderError(message: error.description))
|
||||
|
@ -32,36 +32,28 @@ extension AVCaptureOutput {
|
||||
func setOrientation(_ orientation: Orientation) {
|
||||
// Set orientation for each connection
|
||||
for connection in connections {
|
||||
connection.setOrientation(orientation)
|
||||
#if swift(>=5.9)
|
||||
if #available(iOS 17.0, *) {
|
||||
// Camera Sensors are always in landscape rotation (90deg).
|
||||
// We are setting the target rotation here, so we need to rotate by landscape once.
|
||||
let cameraOrientation = orientation.rotateBy(orientation: .landscapeLeft)
|
||||
let degrees = cameraOrientation.toDegrees()
|
||||
|
||||
// TODO: Don't rotate the video output because it adds overhead. Instead just use EXIF flags for the .mp4 file if recording.
|
||||
// Does that work when we flip the camera?
|
||||
if connection.isVideoRotationAngleSupported(degrees) {
|
||||
connection.videoRotationAngle = degrees
|
||||
}
|
||||
} else {
|
||||
if connection.isVideoOrientationSupported {
|
||||
connection.videoOrientation = orientation.toAVCaptureVideoOrientation()
|
||||
}
|
||||
}
|
||||
#else
|
||||
if connection.isVideoOrientationSupported {
|
||||
connection.videoOrientation = orientation.toAVCaptureVideoOrientation()
|
||||
}
|
||||
#endif
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
|
||||
|
||||
extension AVCaptureConnection {
|
||||
func setOrientation(_ orientation: Orientation) {
|
||||
#if swift(>=5.9)
|
||||
if #available(iOS 17.0, *) {
|
||||
// Camera Sensors are always in landscape rotation (90deg).
|
||||
// We are setting the target rotation here, so we need to rotate by landscape once.
|
||||
let cameraOrientation = orientation.rotateBy(orientation: .landscapeLeft)
|
||||
let degrees = cameraOrientation.toDegrees()
|
||||
|
||||
// TODO: Don't rotate the video output because it adds overhead. Instead just use EXIF flags for the .mp4 file if recording.
|
||||
// Does that work when we flip the camera?
|
||||
if isVideoRotationAngleSupported(degrees) {
|
||||
videoRotationAngle = degrees
|
||||
}
|
||||
} else {
|
||||
if isVideoOrientationSupported {
|
||||
videoOrientation = orientation.toAVCaptureVideoOrientation()
|
||||
}
|
||||
}
|
||||
#else
|
||||
if isVideoOrientationSupported {
|
||||
videoOrientation = orientation.toAVCaptureVideoOrientation()
|
||||
}
|
||||
#endif
|
||||
}
|
||||
}
|
||||
|
@ -113,19 +113,5 @@ class ViewController: UIViewController {
|
||||
}
|
||||
}
|
||||
|
||||
override func viewWillTransition(to size: CGSize, with coordinator: any UIViewControllerTransitionCoordinator) {
|
||||
switch UIDevice.current.orientation {
|
||||
case .landscapeLeft:
|
||||
cameraView.orientation = "landscape-right"
|
||||
case .landscapeRight:
|
||||
cameraView.orientation = "landscape-left"
|
||||
default:
|
||||
cameraView.orientation = "portrait"
|
||||
}
|
||||
|
||||
cameraView.didSetProps([])
|
||||
super.viewWillTransition(to: size, with: coordinator)
|
||||
}
|
||||
|
||||
}
|
||||
|
||||
|
Loading…
Reference in New Issue
Block a user