* fix: Add more logs
* perf: Init audio AVAssetWriter in parallel
* chore: Log Audio Session activation
* perf: Init Audio Session asynchronously
* chore: Log when category changed
* fix: Revert lazy audio initialization
* Update Podfile.lock
* Pass `sourceFormatHint` to video and audio AVAssetWriter
* fix: Remove `sourceFormatHint` from Video Asset Writer
* Use default options for Audio Asset Writer
* Update Podfile.lock
* Revert "Use default options for Audio Asset Writer"
This reverts commit e575a14c5342ddc7f9db557d5e3915328ed9e798.
* Add time logs
* fix: Don't synchronize audio buffers, they are already in sync
* shouldOptimizeForNetworkUse = false
* fix: Only update `latestTimestamp` once video buffer has been written
* perf: Use `AVAssetWriterInput` instead of `AVAssetWriterInputPixelBufferAdaptor`
* fix: Fix Audio not being synchronized with Video
* Remove logs add comments
* Format
* feat: Set `.videoRecording` AVAudioSession mode
* Refactor `startRecording()` a bit
* Format
* chore: Throw error directly instead of double-checking
* feat: Split `videoHdr` and `photoHdr` into two settings
* fix: Rename all `hdr`
* fix: Fix HDR on Android
* Update CameraDeviceDetails.kt
* Update CameraDeviceDetails.kt
* fix: Correctly configure `pixelFormat` AFTER `format`
* Update CameraSession+Configuration.swift
* fix: Also after format changed
Moves everything Camera related into `core/` / `Core/` so that it is better encapsulated from React Native.
Benefits:
1. Code is much better organized. Should be easier for collaborators now, and cleaner codebase for me.
2. Locking is fully atomically as you can now only configure the session through a lock/Mutex which is batch-overridable
* On iOS, this makes Camera startup time **MUCH** faster, I measured speedups from **1.5 seconds** to only **240 milliseconds** since we only lock/commit once! 🚀
* On Android, this fixes a few out-of-sync/concurrency issues like "Capture Request contains unconfigured Input/Output Surface!" since it is now a single lock-operation! 💪
3. It is easier to integrate VisionCamera outside of React Native (e.g. Native iOS Apps, NativeScript, Flutter, etc)
With this PR, VisionCamera V3 is up to **7x** faster than V2
* feat: Use new photo dimensions API
* Update AVCaptureDevice.Format+matchesFilter.swift
* fix: Use Pixels instead of Points for video size
* feat: Set `PhotoOutput`'s maximum photo resolution
* fix: Compare dictionaries instead
* chore: Format code
* fix: Try to use hash.... failing atm
* fix: Use rough comparison again
* fix: Also take video HDR into consideration
* chore: Format
* Use contains
* Update AVCaptureDevice.Format+toDictionary.swift
* docs: Add better docs to Camera props
* Update CameraView+AVCaptureSession.swift
* Update CameraView+AVCaptureSession.swift
* feat: Add support for LiDAR, TrueDepth, External (USB) and Continuity Camera Devices (iOS 17)
* Rename `devices` -> `physicalDevices`
* fix: Comment out iOS 17 cameras for now
* fix: Move `supportsDepthCapture` to `format`
* fix: Fall back to `wide-angle-camera` for any unknown types
* Update CameraPage.tsx
* `descriptor` -> `physicalDeviceDescriptor`
* Update CameraDevice.ts
* Format
* feat: Expose `userPreferredCameraDevice`
Uses the new iOS 17 API where the user can prefer a default device, otherwise fall back to the first device of the available ones
* fix: Expose as property
* Add TODO comments
* fix: Format code
* fix: Compile below Swift 5.9