Compare commits
17 Commits
dd9de38a7d
...
main
| Author | SHA1 | Date | |
|---|---|---|---|
| 8a6537c914 | |||
| fde8a47a2d | |||
| b5cde91062 | |||
| 0e665528fe | |||
| 45b16b5189 | |||
| 2d570c9af2 | |||
| 31e7b8bd35 | |||
| 3bb72d5d94 | |||
| ac5dac127f | |||
| e3de8c018c | |||
|
|
dd26812a9c | ||
|
|
b716608379 | ||
| 0ecc3d8210 | |||
| 309e1e9457 | |||
| 71b08e6898 | |||
|
|
699481f6f8 | ||
| 11ce9ba8f6 |
@@ -17,15 +17,18 @@
|
||||
|
||||
namespace vision {
|
||||
|
||||
std::unique_ptr<OpenGLRenderer> OpenGLRenderer::CreateWithWindowSurface(std::shared_ptr<OpenGLContext> context, ANativeWindow* surface) {
|
||||
return std::unique_ptr<OpenGLRenderer>(new OpenGLRenderer(std::move(context), surface));
|
||||
std::unique_ptr<OpenGLRenderer> OpenGLRenderer::CreateWithWindowSurface(std::shared_ptr<OpenGLContext> context,
|
||||
ANativeWindow* surface,
|
||||
int rotationDegrees) {
|
||||
return std::unique_ptr<OpenGLRenderer>(new OpenGLRenderer(std::move(context), surface, rotationDegrees));
|
||||
}
|
||||
|
||||
OpenGLRenderer::OpenGLRenderer(std::shared_ptr<OpenGLContext> context, ANativeWindow* surface) {
|
||||
OpenGLRenderer::OpenGLRenderer(std::shared_ptr<OpenGLContext> context, ANativeWindow* surface, int rotationDegrees) {
|
||||
_context = std::move(context);
|
||||
_outputSurface = surface;
|
||||
_width = ANativeWindow_getWidth(surface);
|
||||
_height = ANativeWindow_getHeight(surface);
|
||||
_rotationDegrees = rotationDegrees;
|
||||
}
|
||||
|
||||
OpenGLRenderer::~OpenGLRenderer() {
|
||||
@@ -66,7 +69,7 @@ void OpenGLRenderer::renderTextureToSurface(const OpenGLTexture& texture, float*
|
||||
glTexParameteri(texture.target, GL_TEXTURE_WRAP_T, GL_CLAMP_TO_EDGE);
|
||||
|
||||
// 4. Draw it using the pass-through shader which also applies transforms
|
||||
_passThroughShader.draw(texture, transformMatrix);
|
||||
_passThroughShader.draw(texture, transformMatrix, _rotationDegrees);
|
||||
|
||||
// 5. Swap buffers to pass it to the window surface
|
||||
eglSwapBuffers(_context->display, _surface);
|
||||
|
||||
@@ -24,7 +24,7 @@ public:
|
||||
* Note: The `surface` is considered moved, and the OpenGL context will release it when it is
|
||||
* being deleted.
|
||||
*/
|
||||
static std::unique_ptr<OpenGLRenderer> CreateWithWindowSurface(std::shared_ptr<OpenGLContext> context, ANativeWindow* surface);
|
||||
static std::unique_ptr<OpenGLRenderer> CreateWithWindowSurface(std::shared_ptr<OpenGLContext> context, ANativeWindow* surface, int rotationDegrees);
|
||||
/**
|
||||
* Destroy the OpenGL Context. This needs to be called on the same thread that `use()` was called.
|
||||
*/
|
||||
@@ -43,10 +43,11 @@ public:
|
||||
void destroy();
|
||||
|
||||
private:
|
||||
explicit OpenGLRenderer(std::shared_ptr<OpenGLContext> context, ANativeWindow* surface);
|
||||
explicit OpenGLRenderer(std::shared_ptr<OpenGLContext> context, ANativeWindow* surface, int rotationDegrees);
|
||||
|
||||
private:
|
||||
int _width = 0, _height = 0;
|
||||
int _rotationDegrees = 0;
|
||||
std::shared_ptr<OpenGLContext> _context;
|
||||
ANativeWindow* _outputSurface;
|
||||
EGLSurface _surface = EGL_NO_SURFACE;
|
||||
|
||||
@@ -7,11 +7,76 @@
|
||||
#include <EGL/egl.h>
|
||||
#include <GLES2/gl2.h>
|
||||
#include <GLES2/gl2ext.h>
|
||||
#include <cmath>
|
||||
#include <memory>
|
||||
#include <string>
|
||||
|
||||
namespace vision {
|
||||
|
||||
namespace {
|
||||
|
||||
void setIdentity(float* matrix) {
|
||||
for (int i = 0; i < 16; i++) {
|
||||
matrix[i] = 0.0f;
|
||||
}
|
||||
matrix[0] = 1.0f;
|
||||
matrix[5] = 1.0f;
|
||||
matrix[10] = 1.0f;
|
||||
matrix[15] = 1.0f;
|
||||
}
|
||||
|
||||
void multiply4x4(const float* left, const float* right, float* out) {
|
||||
for (int column = 0; column < 4; column++) {
|
||||
for (int row = 0; row < 4; row++) {
|
||||
float sum = 0.0f;
|
||||
for (int k = 0; k < 4; k++) {
|
||||
sum += left[k * 4 + row] * right[column * 4 + k];
|
||||
}
|
||||
out[column * 4 + row] = sum;
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
void makeTranslation(float tx, float ty, float* matrix) {
|
||||
setIdentity(matrix);
|
||||
matrix[12] = tx;
|
||||
matrix[13] = ty;
|
||||
}
|
||||
|
||||
void makeRotation(float degrees, float* matrix) {
|
||||
setIdentity(matrix);
|
||||
const float radians = degrees * static_cast<float>(M_PI) / 180.0f;
|
||||
const float cosine = std::cos(radians);
|
||||
const float sine = std::sin(radians);
|
||||
|
||||
matrix[0] = cosine;
|
||||
matrix[1] = sine;
|
||||
matrix[4] = -sine;
|
||||
matrix[5] = cosine;
|
||||
}
|
||||
|
||||
void makeCenteredRotation(int rotationDegrees, float* matrix) {
|
||||
const int normalized = ((rotationDegrees % 360) + 360) % 360;
|
||||
if (normalized == 0) {
|
||||
setIdentity(matrix);
|
||||
return;
|
||||
}
|
||||
|
||||
float translateToOrigin[16];
|
||||
float rotation[16];
|
||||
float translateBack[16];
|
||||
float temp[16];
|
||||
|
||||
makeTranslation(-0.5f, -0.5f, translateToOrigin);
|
||||
makeRotation(static_cast<float>(normalized), rotation);
|
||||
makeTranslation(0.5f, 0.5f, translateBack);
|
||||
|
||||
multiply4x4(rotation, translateToOrigin, temp);
|
||||
multiply4x4(translateBack, temp, matrix);
|
||||
}
|
||||
|
||||
} // namespace
|
||||
|
||||
PassThroughShader::~PassThroughShader() {
|
||||
if (_programId != NO_SHADER) {
|
||||
glDeleteProgram(_programId);
|
||||
@@ -24,7 +89,7 @@ PassThroughShader::~PassThroughShader() {
|
||||
}
|
||||
}
|
||||
|
||||
void PassThroughShader::draw(const OpenGLTexture& texture, float* transformMatrix) {
|
||||
void PassThroughShader::draw(const OpenGLTexture& texture, float* transformMatrix, int rotationDegrees) {
|
||||
// 1. Set up Shader Program
|
||||
if (_programId == NO_SHADER) {
|
||||
_programId = createProgram();
|
||||
@@ -57,7 +122,12 @@ void PassThroughShader::draw(const OpenGLTexture& texture, float* transformMatri
|
||||
glVertexAttribPointer(_vertexParameters.aTexCoord, 2, GL_FLOAT, GL_FALSE, sizeof(Vertex),
|
||||
reinterpret_cast<void*>(offsetof(Vertex, texCoord)));
|
||||
|
||||
glUniformMatrix4fv(_vertexParameters.uTransformMatrix, 1, GL_FALSE, transformMatrix);
|
||||
float outputRotationMatrix[16];
|
||||
float combinedTransformMatrix[16];
|
||||
makeCenteredRotation(rotationDegrees, outputRotationMatrix);
|
||||
multiply4x4(transformMatrix, outputRotationMatrix, combinedTransformMatrix);
|
||||
|
||||
glUniformMatrix4fv(_vertexParameters.uTransformMatrix, 1, GL_FALSE, combinedTransformMatrix);
|
||||
|
||||
// 4. Pass texture to fragment shader
|
||||
glActiveTexture(GL_TEXTURE0);
|
||||
|
||||
@@ -29,7 +29,7 @@ public:
|
||||
* Draw the texture using this shader.
|
||||
* Note: At the moment, only EXTERNAL textures are supported by the Shader.
|
||||
*/
|
||||
void draw(const OpenGLTexture& texture, float* transformMatrix);
|
||||
void draw(const OpenGLTexture& texture, float* transformMatrix, int rotationDegrees);
|
||||
|
||||
private:
|
||||
// Loading
|
||||
|
||||
@@ -47,13 +47,13 @@ void VideoPipeline::removeRecordingSessionOutputSurface() {
|
||||
_recordingSessionOutput = nullptr;
|
||||
}
|
||||
|
||||
void VideoPipeline::setRecordingSessionOutputSurface(jobject surface) {
|
||||
void VideoPipeline::setRecordingSessionOutputSurface(jobject surface, int rotationDegrees) {
|
||||
// 1. Delete existing output surface
|
||||
removeRecordingSessionOutputSurface();
|
||||
|
||||
// 2. Set new output surface if it is not null
|
||||
ANativeWindow* window = ANativeWindow_fromSurface(jni::Environment::current(), surface);
|
||||
_recordingSessionOutput = OpenGLRenderer::CreateWithWindowSurface(_context, window);
|
||||
_recordingSessionOutput = OpenGLRenderer::CreateWithWindowSurface(_context, window, rotationDegrees);
|
||||
}
|
||||
|
||||
int VideoPipeline::getInputTextureId() {
|
||||
@@ -78,7 +78,6 @@ void VideoPipeline::onFrame(jni::alias_ref<jni::JArrayFloat> transformMatrixPara
|
||||
OpenGLTexture& texture = _inputTexture.value();
|
||||
|
||||
if (_recordingSessionOutput) {
|
||||
__android_log_print(ANDROID_LOG_INFO, TAG, "Rendering to RecordingSession..");
|
||||
_recordingSessionOutput->renderTextureToSurface(texture, transformMatrix);
|
||||
}
|
||||
}
|
||||
|
||||
@@ -31,7 +31,7 @@ public:
|
||||
int getInputTextureId();
|
||||
|
||||
// <- MediaRecorder output
|
||||
void setRecordingSessionOutputSurface(jobject surface);
|
||||
void setRecordingSessionOutputSurface(jobject surface, int rotationDegrees);
|
||||
void removeRecordingSessionOutputSurface();
|
||||
|
||||
// Frame callbacks
|
||||
|
||||
@@ -30,8 +30,15 @@ class ChunkedRecordingManager(private val encoder: MediaCodec, private val outpu
|
||||
iFrameInterval: Int = 5
|
||||
): ChunkedRecordingManager {
|
||||
val mimeType = options.videoCodec.toMimeType()
|
||||
val cameraOrientationDegrees = cameraOrientation.toDegrees()
|
||||
val recordingOrientationDegrees = (options.orientation ?: Orientation.PORTRAIT).toDegrees();
|
||||
// Use cameraOrientation (from WindowManager) for rotation metadata
|
||||
// The options.orientation from JavaScript is unreliable on Android when rotating between landscape modes
|
||||
// Note: MediaMuxer.setOrientationHint() uses opposite convention from HlsMuxer's rotation matrix
|
||||
// We need to invert the rotation: 90 <-> 270, while 0 and 180 stay the same
|
||||
val orientationDegrees = when (cameraOrientation.toDegrees()) {
|
||||
90 -> 270
|
||||
270 -> 90
|
||||
else -> cameraOrientation.toDegrees()
|
||||
}
|
||||
val (width, height) = if (cameraOrientation.isLandscape()) {
|
||||
size.height to size.width
|
||||
} else {
|
||||
@@ -55,12 +62,12 @@ class ChunkedRecordingManager(private val encoder: MediaCodec, private val outpu
|
||||
format.setInteger(MediaFormat.KEY_I_FRAME_INTERVAL, iFrameInterval)
|
||||
format.setInteger(MediaFormat.KEY_BIT_RATE, bitRate)
|
||||
|
||||
Log.d(TAG, "Video Format: $format, camera orientation $cameraOrientationDegrees, recordingOrientation: $recordingOrientationDegrees")
|
||||
Log.d(TAG, "Video Format: $format, orientation: $orientationDegrees")
|
||||
// Create a MediaCodec encoder, and configure it with our format. Get a Surface
|
||||
// we can use for input and wrap it with a class that handles the EGL work.
|
||||
codec.configure(format, null, null, MediaCodec.CONFIGURE_FLAG_ENCODE)
|
||||
return ChunkedRecordingManager(
|
||||
codec, outputDirectory, recordingOrientationDegrees, iFrameInterval, callbacks
|
||||
codec, outputDirectory, orientationDegrees, iFrameInterval, callbacks
|
||||
)
|
||||
}
|
||||
}
|
||||
@@ -91,12 +98,13 @@ class ChunkedRecordingManager(private val encoder: MediaCodec, private val outpu
|
||||
muxer.start()
|
||||
}
|
||||
|
||||
fun writeSample(buffer: java.nio.ByteBuffer, bufferInfo: BufferInfo) {
|
||||
muxer.writeSampleData(videoTrack, buffer, bufferInfo)
|
||||
}
|
||||
|
||||
fun finish() {
|
||||
muxer.stop()
|
||||
muxer.release()
|
||||
// Calculate duration from start time - this is approximate
|
||||
// The new FragmentedRecordingManager provides accurate duration
|
||||
callbacks.onVideoChunkReady(filepath, chunkIndex, null)
|
||||
}
|
||||
}
|
||||
@@ -170,7 +178,7 @@ class ChunkedRecordingManager(private val encoder: MediaCodec, private val outpu
|
||||
encoder.releaseOutputBuffer(index, false)
|
||||
return
|
||||
}
|
||||
context.muxer.writeSampleData(context.videoTrack, encodedData, bufferInfo)
|
||||
context.writeSample(encodedData, bufferInfo)
|
||||
encoder.releaseOutputBuffer(index, false)
|
||||
}
|
||||
}
|
||||
|
||||
@@ -39,18 +39,11 @@ class FragmentedRecordingManager(
|
||||
segmentDurationSeconds: Int = DEFAULT_SEGMENT_DURATION_SECONDS
|
||||
): FragmentedRecordingManager {
|
||||
val mimeType = options.videoCodec.toMimeType()
|
||||
// Use cameraOrientation (from WindowManager) for rotation metadata
|
||||
// The options.orientation from JavaScript is unreliable on Android when rotating between landscape modes
|
||||
val orientationDegrees = cameraOrientation.toDegrees()
|
||||
val cameraOrientationDegrees = cameraOrientation.toDegrees()
|
||||
val orientationDegrees = 0
|
||||
val (width, height) = size.width to size.height
|
||||
|
||||
// Swap dimensions based on camera orientation, same as ChunkedRecordingManager
|
||||
val (width, height) = if (cameraOrientation.isLandscape()) {
|
||||
size.height to size.width
|
||||
} else {
|
||||
size.width to size.height
|
||||
}
|
||||
|
||||
Log.d(TAG, "Recording: ${width}x${height}, orientation=$orientationDegrees°")
|
||||
Log.d(TAG, "Recording: ${width}x${height}, orientation=$orientationDegrees° (cameraOrientation=$cameraOrientationDegrees°)")
|
||||
|
||||
val format = MediaFormat.createVideoFormat(mimeType, width, height)
|
||||
val codec = MediaCodec.createEncoderByType(mimeType)
|
||||
|
||||
@@ -753,17 +753,32 @@ class HlsMuxer(
|
||||
dos.writeShort(-1) // pre-defined
|
||||
|
||||
output.write(buildAvcCBox(sps, pps))
|
||||
output.write(buildPaspBox())
|
||||
|
||||
return wrapBox("avc1", output.toByteArray())
|
||||
}
|
||||
|
||||
/**
|
||||
* Builds pixel aspect ratio box to explicitly declare square pixels (1:1).
|
||||
* This helps players correctly interpret video dimensions without SAR scaling.
|
||||
*/
|
||||
private fun buildPaspBox(): ByteArray {
|
||||
val output = ByteArrayOutputStream()
|
||||
val dos = DataOutputStream(output)
|
||||
dos.writeInt(1) // hSpacing (horizontal)
|
||||
dos.writeInt(1) // vSpacing (vertical)
|
||||
return wrapBox("pasp", output.toByteArray())
|
||||
}
|
||||
|
||||
private fun buildAvcCBox(sps: ByteArray, pps: ByteArray): ByteArray {
|
||||
val output = ByteArrayOutputStream()
|
||||
val dos = DataOutputStream(output)
|
||||
|
||||
val profileIdc = if (sps.isNotEmpty()) sps[0].toInt() and 0xFF else 0x42
|
||||
val profileCompat = if (sps.size > 1) sps[1].toInt() and 0xFF else 0x00
|
||||
val levelIdc = if (sps.size > 2) sps[2].toInt() and 0xFF else 0x1F
|
||||
// SPS NAL unit format: [NAL header, profile_idc, constraint_flags, level_idc, ...]
|
||||
// Skip byte 0 (NAL header, typically 0x67) to get the actual profile data
|
||||
val profileIdc = if (sps.size > 1) sps[1].toInt() and 0xFF else 0x42
|
||||
val profileCompat = if (sps.size > 2) sps[2].toInt() and 0xFF else 0x00
|
||||
val levelIdc = if (sps.size > 3) sps[3].toInt() and 0xFF else 0x1F
|
||||
|
||||
dos.writeByte(1) // configuration version
|
||||
dos.writeByte(profileIdc) // AVC profile
|
||||
|
||||
@@ -235,10 +235,19 @@ class PersistentCameraCaptureSession(private val cameraManager: CameraManager, p
|
||||
|
||||
// 1. Run a precapture sequence for AF, AE and AWB.
|
||||
focusJob = coroutineScope.launch {
|
||||
try {
|
||||
val request = repeatingRequest.createCaptureRequest(device, deviceDetails, outputs)
|
||||
val options =
|
||||
PrecaptureOptions(listOf(PrecaptureTrigger.AF, PrecaptureTrigger.AE), Flash.OFF, listOf(point), false, FOCUS_RESET_TIMEOUT)
|
||||
session.precapture(request, deviceDetails, options)
|
||||
} catch (e: CaptureTimedOutError) {
|
||||
// Focus timed out - this is non-fatal, just log and continue
|
||||
Log.w(TAG, "Focus timed out at point $point, continuing without focus lock")
|
||||
} catch (e: IllegalStateException) {
|
||||
Log.w(TAG, "Focus failed, camera device was already closed: ${e.message}")
|
||||
} catch (e: CameraAccessException) {
|
||||
Log.w(TAG, "Focus failed, camera not accessible: ${e.message}")
|
||||
}
|
||||
}
|
||||
focusJob?.join()
|
||||
|
||||
@@ -254,9 +263,15 @@ class PersistentCameraCaptureSession(private val cameraManager: CameraManager, p
|
||||
return@launch
|
||||
}
|
||||
Log.i(TAG, "Resetting focus to auto-focus...")
|
||||
try {
|
||||
repeatingRequest.createCaptureRequest(device, deviceDetails, outputs).also { request ->
|
||||
session.setRepeatingRequest(request.build(), null, null)
|
||||
}
|
||||
} catch (e: IllegalStateException) {
|
||||
Log.w(TAG, "Failed to reset focus, camera device was already closed: ${e.message}")
|
||||
} catch (e: CameraAccessException) {
|
||||
Log.w(TAG, "Failed to reset focus, camera not accessible: ${e.message}")
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
@@ -5,7 +5,7 @@ import android.content.Context
|
||||
import android.content.res.Configuration
|
||||
import android.graphics.Point
|
||||
import android.os.Handler
|
||||
import android.os.Looper
|
||||
import android.os.HandlerThread
|
||||
import android.util.Log
|
||||
import android.util.Size
|
||||
import android.view.PixelCopy
|
||||
@@ -25,58 +25,72 @@ import kotlinx.coroutines.Dispatchers
|
||||
import kotlinx.coroutines.suspendCancellableCoroutine
|
||||
import kotlinx.coroutines.withContext
|
||||
import android.graphics.Bitmap
|
||||
import android.graphics.Matrix
|
||||
|
||||
fun rotateBitmap90CounterClockwise(source: Bitmap): Bitmap {
|
||||
val width = source.width
|
||||
val height = source.height
|
||||
|
||||
// Create a new Bitmap with swapped width and height
|
||||
val rotatedBitmap = Bitmap.createBitmap(height, width, source.config ?: Bitmap.Config.ARGB_8888)
|
||||
|
||||
for (y in 0 until height) {
|
||||
for (x in 0 until width) {
|
||||
// Set the pixel in the new position
|
||||
rotatedBitmap.setPixel(y, width - 1 - x, source.getPixel(x, y))
|
||||
}
|
||||
}
|
||||
|
||||
return rotatedBitmap
|
||||
}
|
||||
|
||||
|
||||
fun Bitmap.transformBitmap(orientation: Orientation): Bitmap {
|
||||
return when (orientation) {
|
||||
Orientation.PORTRAIT -> this // No transformation needed
|
||||
Orientation.LANDSCAPE_LEFT -> {
|
||||
// Transpose (swap width and height)
|
||||
val transposedBitmap = Bitmap.createBitmap(height, width, config ?: Bitmap.Config.ARGB_8888)
|
||||
for (y in 0 until height) {
|
||||
for (x in 0 until width) {
|
||||
transposedBitmap.setPixel(y, width - 1 - x, getPixel(x, y))
|
||||
val srcWidth = width
|
||||
val srcHeight = height
|
||||
val sourcePixels = IntArray(srcWidth * srcHeight)
|
||||
getPixels(sourcePixels, 0, srcWidth, 0, 0, srcWidth, srcHeight)
|
||||
|
||||
val dstWidth = srcHeight
|
||||
val dstHeight = srcWidth
|
||||
val destinationPixels = IntArray(dstWidth * dstHeight)
|
||||
for (y in 0 until srcHeight) {
|
||||
for (x in 0 until srcWidth) {
|
||||
val dstX = y
|
||||
val dstY = srcWidth - 1 - x
|
||||
destinationPixels[dstY * dstWidth + dstX] = sourcePixels[y * srcWidth + x]
|
||||
}
|
||||
}
|
||||
transposedBitmap
|
||||
|
||||
val transformedBitmap = Bitmap.createBitmap(dstWidth, dstHeight, config ?: Bitmap.Config.ARGB_8888)
|
||||
transformedBitmap.setPixels(destinationPixels, 0, dstWidth, 0, 0, dstWidth, dstHeight)
|
||||
transformedBitmap
|
||||
}
|
||||
Orientation.PORTRAIT_UPSIDE_DOWN -> {
|
||||
// Invert vertically and horizontally (180-degree rotation)
|
||||
val invertedBitmap = Bitmap.createBitmap(width, height, config ?: Bitmap.Config.ARGB_8888)
|
||||
for (y in 0 until height) {
|
||||
for (x in 0 until width) {
|
||||
invertedBitmap.setPixel(width - 1 - x, height - 1 - y, getPixel(x, y))
|
||||
val srcWidth = width
|
||||
val srcHeight = height
|
||||
val sourcePixels = IntArray(srcWidth * srcHeight)
|
||||
getPixels(sourcePixels, 0, srcWidth, 0, 0, srcWidth, srcHeight)
|
||||
|
||||
val dstWidth = srcWidth
|
||||
val dstHeight = srcHeight
|
||||
val destinationPixels = IntArray(dstWidth * dstHeight)
|
||||
for (y in 0 until srcHeight) {
|
||||
for (x in 0 until srcWidth) {
|
||||
val dstX = srcWidth - 1 - x
|
||||
val dstY = srcHeight - 1 - y
|
||||
destinationPixels[dstY * dstWidth + dstX] = sourcePixels[y * srcWidth + x]
|
||||
}
|
||||
}
|
||||
invertedBitmap
|
||||
|
||||
val transformedBitmap = Bitmap.createBitmap(dstWidth, dstHeight, config ?: Bitmap.Config.ARGB_8888)
|
||||
transformedBitmap.setPixels(destinationPixels, 0, dstWidth, 0, 0, dstWidth, dstHeight)
|
||||
transformedBitmap
|
||||
}
|
||||
Orientation.LANDSCAPE_RIGHT -> {
|
||||
// Transpose (swap width and height) and invert vertically
|
||||
val transposedBitmap = Bitmap.createBitmap(height, width, config ?: Bitmap.Config.ARGB_8888)
|
||||
for (y in 0 until height) {
|
||||
for (x in 0 until width) {
|
||||
transposedBitmap.setPixel(height - 1 - y, x, getPixel(x, y))
|
||||
val srcWidth = width
|
||||
val srcHeight = height
|
||||
val sourcePixels = IntArray(srcWidth * srcHeight)
|
||||
getPixels(sourcePixels, 0, srcWidth, 0, 0, srcWidth, srcHeight)
|
||||
|
||||
val dstWidth = srcHeight
|
||||
val dstHeight = srcWidth
|
||||
val destinationPixels = IntArray(dstWidth * dstHeight)
|
||||
for (y in 0 until srcHeight) {
|
||||
for (x in 0 until srcWidth) {
|
||||
val dstX = srcHeight - 1 - y
|
||||
val dstY = x
|
||||
destinationPixels[dstY * dstWidth + dstX] = sourcePixels[y * srcWidth + x]
|
||||
}
|
||||
}
|
||||
transposedBitmap
|
||||
|
||||
val transformedBitmap = Bitmap.createBitmap(dstWidth, dstHeight, config ?: Bitmap.Config.ARGB_8888)
|
||||
transformedBitmap.setPixels(destinationPixels, 0, dstWidth, 0, 0, dstWidth, dstHeight)
|
||||
transformedBitmap
|
||||
}
|
||||
}
|
||||
}
|
||||
@@ -186,7 +200,7 @@ class PreviewView(context: Context, callback: SurfaceHolder.Callback) :
|
||||
)
|
||||
}
|
||||
},
|
||||
Handler(Looper.getMainLooper())
|
||||
pixelCopyHandler
|
||||
)
|
||||
}
|
||||
}
|
||||
@@ -198,8 +212,10 @@ class PreviewView(context: Context, callback: SurfaceHolder.Callback) :
|
||||
val viewOrientation = Orientation.PORTRAIT
|
||||
|
||||
val rotated = point.rotatedBy(viewSize, cameraSize, viewOrientation, sensorOrientation)
|
||||
Log.i(TAG, "Converted layer point $point to camera point $rotated! ($sensorOrientation, $cameraSize -> $viewSize)")
|
||||
return rotated
|
||||
// Clamp to valid camera coordinates (must be non-negative for MeteringRectangle)
|
||||
val clamped = Point(maxOf(0, rotated.x), maxOf(0, rotated.y))
|
||||
Log.i(TAG, "Converted layer point $point to camera point $clamped! ($sensorOrientation, $cameraSize -> $viewSize)")
|
||||
return clamped
|
||||
}
|
||||
|
||||
private fun updateLayout() {
|
||||
@@ -254,5 +270,10 @@ class PreviewView(context: Context, callback: SurfaceHolder.Callback) :
|
||||
|
||||
companion object {
|
||||
private const val TAG = "PreviewView"
|
||||
private val pixelCopyHandler: Handler by lazy {
|
||||
val handlerThread = HandlerThread("VisionCamera.PixelCopy")
|
||||
handlerThread.start()
|
||||
Handler(handlerThread.looper)
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
@@ -8,6 +8,7 @@ import com.facebook.common.statfs.StatFsHelper
|
||||
import com.mrousavy.camera.extensions.getRecommendedBitRate
|
||||
import com.mrousavy.camera.types.Orientation
|
||||
import com.mrousavy.camera.types.RecordVideoOptions
|
||||
import com.mrousavy.camera.types.StreamSegmentType
|
||||
import com.mrousavy.camera.utils.FileUtils
|
||||
import java.io.File
|
||||
import android.os.Environment
|
||||
@@ -27,9 +28,7 @@ class RecordingSession(
|
||||
private val filePath: String,
|
||||
private val callback: (video: Video) -> Unit,
|
||||
private val onError: (error: CameraError) -> Unit,
|
||||
private val allCallbacks: CameraSession.Callback,
|
||||
// Use FragmentedRecordingManager for HLS-compatible fMP4 output
|
||||
private val useFragmentedMp4: Boolean = true
|
||||
private val allCallbacks: CameraSession.Callback
|
||||
) {
|
||||
companion object {
|
||||
private const val TAG = "RecordingSession"
|
||||
@@ -44,14 +43,26 @@ class RecordingSession(
|
||||
|
||||
data class Video(val path: String, val durationMs: Long, val size: Size)
|
||||
|
||||
val outputRotationDegrees: Int =
|
||||
if (options.streamSegmentType == StreamSegmentType.FRAGMENTED_MP4) {
|
||||
when (cameraOrientation.toDegrees()) {
|
||||
90 -> 270
|
||||
270 -> 90
|
||||
else -> cameraOrientation.toDegrees()
|
||||
}
|
||||
} else {
|
||||
0
|
||||
}
|
||||
|
||||
// Normalize path - expo-file-system passes file:// URIs but File expects raw paths
|
||||
private val outputPath: File = File(filePath.removePrefix("file://"))
|
||||
// Handle both file:// and file:/ variants
|
||||
private val outputPath: File = File(filePath.replace(Regex("^file:/+"), "/"))
|
||||
|
||||
private val bitRate = getBitRate()
|
||||
|
||||
// Use FragmentedRecordingManager for HLS-compatible fMP4 output,
|
||||
// or fall back to ChunkedRecordingManager for regular MP4 chunks
|
||||
private val recorder: ChunkedRecorderInterface = if (useFragmentedMp4) {
|
||||
private val recorder: ChunkedRecorderInterface = if (options.streamSegmentType == StreamSegmentType.FRAGMENTED_MP4) {
|
||||
FragmentedRecordingManager.fromParams(
|
||||
allCallbacks,
|
||||
size,
|
||||
@@ -83,7 +94,7 @@ class RecordingSession(
|
||||
|
||||
fun start() {
|
||||
synchronized(this) {
|
||||
Log.i(TAG, "Starting RecordingSession..")
|
||||
Log.i(TAG, "Starting RecordingSession with ${options.streamSegmentType} recorder..")
|
||||
startTime = System.currentTimeMillis()
|
||||
recorder.start()
|
||||
}
|
||||
|
||||
@@ -186,7 +186,7 @@ class VideoPipeline(
|
||||
if (recordingSession != null) {
|
||||
// Configure OpenGL pipeline to stream Frames into the Recording Session's surface
|
||||
Log.i(TAG, "Setting ${recordingSession.size} RecordingSession Output...")
|
||||
setRecordingSessionOutputSurface(recordingSession.surface)
|
||||
setRecordingSessionOutputSurface(recordingSession.surface, recordingSession.outputRotationDegrees)
|
||||
this.recordingSession = recordingSession
|
||||
} else {
|
||||
// Configure OpenGL pipeline to stop streaming Frames into the Recording Session's surface
|
||||
@@ -250,7 +250,7 @@ class VideoPipeline(
|
||||
private external fun getInputTextureId(): Int
|
||||
private external fun onBeforeFrame()
|
||||
private external fun onFrame(transformMatrix: FloatArray)
|
||||
private external fun setRecordingSessionOutputSurface(surface: Any)
|
||||
private external fun setRecordingSessionOutputSurface(surface: Any, rotationDegrees: Int)
|
||||
private external fun removeRecordingSessionOutputSurface()
|
||||
private external fun initHybrid(width: Int, height: Int): HybridData
|
||||
}
|
||||
|
||||
@@ -9,6 +9,7 @@ class RecordVideoOptions(map: ReadableMap) {
|
||||
var videoBitRateOverride: Double? = null
|
||||
var videoBitRateMultiplier: Double? = null
|
||||
var orientation: Orientation? = null
|
||||
var streamSegmentType: StreamSegmentType = StreamSegmentType.FRAGMENTED_MP4
|
||||
|
||||
init {
|
||||
if (map.hasKey("fileType")) {
|
||||
@@ -29,5 +30,8 @@ class RecordVideoOptions(map: ReadableMap) {
|
||||
if (map.hasKey("orientation")) {
|
||||
orientation = Orientation.fromUnionValue(map.getString("orientation"))
|
||||
}
|
||||
if (map.hasKey("streamSegmentType")) {
|
||||
streamSegmentType = StreamSegmentType.fromUnionValue(map.getString("streamSegmentType"))
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
@@ -0,0 +1,15 @@
|
||||
package com.mrousavy.camera.types
|
||||
|
||||
enum class StreamSegmentType(override val unionValue: String) : JSUnionValue {
|
||||
FRAGMENTED_MP4("FRAGMENTED_MP4"),
|
||||
RB_CHUNKED_MP4("RB_CHUNKED_MP4");
|
||||
|
||||
companion object : JSUnionValue.Companion<StreamSegmentType> {
|
||||
override fun fromUnionValue(unionValue: String?): StreamSegmentType =
|
||||
when (unionValue) {
|
||||
"FRAGMENTED_MP4" -> FRAGMENTED_MP4
|
||||
"RB_CHUNKED_MP4" -> RB_CHUNKED_MP4
|
||||
else -> FRAGMENTED_MP4 // Default to fMP4
|
||||
}
|
||||
}
|
||||
}
|
||||
@@ -41,6 +41,17 @@ export interface RecordVideoOptions {
|
||||
* @default 'normal'
|
||||
*/
|
||||
videoBitRate?: 'extra-low' | 'low' | 'normal' | 'high' | 'extra-high' | number
|
||||
/**
|
||||
* The stream segment type for recording on Android.
|
||||
* - `FRAGMENTED_MP4`: HLS-compatible segments (init.mp4 + numbered segments)
|
||||
* - `RB_CHUNKED_MP4`: Legacy chunked MP4 format
|
||||
*
|
||||
* iOS always uses FRAGMENTED_MP4 regardless of this setting.
|
||||
*
|
||||
* @platform android
|
||||
* @default 'FRAGMENTED_MP4'
|
||||
*/
|
||||
streamSegmentType?: 'FRAGMENTED_MP4' | 'RB_CHUNKED_MP4'
|
||||
}
|
||||
|
||||
/**
|
||||
|
||||
Reference in New Issue
Block a user