Universal UVCCamera library,supporting recording, pushing, etc
You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.

1044 lines
30 KiB

8 years ago
package com.serenegiant.usb.common;
import android.annotation.TargetApi;
8 years ago
import android.app.Activity;
import android.graphics.Bitmap;
8 years ago
import android.graphics.SurfaceTexture;
import android.hardware.usb.UsbDevice;
import android.media.MediaScannerConnection;
import android.os.Build;
import android.os.Environment;
8 years ago
import android.os.Handler;
import android.os.Looper;
import android.os.Message;
import android.text.TextUtils;
8 years ago
import android.util.Log;
import android.view.Surface;
import android.view.SurfaceHolder;
import com.jiangdg.usbcamera.FileUtils;
8 years ago
import com.serenegiant.usb.IFrameCallback;
import com.serenegiant.usb.USBMonitor;
import com.serenegiant.usb.UVCCamera;
import com.serenegiant.usb.encoder.MediaAudioEncoder;
import com.serenegiant.usb.encoder.MediaEncoder;
import com.serenegiant.usb.encoder.MediaMuxerWrapper;
import com.serenegiant.usb.encoder.MediaSurfaceEncoder;
import com.serenegiant.usb.encoder.MediaVideoBufferEncoder;
import com.serenegiant.usb.encoder.MediaVideoEncoder;
import com.serenegiant.usb.encoder.RecordParams;
import com.serenegiant.usb.encoder.biz.AACEncodeConsumer;
import com.serenegiant.usb.encoder.biz.H264EncodeConsumer;
import com.serenegiant.usb.encoder.biz.Mp4MediaMuxer;
8 years ago
import com.serenegiant.usb.widget.CameraViewInterface;
import java.io.BufferedOutputStream;
import java.io.File;
import java.io.FileOutputStream;
import java.io.IOException;
8 years ago
import java.lang.ref.WeakReference;
import java.lang.reflect.Constructor;
import java.lang.reflect.InvocationTargetException;
import java.nio.ByteBuffer;
import java.text.SimpleDateFormat;
import java.util.Date;
8 years ago
import java.util.Set;
import java.util.concurrent.CopyOnWriteArraySet;
/**
* Camera业务处理抽象类
*
* */
public abstract class AbstractUVCCameraHandler extends Handler {
private static final boolean DEBUG = true; // TODO set false on release
private static final String TAG = "AbsUVCCameraHandler";
// 对外回调接口
public interface CameraCallback {
public void onOpen();
public void onClose();
public void onStartPreview();
public void onStopPreview();
public void onStartRecording();
public void onStopRecording();
public void onError(final Exception e);
}
public static OnEncodeResultListener mListener;
public static OnPreViewResultListener mPreviewListener;
public interface OnEncodeResultListener{
void onEncodeResult(byte[] data, int offset, int length, long timestamp, int type);
}
public interface OnPreViewResultListener{
void onPreviewResult(boolean result);
}
private static final int MSG_OPEN = 0;
private static final int MSG_CLOSE = 1;
private static final int MSG_PREVIEW_START = 2;
private static final int MSG_PREVIEW_STOP = 3;
private static final int MSG_CAPTURE_STILL = 4;
private static final int MSG_CAPTURE_START = 5;
private static final int MSG_CAPTURE_STOP = 6;
private static final int MSG_MEDIA_UPDATE = 7;
private static final int MSG_RELEASE = 9;
// 音频线程
// private static final int MSG_AUDIO_START = 10;
// private static final int MSG_AUDIO_STOP = 11;
private final WeakReference<CameraThread> mWeakThread;
private volatile boolean mReleased;
protected AbstractUVCCameraHandler(final CameraThread thread) {
mWeakThread = new WeakReference<CameraThread>(thread);
}
public int getWidth() {
final CameraThread thread = mWeakThread.get();
return thread != null ? thread.getWidth() : 0;
}
public int getHeight() {
final CameraThread thread = mWeakThread.get();
return thread != null ? thread.getHeight() : 0;
}
public boolean isOpened() {
final CameraThread thread = mWeakThread.get();
return thread != null && thread.isCameraOpened();
}
public boolean isPreviewing() {
final CameraThread thread = mWeakThread.get();
return thread != null && thread.isPreviewing();
}
public boolean isRecording() {
final CameraThread thread = mWeakThread.get();
return thread != null && thread.isRecording();
}
// public boolean isAudioThreadStart() {
// final CameraThread thread = mWeakThread.get();
// return thread != null && thread.isAudioRecording();
// }
public boolean isEqual(final UsbDevice device) {
final CameraThread thread = mWeakThread.get();
return (thread != null) && thread.isEqual(device);
}
protected boolean isCameraThread() {
final CameraThread thread = mWeakThread.get();
return thread != null && (thread.getId() == Thread.currentThread().getId());
}
protected boolean isReleased() {
final CameraThread thread = mWeakThread.get();
return mReleased || (thread == null);
}
protected void checkReleased() {
if (isReleased()) {
throw new IllegalStateException("already released");
}
}
public void open(final USBMonitor.UsbControlBlock ctrlBlock) {
checkReleased();
sendMessage(obtainMessage(MSG_OPEN, ctrlBlock));
}
public void close() {
if (DEBUG) Log.v(TAG, "close:");
if (isOpened()) {
stopPreview();
sendEmptyMessage(MSG_CLOSE);
}
if (DEBUG) Log.v(TAG, "close:finished");
}
// 切换分辨率
public void resize(final int width, final int height) {
checkReleased();
throw new UnsupportedOperationException("does not support now");
}
// 开启Camera预览
protected void startPreview(final Object surface,OnPreViewResultListener listener) {
checkReleased();
if (!((surface instanceof SurfaceHolder) || (surface instanceof Surface) || (surface instanceof SurfaceTexture))) {
throw new IllegalArgumentException("surface should be one of SurfaceHolder, Surface or SurfaceTexture");
}
this.mPreviewListener = listener;
sendMessage(obtainMessage(MSG_PREVIEW_START, surface));
}
// 关闭Camera预览
public void stopPreview() {
if (DEBUG) Log.v(TAG, "stopPreview:");
removeMessages(MSG_PREVIEW_START);
stopRecording();
if (isPreviewing()) {
final CameraThread thread = mWeakThread.get();
if (thread == null) return;
synchronized (thread.mSync) {
sendEmptyMessage(MSG_PREVIEW_STOP);
if (!isCameraThread()) {
// wait for actually preview stopped to avoid releasing Surface/SurfaceTexture
// while preview is still running.
// therefore this method will take a time to execute
try {
thread.mSync.wait();
} catch (final InterruptedException e) {
}
}
}
}
if (DEBUG) Log.v(TAG, "stopPreview:finished");
}
// 捕获图像
protected void captureStill() {
checkReleased();
sendEmptyMessage(MSG_CAPTURE_STILL);
}
protected void captureStill(final String path) {
checkReleased();
sendMessage(obtainMessage(MSG_CAPTURE_STILL, path));
}
// 开始录制
public void startRecording(final RecordParams params, OnEncodeResultListener listener) {
AbstractUVCCameraHandler.mListener = listener;
checkReleased();
sendMessage(obtainMessage(MSG_CAPTURE_START, params));
}
// 停止录制
public void stopRecording() {
sendEmptyMessage(MSG_CAPTURE_STOP);
}
// // 启动音频线程
// public void startAudioThread(){
// sendEmptyMessage(MSG_AUDIO_START);
// }
//
// // 关闭音频线程
// public void stopAudioThread(){
// sendEmptyMessage(MSG_AUDIO_STOP);
// }
public void release() {
mReleased = true;
close();
sendEmptyMessage(MSG_RELEASE);
}
// 对外注册监听事件
public void addCallback(final CameraCallback callback) {
checkReleased();
if (!mReleased && (callback != null)) {
final CameraThread thread = mWeakThread.get();
if (thread != null) {
thread.mCallbacks.add(callback);
}
}
}
public void removeCallback(final CameraCallback callback) {
if (callback != null) {
final CameraThread thread = mWeakThread.get();
if (thread != null) {
thread.mCallbacks.remove(callback);
}
}
}
protected void updateMedia(final String path) {
sendMessage(obtainMessage(MSG_MEDIA_UPDATE, path));
}
public boolean checkSupportFlag(final long flag) {
checkReleased();
final CameraThread thread = mWeakThread.get();
return thread != null && thread.mUVCCamera != null && thread.mUVCCamera.checkSupportFlag(flag);
}
public int getValue(final int flag) {
checkReleased();
final CameraThread thread = mWeakThread.get();
final UVCCamera camera = thread != null ? thread.mUVCCamera : null;
if (camera != null) {
if (flag == UVCCamera.PU_BRIGHTNESS) {
return camera.getBrightness();
} else if (flag == UVCCamera.PU_CONTRAST) {
return camera.getContrast();
}
}
throw new IllegalStateException();
}
public int setValue(final int flag, final int value) {
checkReleased();
final CameraThread thread = mWeakThread.get();
final UVCCamera camera = thread != null ? thread.mUVCCamera : null;
if (camera != null) {
if (flag == UVCCamera.PU_BRIGHTNESS) {
camera.setBrightness(value);
return camera.getBrightness();
} else if (flag == UVCCamera.PU_CONTRAST) {
camera.setContrast(value);
return camera.getContrast();
}
}
throw new IllegalStateException();
}
public int resetValue(final int flag) {
checkReleased();
final CameraThread thread = mWeakThread.get();
final UVCCamera camera = thread != null ? thread.mUVCCamera : null;
if (camera != null) {
if (flag == UVCCamera.PU_BRIGHTNESS) {
camera.resetBrightness();
return camera.getBrightness();
} else if (flag == UVCCamera.PU_CONTRAST) {
camera.resetContrast();
return camera.getContrast();
}
}
throw new IllegalStateException();
}
@Override
public void handleMessage(final Message msg) {
final CameraThread thread = mWeakThread.get();
if (thread == null) return;
switch (msg.what) {
case MSG_OPEN:
thread.handleOpen((USBMonitor.UsbControlBlock)msg.obj);
break;
case MSG_CLOSE:
thread.handleClose();
break;
case MSG_PREVIEW_START:
thread.handleStartPreview(msg.obj);
break;
case MSG_PREVIEW_STOP:
thread.handleStopPreview();
break;
case MSG_CAPTURE_STILL:
thread.handleCaptureStill((String)msg.obj);
break;
case MSG_CAPTURE_START:
// thread.handleStartRecording((String)msg.obj);
thread.handleStartRecording((RecordParams)msg.obj);
break;
case MSG_CAPTURE_STOP:
thread.handleStopRecording();
break;
case MSG_MEDIA_UPDATE:
thread.handleUpdateMedia((String)msg.obj);
break;
case MSG_RELEASE:
thread.handleRelease();
break;
// 音频线程
// case MSG_AUDIO_START:
// thread.startAudioRecord();
// break;
// case MSG_AUDIO_STOP:
// thread.stopAudioRecord();
// break;
default:
throw new RuntimeException("unsupported message:what=" + msg.what);
}
}
public static final class CameraThread extends Thread {
private static final String TAG_THREAD = "CameraThread";
private final Object mSync = new Object();
private final Class<? extends AbstractUVCCameraHandler> mHandlerClass;
private final WeakReference<Activity> mWeakParent;
private final WeakReference<CameraViewInterface> mWeakCameraView;
private final int mEncoderType;
private final Set<CameraCallback> mCallbacks = new CopyOnWriteArraySet<CameraCallback>();
private int mWidth, mHeight, mPreviewMode;
private float mBandwidthFactor;
private boolean mIsPreviewing;
private boolean mIsRecording;
// 播放声音
// private SoundPool mSoundPool;
// private int mSoundId;
private AbstractUVCCameraHandler mHandler;
// 处理与Camera相关的逻辑,比如获取byte数据流等
private UVCCamera mUVCCamera;
// private MediaMuxerWrapper mMuxer;
private MediaVideoBufferEncoder mVideoEncoder;
private Mp4MediaMuxer mMuxer;
// private boolean isAudioThreadStart;
/** 构造方法
*
* clazz 继承于AbstractUVCCameraHandler
* parent Activity子类
* cameraView 用于捕获静止图像
* encoderType 0表示使用MediaSurfaceEncoder;1表示使用MediaVideoEncoder, 2表示使用MediaVideoBufferEncoder
* width 分辨率的宽
* height 分辨率的高
* format 颜色格式0为FRAME_FORMAT_YUYV1为FRAME_FORMAT_MJPEG
* bandwidthFactor
*/
CameraThread(final Class<? extends AbstractUVCCameraHandler> clazz,
final Activity parent, final CameraViewInterface cameraView,
final int encoderType, final int width, final int height, final int format,
final float bandwidthFactor) {
super("CameraThread");
mHandlerClass = clazz;
mEncoderType = encoderType;
mWidth = width;
mHeight = height;
mPreviewMode = format;
mBandwidthFactor = bandwidthFactor;
mWeakParent = new WeakReference<>(parent);
mWeakCameraView = new WeakReference<>(cameraView);
// loadShutterSound(parent);
}
@Override
protected void finalize() throws Throwable {
Log.i(TAG, "CameraThread#finalize");
super.finalize();
}
public AbstractUVCCameraHandler getHandler() {
if (DEBUG) Log.v(TAG_THREAD, "getHandler:");
synchronized (mSync) {
if (mHandler == null)
try {
mSync.wait();
} catch (final InterruptedException e) {
}
}
return mHandler;
}
public int getWidth() {
synchronized (mSync) {
return mWidth;
}
}
public int getHeight() {
synchronized (mSync) {
return mHeight;
}
}
public boolean isCameraOpened() {
synchronized (mSync) {
return mUVCCamera != null;
}
}
public boolean isPreviewing() {
synchronized (mSync) {
return mUVCCamera != null && mIsPreviewing;
}
}
public boolean isRecording() {
synchronized (mSync) {
return (mUVCCamera != null) && (mMuxer != null);
}
}
// public boolean isAudioRecording(){
// synchronized (mSync){
// return isAudioThreadStart;
// }
// }
public boolean isEqual(final UsbDevice device) {
return (mUVCCamera != null) && (mUVCCamera.getDevice() != null) && mUVCCamera.getDevice().equals(device);
}
public void handleOpen(final USBMonitor.UsbControlBlock ctrlBlock) {
if (DEBUG) Log.v(TAG_THREAD, "handleOpen:");
handleClose();
try {
final UVCCamera camera = new UVCCamera();
camera.open(ctrlBlock);
synchronized (mSync) {
mUVCCamera = camera;
}
callOnOpen();
} catch (final Exception e) {
callOnError(e);
}
if (DEBUG) Log.i(TAG, "supportedSize:" + (mUVCCamera != null ? mUVCCamera.getSupportedSize() : null));
}
public void handleClose() {
if (DEBUG) Log.v(TAG_THREAD, "handleClose:");
handleStopRecording();
final UVCCamera camera;
synchronized (mSync) {
camera = mUVCCamera;
mUVCCamera = null;
}
if (camera != null) {
camera.stopPreview();
camera.destroy();
callOnClose();
}
}
public void handleStartPreview(final Object surface) {
if (DEBUG) Log.v(TAG_THREAD, "handleStartPreview:");
if ((mUVCCamera == null) || mIsPreviewing) return;
try {
mUVCCamera.setPreviewSize(mWidth, mHeight, 1, 31, mPreviewMode, mBandwidthFactor);
if(mPreviewListener != null){
mPreviewListener.onPreviewResult(true);
}
} catch (final IllegalArgumentException e) {
// 添加分辨率参数合法性检测
if(mPreviewListener != null){
mPreviewListener.onPreviewResult(false);
}
// try {
// // fallback to YUV mode
// mUVCCamera.setPreviewSize(mWidth, mHeight, 1, 31, UVCCamera.DEFAULT_PREVIEW_MODE, mBandwidthFactor);
// } catch (final IllegalArgumentException e1) {
// callOnError(e1);
// return;
// }
}
if (surface instanceof SurfaceHolder) {
mUVCCamera.setPreviewDisplay((SurfaceHolder)surface);
} if (surface instanceof Surface) {
mUVCCamera.setPreviewDisplay((Surface)surface);
} else {
mUVCCamera.setPreviewTexture((SurfaceTexture)surface);
}
mUVCCamera.startPreview();
mUVCCamera.updateCameraParams();
synchronized (mSync) {
mIsPreviewing = true;
}
callOnStartPreview();
}
public void handleStopPreview() {
if (DEBUG) Log.v(TAG_THREAD, "handleStopPreview:");
if (mIsPreviewing) {
if (mUVCCamera != null) {
mUVCCamera.stopPreview();
}
synchronized (mSync) {
mIsPreviewing = false;
mSync.notifyAll();
}
callOnStopPreview();
}
if (DEBUG) Log.v(TAG_THREAD, "handleStopPreview:finished");
}
// 捕获静态图片
public void handleCaptureStill(final String path) {
if (DEBUG) Log.v(TAG_THREAD, "handleCaptureStill:");
final Activity parent = mWeakParent.get();
if (parent == null) return;
// mSoundPool.play(mSoundId, 0.2f, 0.2f, 0, 0, 1.0f); // play shutter sound
try {
final Bitmap bitmap = mWeakCameraView.get().captureStillImage();
// get buffered output stream for saving a captured still image as a file on external storage.
// the file name is came from current time.
// You should use extension name as same as CompressFormat when calling Bitmap#compress.
final File outputFile = TextUtils.isEmpty(path)
? MediaMuxerWrapper.getCaptureFile(Environment.DIRECTORY_DCIM, ".png")
: new File(path);
final BufferedOutputStream os = new BufferedOutputStream(new FileOutputStream(outputFile));
try {
try {
bitmap.compress(Bitmap.CompressFormat.PNG, 100, os);
os.flush();
mHandler.sendMessage(mHandler.obtainMessage(MSG_MEDIA_UPDATE, outputFile.getPath()));
} catch (final IOException e) {
}
} finally {
os.close();
}
} catch (final Exception e) {
callOnError(e);
}
}
// 开始录制视频
// public void handleStartRecording2(String path) {
// if (DEBUG) Log.v(TAG_THREAD, "handleStartRecording:");
// try {
// if ((mUVCCamera == null) || (mMuxer != null)) return;
//// final MediaMuxerWrapper muxer = new MediaMuxerWrapper(".mp4"); // if you record audio only, ".m4a" is also OK.
// final MediaMuxerWrapper muxer = new MediaMuxerWrapper(path);
// MediaVideoBufferEncoder videoEncoder = null;
// switch (mEncoderType) {
// case 1: // for video capturing using MediaVideoEncoder
// // 开启视频编码线程
// new MediaVideoEncoder(muxer,getWidth(), getHeight(), mMediaEncoderListener);
// break;
// case 2: // for video capturing using MediaVideoBufferEncoder
// videoEncoder = new MediaVideoBufferEncoder(muxer, getWidth(), getHeight(), mMediaEncoderListener);
// break;
// // case 0: // for video capturing using MediaSurfaceEncoder
// default:
// new MediaSurfaceEncoder(muxer, getWidth(), getHeight(), mMediaEncoderListener);
// break;
// }
// // 开启音频编码线程
// if (true) {
// // for audio capturing
//// new MediaAudioEncoder(muxer, mMediaEncoderListener);
// }
// muxer.prepare();
// muxer.startRecording();
// if (videoEncoder != null) {
// mUVCCamera.setFrameCallback(mIFrameCallback, UVCCamera.PIXEL_FORMAT_NV21);
// }
// synchronized (mSync) {
// mMuxer = muxer;
// mVideoEncoder = videoEncoder;
// }
// callOnStartRecording();
// } catch (final IOException e) {
// callOnError(e);
// Log.e(TAG, "startCapture:", e);
// }
// }
private AACEncodeConsumer mAacConsumer;
private H264EncodeConsumer mH264Consumer;
public void handleStartRecording(RecordParams params){
if ((mUVCCamera == null) || (mMuxer != null))
return;
if (params == null)
throw new NullPointerException("RecordParams can not be null!");
// 获取USB Camera预览数据
mUVCCamera.setFrameCallback(mIFrameCallback, UVCCamera.PIXEL_FORMAT_NV21);
// 初始化混合器
mMuxer = new Mp4MediaMuxer(params.getRecordPath(),
params.getRecordDuration() * 60 * 1000);
// 启动视频编码线程
startVideoRecord();
// 启动音频编码线程
if(! params.isVoiceClose()) {
startAudioRecord();
}
callOnStartRecording();
}
public void handleStopRecording(){
// 停止混合器
if (mMuxer != null){
mMuxer.release();
mMuxer = null;
Log.i(TAG,TAG+"---->停止本地录制");
}
// 停止音视频编码线程
stopAudioRecord();
stopVideoRecord();
// 停止捕获视频数据
if (mUVCCamera != null) {
mUVCCamera.stopCapture();
mUVCCamera.setFrameCallback(null, 0);
}
mWeakCameraView.get().setVideoEncoder(null);
// you should not wait here
callOnStopRecording();
}
private void startVideoRecord() {
mH264Consumer = new H264EncodeConsumer(getWidth(),getHeight());
mH264Consumer.setOnH264EncodeResultListener(new H264EncodeConsumer.OnH264EncodeResultListener() {
@Override
public void onEncodeResult(byte[] data, int offset, int length, long timestamp) {
if(mListener != null){
mListener.onEncodeResult(data,offset,length,timestamp,1);
}
}
});
mH264Consumer.start();
// 添加混合器
if(mH264Consumer != null){
mH264Consumer.setTmpuMuxer(mMuxer);
}
}
private void stopVideoRecord(){
if(mH264Consumer != null){
mH264Consumer.exit();
mH264Consumer.setTmpuMuxer(null);
try {
Thread t2 = mH264Consumer;
mH264Consumer = null;
if(t2 != null){
t2.interrupt();
t2.join();
}
} catch (InterruptedException e) {
e.printStackTrace();
}
}
}
private void startAudioRecord(){
mAacConsumer = new AACEncodeConsumer();
mAacConsumer.setOnAACEncodeResultListener(new AACEncodeConsumer.OnAACEncodeResultListener() {
@Override
public void onEncodeResult(byte[] data, int offset, int length, long timestamp) {
if(mListener != null){
mListener.onEncodeResult(data,offset,length,timestamp,0);
}
}
});
mAacConsumer.start();
// 添加混合器
if(mAacConsumer != null){
mAacConsumer.setTmpuMuxer(mMuxer);
}
// isAudioThreadStart = true;
}
private void stopAudioRecord(){
if(mAacConsumer != null){
mAacConsumer.exit();
mAacConsumer.setTmpuMuxer(null);
try {
Thread t1 = mAacConsumer;
mAacConsumer = null;
if(t1 != null){
t1.interrupt();
t1.join();
}
} catch (InterruptedException e) {
e.printStackTrace();
}
}
// isAudioThreadStart = false;
}
// 停止录制视频
// public void handleStopRecording2() {
// if (DEBUG) Log.v(TAG_THREAD, "handleStopRecording:mMuxer=" + mMuxer);
// final MediaMuxerWrapper muxer;
// synchronized (mSync) {
// muxer = mMuxer;
// mMuxer = null;
// mVideoEncoder = null;
// if (mUVCCamera != null) {
// mUVCCamera.stopCapture();
// }
// }
// try {
// mWeakCameraView.get().setVideoEncoder(null);
// } catch (final Exception e) {
// // ignore
// }
// if (muxer != null) {
// muxer.stopRecording();
// mUVCCamera.setFrameCallback(null, 0);
// // you should not wait here
// callOnStopRecording();
// }
// }
private final IFrameCallback mIFrameCallback = new IFrameCallback() {
@Override
public void onFrame(final ByteBuffer frame) {
// final MediaVideoBufferEncoder videoEncoder;
// synchronized (mSync) {
// videoEncoder = mVideoEncoder;
// }
// if (videoEncoder != null) {
// videoEncoder.frameAvailableSoon();
// videoEncoder.encode(frame);
// }
int len = frame.capacity();
byte[] yuv = new byte[len];
frame.get(yuv);
if(mH264Consumer != null){
mH264Consumer.setRawYuv(yuv,640,480);
}
}
};
@TargetApi(Build.VERSION_CODES.JELLY_BEAN_MR1)
public void handleUpdateMedia(final String path) {
if (DEBUG) Log.v(TAG_THREAD, "handleUpdateMedia:path=" + path);
final Activity parent = mWeakParent.get();
final boolean released = (mHandler == null) || mHandler.mReleased;
if (parent != null && parent.getApplicationContext() != null) {
try {
if (DEBUG) Log.i(TAG, "MediaScannerConnection#scanFile");
MediaScannerConnection.scanFile(parent.getApplicationContext(), new String[]{ path }, null, null);
} catch (final Exception e) {
Log.e(TAG, "handleUpdateMedia:", e);
}
if (released || parent.isDestroyed())
handleRelease();
} else {
Log.w(TAG, "MainActivity already destroyed");
// give up to add this movie to MediaStore now.
// Seeing this movie on Gallery app etc. will take a lot of time.
handleRelease();
}
}
public void handleRelease() {
if (DEBUG) Log.v(TAG_THREAD, "handleRelease:mIsRecording=" + mIsRecording);
handleClose();
mCallbacks.clear();
if (!mIsRecording) {
mHandler.mReleased = true;
Looper.myLooper().quit();
}
if (DEBUG) Log.v(TAG_THREAD, "handleRelease:finished");
}
private final MediaEncoder.MediaEncoderListener mMediaEncoderListener = new MediaEncoder.MediaEncoderListener() {
@Override
public void onPrepared(final MediaEncoder encoder) {
if (DEBUG) Log.v(TAG, "onPrepared:encoder=" + encoder);
mIsRecording = true;
if (encoder instanceof MediaVideoEncoder)
try {
mWeakCameraView.get().setVideoEncoder((MediaVideoEncoder)encoder);
} catch (final Exception e) {
Log.e(TAG, "onPrepared:", e);
}
if (encoder instanceof MediaSurfaceEncoder)
try {
mWeakCameraView.get().setVideoEncoder((MediaSurfaceEncoder)encoder);
mUVCCamera.startCapture(((MediaSurfaceEncoder)encoder).getInputSurface());
} catch (final Exception e) {
Log.e(TAG, "onPrepared:", e);
}
}
@TargetApi(Build.VERSION_CODES.JELLY_BEAN_MR1)
@Override
public void onStopped(final MediaEncoder encoder) {
if (DEBUG) Log.v(TAG_THREAD, "onStopped:encoder=" + encoder);
if ((encoder instanceof MediaVideoEncoder)
|| (encoder instanceof MediaSurfaceEncoder))
try {
mIsRecording = false;
final Activity parent = mWeakParent.get();
mWeakCameraView.get().setVideoEncoder(null);
synchronized (mSync) {
if (mUVCCamera != null) {
mUVCCamera.stopCapture();
}
}
final String path = encoder.getOutputPath();
if (!TextUtils.isEmpty(path)) {
mHandler.sendMessageDelayed(mHandler.obtainMessage(MSG_MEDIA_UPDATE, path), 1000);
} else {
final boolean released = (mHandler == null) || mHandler.mReleased;
if (released || parent == null || parent.isDestroyed()) {
handleRelease();
}
}
} catch (final Exception e) {
Log.e(TAG, "onPrepared:", e);
}
}
@Override
public void onEncodeResult(byte[] data, int offset, int length, long timestamp, int type) {
if(mListener != null){
mListener.onEncodeResult(data, offset, length, timestamp, type);
}
}
};
// private void loadShutterSound(final Context context) {
// // get system stream type using reflection
// int streamType;
// try {
// final Class<?> audioSystemClass = Class.forName("android.media.AudioSystem");
// final Field sseField = audioSystemClass.getDeclaredField("STREAM_SYSTEM_ENFORCED");
// streamType = sseField.getInt(null);
// } catch (final Exception e) {
// streamType = AudioManager.STREAM_SYSTEM; // set appropriate according to your app policy
// }
// if (mSoundPool != null) {
// try {
// mSoundPool.release();
// } catch (final Exception e) {
// }
// mSoundPool = null;
// }
// // load shutter sound from resource
// mSoundPool = new SoundPool(2, streamType, 0);
// mSoundId = mSoundPool.load(context, R.raw.camera_click, 1);
// }
/**
* prepare and load shutter sound for still image capturing
*/
@SuppressWarnings("deprecation")
// private void loadShutterSound(final Context context) {
// // get system stream type using reflection
// int streamType;
// try {
// final Class<?> audioSystemClass = Class.forName("android.media.AudioSystem");
// final Field sseField = audioSystemClass.getDeclaredField("STREAM_SYSTEM_ENFORCED");
// streamType = sseField.getInt(null);
// } catch (final Exception e) {
// streamType = AudioManager.STREAM_SYSTEM; // set appropriate according to your app policy
// }
// if (mSoundPool != null) {
// try {
// mSoundPool.release();
// } catch (final Exception e) {
// }
// mSoundPool = null;
// }
// // load shutter sound from resource
// mSoundPool = new SoundPool(2, streamType, 0);
// mSoundId = mSoundPool.load(context, R.raw.camera_click, 1);
// }
@Override
public void run() {
Looper.prepare();
AbstractUVCCameraHandler handler = null;
try {
final Constructor<? extends AbstractUVCCameraHandler> constructor = mHandlerClass.getDeclaredConstructor(CameraThread.class);
handler = constructor.newInstance(this);
} catch (final NoSuchMethodException e) {
Log.w(TAG, e);
} catch (final IllegalAccessException e) {
Log.w(TAG, e);
} catch (final InstantiationException e) {
Log.w(TAG, e);
} catch (final InvocationTargetException e) {
Log.w(TAG, e);
}
if (handler != null) {
synchronized (mSync) {
mHandler = handler;
mSync.notifyAll();
}
Looper.loop();
// if (mSoundPool != null) {
// mSoundPool.release();
// mSoundPool = null;
// }
if (mHandler != null) {
mHandler.mReleased = true;
}
}
mCallbacks.clear();
synchronized (mSync) {
mHandler = null;
mSync.notifyAll();
}
}
private void callOnOpen() {
for (final CameraCallback callback: mCallbacks) {
try {
callback.onOpen();
} catch (final Exception e) {
mCallbacks.remove(callback);
Log.w(TAG, e);
}
}
}
private void callOnClose() {
for (final CameraCallback callback: mCallbacks) {
try {
callback.onClose();
} catch (final Exception e) {
mCallbacks.remove(callback);
Log.w(TAG, e);
}
}
}
private void callOnStartPreview() {
for (final CameraCallback callback: mCallbacks) {
try {
callback.onStartPreview();
} catch (final Exception e) {
mCallbacks.remove(callback);
Log.w(TAG, e);
}
}
}
private void callOnStopPreview() {
for (final CameraCallback callback: mCallbacks) {
try {
callback.onStopPreview();
} catch (final Exception e) {
mCallbacks.remove(callback);
Log.w(TAG, e);
}
}
}
private void callOnStartRecording() {
for (final CameraCallback callback: mCallbacks) {
try {
callback.onStartRecording();
} catch (final Exception e) {
mCallbacks.remove(callback);
Log.w(TAG, e);
}
}
}
private void callOnStopRecording() {
for (final CameraCallback callback: mCallbacks) {
try {
callback.onStopRecording();
} catch (final Exception e) {
mCallbacks.remove(callback);
Log.w(TAG, e);
}
}
}
private void callOnError(final Exception e) {
for (final CameraCallback callback: mCallbacks) {
try {
callback.onError(e);
} catch (final Exception e1) {
mCallbacks.remove(callback);
Log.w(TAG, e);
}
}
}
}
8 years ago
}