1 /*
2 * Copyright (C) 2010 The Android Open Source Project
3 *
4 * Licensed under the Apache License, Version 2.0 (the "License");
5 * you may not use this file except in compliance with the License.
6 * You may obtain a copy of the License at
7 *
8 * http://www.apache.org/licenses/LICENSE-2.0
9 *
10 * Unless required by applicable law or agreed to in writing, software
11 * distributed under the License is distributed on an "AS IS" BASIS,
12 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
13 * See the License for the specific language governing permissions and
14 * limitations under the License.
15 */
16
17 #include <inttypes.h>
18
19 //#define LOG_NDEBUG 0
20 #define LOG_TAG "CameraSourceTimeLapse"
21
22 #include <media/hardware/HardwareAPI.h>
23 #include <binder/IPCThreadState.h>
24 #include <binder/MemoryBase.h>
25 #include <binder/MemoryHeapBase.h>
26 #include <media/stagefright/foundation/ADebug.h>
27 #include <media/stagefright/CameraSource.h>
28 #include <media/stagefright/CameraSourceTimeLapse.h>
29 #include <media/stagefright/MetaData.h>
30 #include <camera/Camera.h>
31 #include <camera/CameraParameters.h>
32 #include <utils/String8.h>
33 #include <utils/Vector.h>
34
35 namespace android {
36
37 // static
CreateFromCamera(const sp<hardware::ICamera> & camera,const sp<ICameraRecordingProxy> & proxy,int32_t cameraId,const String16 & clientName,uid_t clientUid,pid_t clientPid,Size videoSize,int32_t videoFrameRate,const sp<IGraphicBufferProducer> & surface,int64_t timeBetweenFrameCaptureUs,bool storeMetaDataInVideoBuffers)38 CameraSourceTimeLapse *CameraSourceTimeLapse::CreateFromCamera(
39 const sp<hardware::ICamera> &camera,
40 const sp<ICameraRecordingProxy> &proxy,
41 int32_t cameraId,
42 const String16& clientName,
43 uid_t clientUid,
44 pid_t clientPid,
45 Size videoSize,
46 int32_t videoFrameRate,
47 const sp<IGraphicBufferProducer>& surface,
48 int64_t timeBetweenFrameCaptureUs,
49 bool storeMetaDataInVideoBuffers) {
50
51 CameraSourceTimeLapse *source = new
52 CameraSourceTimeLapse(camera, proxy, cameraId,
53 clientName, clientUid, clientPid,
54 videoSize, videoFrameRate, surface,
55 timeBetweenFrameCaptureUs,
56 storeMetaDataInVideoBuffers);
57
58 if (source != NULL) {
59 if (source->initCheck() != OK) {
60 delete source;
61 return NULL;
62 }
63 }
64 return source;
65 }
66
CameraSourceTimeLapse(const sp<hardware::ICamera> & camera,const sp<ICameraRecordingProxy> & proxy,int32_t cameraId,const String16 & clientName,uid_t clientUid,pid_t clientPid,Size videoSize,int32_t videoFrameRate,const sp<IGraphicBufferProducer> & surface,int64_t timeBetweenFrameCaptureUs,bool storeMetaDataInVideoBuffers)67 CameraSourceTimeLapse::CameraSourceTimeLapse(
68 const sp<hardware::ICamera>& camera,
69 const sp<ICameraRecordingProxy>& proxy,
70 int32_t cameraId,
71 const String16& clientName,
72 uid_t clientUid,
73 pid_t clientPid,
74 Size videoSize,
75 int32_t videoFrameRate,
76 const sp<IGraphicBufferProducer>& surface,
77 int64_t timeBetweenFrameCaptureUs,
78 bool storeMetaDataInVideoBuffers)
79 : CameraSource(camera, proxy, cameraId, clientName, clientUid, clientPid,
80 videoSize, videoFrameRate, surface,
81 storeMetaDataInVideoBuffers),
82 mTimeBetweenTimeLapseVideoFramesUs(1E6/videoFrameRate),
83 mLastTimeLapseFrameRealTimestampUs(0),
84 mSkipCurrentFrame(false) {
85
86 mTimeBetweenFrameCaptureUs = timeBetweenFrameCaptureUs;
87 ALOGD("starting time lapse mode: %" PRId64 " us",
88 mTimeBetweenFrameCaptureUs);
89
90 mVideoWidth = videoSize.width;
91 mVideoHeight = videoSize.height;
92
93 if (OK == mInitCheck && !trySettingVideoSize(videoSize.width, videoSize.height)) {
94 releaseCamera();
95 mInitCheck = NO_INIT;
96 }
97
98 // Initialize quick stop variables.
99 mQuickStop = false;
100 mForceRead = false;
101 mLastReadBufferCopy = NULL;
102 mStopWaitingForIdleCamera = false;
103 }
104
~CameraSourceTimeLapse()105 CameraSourceTimeLapse::~CameraSourceTimeLapse() {
106 if (mLastReadBufferCopy) {
107 mLastReadBufferCopy->release();
108 mLastReadBufferCopy = NULL;
109 }
110 }
111
startQuickReadReturns()112 void CameraSourceTimeLapse::startQuickReadReturns() {
113 ALOGV("startQuickReadReturns");
114 Mutex::Autolock autoLock(mQuickStopLock);
115
116 // Enable quick stop mode.
117 mQuickStop = true;
118
119 // Force dataCallbackTimestamp() coming from the video camera to
120 // not skip the next frame as we want read() to get a get a frame
121 // right away.
122 mForceRead = true;
123 }
124
trySettingVideoSize(int32_t width,int32_t height)125 bool CameraSourceTimeLapse::trySettingVideoSize(
126 int32_t width, int32_t height) {
127
128 ALOGV("trySettingVideoSize");
129 int64_t token = IPCThreadState::self()->clearCallingIdentity();
130 String8 s = mCamera->getParameters();
131
132 CameraParameters params(s);
133 Vector<Size> supportedSizes;
134 params.getSupportedVideoSizes(supportedSizes);
135 bool videoOutputSupported = false;
136 if (supportedSizes.size() == 0) {
137 params.getSupportedPreviewSizes(supportedSizes);
138 } else {
139 videoOutputSupported = true;
140 }
141
142 bool videoSizeSupported = false;
143 for (size_t i = 0; i < supportedSizes.size(); ++i) {
144 int32_t pictureWidth = supportedSizes[i].width;
145 int32_t pictureHeight = supportedSizes[i].height;
146
147 if ((pictureWidth == width) && (pictureHeight == height)) {
148 videoSizeSupported = true;
149 }
150 }
151
152 bool isSuccessful = false;
153 if (videoSizeSupported) {
154 ALOGV("Video size (%d, %d) is supported", width, height);
155 if (videoOutputSupported) {
156 params.setVideoSize(width, height);
157 } else {
158 params.setPreviewSize(width, height);
159 }
160 if (mCamera->setParameters(params.flatten()) == OK) {
161 isSuccessful = true;
162 } else {
163 ALOGE("Failed to set preview size to %dx%d", width, height);
164 isSuccessful = false;
165 }
166 }
167
168 IPCThreadState::self()->restoreCallingIdentity(token);
169 return isSuccessful;
170 }
171
signalBufferReturned(MediaBufferBase * buffer)172 void CameraSourceTimeLapse::signalBufferReturned(MediaBufferBase* buffer) {
173 ALOGV("signalBufferReturned");
174 Mutex::Autolock autoLock(mQuickStopLock);
175 if (mQuickStop && (buffer == mLastReadBufferCopy)) {
176 if (metaDataStoredInVideoBuffers() == kMetadataBufferTypeNativeHandleSource) {
177 native_handle_t* handle = (
178 (VideoNativeHandleMetadata*)(mLastReadBufferCopy->data()))->pHandle;
179 native_handle_close(handle);
180 native_handle_delete(handle);
181 }
182 buffer->setObserver(NULL);
183 buffer->release();
184 mLastReadBufferCopy = NULL;
185 mForceRead = true;
186 } else {
187 return CameraSource::signalBufferReturned(buffer);
188 }
189 }
190
createMediaBufferCopy(const MediaBufferBase & sourceBuffer,int64_t frameTime,MediaBufferBase ** newBuffer,int32_t videoBufferMode)191 void createMediaBufferCopy(
192 const MediaBufferBase& sourceBuffer,
193 int64_t frameTime,
194 MediaBufferBase **newBuffer,
195 int32_t videoBufferMode) {
196
197 ALOGV("createMediaBufferCopy");
198 size_t sourceSize = sourceBuffer.size();
199 void* sourcePointer = sourceBuffer.data();
200
201 (*newBuffer) = new MediaBuffer(sourceSize);
202 memcpy((*newBuffer)->data(), sourcePointer, sourceSize);
203
204 (*newBuffer)->meta_data().setInt64(kKeyTime, frameTime);
205
206 if (videoBufferMode == kMetadataBufferTypeNativeHandleSource) {
207 ((VideoNativeHandleMetadata*)((*newBuffer)->data()))->pHandle =
208 native_handle_clone(
209 ((VideoNativeHandleMetadata*)(sourceBuffer.data()))->pHandle);
210 }
211 }
212
fillLastReadBufferCopy(MediaBufferBase & sourceBuffer)213 void CameraSourceTimeLapse::fillLastReadBufferCopy(MediaBufferBase& sourceBuffer) {
214 ALOGV("fillLastReadBufferCopy");
215 int64_t frameTime;
216 CHECK(sourceBuffer.meta_data().findInt64(kKeyTime, &frameTime));
217 createMediaBufferCopy(sourceBuffer, frameTime, &mLastReadBufferCopy,
218 metaDataStoredInVideoBuffers());
219 mLastReadBufferCopy->add_ref();
220 mLastReadBufferCopy->setObserver(this);
221 }
222
read(MediaBufferBase ** buffer,const ReadOptions * options)223 status_t CameraSourceTimeLapse::read(
224 MediaBufferBase **buffer, const ReadOptions *options) {
225 ALOGV("read");
226 if (mLastReadBufferCopy == NULL) {
227 mLastReadStatus = CameraSource::read(buffer, options);
228
229 // mQuickStop may have turned to true while read was blocked.
230 // Make a copy of the buffer in that case.
231 Mutex::Autolock autoLock(mQuickStopLock);
232 if (mQuickStop && *buffer) {
233 fillLastReadBufferCopy(**buffer);
234 }
235 return mLastReadStatus;
236 } else {
237 (*buffer) = mLastReadBufferCopy;
238 (*buffer)->add_ref();
239 return mLastReadStatus;
240 }
241 }
242
createIMemoryCopy(const sp<IMemory> & source_data)243 sp<IMemory> CameraSourceTimeLapse::createIMemoryCopy(
244 const sp<IMemory> &source_data) {
245
246 ALOGV("createIMemoryCopy");
247 size_t source_size = source_data->size();
248 void* source_pointer = source_data->pointer();
249
250 sp<MemoryHeapBase> newMemoryHeap = new MemoryHeapBase(source_size);
251 sp<MemoryBase> newMemory = new MemoryBase(newMemoryHeap, 0, source_size);
252 memcpy(newMemory->pointer(), source_pointer, source_size);
253 return newMemory;
254 }
255
skipCurrentFrame(int64_t)256 bool CameraSourceTimeLapse::skipCurrentFrame(int64_t /* timestampUs */) {
257 ALOGV("skipCurrentFrame");
258 if (mSkipCurrentFrame) {
259 mSkipCurrentFrame = false;
260 return true;
261 } else {
262 return false;
263 }
264 }
265
skipFrameAndModifyTimeStamp(int64_t * timestampUs)266 bool CameraSourceTimeLapse::skipFrameAndModifyTimeStamp(int64_t *timestampUs) {
267 ALOGV("skipFrameAndModifyTimeStamp");
268 if (mLastTimeLapseFrameRealTimestampUs == 0) {
269 // First time lapse frame. Initialize mLastTimeLapseFrameRealTimestampUs
270 // to current time (timestampUs) and save frame data.
271 ALOGV("dataCallbackTimestamp timelapse: initial frame");
272
273 mLastTimeLapseFrameRealTimestampUs = *timestampUs;
274 return false;
275 }
276
277 {
278 Mutex::Autolock autoLock(mQuickStopLock);
279
280 // mForceRead may be set to true by startQuickReadReturns(). In that
281 // case don't skip this frame.
282 if (mForceRead) {
283 ALOGV("dataCallbackTimestamp timelapse: forced read");
284 mForceRead = false;
285 *timestampUs =
286 mLastFrameTimestampUs + mTimeBetweenTimeLapseVideoFramesUs;
287
288 // Really make sure that this video recording frame will not be dropped.
289 if (*timestampUs < mStartTimeUs) {
290 ALOGI("set timestampUs to start time stamp %" PRId64 " us", mStartTimeUs);
291 *timestampUs = mStartTimeUs;
292 }
293 return false;
294 }
295 }
296
297 // Workaround to bypass the first 2 input frames for skipping.
298 // The first 2 output frames from the encoder are: decoder specific info and
299 // the compressed video frame data for the first input video frame.
300 if (mNumFramesEncoded >= 1 && *timestampUs <
301 (mLastTimeLapseFrameRealTimestampUs + mTimeBetweenFrameCaptureUs)) {
302 // Skip all frames from last encoded frame until
303 // sufficient time (mTimeBetweenFrameCaptureUs) has passed.
304 // Tell the camera to release its recording frame and return.
305 ALOGV("dataCallbackTimestamp timelapse: skipping intermediate frame");
306 return true;
307 } else {
308 // Desired frame has arrived after mTimeBetweenFrameCaptureUs time:
309 // - Reset mLastTimeLapseFrameRealTimestampUs to current time.
310 // - Artificially modify timestampUs to be one frame time (1/framerate) ahead
311 // of the last encoded frame's time stamp.
312 ALOGV("dataCallbackTimestamp timelapse: got timelapse frame");
313
314 mLastTimeLapseFrameRealTimestampUs = *timestampUs;
315 *timestampUs = mLastFrameTimestampUs + mTimeBetweenTimeLapseVideoFramesUs;
316 return false;
317 }
318 return false;
319 }
320
dataCallbackTimestamp(int64_t timestampUs,int32_t msgType,const sp<IMemory> & data)321 void CameraSourceTimeLapse::dataCallbackTimestamp(int64_t timestampUs, int32_t msgType,
322 const sp<IMemory> &data) {
323 ALOGV("dataCallbackTimestamp");
324 mSkipCurrentFrame = skipFrameAndModifyTimeStamp(×tampUs);
325 CameraSource::dataCallbackTimestamp(timestampUs, msgType, data);
326 }
327
recordingFrameHandleCallbackTimestamp(int64_t timestampUs,native_handle_t * handle)328 void CameraSourceTimeLapse::recordingFrameHandleCallbackTimestamp(int64_t timestampUs,
329 native_handle_t* handle) {
330 ALOGV("recordingFrameHandleCallbackTimestamp");
331 mSkipCurrentFrame = skipFrameAndModifyTimeStamp(×tampUs);
332 CameraSource::recordingFrameHandleCallbackTimestamp(timestampUs, handle);
333 }
334
recordingFrameHandleCallbackTimestampBatch(const std::vector<int64_t> & timestampsUs,const std::vector<native_handle_t * > & handles)335 void CameraSourceTimeLapse::recordingFrameHandleCallbackTimestampBatch(
336 const std::vector<int64_t>& timestampsUs,
337 const std::vector<native_handle_t*>& handles) {
338 ALOGV("recordingFrameHandleCallbackTimestampBatch");
339 int n = timestampsUs.size();
340 for (int i = 0; i < n; i++) {
341 // Don't do batching for CameraSourceTimeLapse for now
342 recordingFrameHandleCallbackTimestamp(timestampsUs[i], handles[i]);
343 }
344 }
345
processBufferQueueFrame(BufferItem & buffer)346 void CameraSourceTimeLapse::processBufferQueueFrame(BufferItem& buffer) {
347 ALOGV("processBufferQueueFrame");
348 int64_t timestampUs = buffer.mTimestamp / 1000;
349 mSkipCurrentFrame = skipFrameAndModifyTimeStamp(×tampUs);
350 buffer.mTimestamp = timestampUs * 1000;
351 CameraSource::processBufferQueueFrame(buffer);
352 }
353
354 } // namespace android
355