/hardware/libhardware/modules/audio_remote_submix/tests/ |
D | remote_submix_tests.cpp | 246 uint64_t frames; in TEST_F() local 248 EXPECT_EQ(0, streamOut->get_presentation_position(streamOut, &frames, ×tamp)); in TEST_F() 249 EXPECT_EQ(uint64_t{0}, frames); in TEST_F() 250 uint64_t prevFrames = frames; in TEST_F() 253 EXPECT_EQ(0, streamOut->get_presentation_position(streamOut, &frames, ×tamp)); in TEST_F() 254 EXPECT_LE(prevFrames, frames); in TEST_F() 255 prevFrames = frames; in TEST_F() 264 uint32_t frames; in TEST_F() local 265 EXPECT_EQ(0, streamOut->get_render_position(streamOut, &frames)); in TEST_F() 266 EXPECT_EQ(0U, frames); in TEST_F() [all …]
|
/hardware/interfaces/automotive/evs/1.0/ |
D | IEvsCamera.hal | 39 * Up to this many frames may be held concurrently by the client of IEvsCamera. 40 * If this many frames have been delivered to the receiver without being returned 41 * by doneWithFrame, the stream must skip frames until a buffer is returned for reuse. 47 * requested number of concurrent frames. 52 * Request delivery of EVS camera frames from this camera. 55 * frames until stopVideoStream() is called. 65 * as one), and if the supply is exhausted, no further frames may be 71 * Stop the delivery of EVS camera frames. 73 * Because delivery is asynchronous, frames may continue to arrive for
|
/hardware/qcom/audio/legacy/libalsa-intf/ |
D | arec.c | 185 long frames; in record_file() local 269 frames = bufsize / 2; in record_file() 271 frames = bufsize / 8; in record_file() 273 frames = bufsize / 12; in record_file() 275 frames = bufsize / 4; in record_file() 277 x.frames = frames; in record_file() 300 fprintf(stderr, "Arec:avail 1 = %d frames = %ld\n",avail, frames); in record_file() 307 if (x.frames > avail) in record_file() 308 frames = avail; in record_file() 324 x.frames -= frames; in record_file() [all …]
|
D | alsa_pcm.c | 468 long frames) in mmap_transfer() argument 478 frames = frames * channels *2 ; in mmap_transfer() 480 while (frames-- > 0) { in mmap_transfer() 489 long frames) in mmap_transfer_capture() argument 502 frames = frames * channels *2 ; in mmap_transfer_capture() 504 while (frames-- > 0) { in mmap_transfer_capture() 524 long frames; in pcm_write_mmap() local 528 frames = (pcm->flags & PCM_MONO) ? (count / 2) : (count / 4); in pcm_write_mmap() 539 pcm->sync_ptr->c.control.appl_ptr += frames; in pcm_write_mmap() 580 x.frames = (count / (channels * 2)) ; in pcm_write_nmmap() [all …]
|
D | aplay.c | 173 long frames; in play_file() local 281 frames = (pcm->flags & PCM_MONO) ? (bufsize / 2) : (bufsize / 4); in play_file() 325 fprintf(stderr, "Aplay:avail = %d frames = %d\n",avail, frames); in play_file() 341 frames = (pcm->flags & PCM_MONO) ? (remainingData / 2) : (remainingData / 4); in play_file() 361 pcm->sync_ptr->c.control.appl_ptr += frames; in play_file() 410 offset += frames; in play_file()
|
/hardware/interfaces/audio/2.0/ |
D | IStreamIn.hal | 78 uint64_t frames; 101 * @param framesCount the number of frames in a buffer. 123 * Return the amount of input frames lost in the audio driver since the last 131 * @return framesLost the number of input audio frames lost. 136 * Return a recent count of the number of audio frames received and the 142 * @return frames the total frame count received. This must be as early in 143 * the capture pipeline as possible. In general, frames 145 * @return time is the clock monotonic time when frames was measured. In 150 generates (Result retval, uint64_t frames, uint64_t time);
|
D | IStreamOut.hal | 71 uint64_t frames; // getPresentationPosition. 95 * @param framesCount the number of frames in a buffer. 116 * Return the number of audio frames written by the audio DSP to DAC since 120 * @return dspFrames number of audio frames written. 234 * Return a recent count of the number of audio frames presented to an 235 * external observer. This excludes frames which have been written but are 242 * Example: assume that N+M frames have been presented, where M is a 'small' 248 * @return frames count of presented audio frames. 252 generates (Result retval, uint64_t frames, TimeSpec timeStamp);
|
D | types.hal | 82 int32_t bufferSizeFrames; // total buffer size in frames 83 int32_t burstSizeFrames; // transfer size granularity in frames
|
/hardware/interfaces/audio/4.0/ |
D | IStreamIn.hal | 78 uint64_t frames; 107 * @param framesCount the number of frames in a buffer. 129 * Return the amount of input frames lost in the audio driver since the last 137 * @return framesLost the number of input audio frames lost. 142 * Return a recent count of the number of audio frames received and the 148 * @return frames the total frame count received. This must be as early in 149 * the capture pipeline as possible. In general, frames 151 * @return time is the clock monotonic time when frames was measured. In 156 generates (Result retval, uint64_t frames, uint64_t time);
|
D | IStreamOut.hal | 71 uint64_t frames; // getPresentationPosition. 101 * @param framesCount the number of frames in a buffer. 122 * Return the number of audio frames written by the audio DSP to DAC since 127 * @return dspFrames number of audio frames written. 246 * Return a recent count of the number of audio frames presented to an 247 * external observer. This excludes frames which have been written but are 254 * Example: assume that N+M frames have been presented, where M is a 'small' 262 * @return frames count of presented audio frames. 266 generates (Result retval, uint64_t frames, TimeSpec timeStamp);
|
/hardware/qcom/audio/legacy/alsa_sound/ |
D | AudioUsbALSA.cpp | 383 long frames; in RecordingThreadEntry() local 432 frames = (musbRecordingHandle->flags & PCM_MONO) ? (bufsize / 2) : (bufsize / 4); in RecordingThreadEntry() 433 x.frames = (musbRecordingHandle->flags & PCM_MONO) ? (bufsize / 2) : (bufsize / 4); in RecordingThreadEntry() 492 if (x.frames > avail) in RecordingThreadEntry() 493 frames = avail; in RecordingThreadEntry() 537 musbRecordingHandle->sync_ptr->c.control.appl_ptr += frames; in RecordingThreadEntry() 549 mproxyRecordingHandle->sync_ptr->c.control.appl_ptr += frames; in RecordingThreadEntry() 772 long frames; in PlaybackThreadEntry() local 839 frames = (mproxyPlaybackHandle->flags & PCM_MONO) ? (proxyPeriod / 2) : (proxyPeriod / 4); in PlaybackThreadEntry() 840 x.frames = (mproxyPlaybackHandle->flags & PCM_MONO) ? (proxyPeriod / 2) : (proxyPeriod / 4); in PlaybackThreadEntry() [all …]
|
/hardware/interfaces/audio/6.0/ |
D | IStreamIn.hal | 78 uint64_t frames; 107 * @param framesCount the number of frames in a buffer. 129 * Return the amount of input frames lost in the audio driver since the last 137 * @return framesLost the number of input audio frames lost. 142 * Return a recent count of the number of audio frames received and the 148 * @return frames the total frame count received. This must be as early in 149 * the capture pipeline as possible. In general, frames 151 * @return time is the clock monotonic time when frames was measured. In 156 generates (Result retval, uint64_t frames, uint64_t time);
|
D | IStreamOut.hal | 71 uint64_t frames; // getPresentationPosition. 101 * @param framesCount the number of frames in a buffer. 122 * Return the number of audio frames written by the audio DSP to DAC since 127 * @return dspFrames number of audio frames written. 246 * Return a recent count of the number of audio frames presented to an 247 * external observer. This excludes frames which have been written but are 254 * Example: assume that N+M frames have been presented, where M is a 'small' 262 * @return frames count of presented audio frames. 266 generates (Result retval, uint64_t frames, TimeSpec timeStamp);
|
/hardware/interfaces/audio/5.0/ |
D | IStreamIn.hal | 78 uint64_t frames; 107 * @param framesCount the number of frames in a buffer. 129 * Return the amount of input frames lost in the audio driver since the last 137 * @return framesLost the number of input audio frames lost. 142 * Return a recent count of the number of audio frames received and the 148 * @return frames the total frame count received. This must be as early in 149 * the capture pipeline as possible. In general, frames 151 * @return time is the clock monotonic time when frames was measured. In 156 generates (Result retval, uint64_t frames, uint64_t time);
|
D | IStreamOut.hal | 71 uint64_t frames; // getPresentationPosition. 101 * @param framesCount the number of frames in a buffer. 122 * Return the number of audio frames written by the audio DSP to DAC since 127 * @return dspFrames number of audio frames written. 246 * Return a recent count of the number of audio frames presented to an 247 * external observer. This excludes frames which have been written but are 254 * Example: assume that N+M frames have been presented, where M is a 'small' 262 * @return frames count of presented audio frames. 266 generates (Result retval, uint64_t frames, TimeSpec timeStamp);
|
/hardware/interfaces/input/classifier/1.0/vts/functional/ |
D | VtsHalInputClassifierV1_0TargetTest.cpp | 55 event.frames = {}; in getSimpleMotionEvent() 147 event.frames = {frame}; in TEST_F() 176 event.frames = {frame1, frame2}; in TEST_F()
|
/hardware/interfaces/camera/device/1.0/ |
D | ICameraDevice.hal | 187 /** Set the callback interface through which preview frames are sent */ 203 * release video recording frames sent out by the cameral HAL before and 280 * responsibility to manage the life-cycle of the video recording frames, 281 * and the client must not modify/access any video recording frames. 304 * frames sent out by the camera HAL before the camera HAL receives a call 307 * responsibility to manage the life-cycle of the video recording frames. 319 * frames sent out by the camera HAL before the camera HAL receives a call 322 * responsibility to manage the life-cycle of the video recording frames. 331 * Release a batch of record frames previously returned by CAMERA_MSG_VIDEO_FRAME 335 * frames sent out by the camera HAL before the camera HAL receives a call [all …]
|
D | types.hal | 63 * A set of bit masks for specifying how the received preview frames are 76 * corrupted by subsequent preview frames filled by the camera. This flag is 84 * 2. 0x01 enables a callback without copying out the received frames. A 278 * Used by camera framework to send a batch of recording frames back to camera HAL.
|
/hardware/google/interfaces/media/c2/1.0/ |
D | IComponentListener.hal | 64 * Notify the listener that frames are rendered. 66 * @param renderedFrames List of information about renderings of frames.
|
/hardware/interfaces/audio/core/all-versions/default/ |
D | StreamOut.cpp | 92 StreamOut::getPresentationPositionImpl(mStream, &mStatus.reply.presentationPosition.frames, in doGetPresentationPosition() 510 Result StreamOut::getPresentationPositionImpl(audio_stream_out_t* stream, uint64_t* frames, in getPresentationPositionImpl() argument 521 stream->get_presentation_position(stream, frames, &halTimeStamp), in getPresentationPositionImpl() 531 uint64_t frames = 0; in getPresentationPosition() local 533 Result retval = getPresentationPositionImpl(mStream, &frames, &timeStamp); in getPresentationPosition() 534 _hidl_cb(retval, frames, timeStamp); in getPresentationPosition()
|
D | StreamIn.cpp | 101 mStream, &mStatus.reply.capturePosition.frames, &mStatus.reply.capturePosition.time); in doGetCapturePosition() 425 Result StreamIn::getCapturePositionImpl(audio_stream_in_t* stream, uint64_t* frames, in getCapturePositionImpl() argument 437 *frames = halFrames; in getCapturePositionImpl() 444 uint64_t frames = 0, time = 0; in getCapturePosition() local 445 Result retval = getCapturePositionImpl(mStream, &frames, &time); in getCapturePosition() 446 _hidl_cb(retval, frames, time); in getCapturePosition()
|
/hardware/libhardware/modules/audio_remote_submix/ |
D | audio_hw.cpp | 795 const size_t frames = bytes / frame_size; in audio_config_compare() local 809 usleep(frames * 1000000 / out_get_sample_rate(&stream->common)); in audio_config_compare() 812 out->frames_written += frames; in audio_config_compare() 813 out->frames_written_since_standby += frames; in audio_config_compare() 838 if (dont_block && availableToWrite < frames) { in audio_config_compare() 841 size_t frames_to_flush_from_source = frames - availableToWrite; in audio_config_compare() 855 written_frames = sink->write(buffer, frames); in audio_config_compare() 874 written_frames = sink->write(buffer, frames); in audio_config_compare() 896 uint64_t *frames, struct timespec *timestamp) in audio_config_compare() argument 898 if (stream == NULL || frames == NULL || timestamp == NULL) { in audio_config_compare() [all …]
|
/hardware/libhardware/include/hardware/ |
D | audio.h | 357 uint64_t *frames, struct timespec *timestamp); 474 int64_t *frames, int64_t *time);
|
/hardware/interfaces/neuralnetworks/1.1/ |
D | types.t | 154 * Prefer maximizing the throughput of successive frames, for example when 155 * processing successive frames coming from the camera.
|
/hardware/interfaces/graphics/composer/2.3/ |
D | IComposerClient.hal | 318 * @param maxFrames is the maximum number of frames that should be stored before discard. 319 * The sample represents the most-recently posted frames. 339 * @param maxFrames is the maximum number of frames that should be represented in the sample. 340 * The sample represents the most-recently posted frames. 341 * If maxFrames is 0, all frames are to be represented by the sample. 342 * @param timestamp is the timestamp after which any frames were posted that should be 349 * @return frameCount The number of frames represented by this sample.
|