| /* |
| * Copyright (C) 2012 The Android Open Source Project |
| * |
| * Licensed under the Apache License, Version 2.0 (the "License"); |
| * you may not use this file except in compliance with the License. |
| * You may obtain a copy of the License at |
| * |
| * http://www.apache.org/licenses/LICENSE-2.0 |
| * |
| * Unless required by applicable law or agreed to in writing, software |
| * distributed under the License is distributed on an "AS IS" BASIS, |
| * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
| * See the License for the specific language governing permissions and |
| * limitations under the License. |
| */ |
| |
| #define LOG_TAG "r_submix" |
| //#define LOG_NDEBUG 0 |
| |
| #include <errno.h> |
| #include <pthread.h> |
| #include <stdint.h> |
| #include <stdlib.h> |
| #include <sys/param.h> |
| #include <sys/time.h> |
| #include <sys/limits.h> |
| #include <unistd.h> |
| |
| #include <cutils/compiler.h> |
| #include <cutils/properties.h> |
| #include <cutils/str_parms.h> |
| #include <log/log.h> |
| #include <utils/String8.h> |
| |
| #include <hardware/audio.h> |
| #include <hardware/hardware.h> |
| #include <system/audio.h> |
| |
| #include <media/AudioParameter.h> |
| #include <media/AudioBufferProvider.h> |
| #include <media/nbaio/MonoPipe.h> |
| #include <media/nbaio/MonoPipeReader.h> |
| |
| #define LOG_STREAMS_TO_FILES 0 |
| #if LOG_STREAMS_TO_FILES |
| #include <fcntl.h> |
| #include <stdio.h> |
| #include <sys/stat.h> |
| #endif // LOG_STREAMS_TO_FILES |
| |
| extern "C" { |
| |
| namespace android { |
| |
| // Uncomment to enable extremely verbose logging in this module. |
| // #define SUBMIX_VERBOSE_LOGGING |
| #if defined(SUBMIX_VERBOSE_LOGGING) |
| #define SUBMIX_ALOGV(...) ALOGV(__VA_ARGS__) |
| #define SUBMIX_ALOGE(...) ALOGE(__VA_ARGS__) |
| #else |
| #define SUBMIX_ALOGV(...) |
| #define SUBMIX_ALOGE(...) |
| #endif // SUBMIX_VERBOSE_LOGGING |
| |
| // NOTE: This value will be rounded up to the nearest power of 2 by MonoPipe(). |
| #define DEFAULT_PIPE_SIZE_IN_FRAMES (1024*4) // size at default sample rate |
| // Value used to divide the MonoPipe() buffer into segments that are written to the source and |
| // read from the sink. The maximum latency of the device is the size of the MonoPipe's buffer |
| // the minimum latency is the MonoPipe buffer size divided by this value. |
| #define DEFAULT_PIPE_PERIOD_COUNT 4 |
| // The duration of MAX_READ_ATTEMPTS * READ_ATTEMPT_SLEEP_MS must be stricly inferior to |
| // the duration of a record buffer at the current record sample rate (of the device, not of |
| // the recording itself). Here we have: |
| // 3 * 5ms = 15ms < 1024 frames * 1000 / 48000 = 21.333ms |
| #define MAX_READ_ATTEMPTS 3 |
| #define READ_ATTEMPT_SLEEP_MS 5 // 5ms between two read attempts when pipe is empty |
| #define DEFAULT_SAMPLE_RATE_HZ 48000 // default sample rate |
| // See NBAIO_Format frameworks/av/include/media/nbaio/NBAIO.h. |
| #define DEFAULT_FORMAT AUDIO_FORMAT_PCM_16_BIT |
| // A legacy user of this device does not close the input stream when it shuts down, which |
| // results in the application opening a new input stream before closing the old input stream |
| // handle it was previously using. Setting this value to 1 allows multiple clients to open |
| // multiple input streams from this device. If this option is enabled, each input stream returned |
| // is *the same stream* which means that readers will race to read data from these streams. |
| #define ENABLE_LEGACY_INPUT_OPEN 1 |
| |
| #if LOG_STREAMS_TO_FILES |
| // Folder to save stream log files to. |
| #define LOG_STREAM_FOLDER "/data/misc/audioserver" |
| // Log filenames for input and output streams. |
| #define LOG_STREAM_OUT_FILENAME LOG_STREAM_FOLDER "/r_submix_out.raw" |
| #define LOG_STREAM_IN_FILENAME LOG_STREAM_FOLDER "/r_submix_in.raw" |
| // File permissions for stream log files. |
| #define LOG_STREAM_FILE_PERMISSIONS (S_IRUSR | S_IWUSR | S_IRGRP | S_IROTH) |
| #endif // LOG_STREAMS_TO_FILES |
| // limit for number of read error log entries to avoid spamming the logs |
| #define MAX_READ_ERROR_LOGS 5 |
| |
| // Common limits macros. |
| #ifndef min |
| #define min(a, b) ((a) < (b) ? (a) : (b)) |
| #endif // min |
| #ifndef max |
| #define max(a, b) ((a) > (b) ? (a) : (b)) |
| #endif // max |
| |
| // Set *result_variable_ptr to true if value_to_find is present in the array array_to_search, |
| // otherwise set *result_variable_ptr to false. |
| #define SUBMIX_VALUE_IN_SET(value_to_find, array_to_search, result_variable_ptr) \ |
| { \ |
| size_t i; \ |
| *(result_variable_ptr) = false; \ |
| for (i = 0; i < sizeof(array_to_search) / sizeof((array_to_search)[0]); i++) { \ |
| if ((value_to_find) == (array_to_search)[i]) { \ |
| *(result_variable_ptr) = true; \ |
| break; \ |
| } \ |
| } \ |
| } |
| |
| // Configuration of the submix pipe. |
| struct submix_config { |
| // Channel mask field in this data structure is set to either input_channel_mask or |
| // output_channel_mask depending upon the last stream to be opened on this device. |
| struct audio_config common; |
| // Input stream and output stream channel masks. This is required since input and output |
| // channel bitfields are not equivalent. |
| audio_channel_mask_t input_channel_mask; |
| audio_channel_mask_t output_channel_mask; |
| size_t pipe_frame_size; // Number of bytes in each audio frame in the pipe. |
| size_t buffer_size_frames; // Size of the audio pipe in frames. |
| // Maximum number of frames buffered by the input and output streams. |
| size_t buffer_period_size_frames; |
| }; |
| |
| #define MAX_ROUTES 10 |
| typedef struct route_config { |
| struct submix_config config; |
| char address[AUDIO_DEVICE_MAX_ADDRESS_LEN]; |
| // Pipe variables: they handle the ring buffer that "pipes" audio: |
| // - from the submix virtual audio output == what needs to be played |
| // remotely, seen as an output for AudioFlinger |
| // - to the virtual audio source == what is captured by the component |
| // which "records" the submix / virtual audio source, and handles it as needed. |
| // A usecase example is one where the component capturing the audio is then sending it over |
| // Wifi for presentation on a remote Wifi Display device (e.g. a dongle attached to a TV, or a |
| // TV with Wifi Display capabilities), or to a wireless audio player. |
| sp<MonoPipe> rsxSink; |
| sp<MonoPipeReader> rsxSource; |
| // Pointers to the current input and output stream instances. rsxSink and rsxSource are |
| // destroyed if both and input and output streams are destroyed. |
| struct submix_stream_out *output; |
| struct submix_stream_in *input; |
| } route_config_t; |
| |
| struct submix_audio_device { |
| struct audio_hw_device device; |
| route_config_t routes[MAX_ROUTES]; |
| // Device lock, also used to protect access to submix_audio_device from the input and output |
| // streams. |
| pthread_mutex_t lock; |
| }; |
| |
| struct submix_stream_out { |
| struct audio_stream_out stream; |
| struct submix_audio_device *dev; |
| int route_handle; |
| bool output_standby; |
| uint64_t frames_written; |
| uint64_t frames_written_since_standby; |
| #if LOG_STREAMS_TO_FILES |
| int log_fd; |
| #endif // LOG_STREAMS_TO_FILES |
| }; |
| |
| struct submix_stream_in { |
| struct audio_stream_in stream; |
| struct submix_audio_device *dev; |
| int route_handle; |
| bool input_standby; |
| bool output_standby_rec_thr; // output standby state as seen from record thread |
| // wall clock when recording starts |
| struct timespec record_start_time; |
| // how many frames have been requested to be read |
| uint64_t read_counter_frames; |
| uint64_t read_counter_frames_since_standby; |
| |
| #if ENABLE_LEGACY_INPUT_OPEN |
| // Number of references to this input stream. |
| volatile int32_t ref_count; |
| #endif // ENABLE_LEGACY_INPUT_OPEN |
| #if LOG_STREAMS_TO_FILES |
| int log_fd; |
| #endif // LOG_STREAMS_TO_FILES |
| |
| volatile uint16_t read_error_count; |
| }; |
| |
| // Determine whether the specified sample rate is supported by the submix module. |
| static bool sample_rate_supported(const uint32_t sample_rate) |
| { |
| // Set of sample rates supported by Format_from_SR_C() frameworks/av/media/libnbaio/NAIO.cpp. |
| static const unsigned int supported_sample_rates[] = { |
| 8000, 11025, 12000, 16000, 22050, 24000, 32000, 44100, 48000, |
| }; |
| bool return_value; |
| SUBMIX_VALUE_IN_SET(sample_rate, supported_sample_rates, &return_value); |
| return return_value; |
| } |
| |
| static size_t pipe_size_in_frames(const uint32_t sample_rate) |
| { |
| return DEFAULT_PIPE_SIZE_IN_FRAMES * ((float) sample_rate / DEFAULT_SAMPLE_RATE_HZ); |
| } |
| |
| // Determine whether the specified sample rate is supported, if it is return the specified sample |
| // rate, otherwise return the default sample rate for the submix module. |
| static uint32_t get_supported_sample_rate(uint32_t sample_rate) |
| { |
| return sample_rate_supported(sample_rate) ? sample_rate : DEFAULT_SAMPLE_RATE_HZ; |
| } |
| |
| // Determine whether the specified channel in mask is supported by the submix module. |
| static bool channel_in_mask_supported(const audio_channel_mask_t channel_in_mask) |
| { |
| // Set of channel in masks supported by Format_from_SR_C() |
| // frameworks/av/media/libnbaio/NAIO.cpp. |
| static const audio_channel_mask_t supported_channel_in_masks[] = { |
| AUDIO_CHANNEL_IN_MONO, AUDIO_CHANNEL_IN_STEREO, |
| }; |
| bool return_value; |
| SUBMIX_VALUE_IN_SET(channel_in_mask, supported_channel_in_masks, &return_value); |
| return return_value; |
| } |
| |
| // Determine whether the specified channel in mask is supported, if it is return the specified |
| // channel in mask, otherwise return the default channel in mask for the submix module. |
| static audio_channel_mask_t get_supported_channel_in_mask( |
| const audio_channel_mask_t channel_in_mask) |
| { |
| return channel_in_mask_supported(channel_in_mask) ? channel_in_mask : |
| static_cast<audio_channel_mask_t>(AUDIO_CHANNEL_IN_STEREO); |
| } |
| |
| // Determine whether the specified channel out mask is supported by the submix module. |
| static bool channel_out_mask_supported(const audio_channel_mask_t channel_out_mask) |
| { |
| // Set of channel out masks supported by Format_from_SR_C() |
| // frameworks/av/media/libnbaio/NAIO.cpp. |
| static const audio_channel_mask_t supported_channel_out_masks[] = { |
| AUDIO_CHANNEL_OUT_MONO, AUDIO_CHANNEL_OUT_STEREO, |
| }; |
| bool return_value; |
| SUBMIX_VALUE_IN_SET(channel_out_mask, supported_channel_out_masks, &return_value); |
| return return_value; |
| } |
| |
| // Determine whether the specified channel out mask is supported, if it is return the specified |
| // channel out mask, otherwise return the default channel out mask for the submix module. |
| static audio_channel_mask_t get_supported_channel_out_mask( |
| const audio_channel_mask_t channel_out_mask) |
| { |
| return channel_out_mask_supported(channel_out_mask) ? channel_out_mask : |
| static_cast<audio_channel_mask_t>(AUDIO_CHANNEL_OUT_STEREO); |
| } |
| |
| // Get a pointer to submix_stream_out given an audio_stream_out that is embedded within the |
| // structure. |
| static struct submix_stream_out * audio_stream_out_get_submix_stream_out( |
| struct audio_stream_out * const stream) |
| { |
| ALOG_ASSERT(stream); |
| return reinterpret_cast<struct submix_stream_out *>(reinterpret_cast<uint8_t *>(stream) - |
| offsetof(struct submix_stream_out, stream)); |
| } |
| |
| // Get a pointer to submix_stream_out given an audio_stream that is embedded within the structure. |
| static struct submix_stream_out * audio_stream_get_submix_stream_out( |
| struct audio_stream * const stream) |
| { |
| ALOG_ASSERT(stream); |
| return audio_stream_out_get_submix_stream_out( |
| reinterpret_cast<struct audio_stream_out *>(stream)); |
| } |
| |
| // Get a pointer to submix_stream_in given an audio_stream_in that is embedded within the |
| // structure. |
| static struct submix_stream_in * audio_stream_in_get_submix_stream_in( |
| struct audio_stream_in * const stream) |
| { |
| ALOG_ASSERT(stream); |
| return reinterpret_cast<struct submix_stream_in *>(reinterpret_cast<uint8_t *>(stream) - |
| offsetof(struct submix_stream_in, stream)); |
| } |
| |
| // Get a pointer to submix_stream_in given an audio_stream that is embedded within the structure. |
| static struct submix_stream_in * audio_stream_get_submix_stream_in( |
| struct audio_stream * const stream) |
| { |
| ALOG_ASSERT(stream); |
| return audio_stream_in_get_submix_stream_in( |
| reinterpret_cast<struct audio_stream_in *>(stream)); |
| } |
| |
| // Get a pointer to submix_audio_device given a pointer to an audio_device that is embedded within |
| // the structure. |
| static struct submix_audio_device * audio_hw_device_get_submix_audio_device( |
| struct audio_hw_device *device) |
| { |
| ALOG_ASSERT(device); |
| return reinterpret_cast<struct submix_audio_device *>(reinterpret_cast<uint8_t *>(device) - |
| offsetof(struct submix_audio_device, device)); |
| } |
| |
| // Compare an audio_config with input channel mask and an audio_config with output channel mask |
| // returning false if they do *not* match, true otherwise. |
| static bool audio_config_compare(const audio_config * const input_config, |
| const audio_config * const output_config) |
| { |
| const uint32_t input_channels = audio_channel_count_from_in_mask(input_config->channel_mask); |
| const uint32_t output_channels = audio_channel_count_from_out_mask(output_config->channel_mask); |
| if (input_channels != output_channels) { |
| ALOGE("audio_config_compare() channel count mismatch input=%d vs. output=%d", |
| input_channels, output_channels); |
| return false; |
| } |
| |
| if (input_config->sample_rate != output_config->sample_rate) { |
| ALOGE("audio_config_compare() sample rate mismatch %ul vs. %ul", |
| input_config->sample_rate, output_config->sample_rate); |
| return false; |
| } |
| if (input_config->format != output_config->format) { |
| ALOGE("audio_config_compare() format mismatch %x vs. %x", |
| input_config->format, output_config->format); |
| return false; |
| } |
| // This purposely ignores offload_info as it's not required for the submix device. |
| return true; |
| } |
| |
| // If one doesn't exist, create a pipe for the submix audio device rsxadev of size |
| // buffer_size_frames and optionally associate "in" or "out" with the submix audio device. |
| // Must be called with lock held on the submix_audio_device |
| static void submix_audio_device_create_pipe_l(struct submix_audio_device * const rsxadev, |
| const struct audio_config * const config, |
| const size_t buffer_size_frames, |
| const uint32_t buffer_period_count, |
| struct submix_stream_in * const in, |
| struct submix_stream_out * const out, |
| const char *address, |
| int route_idx) |
| { |
| ALOG_ASSERT(in || out); |
| ALOG_ASSERT(route_idx > -1); |
| ALOG_ASSERT(route_idx < MAX_ROUTES); |
| ALOGD("submix_audio_device_create_pipe_l(addr=%s, idx=%d)", address, route_idx); |
| |
| // Save a reference to the specified input or output stream and the associated channel |
| // mask. |
| if (in) { |
| in->route_handle = route_idx; |
| rsxadev->routes[route_idx].input = in; |
| rsxadev->routes[route_idx].config.input_channel_mask = config->channel_mask; |
| } |
| if (out) { |
| out->route_handle = route_idx; |
| rsxadev->routes[route_idx].output = out; |
| rsxadev->routes[route_idx].config.output_channel_mask = config->channel_mask; |
| } |
| // Save the address |
| strncpy(rsxadev->routes[route_idx].address, address, AUDIO_DEVICE_MAX_ADDRESS_LEN); |
| ALOGD(" now using address %s for route %d", rsxadev->routes[route_idx].address, route_idx); |
| // If a pipe isn't associated with the device, create one. |
| if (rsxadev->routes[route_idx].rsxSink == NULL || rsxadev->routes[route_idx].rsxSource == NULL) |
| { |
| struct submix_config * const device_config = &rsxadev->routes[route_idx].config; |
| uint32_t channel_count; |
| if (out) { |
| channel_count = audio_channel_count_from_out_mask(config->channel_mask); |
| } else { |
| channel_count = audio_channel_count_from_in_mask(config->channel_mask); |
| } |
| |
| const uint32_t pipe_channel_count = channel_count; |
| |
| const NBAIO_Format format = Format_from_SR_C(config->sample_rate, pipe_channel_count, |
| config->format); |
| const NBAIO_Format offers[1] = {format}; |
| size_t numCounterOffers = 0; |
| // Create a MonoPipe with optional blocking set to true. |
| MonoPipe* sink = new MonoPipe(buffer_size_frames, format, true /*writeCanBlock*/); |
| // Negotiation between the source and sink cannot fail as the device open operation |
| // creates both ends of the pipe using the same audio format. |
| ssize_t index = sink->negotiate(offers, 1, NULL, numCounterOffers); |
| ALOG_ASSERT(index == 0); |
| MonoPipeReader* source = new MonoPipeReader(sink); |
| numCounterOffers = 0; |
| index = source->negotiate(offers, 1, NULL, numCounterOffers); |
| ALOG_ASSERT(index == 0); |
| ALOGV("submix_audio_device_create_pipe_l(): created pipe"); |
| |
| // Save references to the source and sink. |
| ALOG_ASSERT(rsxadev->routes[route_idx].rsxSink == NULL); |
| ALOG_ASSERT(rsxadev->routes[route_idx].rsxSource == NULL); |
| rsxadev->routes[route_idx].rsxSink = sink; |
| rsxadev->routes[route_idx].rsxSource = source; |
| // Store the sanitized audio format in the device so that it's possible to determine |
| // the format of the pipe source when opening the input device. |
| memcpy(&device_config->common, config, sizeof(device_config->common)); |
| device_config->buffer_size_frames = sink->maxFrames(); |
| device_config->buffer_period_size_frames = device_config->buffer_size_frames / |
| buffer_period_count; |
| if (in) device_config->pipe_frame_size = audio_stream_in_frame_size(&in->stream); |
| if (out) device_config->pipe_frame_size = audio_stream_out_frame_size(&out->stream); |
| |
| SUBMIX_ALOGV("submix_audio_device_create_pipe_l(): pipe frame size %zd, pipe size %zd, " |
| "period size %zd", device_config->pipe_frame_size, |
| device_config->buffer_size_frames, device_config->buffer_period_size_frames); |
| } |
| } |
| |
| // Release references to the sink and source. Input and output threads may maintain references |
| // to these objects via StrongPointer (sp<MonoPipe> and sp<MonoPipeReader>) which they can use |
| // before they shutdown. |
| // Must be called with lock held on the submix_audio_device |
| static void submix_audio_device_release_pipe_l(struct submix_audio_device * const rsxadev, |
| int route_idx) |
| { |
| ALOG_ASSERT(route_idx > -1); |
| ALOG_ASSERT(route_idx < MAX_ROUTES); |
| ALOGD("submix_audio_device_release_pipe_l(idx=%d) addr=%s", route_idx, |
| rsxadev->routes[route_idx].address); |
| if (rsxadev->routes[route_idx].rsxSink != 0) { |
| rsxadev->routes[route_idx].rsxSink.clear(); |
| } |
| if (rsxadev->routes[route_idx].rsxSource != 0) { |
| rsxadev->routes[route_idx].rsxSource.clear(); |
| } |
| memset(rsxadev->routes[route_idx].address, 0, AUDIO_DEVICE_MAX_ADDRESS_LEN); |
| } |
| |
| // Remove references to the specified input and output streams. When the device no longer |
| // references input and output streams destroy the associated pipe. |
| // Must be called with lock held on the submix_audio_device |
| static void submix_audio_device_destroy_pipe_l(struct submix_audio_device * const rsxadev, |
| const struct submix_stream_in * const in, |
| const struct submix_stream_out * const out) |
| { |
| ALOGV("submix_audio_device_destroy_pipe_l()"); |
| int route_idx = -1; |
| if (in != NULL) { |
| bool shut_down = false; |
| #if ENABLE_LEGACY_INPUT_OPEN |
| const_cast<struct submix_stream_in*>(in)->ref_count--; |
| route_idx = in->route_handle; |
| ALOG_ASSERT(rsxadev->routes[route_idx].input == in); |
| if (in->ref_count == 0) { |
| rsxadev->routes[route_idx].input = NULL; |
| shut_down = true; |
| } |
| ALOGV("submix_audio_device_destroy_pipe_l(): input ref_count %d", in->ref_count); |
| #else |
| route_idx = in->route_handle; |
| ALOG_ASSERT(rsxadev->routes[route_idx].input == in); |
| rsxadev->routes[route_idx].input = NULL; |
| shut_down = true; |
| #endif // ENABLE_LEGACY_INPUT_OPEN |
| if (shut_down) { |
| sp <MonoPipe> sink = rsxadev->routes[in->route_handle].rsxSink; |
| if (sink != NULL) { |
| sink->shutdown(true); |
| } |
| } |
| } |
| if (out != NULL) { |
| route_idx = out->route_handle; |
| ALOG_ASSERT(rsxadev->routes[route_idx].output == out); |
| rsxadev->routes[route_idx].output = NULL; |
| } |
| if (route_idx != -1 && |
| rsxadev->routes[route_idx].input == NULL && rsxadev->routes[route_idx].output == NULL) { |
| submix_audio_device_release_pipe_l(rsxadev, route_idx); |
| ALOGD("submix_audio_device_destroy_pipe_l(): pipe destroyed"); |
| } |
| } |
| |
| // Sanitize the user specified audio config for a submix input / output stream. |
| static void submix_sanitize_config(struct audio_config * const config, const bool is_input_format) |
| { |
| config->channel_mask = is_input_format ? get_supported_channel_in_mask(config->channel_mask) : |
| get_supported_channel_out_mask(config->channel_mask); |
| config->sample_rate = get_supported_sample_rate(config->sample_rate); |
| config->format = DEFAULT_FORMAT; |
| } |
| |
| // Verify a submix input or output stream can be opened. |
| // Must be called with lock held on the submix_audio_device |
| static bool submix_open_validate_l(const struct submix_audio_device * const rsxadev, |
| int route_idx, |
| const struct audio_config * const config, |
| const bool opening_input) |
| { |
| bool input_open; |
| bool output_open; |
| audio_config pipe_config; |
| |
| // Query the device for the current audio config and whether input and output streams are open. |
| output_open = rsxadev->routes[route_idx].output != NULL; |
| input_open = rsxadev->routes[route_idx].input != NULL; |
| memcpy(&pipe_config, &rsxadev->routes[route_idx].config.common, sizeof(pipe_config)); |
| |
| // If the stream is already open, don't open it again. |
| if (opening_input ? !ENABLE_LEGACY_INPUT_OPEN && input_open : output_open) { |
| ALOGE("submix_open_validate_l(): %s stream already open.", opening_input ? "Input" : |
| "Output"); |
| return false; |
| } |
| |
| SUBMIX_ALOGV("submix_open_validate_l(): sample rate=%d format=%x " |
| "%s_channel_mask=%x", config->sample_rate, config->format, |
| opening_input ? "in" : "out", config->channel_mask); |
| |
| // If either stream is open, verify the existing audio config the pipe matches the user |
| // specified config. |
| if (input_open || output_open) { |
| const audio_config * const input_config = opening_input ? config : &pipe_config; |
| const audio_config * const output_config = opening_input ? &pipe_config : config; |
| // Get the channel mask of the open device. |
| pipe_config.channel_mask = |
| opening_input ? rsxadev->routes[route_idx].config.output_channel_mask : |
| rsxadev->routes[route_idx].config.input_channel_mask; |
| if (!audio_config_compare(input_config, output_config)) { |
| ALOGE("submix_open_validate_l(): Unsupported format."); |
| return false; |
| } |
| } |
| return true; |
| } |
| |
| // Must be called with lock held on the submix_audio_device |
| static status_t submix_get_route_idx_for_address_l(const struct submix_audio_device * const rsxadev, |
| const char* address, /*in*/ |
| int *idx /*out*/) |
| { |
| // Do we already have a route for this address |
| int route_idx = -1; |
| int route_empty_idx = -1; // index of an empty route slot that can be used if needed |
| for (int i=0 ; i < MAX_ROUTES ; i++) { |
| if (strcmp(rsxadev->routes[i].address, "") == 0) { |
| route_empty_idx = i; |
| } |
| if (strncmp(rsxadev->routes[i].address, address, AUDIO_DEVICE_MAX_ADDRESS_LEN) == 0) { |
| route_idx = i; |
| break; |
| } |
| } |
| |
| if ((route_idx == -1) && (route_empty_idx == -1)) { |
| ALOGE("Cannot create new route for address %s, max number of routes reached", address); |
| return -ENOMEM; |
| } |
| if (route_idx == -1) { |
| route_idx = route_empty_idx; |
| } |
| *idx = route_idx; |
| return OK; |
| } |
| |
| |
| // Calculate the maximum size of the pipe buffer in frames for the specified stream. |
| static size_t calculate_stream_pipe_size_in_frames(const struct audio_stream *stream, |
| const struct submix_config *config, |
| const size_t pipe_frames, |
| const size_t stream_frame_size) |
| { |
| const size_t pipe_frame_size = config->pipe_frame_size; |
| const size_t max_frame_size = max(stream_frame_size, pipe_frame_size); |
| return (pipe_frames * config->pipe_frame_size) / max_frame_size; |
| } |
| |
| /* audio HAL functions */ |
| |
| static uint32_t out_get_sample_rate(const struct audio_stream *stream) |
| { |
| const struct submix_stream_out * const out = audio_stream_get_submix_stream_out( |
| const_cast<struct audio_stream *>(stream)); |
| const uint32_t out_rate = out->dev->routes[out->route_handle].config.common.sample_rate; |
| SUBMIX_ALOGV("out_get_sample_rate() returns %u for addr %s", |
| out_rate, out->dev->routes[out->route_handle].address); |
| return out_rate; |
| } |
| |
| static int out_set_sample_rate(struct audio_stream *stream, uint32_t rate) |
| { |
| struct submix_stream_out * const out = audio_stream_get_submix_stream_out(stream); |
| if (!sample_rate_supported(rate)) { |
| ALOGE("out_set_sample_rate(rate=%u) rate unsupported", rate); |
| return -ENOSYS; |
| } |
| SUBMIX_ALOGV("out_set_sample_rate(rate=%u)", rate); |
| out->dev->routes[out->route_handle].config.common.sample_rate = rate; |
| return 0; |
| } |
| |
| static size_t out_get_buffer_size(const struct audio_stream *stream) |
| { |
| const struct submix_stream_out * const out = audio_stream_get_submix_stream_out( |
| const_cast<struct audio_stream *>(stream)); |
| const struct submix_config * const config = &out->dev->routes[out->route_handle].config; |
| const size_t stream_frame_size = |
| audio_stream_out_frame_size((const struct audio_stream_out *)stream); |
| const size_t buffer_size_frames = calculate_stream_pipe_size_in_frames( |
| stream, config, config->buffer_period_size_frames, stream_frame_size); |
| const size_t buffer_size_bytes = buffer_size_frames * stream_frame_size; |
| SUBMIX_ALOGV("out_get_buffer_size() returns %zu bytes, %zu frames", |
| buffer_size_bytes, buffer_size_frames); |
| return buffer_size_bytes; |
| } |
| |
| static audio_channel_mask_t out_get_channels(const struct audio_stream *stream) |
| { |
| const struct submix_stream_out * const out = audio_stream_get_submix_stream_out( |
| const_cast<struct audio_stream *>(stream)); |
| audio_channel_mask_t channel_mask = |
| out->dev->routes[out->route_handle].config.output_channel_mask; |
| SUBMIX_ALOGV("out_get_channels() returns %08x", channel_mask); |
| return channel_mask; |
| } |
| |
| static audio_format_t out_get_format(const struct audio_stream *stream) |
| { |
| const struct submix_stream_out * const out = audio_stream_get_submix_stream_out( |
| const_cast<struct audio_stream *>(stream)); |
| const audio_format_t format = out->dev->routes[out->route_handle].config.common.format; |
| SUBMIX_ALOGV("out_get_format() returns %x", format); |
| return format; |
| } |
| |
| static int out_set_format(struct audio_stream *stream, audio_format_t format) |
| { |
| const struct submix_stream_out * const out = audio_stream_get_submix_stream_out(stream); |
| if (format != out->dev->routes[out->route_handle].config.common.format) { |
| ALOGE("out_set_format(format=%x) format unsupported", format); |
| return -ENOSYS; |
| } |
| SUBMIX_ALOGV("out_set_format(format=%x)", format); |
| return 0; |
| } |
| |
| static int out_standby(struct audio_stream *stream) |
| { |
| ALOGI("out_standby()"); |
| struct submix_stream_out * const out = audio_stream_get_submix_stream_out(stream); |
| struct submix_audio_device * const rsxadev = out->dev; |
| |
| pthread_mutex_lock(&rsxadev->lock); |
| |
| out->output_standby = true; |
| out->frames_written_since_standby = 0; |
| |
| pthread_mutex_unlock(&rsxadev->lock); |
| |
| return 0; |
| } |
| |
| static int out_dump(const struct audio_stream *stream, int fd) |
| { |
| (void)stream; |
| (void)fd; |
| return 0; |
| } |
| |
| static int out_set_parameters(struct audio_stream *stream, const char *kvpairs) |
| { |
| int exiting = -1; |
| AudioParameter parms = AudioParameter(String8(kvpairs)); |
| SUBMIX_ALOGV("out_set_parameters() kvpairs='%s'", kvpairs); |
| |
| // FIXME this is using hard-coded strings but in the future, this functionality will be |
| // converted to use audio HAL extensions required to support tunneling |
| if ((parms.getInt(String8(AUDIO_PARAMETER_KEY_EXITING), exiting) == NO_ERROR) |
| && (exiting > 0)) { |
| struct submix_audio_device * const rsxadev = |
| audio_stream_get_submix_stream_out(stream)->dev; |
| pthread_mutex_lock(&rsxadev->lock); |
| { // using the sink |
| sp<MonoPipe> sink = |
| rsxadev->routes[audio_stream_get_submix_stream_out(stream)->route_handle] |
| .rsxSink; |
| if (sink == NULL) { |
| pthread_mutex_unlock(&rsxadev->lock); |
| return 0; |
| } |
| |
| ALOGD("out_set_parameters(): shutting down MonoPipe sink"); |
| sink->shutdown(true); |
| } // done using the sink |
| pthread_mutex_unlock(&rsxadev->lock); |
| } |
| return 0; |
| } |
| |
| static char * out_get_parameters(const struct audio_stream *stream, const char *keys) |
| { |
| (void)stream; |
| (void)keys; |
| return strdup(""); |
| } |
| |
| static uint32_t out_get_latency(const struct audio_stream_out *stream) |
| { |
| const struct submix_stream_out * const out = audio_stream_out_get_submix_stream_out( |
| const_cast<struct audio_stream_out *>(stream)); |
| const struct submix_config * const config = &out->dev->routes[out->route_handle].config; |
| const size_t stream_frame_size = |
| audio_stream_out_frame_size(stream); |
| const size_t buffer_size_frames = calculate_stream_pipe_size_in_frames( |
| &stream->common, config, config->buffer_size_frames, stream_frame_size); |
| const uint32_t sample_rate = out_get_sample_rate(&stream->common); |
| const uint32_t latency_ms = (buffer_size_frames * 1000) / sample_rate; |
| SUBMIX_ALOGV("out_get_latency() returns %u ms, size in frames %zu, sample rate %u", |
| latency_ms, buffer_size_frames, sample_rate); |
| return latency_ms; |
| } |
| |
| static int out_set_volume(struct audio_stream_out *stream, float left, |
| float right) |
| { |
| (void)stream; |
| (void)left; |
| (void)right; |
| return -ENOSYS; |
| } |
| |
| static ssize_t out_write(struct audio_stream_out *stream, const void* buffer, |
| size_t bytes) |
| { |
| SUBMIX_ALOGV("out_write(bytes=%zd)", bytes); |
| ssize_t written_frames = 0; |
| const size_t frame_size = audio_stream_out_frame_size(stream); |
| struct submix_stream_out * const out = audio_stream_out_get_submix_stream_out(stream); |
| struct submix_audio_device * const rsxadev = out->dev; |
| const size_t frames = bytes / frame_size; |
| |
| pthread_mutex_lock(&rsxadev->lock); |
| |
| out->output_standby = false; |
| |
| sp<MonoPipe> sink = rsxadev->routes[out->route_handle].rsxSink; |
| if (sink != NULL) { |
| if (sink->isShutdown()) { |
| sink.clear(); |
| pthread_mutex_unlock(&rsxadev->lock); |
| SUBMIX_ALOGV("out_write(): pipe shutdown, ignoring the write."); |
| // the pipe has already been shutdown, this buffer will be lost but we must |
| // simulate timing so we don't drain the output faster than realtime |
| usleep(frames * 1000000 / out_get_sample_rate(&stream->common)); |
| |
| pthread_mutex_lock(&rsxadev->lock); |
| out->frames_written += frames; |
| out->frames_written_since_standby += frames; |
| pthread_mutex_unlock(&rsxadev->lock); |
| return bytes; |
| } |
| } else { |
| pthread_mutex_unlock(&rsxadev->lock); |
| ALOGE("out_write without a pipe!"); |
| ALOG_ASSERT("out_write without a pipe!"); |
| return 0; |
| } |
| |
| // If the write to the sink would block, flush enough frames |
| // from the pipe to make space to write the most recent data. |
| // We DO NOT block if: |
| // - no peer input stream is present |
| // - the peer input is in standby AFTER having been active. |
| // We DO block if: |
| // - the input was never activated to avoid discarding first frames |
| // in the pipe in case capture start was delayed |
| { |
| const size_t availableToWrite = sink->availableToWrite(); |
| // NOTE: rsxSink has been checked above and sink and source life cycles are synchronized |
| sp<MonoPipeReader> source = rsxadev->routes[out->route_handle].rsxSource; |
| const struct submix_stream_in *in = rsxadev->routes[out->route_handle].input; |
| const bool dont_block = (in == NULL) |
| || (in->input_standby && (in->read_counter_frames_since_standby != 0)); |
| if (dont_block && availableToWrite < frames) { |
| static uint8_t flush_buffer[64]; |
| const size_t flushBufferSizeFrames = sizeof(flush_buffer) / frame_size; |
| size_t frames_to_flush_from_source = frames - availableToWrite; |
| SUBMIX_ALOGV("out_write(): flushing %llu frames from the pipe to avoid blocking", |
| (unsigned long long)frames_to_flush_from_source); |
| while (frames_to_flush_from_source) { |
| const size_t flush_size = min(frames_to_flush_from_source, flushBufferSizeFrames); |
| frames_to_flush_from_source -= flush_size; |
| // read does not block |
| source->read(flush_buffer, flush_size); |
| } |
| } |
| } |
| |
| pthread_mutex_unlock(&rsxadev->lock); |
| |
| written_frames = sink->write(buffer, frames); |
| |
| #if LOG_STREAMS_TO_FILES |
| if (out->log_fd >= 0) write(out->log_fd, buffer, written_frames * frame_size); |
| #endif // LOG_STREAMS_TO_FILES |
| |
| if (written_frames < 0) { |
| if (written_frames == (ssize_t)NEGOTIATE) { |
| ALOGE("out_write() write to pipe returned NEGOTIATE"); |
| |
| pthread_mutex_lock(&rsxadev->lock); |
| sink.clear(); |
| pthread_mutex_unlock(&rsxadev->lock); |
| |
| written_frames = 0; |
| return 0; |
| } else { |
| // write() returned UNDERRUN or WOULD_BLOCK, retry |
| ALOGE("out_write() write to pipe returned unexpected %zd", written_frames); |
| written_frames = sink->write(buffer, frames); |
| } |
| } |
| |
| pthread_mutex_lock(&rsxadev->lock); |
| sink.clear(); |
| if (written_frames > 0) { |
| out->frames_written_since_standby += written_frames; |
| out->frames_written += written_frames; |
| } |
| pthread_mutex_unlock(&rsxadev->lock); |
| |
| if (written_frames < 0) { |
| ALOGE("out_write() failed writing to pipe with %zd", written_frames); |
| return 0; |
| } |
| const ssize_t written_bytes = written_frames * frame_size; |
| SUBMIX_ALOGV("out_write() wrote %zd bytes %zd frames", written_bytes, written_frames); |
| return written_bytes; |
| } |
| |
| static int out_get_presentation_position(const struct audio_stream_out *stream, |
| uint64_t *frames, struct timespec *timestamp) |
| { |
| if (stream == NULL || frames == NULL || timestamp == NULL) { |
| return -EINVAL; |
| } |
| |
| const submix_stream_out *out = audio_stream_out_get_submix_stream_out( |
| const_cast<struct audio_stream_out *>(stream)); |
| struct submix_audio_device * const rsxadev = out->dev; |
| |
| int ret = -EWOULDBLOCK; |
| pthread_mutex_lock(&rsxadev->lock); |
| sp<MonoPipeReader> source = rsxadev->routes[out->route_handle].rsxSource; |
| if (source == NULL) { |
| ALOGW("%s called on released output", __FUNCTION__); |
| pthread_mutex_unlock(&rsxadev->lock); |
| return -ENODEV; |
| } |
| |
| const ssize_t frames_in_pipe = source->availableToRead(); |
| if (CC_UNLIKELY(frames_in_pipe < 0)) { |
| *frames = out->frames_written; |
| ret = 0; |
| } else if (out->frames_written >= (uint64_t)frames_in_pipe) { |
| *frames = out->frames_written - frames_in_pipe; |
| ret = 0; |
| } |
| pthread_mutex_unlock(&rsxadev->lock); |
| |
| if (ret == 0) { |
| clock_gettime(CLOCK_MONOTONIC, timestamp); |
| } |
| |
| SUBMIX_ALOGV("out_get_presentation_position() got frames=%llu timestamp sec=%llu", |
| frames ? (unsigned long long)*frames : -1ULL, |
| timestamp ? (unsigned long long)timestamp->tv_sec : -1ULL); |
| |
| return ret; |
| } |
| |
| static int out_get_render_position(const struct audio_stream_out *stream, |
| uint32_t *dsp_frames) |
| { |
| if (stream == NULL || dsp_frames == NULL) { |
| return -EINVAL; |
| } |
| |
| const submix_stream_out *out = audio_stream_out_get_submix_stream_out( |
| const_cast<struct audio_stream_out *>(stream)); |
| struct submix_audio_device * const rsxadev = out->dev; |
| |
| pthread_mutex_lock(&rsxadev->lock); |
| sp<MonoPipeReader> source = rsxadev->routes[out->route_handle].rsxSource; |
| if (source == NULL) { |
| ALOGW("%s called on released output", __FUNCTION__); |
| pthread_mutex_unlock(&rsxadev->lock); |
| return -ENODEV; |
| } |
| |
| const ssize_t frames_in_pipe = source->availableToRead(); |
| if (CC_UNLIKELY(frames_in_pipe < 0)) { |
| *dsp_frames = (uint32_t)out->frames_written_since_standby; |
| } else { |
| *dsp_frames = out->frames_written_since_standby > (uint64_t) frames_in_pipe ? |
| (uint32_t)(out->frames_written_since_standby - frames_in_pipe) : 0; |
| } |
| pthread_mutex_unlock(&rsxadev->lock); |
| |
| return 0; |
| } |
| |
| static int out_add_audio_effect(const struct audio_stream *stream, effect_handle_t effect) |
| { |
| (void)stream; |
| (void)effect; |
| return 0; |
| } |
| |
| static int out_remove_audio_effect(const struct audio_stream *stream, effect_handle_t effect) |
| { |
| (void)stream; |
| (void)effect; |
| return 0; |
| } |
| |
| static int out_get_next_write_timestamp(const struct audio_stream_out *stream, |
| int64_t *timestamp) |
| { |
| (void)stream; |
| (void)timestamp; |
| return -ENOSYS; |
| } |
| |
| /** audio_stream_in implementation **/ |
| static uint32_t in_get_sample_rate(const struct audio_stream *stream) |
| { |
| const struct submix_stream_in * const in = audio_stream_get_submix_stream_in( |
| const_cast<struct audio_stream*>(stream)); |
| const uint32_t rate = in->dev->routes[in->route_handle].config.common.sample_rate; |
| SUBMIX_ALOGV("in_get_sample_rate() returns %u", rate); |
| return rate; |
| } |
| |
| static int in_set_sample_rate(struct audio_stream *stream, uint32_t rate) |
| { |
| const struct submix_stream_in * const in = audio_stream_get_submix_stream_in(stream); |
| if (!sample_rate_supported(rate)) { |
| ALOGE("in_set_sample_rate(rate=%u) rate unsupported", rate); |
| return -ENOSYS; |
| } |
| in->dev->routes[in->route_handle].config.common.sample_rate = rate; |
| SUBMIX_ALOGV("in_set_sample_rate() set %u", rate); |
| return 0; |
| } |
| |
| static size_t in_get_buffer_size(const struct audio_stream *stream) |
| { |
| const struct submix_stream_in * const in = audio_stream_get_submix_stream_in( |
| const_cast<struct audio_stream*>(stream)); |
| const struct submix_config * const config = &in->dev->routes[in->route_handle].config; |
| const size_t stream_frame_size = |
| audio_stream_in_frame_size((const struct audio_stream_in *)stream); |
| size_t buffer_size_frames = calculate_stream_pipe_size_in_frames( |
| stream, config, config->buffer_period_size_frames, stream_frame_size); |
| const size_t buffer_size_bytes = buffer_size_frames * stream_frame_size; |
| SUBMIX_ALOGV("in_get_buffer_size() returns %zu bytes, %zu frames", buffer_size_bytes, |
| buffer_size_frames); |
| return buffer_size_bytes; |
| } |
| |
| static audio_channel_mask_t in_get_channels(const struct audio_stream *stream) |
| { |
| const struct submix_stream_in * const in = audio_stream_get_submix_stream_in( |
| const_cast<struct audio_stream*>(stream)); |
| const audio_channel_mask_t channel_mask = |
| in->dev->routes[in->route_handle].config.input_channel_mask; |
| SUBMIX_ALOGV("in_get_channels() returns %x", channel_mask); |
| return channel_mask; |
| } |
| |
| static audio_format_t in_get_format(const struct audio_stream *stream) |
| { |
| const struct submix_stream_in * const in = audio_stream_get_submix_stream_in( |
| const_cast<struct audio_stream*>(stream)); |
| const audio_format_t format = in->dev->routes[in->route_handle].config.common.format; |
| SUBMIX_ALOGV("in_get_format() returns %x", format); |
| return format; |
| } |
| |
| static int in_set_format(struct audio_stream *stream, audio_format_t format) |
| { |
| const struct submix_stream_in * const in = audio_stream_get_submix_stream_in(stream); |
| if (format != in->dev->routes[in->route_handle].config.common.format) { |
| ALOGE("in_set_format(format=%x) format unsupported", format); |
| return -ENOSYS; |
| } |
| SUBMIX_ALOGV("in_set_format(format=%x)", format); |
| return 0; |
| } |
| |
| static int in_standby(struct audio_stream *stream) |
| { |
| ALOGI("in_standby()"); |
| struct submix_stream_in * const in = audio_stream_get_submix_stream_in(stream); |
| struct submix_audio_device * const rsxadev = in->dev; |
| |
| pthread_mutex_lock(&rsxadev->lock); |
| |
| in->input_standby = true; |
| |
| pthread_mutex_unlock(&rsxadev->lock); |
| |
| return 0; |
| } |
| |
| static int in_dump(const struct audio_stream *stream, int fd) |
| { |
| (void)stream; |
| (void)fd; |
| return 0; |
| } |
| |
| static int in_set_parameters(struct audio_stream *stream, const char *kvpairs) |
| { |
| (void)stream; |
| (void)kvpairs; |
| return 0; |
| } |
| |
| static char * in_get_parameters(const struct audio_stream *stream, |
| const char *keys) |
| { |
| (void)stream; |
| (void)keys; |
| return strdup(""); |
| } |
| |
| static int in_set_gain(struct audio_stream_in *stream, float gain) |
| { |
| (void)stream; |
| (void)gain; |
| return 0; |
| } |
| |
| static ssize_t in_read(struct audio_stream_in *stream, void* buffer, |
| size_t bytes) |
| { |
| struct submix_stream_in * const in = audio_stream_in_get_submix_stream_in(stream); |
| struct submix_audio_device * const rsxadev = in->dev; |
| const size_t frame_size = audio_stream_in_frame_size(stream); |
| const size_t frames_to_read = bytes / frame_size; |
| |
| SUBMIX_ALOGV("in_read bytes=%zu", bytes); |
| pthread_mutex_lock(&rsxadev->lock); |
| |
| const bool output_standby = rsxadev->routes[in->route_handle].output == NULL |
| ? true : rsxadev->routes[in->route_handle].output->output_standby; |
| const bool output_standby_transition = (in->output_standby_rec_thr != output_standby); |
| in->output_standby_rec_thr = output_standby; |
| |
| if (in->input_standby || output_standby_transition) { |
| in->input_standby = false; |
| // keep track of when we exit input standby (== first read == start "real recording") |
| // or when we start recording silence, and reset projected time |
| int rc = clock_gettime(CLOCK_MONOTONIC, &in->record_start_time); |
| if (rc == 0) { |
| in->read_counter_frames_since_standby = 0; |
| } |
| } |
| |
| in->read_counter_frames += frames_to_read; |
| in->read_counter_frames_since_standby += frames_to_read; |
| size_t remaining_frames = frames_to_read; |
| |
| { |
| // about to read from audio source |
| sp<MonoPipeReader> source = rsxadev->routes[in->route_handle].rsxSource; |
| if (source == NULL) { |
| in->read_error_count++;// ok if it rolls over |
| ALOGE_IF(in->read_error_count < MAX_READ_ERROR_LOGS, |
| "no audio pipe yet we're trying to read! (not all errors will be logged)"); |
| pthread_mutex_unlock(&rsxadev->lock); |
| usleep(frames_to_read * 1000000 / in_get_sample_rate(&stream->common)); |
| memset(buffer, 0, bytes); |
| return bytes; |
| } |
| |
| pthread_mutex_unlock(&rsxadev->lock); |
| |
| // read the data from the pipe (it's non blocking) |
| int attempts = 0; |
| char* buff = (char*)buffer; |
| |
| while ((remaining_frames > 0) && (attempts < MAX_READ_ATTEMPTS)) { |
| ssize_t frames_read = -1977; |
| size_t read_frames = remaining_frames; |
| |
| SUBMIX_ALOGV("in_read(): frames available to read %zd", source->availableToRead()); |
| |
| frames_read = source->read(buff, read_frames); |
| |
| SUBMIX_ALOGV("in_read(): frames read %zd", frames_read); |
| |
| if (frames_read > 0) { |
| #if LOG_STREAMS_TO_FILES |
| if (in->log_fd >= 0) write(in->log_fd, buff, frames_read * frame_size); |
| #endif // LOG_STREAMS_TO_FILES |
| |
| remaining_frames -= frames_read; |
| buff += frames_read * frame_size; |
| SUBMIX_ALOGV(" in_read (att=%d) got %zd frames, remaining=%zu", |
| attempts, frames_read, remaining_frames); |
| } else { |
| attempts++; |
| SUBMIX_ALOGE(" in_read read returned %zd", frames_read); |
| usleep(READ_ATTEMPT_SLEEP_MS * 1000); |
| } |
| } |
| // done using the source |
| pthread_mutex_lock(&rsxadev->lock); |
| source.clear(); |
| pthread_mutex_unlock(&rsxadev->lock); |
| } |
| |
| if (remaining_frames > 0) { |
| const size_t remaining_bytes = remaining_frames * frame_size; |
| SUBMIX_ALOGV(" clearing remaining_frames = %zu", remaining_frames); |
| memset(((char*)buffer)+ bytes - remaining_bytes, 0, remaining_bytes); |
| } |
| |
| // compute how much we need to sleep after reading the data by comparing the wall clock with |
| // the projected time at which we should return. |
| struct timespec time_after_read;// wall clock after reading from the pipe |
| struct timespec record_duration;// observed record duration |
| int rc = clock_gettime(CLOCK_MONOTONIC, &time_after_read); |
| const uint32_t sample_rate = in_get_sample_rate(&stream->common); |
| if (rc == 0) { |
| // for how long have we been recording? |
| record_duration.tv_sec = time_after_read.tv_sec - in->record_start_time.tv_sec; |
| record_duration.tv_nsec = time_after_read.tv_nsec - in->record_start_time.tv_nsec; |
| if (record_duration.tv_nsec < 0) { |
| record_duration.tv_sec--; |
| record_duration.tv_nsec += 1000000000; |
| } |
| |
| // read_counter_frames_since_standby contains the number of frames that have been read since |
| // the beginning of recording (including this call): it's converted to usec and compared to |
| // how long we've been recording for, which gives us how long we must wait to sync the |
| // projected recording time, and the observed recording time. |
| long projected_vs_observed_offset_us = |
| ((int64_t)(in->read_counter_frames_since_standby |
| - (record_duration.tv_sec*sample_rate))) |
| * 1000000 / sample_rate |
| - (record_duration.tv_nsec / 1000); |
| |
| SUBMIX_ALOGV(" record duration %5lds %3ldms, will wait: %7ldus", |
| record_duration.tv_sec, record_duration.tv_nsec/1000000, |
| projected_vs_observed_offset_us); |
| if (projected_vs_observed_offset_us > 0) { |
| usleep(projected_vs_observed_offset_us); |
| } |
| } |
| |
| SUBMIX_ALOGV("in_read returns %zu", bytes); |
| return bytes; |
| |
| } |
| |
| static uint32_t in_get_input_frames_lost(struct audio_stream_in *stream) |
| { |
| (void)stream; |
| return 0; |
| } |
| |
| static int in_get_capture_position(const struct audio_stream_in *stream, |
| int64_t *frames, int64_t *time) |
| { |
| if (stream == NULL || frames == NULL || time == NULL) { |
| return -EINVAL; |
| } |
| |
| struct submix_stream_in * const in = audio_stream_in_get_submix_stream_in( |
| (struct audio_stream_in*)stream); |
| struct submix_audio_device * const rsxadev = in->dev; |
| |
| pthread_mutex_lock(&rsxadev->lock); |
| sp<MonoPipeReader> source = rsxadev->routes[in->route_handle].rsxSource; |
| if (source == NULL) { |
| ALOGW("%s called on released input", __FUNCTION__); |
| pthread_mutex_unlock(&rsxadev->lock); |
| return -ENODEV; |
| } |
| *frames = in->read_counter_frames; |
| const ssize_t frames_in_pipe = source->availableToRead(); |
| pthread_mutex_unlock(&rsxadev->lock); |
| if (frames_in_pipe > 0) { |
| *frames += frames_in_pipe; |
| } |
| |
| struct timespec timestamp; |
| clock_gettime(CLOCK_MONOTONIC, ×tamp); |
| *time = timestamp.tv_sec * 1000000000LL + timestamp.tv_nsec; |
| return 0; |
| } |
| |
| static int in_add_audio_effect(const struct audio_stream *stream, effect_handle_t effect) |
| { |
| (void)stream; |
| (void)effect; |
| return 0; |
| } |
| |
| static int in_remove_audio_effect(const struct audio_stream *stream, effect_handle_t effect) |
| { |
| (void)stream; |
| (void)effect; |
| return 0; |
| } |
| |
| static int adev_open_output_stream(struct audio_hw_device *dev, |
| audio_io_handle_t handle, |
| audio_devices_t devices, |
| audio_output_flags_t flags, |
| struct audio_config *config, |
| struct audio_stream_out **stream_out, |
| const char *address) |
| { |
| struct submix_audio_device * const rsxadev = audio_hw_device_get_submix_audio_device(dev); |
| ALOGD("adev_open_output_stream(address=%s)", address); |
| struct submix_stream_out *out; |
| (void)handle; |
| (void)devices; |
| (void)flags; |
| |
| *stream_out = NULL; |
| |
| // Make sure it's possible to open the device given the current audio config. |
| submix_sanitize_config(config, false); |
| |
| int route_idx = -1; |
| |
| pthread_mutex_lock(&rsxadev->lock); |
| |
| status_t res = submix_get_route_idx_for_address_l(rsxadev, address, &route_idx); |
| if (res != OK) { |
| ALOGE("Error %d looking for address=%s in adev_open_output_stream", res, address); |
| pthread_mutex_unlock(&rsxadev->lock); |
| return res; |
| } |
| |
| if (!submix_open_validate_l(rsxadev, route_idx, config, false)) { |
| ALOGE("adev_open_output_stream(): Unable to open output stream for address %s", address); |
| pthread_mutex_unlock(&rsxadev->lock); |
| return -EINVAL; |
| } |
| |
| out = (struct submix_stream_out *)calloc(1, sizeof(struct submix_stream_out)); |
| if (!out) { |
| pthread_mutex_unlock(&rsxadev->lock); |
| return -ENOMEM; |
| } |
| |
| // Initialize the function pointer tables (v-tables). |
| out->stream.common.get_sample_rate = out_get_sample_rate; |
| out->stream.common.set_sample_rate = out_set_sample_rate; |
| out->stream.common.get_buffer_size = out_get_buffer_size; |
| out->stream.common.get_channels = out_get_channels; |
| out->stream.common.get_format = out_get_format; |
| out->stream.common.set_format = out_set_format; |
| out->stream.common.standby = out_standby; |
| out->stream.common.dump = out_dump; |
| out->stream.common.set_parameters = out_set_parameters; |
| out->stream.common.get_parameters = out_get_parameters; |
| out->stream.common.add_audio_effect = out_add_audio_effect; |
| out->stream.common.remove_audio_effect = out_remove_audio_effect; |
| out->stream.get_latency = out_get_latency; |
| out->stream.set_volume = out_set_volume; |
| out->stream.write = out_write; |
| out->stream.get_render_position = out_get_render_position; |
| out->stream.get_next_write_timestamp = out_get_next_write_timestamp; |
| out->stream.get_presentation_position = out_get_presentation_position; |
| |
| // If the sink has been shutdown or pipe recreation is forced (see above), delete the pipe so |
| // that it's recreated. |
| if ((rsxadev->routes[route_idx].rsxSink != NULL |
| && rsxadev->routes[route_idx].rsxSink->isShutdown())) { |
| submix_audio_device_release_pipe_l(rsxadev, route_idx); |
| } |
| |
| // Store a pointer to the device from the output stream. |
| out->dev = rsxadev; |
| // Initialize the pipe. |
| const size_t pipeSizeInFrames = pipe_size_in_frames(config->sample_rate); |
| ALOGI("adev_open_output_stream(): about to create pipe at index %d, rate %u, pipe size %zu", |
| route_idx, config->sample_rate, pipeSizeInFrames); |
| submix_audio_device_create_pipe_l(rsxadev, config, pipeSizeInFrames, |
| DEFAULT_PIPE_PERIOD_COUNT, NULL, out, address, route_idx); |
| #if LOG_STREAMS_TO_FILES |
| out->log_fd = open(LOG_STREAM_OUT_FILENAME, O_CREAT | O_TRUNC | O_WRONLY, |
| LOG_STREAM_FILE_PERMISSIONS); |
| ALOGE_IF(out->log_fd < 0, "adev_open_output_stream(): log file open failed %s", |
| strerror(errno)); |
| ALOGV("adev_open_output_stream(): log_fd = %d", out->log_fd); |
| #endif // LOG_STREAMS_TO_FILES |
| // Return the output stream. |
| *stream_out = &out->stream; |
| |
| pthread_mutex_unlock(&rsxadev->lock); |
| return 0; |
| } |
| |
| static void adev_close_output_stream(struct audio_hw_device *dev, |
| struct audio_stream_out *stream) |
| { |
| struct submix_audio_device * rsxadev = audio_hw_device_get_submix_audio_device( |
| const_cast<struct audio_hw_device*>(dev)); |
| struct submix_stream_out * const out = audio_stream_out_get_submix_stream_out(stream); |
| |
| pthread_mutex_lock(&rsxadev->lock); |
| ALOGD("adev_close_output_stream() addr = %s", rsxadev->routes[out->route_handle].address); |
| submix_audio_device_destroy_pipe_l(audio_hw_device_get_submix_audio_device(dev), NULL, out); |
| #if LOG_STREAMS_TO_FILES |
| if (out->log_fd >= 0) close(out->log_fd); |
| #endif // LOG_STREAMS_TO_FILES |
| |
| pthread_mutex_unlock(&rsxadev->lock); |
| free(out); |
| } |
| |
| static int adev_set_parameters(struct audio_hw_device *dev, const char *kvpairs) |
| { |
| (void)dev; |
| (void)kvpairs; |
| return -ENOSYS; |
| } |
| |
| static char * adev_get_parameters(const struct audio_hw_device *dev, |
| const char *keys) |
| { |
| (void)dev; |
| (void)keys; |
| return strdup("");; |
| } |
| |
| static int adev_init_check(const struct audio_hw_device *dev) |
| { |
| ALOGI("adev_init_check()"); |
| (void)dev; |
| return 0; |
| } |
| |
| static int adev_set_voice_volume(struct audio_hw_device *dev, float volume) |
| { |
| (void)dev; |
| (void)volume; |
| return -ENOSYS; |
| } |
| |
| static int adev_set_master_volume(struct audio_hw_device *dev, float volume) |
| { |
| (void)dev; |
| (void)volume; |
| return -ENOSYS; |
| } |
| |
| static int adev_get_master_volume(struct audio_hw_device *dev, float *volume) |
| { |
| (void)dev; |
| (void)volume; |
| return -ENOSYS; |
| } |
| |
| static int adev_set_master_mute(struct audio_hw_device *dev, bool muted) |
| { |
| (void)dev; |
| (void)muted; |
| return -ENOSYS; |
| } |
| |
| static int adev_get_master_mute(struct audio_hw_device *dev, bool *muted) |
| { |
| (void)dev; |
| (void)muted; |
| return -ENOSYS; |
| } |
| |
| static int adev_set_mode(struct audio_hw_device *dev, audio_mode_t mode) |
| { |
| (void)dev; |
| (void)mode; |
| return 0; |
| } |
| |
| static int adev_set_mic_mute(struct audio_hw_device *dev, bool state) |
| { |
| (void)dev; |
| (void)state; |
| return -ENOSYS; |
| } |
| |
| static int adev_get_mic_mute(const struct audio_hw_device *dev, bool *state) |
| { |
| (void)dev; |
| (void)state; |
| return -ENOSYS; |
| } |
| |
| static size_t adev_get_input_buffer_size(const struct audio_hw_device *dev, |
| const struct audio_config *config) |
| { |
| if (audio_is_linear_pcm(config->format)) { |
| size_t max_buffer_period_size_frames = 0; |
| struct submix_audio_device * rsxadev = audio_hw_device_get_submix_audio_device( |
| const_cast<struct audio_hw_device*>(dev)); |
| // look for the largest buffer period size |
| for (int i = 0 ; i < MAX_ROUTES ; i++) { |
| if (rsxadev->routes[i].config.buffer_period_size_frames > max_buffer_period_size_frames) |
| { |
| max_buffer_period_size_frames = rsxadev->routes[i].config.buffer_period_size_frames; |
| } |
| } |
| const size_t frame_size_in_bytes = audio_channel_count_from_in_mask(config->channel_mask) * |
| audio_bytes_per_sample(config->format); |
| if (max_buffer_period_size_frames == 0) { |
| max_buffer_period_size_frames = |
| pipe_size_in_frames(get_supported_sample_rate(config->sample_rate));; |
| } |
| const size_t buffer_size = max_buffer_period_size_frames * frame_size_in_bytes; |
| SUBMIX_ALOGV("adev_get_input_buffer_size() returns %zu bytes, %zu frames", |
| buffer_size, max_buffer_period_size_frames); |
| return buffer_size; |
| } |
| return 0; |
| } |
| |
| static int adev_open_input_stream(struct audio_hw_device *dev, |
| audio_io_handle_t handle, |
| audio_devices_t devices, |
| struct audio_config *config, |
| struct audio_stream_in **stream_in, |
| audio_input_flags_t flags __unused, |
| const char *address, |
| audio_source_t source __unused) |
| { |
| struct submix_audio_device *rsxadev = audio_hw_device_get_submix_audio_device(dev); |
| struct submix_stream_in *in; |
| ALOGD("adev_open_input_stream(addr=%s)", address); |
| (void)handle; |
| (void)devices; |
| |
| *stream_in = NULL; |
| |
| // Do we already have a route for this address |
| int route_idx = -1; |
| |
| pthread_mutex_lock(&rsxadev->lock); |
| |
| status_t res = submix_get_route_idx_for_address_l(rsxadev, address, &route_idx); |
| if (res != OK) { |
| ALOGE("Error %d looking for address=%s in adev_open_input_stream", res, address); |
| pthread_mutex_unlock(&rsxadev->lock); |
| return res; |
| } |
| |
| // Make sure it's possible to open the device given the current audio config. |
| submix_sanitize_config(config, true); |
| if (!submix_open_validate_l(rsxadev, route_idx, config, true)) { |
| ALOGE("adev_open_input_stream(): Unable to open input stream."); |
| pthread_mutex_unlock(&rsxadev->lock); |
| return -EINVAL; |
| } |
| |
| #if ENABLE_LEGACY_INPUT_OPEN |
| in = rsxadev->routes[route_idx].input; |
| if (in) { |
| in->ref_count++; |
| sp<MonoPipe> sink = rsxadev->routes[route_idx].rsxSink; |
| ALOG_ASSERT(sink != NULL); |
| // If the sink has been shutdown, delete the pipe. |
| if (sink != NULL) { |
| if (sink->isShutdown()) { |
| ALOGD(" Non-NULL shut down sink when opening input stream, releasing, refcount=%d", |
| in->ref_count); |
| submix_audio_device_release_pipe_l(rsxadev, in->route_handle); |
| } else { |
| ALOGD(" Non-NULL sink when opening input stream, refcount=%d", in->ref_count); |
| } |
| } else { |
| ALOGE("NULL sink when opening input stream, refcount=%d", in->ref_count); |
| } |
| } |
| #else |
| in = NULL; |
| #endif // ENABLE_LEGACY_INPUT_OPEN |
| |
| if (!in) { |
| in = (struct submix_stream_in *)calloc(1, sizeof(struct submix_stream_in)); |
| if (!in) return -ENOMEM; |
| #if ENABLE_LEGACY_INPUT_OPEN |
| in->ref_count = 1; |
| #endif |
| |
| // Initialize the function pointer tables (v-tables). |
| in->stream.common.get_sample_rate = in_get_sample_rate; |
| in->stream.common.set_sample_rate = in_set_sample_rate; |
| in->stream.common.get_buffer_size = in_get_buffer_size; |
| in->stream.common.get_channels = in_get_channels; |
| in->stream.common.get_format = in_get_format; |
| in->stream.common.set_format = in_set_format; |
| in->stream.common.standby = in_standby; |
| in->stream.common.dump = in_dump; |
| in->stream.common.set_parameters = in_set_parameters; |
| in->stream.common.get_parameters = in_get_parameters; |
| in->stream.common.add_audio_effect = in_add_audio_effect; |
| in->stream.common.remove_audio_effect = in_remove_audio_effect; |
| in->stream.set_gain = in_set_gain; |
| in->stream.read = in_read; |
| in->stream.get_input_frames_lost = in_get_input_frames_lost; |
| in->stream.get_capture_position = in_get_capture_position; |
| |
| in->dev = rsxadev; |
| #if LOG_STREAMS_TO_FILES |
| in->log_fd = -1; |
| #endif |
| } |
| |
| // Initialize the input stream. |
| in->read_counter_frames = 0; |
| in->read_counter_frames_since_standby = 0; |
| in->input_standby = true; |
| if (rsxadev->routes[route_idx].output != NULL) { |
| in->output_standby_rec_thr = rsxadev->routes[route_idx].output->output_standby; |
| } else { |
| in->output_standby_rec_thr = true; |
| } |
| |
| in->read_error_count = 0; |
| // Initialize the pipe. |
| const size_t pipeSizeInFrames = pipe_size_in_frames(config->sample_rate); |
| ALOGI("adev_open_input_stream(): about to create pipe at index %d, rate %u, pipe size %zu", |
| route_idx, config->sample_rate, pipeSizeInFrames); |
| submix_audio_device_create_pipe_l(rsxadev, config, pipeSizeInFrames, |
| DEFAULT_PIPE_PERIOD_COUNT, in, NULL, address, route_idx); |
| |
| sp <MonoPipe> sink = rsxadev->routes[route_idx].rsxSink; |
| if (sink != NULL) { |
| sink->shutdown(false); |
| } |
| |
| #if LOG_STREAMS_TO_FILES |
| if (in->log_fd >= 0) close(in->log_fd); |
| in->log_fd = open(LOG_STREAM_IN_FILENAME, O_CREAT | O_TRUNC | O_WRONLY, |
| LOG_STREAM_FILE_PERMISSIONS); |
| ALOGE_IF(in->log_fd < 0, "adev_open_input_stream(): log file open failed %s", |
| strerror(errno)); |
| ALOGV("adev_open_input_stream(): log_fd = %d", in->log_fd); |
| #endif // LOG_STREAMS_TO_FILES |
| // Return the input stream. |
| *stream_in = &in->stream; |
| |
| pthread_mutex_unlock(&rsxadev->lock); |
| return 0; |
| } |
| |
| static void adev_close_input_stream(struct audio_hw_device *dev, |
| struct audio_stream_in *stream) |
| { |
| struct submix_audio_device * const rsxadev = audio_hw_device_get_submix_audio_device(dev); |
| |
| struct submix_stream_in * const in = audio_stream_in_get_submix_stream_in(stream); |
| ALOGD("adev_close_input_stream()"); |
| pthread_mutex_lock(&rsxadev->lock); |
| submix_audio_device_destroy_pipe_l(rsxadev, in, NULL); |
| #if LOG_STREAMS_TO_FILES |
| if (in->log_fd >= 0) close(in->log_fd); |
| #endif // LOG_STREAMS_TO_FILES |
| #if ENABLE_LEGACY_INPUT_OPEN |
| if (in->ref_count == 0) free(in); |
| #else |
| free(in); |
| #endif // ENABLE_LEGACY_INPUT_OPEN |
| |
| pthread_mutex_unlock(&rsxadev->lock); |
| } |
| |
| static int adev_dump(const audio_hw_device_t *device, int fd) |
| { |
| const struct submix_audio_device * rsxadev = //audio_hw_device_get_submix_audio_device(device); |
| reinterpret_cast<const struct submix_audio_device *>( |
| reinterpret_cast<const uint8_t *>(device) - |
| offsetof(struct submix_audio_device, device)); |
| char msg[100]; |
| int n = snprintf(msg, sizeof(msg), "\nReroute submix audio module:\n"); |
| write(fd, &msg, n); |
| for (int i=0 ; i < MAX_ROUTES ; i++) { |
| n = snprintf(msg, sizeof(msg), " route[%d], rate=%d addr=[%s]\n", i, |
| rsxadev->routes[i].config.common.sample_rate, |
| rsxadev->routes[i].address); |
| write(fd, &msg, n); |
| } |
| return 0; |
| } |
| |
| static int adev_close(hw_device_t *device) |
| { |
| ALOGI("adev_close()"); |
| free(device); |
| return 0; |
| } |
| |
| static int adev_open(const hw_module_t* module, const char* name, |
| hw_device_t** device) |
| { |
| ALOGI("adev_open(name=%s)", name); |
| struct submix_audio_device *rsxadev; |
| |
| if (strcmp(name, AUDIO_HARDWARE_INTERFACE) != 0) |
| return -EINVAL; |
| |
| rsxadev = (submix_audio_device*) calloc(1, sizeof(struct submix_audio_device)); |
| if (!rsxadev) |
| return -ENOMEM; |
| |
| rsxadev->device.common.tag = HARDWARE_DEVICE_TAG; |
| rsxadev->device.common.version = AUDIO_DEVICE_API_VERSION_2_0; |
| rsxadev->device.common.module = (struct hw_module_t *) module; |
| rsxadev->device.common.close = adev_close; |
| |
| rsxadev->device.init_check = adev_init_check; |
| rsxadev->device.set_voice_volume = adev_set_voice_volume; |
| rsxadev->device.set_master_volume = adev_set_master_volume; |
| rsxadev->device.get_master_volume = adev_get_master_volume; |
| rsxadev->device.set_master_mute = adev_set_master_mute; |
| rsxadev->device.get_master_mute = adev_get_master_mute; |
| rsxadev->device.set_mode = adev_set_mode; |
| rsxadev->device.set_mic_mute = adev_set_mic_mute; |
| rsxadev->device.get_mic_mute = adev_get_mic_mute; |
| rsxadev->device.set_parameters = adev_set_parameters; |
| rsxadev->device.get_parameters = adev_get_parameters; |
| rsxadev->device.get_input_buffer_size = adev_get_input_buffer_size; |
| rsxadev->device.open_output_stream = adev_open_output_stream; |
| rsxadev->device.close_output_stream = adev_close_output_stream; |
| rsxadev->device.open_input_stream = adev_open_input_stream; |
| rsxadev->device.close_input_stream = adev_close_input_stream; |
| rsxadev->device.dump = adev_dump; |
| |
| for (int i=0 ; i < MAX_ROUTES ; i++) { |
| memset(&rsxadev->routes[i], 0, sizeof(route_config)); |
| strcpy(rsxadev->routes[i].address, ""); |
| } |
| |
| *device = &rsxadev->device.common; |
| |
| return 0; |
| } |
| |
| static struct hw_module_methods_t hal_module_methods = { |
| /* open */ adev_open, |
| }; |
| |
| struct audio_module HAL_MODULE_INFO_SYM = { |
| /* common */ { |
| /* tag */ HARDWARE_MODULE_TAG, |
| /* module_api_version */ AUDIO_MODULE_API_VERSION_0_1, |
| /* hal_api_version */ HARDWARE_HAL_API_VERSION, |
| /* id */ AUDIO_HARDWARE_MODULE_ID, |
| /* name */ "Wifi Display audio HAL", |
| /* author */ "The Android Open Source Project", |
| /* methods */ &hal_module_methods, |
| /* dso */ NULL, |
| /* reserved */ { 0 }, |
| }, |
| }; |
| |
| } //namespace android |
| |
| } //extern "C" |