| /* |
| * Copyright (c) 2013-2021, The Linux Foundation. All rights reserved. |
| * Not a Contribution. |
| * |
| * Copyright (C) 2013 The Android Open Source Project |
| * |
| * Licensed under the Apache License, Version 2.0 (the "License"); |
| * you may not use this file except in compliance with the License. |
| * You may obtain a copy of the License at |
| * |
| * http://www.apache.org/licenses/LICENSE-2.0 |
| * |
| * Unless required by applicable law or agreed to in writing, software |
| * distributed under the License is distributed on an "AS IS" BASIS, |
| * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
| * See the License for the specific language governing permissions and |
| * limitations under the License. |
| * |
| * This file was modified by DTS, Inc. The portions of the |
| * code modified by DTS, Inc are copyrighted and |
| * licensed separately, as follows: |
| * |
| * (C) 2014 DTS, Inc. |
| * |
| * Licensed under the Apache License, Version 2.0 (the "License"); |
| * you may not use this file except in compliance with the License. |
| * You may obtain a copy of the License at |
| * |
| * http://www.apache.org/licenses/LICENSE-2.0 |
| * |
| * Unless required by applicable law or agreed to in writing, software |
| * distributed under the License is distributed on an "AS IS" BASIS, |
| * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
| * See the License for the specific language governing permissions and |
| * limitations under the License. |
| * |
| * Copyright (c) 2023 Qualcomm Innovation Center, Inc. All rights reserved. |
| */ |
| |
| #define LOG_TAG "audio_hw_primary" |
| #define ATRACE_TAG (ATRACE_TAG_AUDIO|ATRACE_TAG_HAL) |
| /*#define LOG_NDEBUG 0*/ |
| /*#define VERY_VERY_VERBOSE_LOGGING*/ |
| #ifdef VERY_VERY_VERBOSE_LOGGING |
| #define ALOGVV ALOGV |
| #else |
| #define ALOGVV(a...) do { } while(0) |
| #endif |
| #include <limits.h> |
| #include <errno.h> |
| #include <pthread.h> |
| #include <stdint.h> |
| #include <sys/time.h> |
| #include <stdlib.h> |
| #include <math.h> |
| #include <dlfcn.h> |
| #include <sys/resource.h> |
| #include <sys/prctl.h> |
| |
| #include <log/log.h> |
| #include <cutils/trace.h> |
| #include <cutils/str_parms.h> |
| #include <cutils/properties.h> |
| #include <cutils/atomic.h> |
| #include <cutils/sched_policy.h> |
| #include <hardware/audio_effect.h> |
| #include <hardware/audio_alsaops.h> |
| #include <system/thread_defs.h> |
| #include <tinyalsa/asoundlib.h> |
| #include <utils/Timers.h> // systemTime |
| #include <audio_effects/effect_aec.h> |
| #include <audio_effects/effect_ns.h> |
| #include <audio_utils/format.h> |
| #include "audio_hw.h" |
| #include "audio_perf.h" |
| #include "platform_api.h" |
| #include <platform.h> |
| #include "audio_extn.h" |
| #include "voice_extn.h" |
| #include "ip_hdlr_intf.h" |
| |
| #include "sound/compress_params.h" |
| |
| #ifdef AUDIO_GKI_ENABLED |
| #include "sound/audio_compressed_formats.h" |
| #endif |
| |
| #include "sound/asound.h" |
| |
| #ifdef DYNAMIC_LOG_ENABLED |
| #include <log_xml_parser.h> |
| #define LOG_MASK HAL_MOD_FILE_AUDIO_HW |
| #include <log_utils.h> |
| #endif |
| |
| #define COMPRESS_OFFLOAD_NUM_FRAGMENTS 4 |
| /*DIRECT PCM has same buffer sizes as DEEP Buffer*/ |
| #define DIRECT_PCM_NUM_FRAGMENTS 2 |
| #define COMPRESS_PLAYBACK_VOLUME_MAX 0x2000 |
| #define VOIP_PLAYBACK_VOLUME_MAX 0x2000 |
| #define MMAP_PLAYBACK_VOLUME_MAX 0x2000 |
| #define PCM_PLAYBACK_VOLUME_MAX 0x2000 |
| #define DSD_VOLUME_MIN_DB (-110) |
| #define INVALID_OUT_VOLUME -1 |
| #define AUDIO_IO_PORTS_MAX 32 |
| |
| #define PLAYBACK_GAIN_MAX 1.0f |
| #define RECORD_GAIN_MIN 0.0f |
| #define RECORD_GAIN_MAX 1.0f |
| #define RECORD_VOLUME_CTL_MAX 0x2000 |
| |
| /* treat as unsigned Q1.13 */ |
| #define APP_TYPE_GAIN_DEFAULT 0x2000 |
| |
| #define PROXY_OPEN_RETRY_COUNT 100 |
| #define PROXY_OPEN_WAIT_TIME 20 |
| |
| #define GET_USECASE_AUDIO_PLAYBACK_PRIMARY(db) \ |
| (db)? USECASE_AUDIO_PLAYBACK_DEEP_BUFFER : \ |
| USECASE_AUDIO_PLAYBACK_LOW_LATENCY |
| #define GET_PCM_CONFIG_AUDIO_PLAYBACK_PRIMARY(db) \ |
| (db)? pcm_config_deep_buffer : pcm_config_low_latency |
| |
| #define ULL_PERIOD_SIZE (DEFAULT_OUTPUT_SAMPLING_RATE/1000) |
| #define DEFAULT_VOIP_BUF_DURATION_MS 20 |
| #define DEFAULT_VOIP_BIT_DEPTH_BYTE sizeof(int16_t) |
| #define DEFAULT_VOIP_SAMP_RATE 48000 |
| |
| #define VOIP_IO_BUF_SIZE(SR, DURATION_MS, BIT_DEPTH) (SR)/1000 * DURATION_MS * BIT_DEPTH |
| |
| struct pcm_config default_pcm_config_voip_copp = { |
| .channels = 1, |
| .rate = DEFAULT_VOIP_SAMP_RATE, /* changed when the stream is opened */ |
| .period_size = VOIP_IO_BUF_SIZE(DEFAULT_VOIP_SAMP_RATE, DEFAULT_VOIP_BUF_DURATION_MS, DEFAULT_VOIP_BIT_DEPTH_BYTE)/2, |
| .period_count = 2, |
| .format = PCM_FORMAT_S16_LE, |
| .avail_min = VOIP_IO_BUF_SIZE(DEFAULT_VOIP_SAMP_RATE, DEFAULT_VOIP_BUF_DURATION_MS, DEFAULT_VOIP_BIT_DEPTH_BYTE)/2, |
| .stop_threshold = INT_MAX, |
| }; |
| |
| #define MIN_CHANNEL_COUNT 1 |
| #define DEFAULT_CHANNEL_COUNT 2 |
| #define MAX_HIFI_CHANNEL_COUNT 8 |
| |
| #ifndef MAX_TARGET_SPECIFIC_CHANNEL_CNT |
| #define MAX_CHANNEL_COUNT 1 |
| #else |
| #define MAX_CHANNEL_COUNT atoi(XSTR(MAX_TARGET_SPECIFIC_CHANNEL_CNT)) |
| #define XSTR(x) STR(x) |
| #define STR(x) #x |
| #endif |
| |
| #ifdef LINUX_ENABLED |
| static inline int64_t audio_utils_ns_from_timespec(const struct timespec *ts) |
| { |
| return ts->tv_sec * 1000000000LL + ts->tv_nsec; |
| } |
| #endif |
| |
| static unsigned int configured_low_latency_capture_period_size = |
| LOW_LATENCY_CAPTURE_PERIOD_SIZE; |
| |
| #define MMAP_PERIOD_SIZE (DEFAULT_OUTPUT_SAMPLING_RATE/1000) |
| #define MMAP_PERIOD_COUNT_MIN 32 |
| #define MMAP_PERIOD_COUNT_MAX 512 |
| #define MMAP_PERIOD_COUNT_DEFAULT (MMAP_PERIOD_COUNT_MAX) |
| |
| /* This constant enables extended precision handling. |
| * TODO The flag is off until more testing is done. |
| */ |
| static const bool k_enable_extended_precision = false; |
| extern int AUDIO_DEVICE_IN_ALL_CODEC_BACKEND; |
| |
| struct pcm_config pcm_config_deep_buffer = { |
| .channels = 2, |
| .rate = DEFAULT_OUTPUT_SAMPLING_RATE, |
| .period_size = DEEP_BUFFER_OUTPUT_PERIOD_SIZE, |
| .period_count = DEEP_BUFFER_OUTPUT_PERIOD_COUNT, |
| .format = PCM_FORMAT_S16_LE, |
| .start_threshold = DEEP_BUFFER_OUTPUT_PERIOD_SIZE / 4, |
| .stop_threshold = INT_MAX, |
| .avail_min = DEEP_BUFFER_OUTPUT_PERIOD_SIZE / 4, |
| }; |
| |
| struct pcm_config pcm_config_low_latency = { |
| .channels = 2, |
| .rate = DEFAULT_OUTPUT_SAMPLING_RATE, |
| .period_size = LOW_LATENCY_OUTPUT_PERIOD_SIZE, |
| .period_count = LOW_LATENCY_OUTPUT_PERIOD_COUNT, |
| .format = PCM_FORMAT_S16_LE, |
| .start_threshold = LOW_LATENCY_OUTPUT_PERIOD_SIZE / 4, |
| .stop_threshold = INT_MAX, |
| .avail_min = LOW_LATENCY_OUTPUT_PERIOD_SIZE / 4, |
| }; |
| |
| struct pcm_config pcm_config_haptics_audio = { |
| .channels = 1, |
| .rate = DEFAULT_OUTPUT_SAMPLING_RATE, |
| .period_size = LOW_LATENCY_OUTPUT_PERIOD_SIZE, |
| .period_count = LOW_LATENCY_OUTPUT_PERIOD_COUNT, |
| .format = PCM_FORMAT_S16_LE, |
| .start_threshold = LOW_LATENCY_OUTPUT_PERIOD_SIZE / 4, |
| .stop_threshold = INT_MAX, |
| .avail_min = LOW_LATENCY_OUTPUT_PERIOD_SIZE / 4, |
| }; |
| |
| struct pcm_config pcm_config_haptics = { |
| .channels = 1, |
| .rate = DEFAULT_OUTPUT_SAMPLING_RATE, |
| .period_size = LOW_LATENCY_OUTPUT_PERIOD_SIZE, |
| .period_count = LOW_LATENCY_OUTPUT_PERIOD_COUNT, |
| .format = PCM_FORMAT_S16_LE, |
| .start_threshold = LOW_LATENCY_OUTPUT_PERIOD_SIZE, |
| .stop_threshold = INT_MAX, |
| .avail_min = LOW_LATENCY_OUTPUT_PERIOD_SIZE / 4, |
| }; |
| |
| static int af_period_multiplier = 4; |
| struct pcm_config pcm_config_rt = { |
| .channels = 2, |
| .rate = DEFAULT_OUTPUT_SAMPLING_RATE, |
| .period_size = ULL_PERIOD_SIZE, //1 ms |
| .period_count = 512, //=> buffer size is 512ms |
| .format = PCM_FORMAT_S16_LE, |
| .start_threshold = ULL_PERIOD_SIZE*8, //8ms |
| .stop_threshold = INT_MAX, |
| .silence_threshold = 0, |
| .silence_size = 0, |
| .avail_min = ULL_PERIOD_SIZE, //1 ms |
| }; |
| |
| struct pcm_config pcm_config_hdmi_multi = { |
| .channels = HDMI_MULTI_DEFAULT_CHANNEL_COUNT, /* changed when the stream is opened */ |
| .rate = DEFAULT_OUTPUT_SAMPLING_RATE, /* changed when the stream is opened */ |
| .period_size = HDMI_MULTI_PERIOD_SIZE, |
| .period_count = HDMI_MULTI_PERIOD_COUNT, |
| .format = PCM_FORMAT_S16_LE, |
| .start_threshold = 0, |
| .stop_threshold = INT_MAX, |
| .avail_min = 0, |
| }; |
| |
| struct pcm_config pcm_config_mmap_playback = { |
| .channels = 2, |
| .rate = DEFAULT_OUTPUT_SAMPLING_RATE, |
| .period_size = MMAP_PERIOD_SIZE, |
| .period_count = MMAP_PERIOD_COUNT_DEFAULT, |
| .format = PCM_FORMAT_S16_LE, |
| .start_threshold = MMAP_PERIOD_SIZE*8, |
| .stop_threshold = INT32_MAX, |
| .silence_threshold = 0, |
| .silence_size = 0, |
| .avail_min = MMAP_PERIOD_SIZE, //1 ms |
| }; |
| |
| struct pcm_config pcm_config_hifi = { |
| .channels = DEFAULT_CHANNEL_COUNT, /* changed when the stream is opened */ |
| .rate = DEFAULT_OUTPUT_SAMPLING_RATE, /* changed when the stream is opened */ |
| .period_size = HIFI_BUFFER_OUTPUT_PERIOD_SIZE, /* change #define */ |
| .period_count = HIFI_BUFFER_OUTPUT_PERIOD_COUNT, |
| .format = PCM_FORMAT_S24_3LE, |
| .start_threshold = 0, |
| .stop_threshold = INT_MAX, |
| .avail_min = 0, |
| }; |
| |
| struct pcm_config pcm_config_audio_capture = { |
| .channels = 2, |
| .period_count = AUDIO_CAPTURE_PERIOD_COUNT, |
| .format = PCM_FORMAT_S16_LE, |
| }; |
| |
| struct pcm_config pcm_config_mmap_capture = { |
| .channels = 2, |
| .rate = DEFAULT_OUTPUT_SAMPLING_RATE, |
| .period_size = MMAP_PERIOD_SIZE, |
| .period_count = MMAP_PERIOD_COUNT_DEFAULT, |
| .format = PCM_FORMAT_S16_LE, |
| .start_threshold = 0, |
| .stop_threshold = INT_MAX, |
| .silence_threshold = 0, |
| .silence_size = 0, |
| .avail_min = MMAP_PERIOD_SIZE, //1 ms |
| }; |
| |
| #define AFE_PROXY_CHANNEL_COUNT 2 |
| #define AFE_PROXY_SAMPLING_RATE 48000 |
| |
| #define AFE_PROXY_PLAYBACK_PERIOD_SIZE 768 |
| #define AFE_PROXY_PLAYBACK_PERIOD_COUNT 4 |
| |
| struct pcm_config pcm_config_afe_proxy_playback = { |
| .channels = AFE_PROXY_CHANNEL_COUNT, |
| .rate = AFE_PROXY_SAMPLING_RATE, |
| .period_size = AFE_PROXY_PLAYBACK_PERIOD_SIZE, |
| .period_count = AFE_PROXY_PLAYBACK_PERIOD_COUNT, |
| .format = PCM_FORMAT_S16_LE, |
| .start_threshold = AFE_PROXY_PLAYBACK_PERIOD_SIZE, |
| .stop_threshold = INT_MAX, |
| .avail_min = AFE_PROXY_PLAYBACK_PERIOD_SIZE, |
| }; |
| |
| #define AFE_PROXY_RECORD_PERIOD_SIZE 768 |
| #define AFE_PROXY_RECORD_PERIOD_COUNT 4 |
| |
| struct pcm_config pcm_config_audio_capture_rt = { |
| .channels = 2, |
| .rate = DEFAULT_OUTPUT_SAMPLING_RATE, |
| .period_size = ULL_PERIOD_SIZE, |
| .period_count = 512, |
| .format = PCM_FORMAT_S16_LE, |
| .start_threshold = 0, |
| .stop_threshold = AFE_PROXY_RECORD_PERIOD_SIZE * AFE_PROXY_RECORD_PERIOD_COUNT, |
| .silence_threshold = 0, |
| .silence_size = 0, |
| .avail_min = ULL_PERIOD_SIZE, //1 ms |
| }; |
| |
| struct pcm_config pcm_config_afe_proxy_record = { |
| .channels = AFE_PROXY_CHANNEL_COUNT, |
| .rate = AFE_PROXY_SAMPLING_RATE, |
| .period_size = AFE_PROXY_RECORD_PERIOD_SIZE, |
| .period_count = AFE_PROXY_RECORD_PERIOD_COUNT, |
| .format = PCM_FORMAT_S16_LE, |
| .start_threshold = AFE_PROXY_RECORD_PERIOD_SIZE, |
| .stop_threshold = INT_MAX, |
| .avail_min = AFE_PROXY_RECORD_PERIOD_SIZE, |
| }; |
| |
| #define AUDIO_MAX_PCM_FORMATS 7 |
| |
| const uint32_t format_to_bitwidth_table[AUDIO_MAX_PCM_FORMATS] = { |
| [AUDIO_FORMAT_DEFAULT] = 0, |
| [AUDIO_FORMAT_PCM_16_BIT] = sizeof(uint16_t), |
| [AUDIO_FORMAT_PCM_8_BIT] = sizeof(uint8_t), |
| [AUDIO_FORMAT_PCM_32_BIT] = sizeof(uint32_t), |
| [AUDIO_FORMAT_PCM_8_24_BIT] = sizeof(uint32_t), |
| [AUDIO_FORMAT_PCM_FLOAT] = sizeof(float), |
| [AUDIO_FORMAT_PCM_24_BIT_PACKED] = sizeof(uint8_t) * 3, |
| }; |
| |
| const char * const use_case_table[AUDIO_USECASE_MAX] = { |
| [USECASE_AUDIO_PLAYBACK_DEEP_BUFFER] = "deep-buffer-playback", |
| [USECASE_AUDIO_PLAYBACK_LOW_LATENCY] = "low-latency-playback", |
| [USECASE_AUDIO_PLAYBACK_WITH_HAPTICS] = "audio-with-haptics-playback", |
| [USECASE_AUDIO_PLAYBACK_HAPTICS] = "haptics-playback", |
| [USECASE_AUDIO_PLAYBACK_ULL] = "audio-ull-playback", |
| [USECASE_AUDIO_PLAYBACK_MULTI_CH] = "multi-channel-playback", |
| [USECASE_AUDIO_PLAYBACK_OFFLOAD] = "compress-offload-playback", |
| //Enabled for Direct_PCM |
| [USECASE_AUDIO_PLAYBACK_OFFLOAD2] = "compress-offload-playback2", |
| [USECASE_AUDIO_PLAYBACK_OFFLOAD3] = "compress-offload-playback3", |
| [USECASE_AUDIO_PLAYBACK_OFFLOAD4] = "compress-offload-playback4", |
| [USECASE_AUDIO_PLAYBACK_OFFLOAD5] = "compress-offload-playback5", |
| [USECASE_AUDIO_PLAYBACK_OFFLOAD6] = "compress-offload-playback6", |
| [USECASE_AUDIO_PLAYBACK_OFFLOAD7] = "compress-offload-playback7", |
| [USECASE_AUDIO_PLAYBACK_OFFLOAD8] = "compress-offload-playback8", |
| [USECASE_AUDIO_PLAYBACK_OFFLOAD9] = "compress-offload-playback9", |
| [USECASE_AUDIO_PLAYBACK_FM] = "play-fm", |
| [USECASE_AUDIO_PLAYBACK_MMAP] = "mmap-playback", |
| [USECASE_AUDIO_PLAYBACK_HIFI] = "hifi-playback", |
| [USECASE_AUDIO_PLAYBACK_TTS] = "audio-tts-playback", |
| |
| [USECASE_AUDIO_RECORD] = "audio-record", |
| [USECASE_AUDIO_RECORD_COMPRESS] = "audio-record-compress", |
| [USECASE_AUDIO_RECORD_COMPRESS2] = "audio-record-compress2", |
| [USECASE_AUDIO_RECORD_COMPRESS3] = "audio-record-compress3", |
| [USECASE_AUDIO_RECORD_COMPRESS4] = "audio-record-compress4", |
| [USECASE_AUDIO_RECORD_COMPRESS5] = "audio-record-compress5", |
| [USECASE_AUDIO_RECORD_COMPRESS6] = "audio-record-compress6", |
| [USECASE_AUDIO_RECORD_LOW_LATENCY] = "low-latency-record", |
| [USECASE_AUDIO_RECORD_FM_VIRTUAL] = "fm-virtual-record", |
| [USECASE_AUDIO_RECORD_MMAP] = "mmap-record", |
| [USECASE_AUDIO_RECORD_HIFI] = "hifi-record", |
| |
| [USECASE_AUDIO_HFP_SCO] = "hfp-sco", |
| [USECASE_AUDIO_HFP_SCO_WB] = "hfp-sco-wb", |
| [USECASE_AUDIO_HFP_SCO_DOWNLINK] = "hfp-sco-downlink", |
| [USECASE_AUDIO_HFP_SCO_WB_DOWNLINK] = "hfp-sco-wb-downlink", |
| |
| [USECASE_VOICE_CALL] = "voice-call", |
| [USECASE_VOICE2_CALL] = "voice2-call", |
| [USECASE_VOLTE_CALL] = "volte-call", |
| [USECASE_QCHAT_CALL] = "qchat-call", |
| [USECASE_VOWLAN_CALL] = "vowlan-call", |
| [USECASE_VOICEMMODE1_CALL] = "voicemmode1-call", |
| [USECASE_VOICEMMODE2_CALL] = "voicemmode2-call", |
| [USECASE_COMPRESS_VOIP_CALL] = "compress-voip-call", |
| [USECASE_INCALL_REC_UPLINK] = "incall-rec-uplink", |
| [USECASE_INCALL_REC_DOWNLINK] = "incall-rec-downlink", |
| [USECASE_INCALL_REC_UPLINK_AND_DOWNLINK] = "incall-rec-uplink-and-downlink", |
| [USECASE_INCALL_REC_UPLINK_COMPRESS] = "incall-rec-uplink-compress", |
| [USECASE_INCALL_REC_DOWNLINK_COMPRESS] = "incall-rec-downlink-compress", |
| [USECASE_INCALL_REC_UPLINK_AND_DOWNLINK_COMPRESS] = "incall-rec-uplink-and-downlink-compress", |
| |
| [USECASE_INCALL_MUSIC_UPLINK] = "incall_music_uplink", |
| [USECASE_INCALL_MUSIC_UPLINK2] = "incall_music_uplink2", |
| [USECASE_AUDIO_SPKR_CALIB_RX] = "spkr-rx-calib", |
| [USECASE_AUDIO_SPKR_CALIB_TX] = "spkr-vi-record", |
| |
| [USECASE_AUDIO_PLAYBACK_AFE_PROXY] = "afe-proxy-playback", |
| [USECASE_AUDIO_RECORD_AFE_PROXY] = "afe-proxy-record", |
| [USECASE_AUDIO_RECORD_AFE_PROXY2] = "afe-proxy-record2", |
| [USECASE_AUDIO_PLAYBACK_SILENCE] = "silence-playback", |
| |
| /* Transcode loopback cases */ |
| [USECASE_AUDIO_TRANSCODE_LOOPBACK_RX] = "audio-transcode-loopback-rx", |
| [USECASE_AUDIO_TRANSCODE_LOOPBACK_TX] = "audio-transcode-loopback-tx", |
| |
| [USECASE_AUDIO_PLAYBACK_VOIP] = "audio-playback-voip", |
| [USECASE_AUDIO_RECORD_VOIP] = "audio-record-voip", |
| [USECASE_AUDIO_RECORD_VOIP_LOW_LATENCY] = "audio-record-voip-low-latency", |
| /* For Interactive Audio Streams */ |
| [USECASE_AUDIO_PLAYBACK_INTERACTIVE_STREAM1] = "audio-interactive-stream1", |
| [USECASE_AUDIO_PLAYBACK_INTERACTIVE_STREAM2] = "audio-interactive-stream2", |
| [USECASE_AUDIO_PLAYBACK_INTERACTIVE_STREAM3] = "audio-interactive-stream3", |
| [USECASE_AUDIO_PLAYBACK_INTERACTIVE_STREAM4] = "audio-interactive-stream4", |
| [USECASE_AUDIO_PLAYBACK_INTERACTIVE_STREAM5] = "audio-interactive-stream5", |
| [USECASE_AUDIO_PLAYBACK_INTERACTIVE_STREAM6] = "audio-interactive-stream6", |
| [USECASE_AUDIO_PLAYBACK_INTERACTIVE_STREAM7] = "audio-interactive-stream7", |
| [USECASE_AUDIO_PLAYBACK_INTERACTIVE_STREAM8] = "audio-interactive-stream8", |
| |
| [USECASE_AUDIO_EC_REF_LOOPBACK] = "ec-ref-audio-capture", |
| |
| [USECASE_AUDIO_A2DP_ABR_FEEDBACK] = "a2dp-abr-feedback", |
| |
| [USECASE_AUDIO_PLAYBACK_MEDIA] = "media-playback", |
| [USECASE_AUDIO_PLAYBACK_SYS_NOTIFICATION] = "sys-notification-playback", |
| [USECASE_AUDIO_PLAYBACK_NAV_GUIDANCE] = "nav-guidance-playback", |
| [USECASE_AUDIO_PLAYBACK_PHONE] = "phone-playback", |
| [USECASE_AUDIO_PLAYBACK_FRONT_PASSENGER] = "front-passenger-playback", |
| [USECASE_AUDIO_PLAYBACK_REAR_SEAT] = "rear-seat-playback", |
| [USECASE_AUDIO_FM_TUNER_EXT] = "fm-tuner-ext", |
| [USECASE_ICC_CALL] = "icc-call", |
| |
| [USECASE_AUDIO_RECORD_BUS] = "audio-record", |
| [USECASE_AUDIO_RECORD_BUS_FRONT_PASSENGER] = "front-passenger-record", |
| [USECASE_AUDIO_RECORD_BUS_REAR_SEAT] = "rear-seat-record", |
| [USECASE_AUDIO_PLAYBACK_SYNTHESIZER] = "synth-loopback", |
| [USECASE_AUDIO_RECORD_ECHO_REF_EXT] = "echo-reference-external", |
| }; |
| |
| static const audio_usecase_t offload_usecases[] = { |
| USECASE_AUDIO_PLAYBACK_OFFLOAD, |
| USECASE_AUDIO_PLAYBACK_OFFLOAD2, |
| USECASE_AUDIO_PLAYBACK_OFFLOAD3, |
| USECASE_AUDIO_PLAYBACK_OFFLOAD4, |
| USECASE_AUDIO_PLAYBACK_OFFLOAD5, |
| USECASE_AUDIO_PLAYBACK_OFFLOAD6, |
| USECASE_AUDIO_PLAYBACK_OFFLOAD7, |
| USECASE_AUDIO_PLAYBACK_OFFLOAD8, |
| USECASE_AUDIO_PLAYBACK_OFFLOAD9, |
| }; |
| |
| static const audio_usecase_t interactive_usecases[] = { |
| USECASE_AUDIO_PLAYBACK_INTERACTIVE_STREAM1, |
| USECASE_AUDIO_PLAYBACK_INTERACTIVE_STREAM2, |
| USECASE_AUDIO_PLAYBACK_INTERACTIVE_STREAM3, |
| USECASE_AUDIO_PLAYBACK_INTERACTIVE_STREAM4, |
| USECASE_AUDIO_PLAYBACK_INTERACTIVE_STREAM5, |
| USECASE_AUDIO_PLAYBACK_INTERACTIVE_STREAM6, |
| USECASE_AUDIO_PLAYBACK_INTERACTIVE_STREAM7, |
| USECASE_AUDIO_PLAYBACK_INTERACTIVE_STREAM8, |
| }; |
| |
| #define STRING_TO_ENUM(string) { #string, string } |
| |
| struct string_to_enum { |
| const char *name; |
| uint32_t value; |
| }; |
| |
| static const struct string_to_enum channels_name_to_enum_table[] = { |
| STRING_TO_ENUM(AUDIO_CHANNEL_OUT_STEREO), |
| STRING_TO_ENUM(AUDIO_CHANNEL_OUT_2POINT1), |
| STRING_TO_ENUM(AUDIO_CHANNEL_OUT_QUAD), |
| STRING_TO_ENUM(AUDIO_CHANNEL_OUT_SURROUND), |
| STRING_TO_ENUM(AUDIO_CHANNEL_OUT_PENTA), |
| STRING_TO_ENUM(AUDIO_CHANNEL_OUT_5POINT1), |
| STRING_TO_ENUM(AUDIO_CHANNEL_OUT_6POINT1), |
| STRING_TO_ENUM(AUDIO_CHANNEL_OUT_7POINT1), |
| STRING_TO_ENUM(AUDIO_CHANNEL_IN_MONO), |
| STRING_TO_ENUM(AUDIO_CHANNEL_IN_STEREO), |
| STRING_TO_ENUM(AUDIO_CHANNEL_IN_FRONT_BACK), |
| STRING_TO_ENUM(AUDIO_CHANNEL_INDEX_MASK_1), |
| STRING_TO_ENUM(AUDIO_CHANNEL_INDEX_MASK_2), |
| STRING_TO_ENUM(AUDIO_CHANNEL_INDEX_MASK_3), |
| STRING_TO_ENUM(AUDIO_CHANNEL_INDEX_MASK_4), |
| STRING_TO_ENUM(AUDIO_CHANNEL_INDEX_MASK_5), |
| STRING_TO_ENUM(AUDIO_CHANNEL_INDEX_MASK_6), |
| STRING_TO_ENUM(AUDIO_CHANNEL_INDEX_MASK_7), |
| STRING_TO_ENUM(AUDIO_CHANNEL_INDEX_MASK_8), |
| }; |
| |
| static const struct string_to_enum formats_name_to_enum_table[] = { |
| STRING_TO_ENUM(AUDIO_FORMAT_PCM_16_BIT), |
| STRING_TO_ENUM(AUDIO_FORMAT_PCM_24_BIT_PACKED), |
| STRING_TO_ENUM(AUDIO_FORMAT_PCM_32_BIT), |
| STRING_TO_ENUM(AUDIO_FORMAT_AC3), |
| STRING_TO_ENUM(AUDIO_FORMAT_E_AC3), |
| STRING_TO_ENUM(AUDIO_FORMAT_E_AC3_JOC), |
| STRING_TO_ENUM(AUDIO_FORMAT_DOLBY_TRUEHD), |
| STRING_TO_ENUM(AUDIO_FORMAT_DTS), |
| STRING_TO_ENUM(AUDIO_FORMAT_DTS_HD), |
| STRING_TO_ENUM(AUDIO_FORMAT_IEC61937) |
| }; |
| |
| //list of all supported sample rates by HDMI specification. |
| static const int out_hdmi_sample_rates[] = { |
| 32000, 44100, 48000, 88200, 96000, 176400, 192000, |
| }; |
| |
| static const struct string_to_enum out_sample_rates_name_to_enum_table[] = { |
| STRING_TO_ENUM(32000), |
| STRING_TO_ENUM(44100), |
| STRING_TO_ENUM(48000), |
| STRING_TO_ENUM(88200), |
| STRING_TO_ENUM(96000), |
| STRING_TO_ENUM(176400), |
| STRING_TO_ENUM(192000), |
| STRING_TO_ENUM(352800), |
| STRING_TO_ENUM(384000), |
| }; |
| |
| struct in_effect_list { |
| struct listnode list; |
| effect_handle_t handle; |
| }; |
| |
| static struct audio_device *adev = NULL; |
| static pthread_mutex_t adev_init_lock = PTHREAD_MUTEX_INITIALIZER; |
| static unsigned int audio_device_ref_count; |
| //cache last MBDRC cal step level |
| static int last_known_cal_step = -1 ; |
| |
| static int out_set_compr_volume(struct audio_stream_out *stream, float left, float right); |
| static int out_set_mmap_volume(struct audio_stream_out *stream, float left, float right); |
| static int out_set_voip_volume(struct audio_stream_out *stream, float left, float right); |
| static int out_set_pcm_volume(struct audio_stream_out *stream, float left, float right); |
| |
| static void adev_snd_mon_cb(void *cookie, struct str_parms *parms); |
| static void in_snd_mon_cb(void * stream, struct str_parms * parms); |
| static void out_snd_mon_cb(void * stream, struct str_parms * parms); |
| |
| static int configure_btsco_sample_rate(snd_device_t snd_device); |
| |
| #ifdef AUDIO_FEATURE_ENABLED_GCOV |
| extern void __gcov_flush(); |
| static void enable_gcov() |
| { |
| __gcov_flush(); |
| } |
| #else |
| static void enable_gcov() |
| { |
| } |
| #endif |
| |
| static int in_set_microphone_direction(const struct audio_stream_in *stream, |
| audio_microphone_direction_t dir); |
| static int in_set_microphone_field_dimension(const struct audio_stream_in *stream, float zoom); |
| |
| static bool may_use_noirq_mode(struct audio_device *adev, audio_usecase_t uc_id, |
| int flags __unused) |
| { |
| int dir = 0; |
| switch (uc_id) { |
| case USECASE_AUDIO_RECORD_LOW_LATENCY: |
| case USECASE_AUDIO_RECORD_VOIP_LOW_LATENCY: |
| dir = 1; |
| case USECASE_AUDIO_PLAYBACK_ULL: |
| break; |
| default: |
| return false; |
| } |
| |
| int dev_id = platform_get_pcm_device_id(uc_id, dir == 0 ? |
| PCM_PLAYBACK : PCM_CAPTURE); |
| if (adev->adm_is_noirq_avail) |
| return adev->adm_is_noirq_avail(adev->adm_data, |
| adev->snd_card, dev_id, dir); |
| return false; |
| } |
| |
| static void register_out_stream(struct stream_out *out) |
| { |
| struct audio_device *adev = out->dev; |
| if (is_offload_usecase(out->usecase) || |
| !adev->adm_register_output_stream) |
| return; |
| |
| // register stream first for backward compatibility |
| adev->adm_register_output_stream(adev->adm_data, |
| out->handle, |
| out->flags); |
| |
| if (!adev->adm_set_config) |
| return; |
| |
| if (out->realtime) |
| adev->adm_set_config(adev->adm_data, |
| out->handle, |
| out->pcm, &out->config); |
| } |
| |
| static void register_in_stream(struct stream_in *in) |
| { |
| struct audio_device *adev = in->dev; |
| if (!adev->adm_register_input_stream) |
| return; |
| |
| adev->adm_register_input_stream(adev->adm_data, |
| in->capture_handle, |
| in->flags); |
| |
| if (!adev->adm_set_config) |
| return; |
| |
| if (in->realtime) |
| adev->adm_set_config(adev->adm_data, |
| in->capture_handle, |
| in->pcm, |
| &in->config); |
| } |
| |
| static void request_out_focus(struct stream_out *out, long ns) |
| { |
| struct audio_device *adev = out->dev; |
| |
| if (adev->adm_request_focus_v2) |
| adev->adm_request_focus_v2(adev->adm_data, out->handle, ns); |
| else if (adev->adm_request_focus) |
| adev->adm_request_focus(adev->adm_data, out->handle); |
| } |
| |
| static int request_in_focus(struct stream_in *in, long ns) |
| { |
| struct audio_device *adev = in->dev; |
| int ret = 0; |
| |
| if (adev->adm_request_focus_v2_1) |
| ret = adev->adm_request_focus_v2_1(adev->adm_data, in->capture_handle, ns); |
| else if (adev->adm_request_focus_v2) |
| adev->adm_request_focus_v2(adev->adm_data, in->capture_handle, ns); |
| else if (adev->adm_request_focus) |
| adev->adm_request_focus(adev->adm_data, in->capture_handle); |
| |
| return ret; |
| } |
| |
| static void release_out_focus(struct stream_out *out) |
| { |
| struct audio_device *adev = out->dev; |
| |
| if (adev->adm_abandon_focus) |
| adev->adm_abandon_focus(adev->adm_data, out->handle); |
| } |
| |
| static void release_in_focus(struct stream_in *in) |
| { |
| struct audio_device *adev = in->dev; |
| if (adev->adm_abandon_focus) |
| adev->adm_abandon_focus(adev->adm_data, in->capture_handle); |
| } |
| |
| static int parse_snd_card_status(struct str_parms *parms, int *card, |
| card_status_t *status) |
| { |
| char value[32]={0}; |
| char state[32]={0}; |
| |
| int ret = str_parms_get_str(parms, "SND_CARD_STATUS", value, sizeof(value)); |
| if (ret < 0) |
| return -1; |
| |
| // sscanf should be okay as value is of max length 32. |
| // same as sizeof state. |
| if (sscanf(value, "%d,%s", card, state) < 2) |
| return -1; |
| |
| *status = !strcmp(state, "ONLINE") ? CARD_STATUS_ONLINE : |
| CARD_STATUS_OFFLINE; |
| return 0; |
| } |
| |
| static inline void adjust_frames_for_device_delay(struct stream_out *out, |
| uint32_t *dsp_frames) { |
| // Adjustment accounts for A2dp encoder latency with offload usecases |
| // Note: Encoder latency is returned in ms. |
| if (is_a2dp_out_device_type(&out->device_list)) { |
| unsigned long offset = |
| (audio_extn_a2dp_get_encoder_latency() * out->sample_rate / 1000); |
| *dsp_frames = (*dsp_frames > offset) ? (*dsp_frames - offset) : 0; |
| } |
| } |
| |
| static inline bool free_entry(void *key __unused, |
| void *value, void *context __unused) |
| { |
| free(value); |
| return true; |
| } |
| |
| static inline void free_map(Hashmap *map) |
| { |
| if (map) { |
| hashmapForEach(map, free_entry, (void *) NULL); |
| hashmapFree(map); |
| } |
| } |
| |
| static inline void patch_map_remove_l(struct audio_device *adev, |
| audio_patch_handle_t patch_handle) |
| { |
| if (patch_handle == AUDIO_PATCH_HANDLE_NONE) |
| return; |
| |
| struct audio_patch_info *p_info = |
| hashmapGet(adev->patch_map, (void *) (intptr_t) patch_handle); |
| if (p_info) { |
| ALOGV("%s: Remove patch %d", __func__, patch_handle); |
| hashmapRemove(adev->patch_map, (void *) (intptr_t) patch_handle); |
| free(p_info->patch); |
| free(p_info); |
| } |
| } |
| |
| static inline int io_streams_map_insert(struct audio_device *adev, |
| struct audio_stream *stream, |
| audio_io_handle_t handle, |
| audio_patch_handle_t patch_handle) |
| { |
| struct audio_stream_info *s_info = |
| (struct audio_stream_info *) calloc(1, sizeof(struct audio_stream_info)); |
| |
| if (s_info == NULL) { |
| ALOGE("%s: Could not allocate stream info", __func__); |
| return -ENOMEM; |
| } |
| s_info->stream = stream; |
| s_info->patch_handle = patch_handle; |
| |
| pthread_mutex_lock(&adev->lock); |
| struct audio_stream_info *stream_info = |
| hashmapPut(adev->io_streams_map, (void *) (intptr_t) handle, (void *) s_info); |
| if (stream_info != NULL) |
| free(stream_info); |
| pthread_mutex_unlock(&adev->lock); |
| ALOGV("%s: Added stream in io_streams_map with handle %d", __func__, handle); |
| return 0; |
| } |
| |
| static inline void io_streams_map_remove(struct audio_device *adev, |
| audio_io_handle_t handle) |
| { |
| pthread_mutex_lock(&adev->lock); |
| struct audio_stream_info *s_info = |
| hashmapRemove(adev->io_streams_map, (void *) (intptr_t) handle); |
| if (s_info == NULL) |
| goto done; |
| ALOGV("%s: Removed stream with handle %d", __func__, handle); |
| patch_map_remove_l(adev, s_info->patch_handle); |
| free(s_info); |
| done: |
| pthread_mutex_unlock(&adev->lock); |
| return; |
| } |
| |
| static struct audio_patch_info* fetch_patch_info_l(struct audio_device *adev, |
| audio_patch_handle_t handle) |
| { |
| struct audio_patch_info *p_info = NULL; |
| p_info = (struct audio_patch_info *) |
| hashmapGet(adev->patch_map, (void *) (intptr_t) handle); |
| return p_info; |
| } |
| |
| __attribute__ ((visibility ("default"))) |
| bool audio_hw_send_gain_dep_calibration(int level) { |
| bool ret_val = false; |
| ALOGV("%s: called ...", __func__); |
| |
| pthread_mutex_lock(&adev_init_lock); |
| |
| if (adev != NULL && adev->platform != NULL) { |
| pthread_mutex_lock(&adev->lock); |
| ret_val = platform_send_gain_dep_cal(adev->platform, level); |
| |
| // cache level info for any of the use case which |
| // was not started. |
| last_known_cal_step = level;; |
| |
| pthread_mutex_unlock(&adev->lock); |
| } else { |
| ALOGE("%s: %s is NULL", __func__, adev == NULL ? "adev" : "adev->platform"); |
| } |
| |
| pthread_mutex_unlock(&adev_init_lock); |
| |
| return ret_val; |
| } |
| |
| static int check_and_set_gapless_mode(struct audio_device *adev, bool enable_gapless) |
| { |
| bool gapless_enabled = false; |
| const char *mixer_ctl_name = "Compress Gapless Playback"; |
| struct mixer_ctl *ctl; |
| |
| ALOGV("%s:", __func__); |
| gapless_enabled = property_get_bool("vendor.audio.offload.gapless.enabled", false); |
| |
| /*Disable gapless if its AV playback*/ |
| gapless_enabled = gapless_enabled && enable_gapless; |
| |
| ctl = mixer_get_ctl_by_name(adev->mixer, mixer_ctl_name); |
| if (!ctl) { |
| ALOGE("%s: Could not get ctl for mixer cmd - %s", |
| __func__, mixer_ctl_name); |
| return -EINVAL; |
| } |
| |
| if (mixer_ctl_set_value(ctl, 0, gapless_enabled) < 0) { |
| ALOGE("%s: Could not set gapless mode %d", |
| __func__, gapless_enabled); |
| return -EINVAL; |
| } |
| return 0; |
| } |
| |
| __attribute__ ((visibility ("default"))) |
| int audio_hw_get_gain_level_mapping(struct amp_db_and_gain_table *mapping_tbl, |
| int table_size) { |
| int ret_val = 0; |
| ALOGV("%s: enter ... ", __func__); |
| |
| pthread_mutex_lock(&adev_init_lock); |
| if (adev == NULL) { |
| ALOGW("%s: adev is NULL .... ", __func__); |
| goto done; |
| } |
| |
| pthread_mutex_lock(&adev->lock); |
| ret_val = platform_get_gain_level_mapping(mapping_tbl, table_size); |
| pthread_mutex_unlock(&adev->lock); |
| done: |
| pthread_mutex_unlock(&adev_init_lock); |
| ALOGV("%s: exit ... ", __func__); |
| return ret_val; |
| } |
| |
| bool audio_hw_send_qdsp_parameter(int stream_type, float vol, bool active) |
| { |
| bool ret = false; |
| ALOGV("%s: enter ...", __func__); |
| |
| pthread_mutex_lock(&adev_init_lock); |
| |
| if (adev != NULL && adev->platform != NULL) { |
| pthread_mutex_lock(&adev->lock); |
| ret = audio_extn_qdsp_set_state(adev, stream_type, vol, active); |
| pthread_mutex_unlock(&adev->lock); |
| } |
| |
| pthread_mutex_unlock(&adev_init_lock); |
| |
| ALOGV("%s: exit with ret %d", __func__, ret); |
| return ret; |
| } |
| |
| static bool is_supported_format(audio_format_t format) |
| { |
| if (format == AUDIO_FORMAT_MP3 || |
| format == AUDIO_FORMAT_MP2 || |
| format == AUDIO_FORMAT_AAC_LC || |
| format == AUDIO_FORMAT_AAC_HE_V1 || |
| format == AUDIO_FORMAT_AAC_HE_V2 || |
| format == AUDIO_FORMAT_AAC_ADTS_LC || |
| format == AUDIO_FORMAT_AAC_ADTS_HE_V1 || |
| format == AUDIO_FORMAT_AAC_ADTS_HE_V2 || |
| format == AUDIO_FORMAT_AAC_LATM_LC || |
| format == AUDIO_FORMAT_AAC_LATM_HE_V1 || |
| format == AUDIO_FORMAT_AAC_LATM_HE_V2 || |
| format == AUDIO_FORMAT_PCM_24_BIT_PACKED || |
| format == AUDIO_FORMAT_PCM_8_24_BIT || |
| format == AUDIO_FORMAT_PCM_FLOAT || |
| format == AUDIO_FORMAT_PCM_32_BIT || |
| format == AUDIO_FORMAT_PCM_16_BIT || |
| format == AUDIO_FORMAT_AC3 || |
| format == AUDIO_FORMAT_E_AC3 || |
| format == AUDIO_FORMAT_DOLBY_TRUEHD || |
| format == AUDIO_FORMAT_DTS || |
| format == AUDIO_FORMAT_DTS_HD || |
| format == AUDIO_FORMAT_FLAC || |
| format == AUDIO_FORMAT_ALAC || |
| format == AUDIO_FORMAT_APE || |
| format == AUDIO_FORMAT_DSD || |
| format == AUDIO_FORMAT_VORBIS || |
| format == AUDIO_FORMAT_WMA || |
| format == AUDIO_FORMAT_WMA_PRO || |
| format == AUDIO_FORMAT_APTX || |
| format == AUDIO_FORMAT_IEC61937) |
| return true; |
| |
| return false; |
| } |
| |
| static bool is_supported_conc_usecase_for_power_mode_call(struct audio_device *adev) |
| { |
| struct listnode *node; |
| struct audio_usecase *usecase; |
| |
| list_for_each(node, &adev->usecase_list) { |
| usecase = node_to_item(node, struct audio_usecase, list); |
| if (usecase->id == USECASE_AUDIO_PLAYBACK_FM) { |
| ALOGD("%s: FM usecase is active, not setting power mode", __func__); |
| return false; |
| } |
| } |
| |
| return true; |
| } |
| static inline bool is_mmap_usecase(audio_usecase_t uc_id) |
| { |
| return (uc_id == USECASE_AUDIO_RECORD_AFE_PROXY) || |
| (uc_id == USECASE_AUDIO_RECORD_AFE_PROXY2) || |
| (uc_id == USECASE_AUDIO_PLAYBACK_AFE_PROXY); |
| } |
| |
| static inline bool is_valid_volume(float left, float right) |
| { |
| return ((left >= 0.0f && right >= 0.0f) ? true : false); |
| } |
| |
| static void enable_asrc_mode(struct audio_device *adev) |
| { |
| ALOGV("%s", __func__); |
| audio_route_apply_and_update_path(adev->audio_route, |
| "asrc-mode"); |
| adev->asrc_mode_enabled = true; |
| } |
| |
| static void disable_asrc_mode(struct audio_device *adev) |
| { |
| ALOGV("%s", __func__); |
| audio_route_reset_and_update_path(adev->audio_route, |
| "asrc-mode"); |
| adev->asrc_mode_enabled = false; |
| } |
| |
| static void check_and_configure_headphone(struct audio_device *adev, |
| struct audio_usecase *uc_info, |
| snd_device_t snd_device) |
| { |
| struct listnode *node; |
| struct audio_usecase *usecase; |
| int new_backend_idx, usecase_backend_idx; |
| bool spkr_hph_single_be_native_concurrency; |
| |
| new_backend_idx = platform_get_backend_index(snd_device); |
| spkr_hph_single_be_native_concurrency = platform_get_spkr_hph_single_be_native_concurrency_flag(); |
| if ((spkr_hph_single_be_native_concurrency && (new_backend_idx == DEFAULT_CODEC_BACKEND)) || |
| uc_info->id == USECASE_AUDIO_PLAYBACK_WITH_HAPTICS) { |
| list_for_each(node, &adev->usecase_list) { |
| usecase = node_to_item(node, struct audio_usecase, list); |
| if ((usecase->type != PCM_CAPTURE) && (usecase != uc_info)) { |
| usecase_backend_idx = platform_get_backend_index(usecase->out_snd_device); |
| if (((usecase_backend_idx == HEADPHONE_BACKEND) || |
| (usecase_backend_idx == HEADPHONE_44_1_BACKEND)) && |
| ((usecase->stream.out->sample_rate % OUTPUT_SAMPLING_RATE_44100) == 0)) { |
| disable_audio_route(adev, usecase); |
| disable_snd_device(adev, usecase->out_snd_device); |
| usecase->stream.out->sample_rate = DEFAULT_OUTPUT_SAMPLING_RATE; |
| platform_check_and_set_codec_backend_cfg(adev, usecase, |
| usecase->out_snd_device); |
| enable_snd_device(adev, usecase->out_snd_device); |
| enable_audio_route(adev, usecase); |
| } |
| } |
| else if ((usecase->type != PCM_CAPTURE) && (usecase == uc_info)) { |
| usecase_backend_idx = platform_get_backend_index(usecase->out_snd_device); |
| if (((usecase_backend_idx == HEADPHONE_BACKEND) || |
| (usecase_backend_idx == HEADPHONE_44_1_BACKEND)) && |
| ((usecase->stream.out->sample_rate % OUTPUT_SAMPLING_RATE_44100) == 0)) { |
| usecase->stream.out->sample_rate = DEFAULT_OUTPUT_SAMPLING_RATE; |
| platform_check_and_set_codec_backend_cfg(adev, usecase, |
| usecase->out_snd_device); |
| } |
| } |
| } |
| } |
| } |
| |
| /* |
| * - Enable ASRC mode for incoming mix path use case(Headphone backend)if Headphone |
| * 44.1 or Native DSD backends are enabled for any of current use case. |
| * e.g. 48-> + (Naitve DSD or Headphone 44.1) |
| * - Disable current mix path use case(Headphone backend) and re-enable it with |
| * ASRC mode for incoming Headphone 44.1 or Native DSD use case. |
| * e.g. Naitve DSD or Headphone 44.1 -> + 48 |
| */ |
| static void check_and_set_asrc_mode(struct audio_device *adev, |
| struct audio_usecase *uc_info, |
| snd_device_t snd_device) |
| { |
| ALOGV("%s snd device %d", __func__, snd_device); |
| int i, num_new_devices = 0; |
| snd_device_t split_new_snd_devices[SND_DEVICE_OUT_END]; |
| /* |
| *Split snd device for new combo use case |
| *e.g. Headphopne 44.1-> + Ringtone (Headphone + Speaker) |
| */ |
| if (platform_split_snd_device(adev->platform, |
| snd_device, |
| &num_new_devices, |
| split_new_snd_devices) == 0) { |
| for (i = 0; i < num_new_devices; i++) |
| check_and_set_asrc_mode(adev, uc_info, split_new_snd_devices[i]); |
| } else { |
| int new_backend_idx = platform_get_backend_index(snd_device); |
| if (((new_backend_idx == HEADPHONE_BACKEND) || |
| (new_backend_idx == HEADPHONE_44_1_BACKEND) || |
| (new_backend_idx == DSD_NATIVE_BACKEND)) && |
| !adev->asrc_mode_enabled) { |
| struct listnode *node = NULL; |
| struct audio_usecase *uc = NULL; |
| struct stream_out *curr_out = NULL; |
| int usecase_backend_idx = DEFAULT_CODEC_BACKEND; |
| int i, num_devices, ret = 0; |
| snd_device_t split_snd_devices[SND_DEVICE_OUT_END]; |
| |
| list_for_each(node, &adev->usecase_list) { |
| uc = node_to_item(node, struct audio_usecase, list); |
| curr_out = (struct stream_out*) uc->stream.out; |
| if (curr_out && PCM_PLAYBACK == uc->type && uc != uc_info) { |
| /* |
| *Split snd device for existing combo use case |
| *e.g. Ringtone (Headphone + Speaker) + Headphopne 44.1 |
| */ |
| ret = platform_split_snd_device(adev->platform, |
| uc->out_snd_device, |
| &num_devices, |
| split_snd_devices); |
| if (ret < 0 || num_devices == 0) { |
| ALOGV("%s: Unable to split uc->out_snd_device: %d",__func__, uc->out_snd_device); |
| split_snd_devices[0] = uc->out_snd_device; |
| num_devices = 1; |
| } |
| for (i = 0; i < num_devices; i++) { |
| usecase_backend_idx = platform_get_backend_index(split_snd_devices[i]); |
| ALOGD("%s:snd_dev %d usecase_backend_idx %d",__func__, split_snd_devices[i],usecase_backend_idx); |
| if((new_backend_idx == HEADPHONE_BACKEND) && |
| ((usecase_backend_idx == HEADPHONE_44_1_BACKEND) || |
| (usecase_backend_idx == DSD_NATIVE_BACKEND))) { |
| ALOGV("%s:DSD or native stream detected enabling asrcmode in hardware", |
| __func__); |
| enable_asrc_mode(adev); |
| break; |
| } else if(((new_backend_idx == HEADPHONE_44_1_BACKEND) || |
| (new_backend_idx == DSD_NATIVE_BACKEND)) && |
| (usecase_backend_idx == HEADPHONE_BACKEND)) { |
| ALOGV("%s: 48K stream detected, disabling and enabling it \ |
| with asrcmode in hardware", __func__); |
| disable_audio_route(adev, uc); |
| disable_snd_device(adev, uc->out_snd_device); |
| // Apply true-high-quality-mode if DSD or > 44.1KHz or >=24-bit |
| if (new_backend_idx == DSD_NATIVE_BACKEND) |
| audio_route_apply_and_update_path(adev->audio_route, |
| "hph-true-highquality-mode"); |
| else if ((new_backend_idx == HEADPHONE_44_1_BACKEND) && |
| (curr_out->bit_width >= 24)) |
| audio_route_apply_and_update_path(adev->audio_route, |
| "hph-highquality-mode"); |
| enable_asrc_mode(adev); |
| enable_snd_device(adev, uc->out_snd_device); |
| enable_audio_route(adev, uc); |
| break; |
| } |
| } |
| // reset split devices count |
| num_devices = 0; |
| } |
| if (adev->asrc_mode_enabled) |
| break; |
| } |
| } |
| } |
| } |
| |
| static int send_effect_enable_disable_mixer_ctl(struct audio_device *adev, |
| struct audio_effect_config effect_config, |
| unsigned int param_value) |
| { |
| char mixer_ctl_name[] = "Audio Effect"; |
| struct mixer_ctl *ctl; |
| long set_values[6]; |
| struct stream_in *in = adev_get_active_input(adev); |
| |
| if (in == NULL) { |
| ALOGE("%s: active input stream is NULL", __func__); |
| return -EINVAL; |
| } |
| |
| ctl = mixer_get_ctl_by_name(adev->mixer, mixer_ctl_name); |
| if (!ctl) { |
| ALOGE("%s: Could not get mixer ctl - %s", |
| __func__, mixer_ctl_name); |
| return -EINVAL; |
| } |
| |
| set_values[0] = 1; //0:Rx 1:Tx |
| set_values[1] = in->app_type_cfg.app_type; |
| set_values[2] = (long)effect_config.module_id; |
| set_values[3] = (long)effect_config.instance_id; |
| set_values[4] = (long)effect_config.param_id; |
| set_values[5] = param_value; |
| |
| mixer_ctl_set_array(ctl, set_values, ARRAY_SIZE(set_values)); |
| |
| return 0; |
| |
| } |
| |
| static int update_effect_param_ecns(struct audio_device *adev, unsigned int module_id, |
| int effect_type, unsigned int *param_value) |
| { |
| int ret = 0; |
| struct audio_effect_config other_effect_config; |
| struct audio_usecase *usecase = NULL; |
| struct stream_in *in = adev_get_active_input(adev); |
| |
| if (in == NULL) { |
| ALOGE("%s: active input stream is NULL", __func__); |
| return -EINVAL; |
| } |
| |
| usecase = get_usecase_from_list(adev, in->usecase); |
| if (!usecase) |
| return -EINVAL; |
| |
| ret = platform_get_effect_config_data(usecase->in_snd_device, &other_effect_config, |
| effect_type == EFFECT_AEC ? EFFECT_NS : EFFECT_AEC); |
| if (ret < 0) { |
| ALOGE("%s Failed to get effect params %d", __func__, ret); |
| return ret; |
| } |
| |
| if (module_id == other_effect_config.module_id) { |
| //Same module id for AEC/NS. Values need to be combined |
| if (((effect_type == EFFECT_AEC) && (in->enable_ns)) || |
| ((effect_type == EFFECT_NS) && (in->enable_aec))) { |
| *param_value |= other_effect_config.param_value; |
| } |
| } |
| |
| return ret; |
| } |
| |
| static int enable_disable_effect(struct audio_device *adev, int effect_type, bool enable) |
| { |
| struct audio_effect_config effect_config; |
| struct audio_usecase *usecase = NULL; |
| int ret = 0; |
| unsigned int param_value = 0; |
| struct stream_in *in = adev_get_active_input(adev); |
| |
| if(!voice_extn_is_dynamic_ecns_enabled()) |
| return ENOSYS; |
| |
| if (!in) { |
| ALOGE("%s: Invalid input stream", __func__); |
| return -EINVAL; |
| } |
| |
| ALOGD("%s: effect_type:%d enable:%d", __func__, effect_type, enable); |
| |
| usecase = get_usecase_from_list(adev, in->usecase); |
| if (usecase == NULL) { |
| ALOGE("%s: Could not find the usecase (%d) in the list", |
| __func__, in->usecase); |
| return -EINVAL; |
| } |
| |
| ret = platform_get_effect_config_data(usecase->in_snd_device, &effect_config, effect_type); |
| if (ret < 0) { |
| ALOGE("%s Failed to get module id %d", __func__, ret); |
| return ret; |
| } |
| ALOGV("%s: %d %d usecase->id:%d usecase->in_snd_device:%d", __func__, effect_config.module_id, |
| in->app_type_cfg.app_type, usecase->id, usecase->in_snd_device); |
| |
| if(enable) |
| param_value = effect_config.param_value; |
| |
| /*Special handling for AEC & NS effects Param values need to be |
| updated if module ids are same*/ |
| |
| if ((effect_type == EFFECT_AEC) || (effect_type == EFFECT_NS)) { |
| ret = update_effect_param_ecns(adev, effect_config.module_id, effect_type, ¶m_value); |
| if (ret < 0) |
| return ret; |
| } |
| |
| ret = send_effect_enable_disable_mixer_ctl(adev, effect_config, param_value); |
| |
| return ret; |
| } |
| |
| static void check_and_enable_effect(struct audio_device *adev) |
| { |
| if(!voice_extn_is_dynamic_ecns_enabled()) |
| return; |
| |
| struct stream_in *in = adev_get_active_input(adev); |
| |
| if (in != NULL && !in->standby) { |
| if (in->enable_aec) |
| enable_disable_effect(adev, EFFECT_AEC, true); |
| |
| if (in->enable_ns && |
| in->source == AUDIO_SOURCE_VOICE_COMMUNICATION) { |
| enable_disable_effect(adev, EFFECT_NS, true); |
| } |
| } |
| } |
| |
| int pcm_ioctl(struct pcm *pcm, int request, ...) |
| { |
| va_list ap; |
| void * arg; |
| int pcm_fd = *(int*)pcm; |
| |
| va_start(ap, request); |
| arg = va_arg(ap, void *); |
| va_end(ap); |
| |
| return ioctl(pcm_fd, request, arg); |
| } |
| |
| int enable_audio_route(struct audio_device *adev, |
| struct audio_usecase *usecase) |
| { |
| snd_device_t snd_device; |
| char mixer_path[MIXER_PATH_MAX_LENGTH]; |
| struct stream_out *out = NULL; |
| struct stream_in *in = NULL; |
| struct listnode out_devices; |
| int ret = 0; |
| |
| if (usecase == NULL) |
| return -EINVAL; |
| |
| ALOGV("%s: enter: usecase(%d)", __func__, usecase->id); |
| |
| if (usecase->type == PCM_CAPTURE) { |
| struct stream_in *in = usecase->stream.in; |
| struct audio_usecase *uinfo; |
| snd_device = usecase->in_snd_device; |
| |
| if (in) { |
| if (in->enable_aec || in->enable_ec_port) { |
| list_init(&out_devices); |
| update_device_list(&out_devices, AUDIO_DEVICE_OUT_SPEAKER, "", true); |
| struct listnode *node; |
| struct audio_usecase *voip_usecase = get_usecase_from_list(adev, |
| USECASE_AUDIO_PLAYBACK_VOIP); |
| if (voip_usecase) { |
| assign_devices(&out_devices, |
| &voip_usecase->stream.out->device_list); |
| } else if (adev->primary_output && |
| !adev->primary_output->standby) { |
| assign_devices(&out_devices, |
| &adev->primary_output->device_list); |
| } else { |
| list_for_each(node, &adev->usecase_list) { |
| uinfo = node_to_item(node, struct audio_usecase, list); |
| if (uinfo->type != PCM_CAPTURE) { |
| assign_devices(&out_devices, |
| &uinfo->stream.out->device_list); |
| break; |
| } |
| } |
| } |
| |
| platform_set_echo_reference(adev, true, &out_devices); |
| in->ec_opened = true; |
| } |
| } |
| } else if ((usecase->type == TRANSCODE_LOOPBACK_TX) || ((usecase->type == PCM_HFP_CALL) && |
| ((usecase->id == USECASE_AUDIO_HFP_SCO) || (usecase->id == USECASE_AUDIO_HFP_SCO_WB)) && |
| (usecase->in_snd_device == SND_DEVICE_IN_VOICE_SPEAKER_MIC_HFP_MMSECNS))) { |
| snd_device = usecase->in_snd_device; |
| } else { |
| snd_device = usecase->out_snd_device; |
| } |
| |
| if (usecase->type == PCM_CAPTURE) { |
| if (platform_get_fluence_nn_state(adev->platform) == 0) { |
| platform_set_fluence_nn_state(adev->platform, true); |
| ALOGD("%s: set fluence nn capture state", __func__); |
| } |
| } |
| |
| #ifdef DS1_DOLBY_DAP_ENABLED |
| audio_extn_dolby_set_dmid(adev); |
| audio_extn_dolby_set_endpoint(adev); |
| #endif |
| audio_extn_dolby_ds2_set_endpoint(adev); |
| audio_extn_sound_trigger_update_stream_status(usecase, ST_EVENT_STREAM_BUSY); |
| audio_extn_listen_update_stream_status(usecase, LISTEN_EVENT_STREAM_BUSY); |
| audio_extn_utils_send_app_type_cfg(adev, usecase); |
| if (audio_extn_is_maxx_audio_enabled()) |
| audio_extn_ma_set_device(usecase); |
| audio_extn_utils_send_audio_calibration(adev, usecase); |
| if ((usecase->type == PCM_PLAYBACK) && is_offload_usecase(usecase->id)) { |
| out = usecase->stream.out; |
| if (out && out->compr) |
| audio_extn_utils_compress_set_clk_rec_mode(usecase); |
| } |
| |
| if (usecase->type == PCM_CAPTURE) { |
| if (platform_get_fluence_nn_state(adev->platform) == 1 && |
| adev->fluence_nn_usecase_id == USECASE_INVALID ) { |
| adev->fluence_nn_usecase_id = usecase->id; |
| ALOGD("%s: assign fluence nn usecase %d", __func__, usecase->id); |
| } |
| } |
| |
| if (usecase->type == PCM_CAPTURE) { |
| in = usecase->stream.in; |
| if (in && is_loopback_input_device(get_device_types(&in->device_list))) { |
| ALOGD("%s: set custom mtmx params v1", __func__); |
| audio_extn_set_custom_mtmx_params_v1(adev, usecase, true); |
| } |
| } else { |
| audio_extn_set_custom_mtmx_params_v2(adev, usecase, true); |
| } |
| |
| // we shouldn't truncate mixer_path |
| ALOGW_IF(strlcpy(mixer_path, use_case_table[usecase->id], sizeof(mixer_path)) |
| >= sizeof(mixer_path), "%s: truncation on mixer path", __func__); |
| // this also appends to mixer_path |
| platform_add_backend_name(mixer_path, snd_device, usecase); |
| ALOGD("%s: apply mixer and update path: %s", __func__, mixer_path); |
| ret = audio_route_apply_and_update_path(adev->audio_route, mixer_path); |
| if (!ret && usecase->id == USECASE_AUDIO_PLAYBACK_FM) { |
| struct str_parms *parms = str_parms_create_str("fm_restore_volume=1"); |
| if (parms) { |
| audio_extn_fm_set_parameters(adev, parms); |
| str_parms_destroy(parms); |
| } |
| } |
| ALOGV("%s: exit", __func__); |
| return 0; |
| } |
| |
| int disable_audio_route(struct audio_device *adev, |
| struct audio_usecase *usecase) |
| { |
| snd_device_t snd_device; |
| char mixer_path[MIXER_PATH_MAX_LENGTH]; |
| struct stream_in *in = NULL; |
| |
| if (usecase == NULL || usecase->id == USECASE_INVALID) |
| return -EINVAL; |
| |
| ALOGV("%s: enter: usecase(%d)", __func__, usecase->id); |
| if (usecase->type == PCM_CAPTURE || usecase->type == TRANSCODE_LOOPBACK_TX) |
| snd_device = usecase->in_snd_device; |
| else |
| snd_device = usecase->out_snd_device; |
| |
| /* disable island and power mode on supported device for voice call */ |
| if (usecase->type == VOICE_CALL) { |
| if (usecase->in_snd_device != SND_DEVICE_NONE) { |
| if (platform_get_island_cfg_on_device(adev->platform, usecase->in_snd_device) && |
| platform_get_power_mode_on_device(adev->platform, usecase->in_snd_device)) { |
| platform_set_island_cfg_on_device(adev, usecase->in_snd_device, false); |
| platform_set_power_mode_on_device(adev, usecase->in_snd_device, false); |
| platform_reset_island_power_status(adev->platform, usecase->in_snd_device); |
| if (voice_is_lte_call_active(adev)) |
| platform_set_tx_lpi_mode(adev->platform, false); |
| ALOGD("%s: disable island cfg and power mode in voice tx path", |
| __func__); |
| } |
| } |
| if (usecase->out_snd_device != SND_DEVICE_NONE) { |
| if (platform_get_island_cfg_on_device(adev->platform, usecase->out_snd_device) && |
| platform_get_power_mode_on_device(adev->platform, usecase->out_snd_device)) { |
| platform_set_island_cfg_on_device(adev, usecase->out_snd_device, false); |
| platform_set_power_mode_on_device(adev, usecase->out_snd_device, false); |
| platform_reset_island_power_status(adev->platform, usecase->out_snd_device); |
| ALOGD("%s: disable island cfg and power mode in voice rx path", |
| __func__); |
| } |
| } |
| } |
| |
| // we shouldn't truncate mixer_path |
| ALOGW_IF(strlcpy(mixer_path, use_case_table[usecase->id], sizeof(mixer_path)) |
| >= sizeof(mixer_path), "%s: truncation on mixer path", __func__); |
| // this also appends to mixer_path |
| platform_add_backend_name(mixer_path, snd_device, usecase); |
| ALOGD("%s: reset and update mixer path: %s", __func__, mixer_path); |
| audio_route_reset_and_update_path(adev->audio_route, mixer_path); |
| if (usecase->type == PCM_CAPTURE) { |
| struct stream_in *in = usecase->stream.in; |
| if (in && in->ec_opened) { |
| struct listnode out_devices; |
| list_init(&out_devices); |
| platform_set_echo_reference(in->dev, false, &out_devices); |
| in->ec_opened = false; |
| } |
| } |
| if (usecase->id == adev->fluence_nn_usecase_id) { |
| platform_set_fluence_nn_state(adev->platform, false); |
| adev->fluence_nn_usecase_id = USECASE_INVALID; |
| ALOGD("%s: reset fluence nn capture state", __func__); |
| } |
| audio_extn_sound_trigger_update_stream_status(usecase, ST_EVENT_STREAM_FREE); |
| audio_extn_listen_update_stream_status(usecase, LISTEN_EVENT_STREAM_FREE); |
| |
| if (usecase->type == PCM_CAPTURE) { |
| in = usecase->stream.in; |
| if (in && is_loopback_input_device(get_device_types(&in->device_list))) { |
| ALOGD("%s: reset custom mtmx params v1", __func__); |
| audio_extn_set_custom_mtmx_params_v1(adev, usecase, false); |
| } |
| } else { |
| audio_extn_set_custom_mtmx_params_v2(adev, usecase, false); |
| } |
| |
| if ((usecase->type == PCM_PLAYBACK) && |
| (usecase->stream.out != NULL)) |
| usecase->stream.out->pspd_coeff_sent = false; |
| |
| ALOGV("%s: exit", __func__); |
| return 0; |
| } |
| |
| int enable_snd_device(struct audio_device *adev, |
| snd_device_t snd_device) |
| { |
| int i, num_devices = 0; |
| snd_device_t new_snd_devices[SND_DEVICE_OUT_END]; |
| char device_name[DEVICE_NAME_MAX_SIZE] = {0}; |
| |
| if (snd_device < SND_DEVICE_MIN || |
| snd_device >= SND_DEVICE_MAX) { |
| ALOGE("%s: Invalid sound device %d", __func__, snd_device); |
| return -EINVAL; |
| } |
| |
| if (platform_get_snd_device_name_extn(adev->platform, snd_device, device_name) < 0) { |
| ALOGE("%s: Invalid sound device returned", __func__); |
| return -EINVAL; |
| } |
| |
| adev->snd_dev_ref_cnt[snd_device]++; |
| |
| if ((adev->snd_dev_ref_cnt[snd_device] > 1) && |
| (platform_split_snd_device(adev->platform, |
| snd_device, |
| &num_devices, |
| new_snd_devices) != 0)) { |
| ALOGV("%s: snd_device(%d: %s) is already active", |
| __func__, snd_device, device_name); |
| /* Set backend config for A2DP to ensure slimbus configuration |
| is correct if A2DP is already active and backend is closed |
| and re-opened */ |
| if (snd_device == SND_DEVICE_OUT_BT_A2DP) |
| audio_extn_a2dp_set_source_backend_cfg(); |
| return 0; |
| } |
| |
| if (audio_extn_spkr_prot_is_enabled()) |
| audio_extn_spkr_prot_calib_cancel(adev); |
| |
| audio_extn_dsm_feedback_enable(adev, snd_device, true); |
| |
| if (platform_can_enable_spkr_prot_on_device(snd_device) && |
| audio_extn_spkr_prot_is_enabled()) { |
| if (platform_get_spkr_prot_acdb_id(snd_device) < 0) { |
| goto err; |
| } |
| audio_extn_dev_arbi_acquire(snd_device); |
| if (audio_extn_spkr_prot_start_processing(snd_device)) { |
| ALOGE("%s: spkr_start_processing failed", __func__); |
| audio_extn_dev_arbi_release(snd_device); |
| goto err; |
| } |
| } else if (platform_split_snd_device(adev->platform, |
| snd_device, |
| &num_devices, |
| new_snd_devices) == 0) { |
| for (i = 0; i < num_devices; i++) { |
| enable_snd_device(adev, new_snd_devices[i]); |
| } |
| platform_set_speaker_gain_in_combo(adev, snd_device, true); |
| } else { |
| ALOGD("%s: snd_device(%d: %s)", __func__, snd_device, device_name); |
| |
| /* enable island and power mode on supported device */ |
| if (platform_get_island_cfg_on_device(adev->platform, snd_device) && |
| platform_get_power_mode_on_device(adev->platform, snd_device)) { |
| platform_set_island_cfg_on_device(adev, snd_device, true); |
| platform_set_power_mode_on_device(adev, snd_device, true); |
| if (voice_is_lte_call_active(adev) && |
| (snd_device >= SND_DEVICE_IN_BEGIN && |
| snd_device < SND_DEVICE_IN_END)) |
| platform_set_tx_lpi_mode(adev->platform, true); |
| ALOGD("%s: enable island cfg and power mode on: %s", |
| __func__, device_name); |
| } |
| |
| if (SND_DEVICE_OUT_BT_A2DP == snd_device) { |
| if (audio_extn_a2dp_start_playback() < 0) { |
| ALOGE(" fail to configure A2dp Source control path "); |
| goto err; |
| } else { |
| adev->a2dp_started = true; |
| } |
| } |
| |
| if ((SND_DEVICE_IN_BT_A2DP == snd_device) && |
| (audio_extn_a2dp_start_capture() < 0)) { |
| ALOGE(" fail to configure A2dp Sink control path "); |
| goto err; |
| } |
| |
| if ((SND_DEVICE_OUT_BT_SCO_SWB == snd_device) || |
| (SND_DEVICE_IN_BT_SCO_MIC_SWB_NREC == snd_device) || |
| (SND_DEVICE_IN_BT_SCO_MIC_SWB == snd_device)) { |
| if (!adev->bt_sco_on || (audio_extn_sco_start_configuration() < 0)) { |
| ALOGE(" fail to configure sco control path "); |
| goto err; |
| } |
| } |
| |
| configure_btsco_sample_rate(snd_device); |
| /* due to the possibility of calibration overwrite between listen |
| and audio, notify listen hal before audio calibration is sent */ |
| audio_extn_sound_trigger_update_device_status(snd_device, |
| ST_EVENT_SND_DEVICE_BUSY); |
| audio_extn_listen_update_device_status(snd_device, |
| LISTEN_EVENT_SND_DEVICE_BUSY); |
| if (platform_get_snd_device_acdb_id(snd_device) < 0) { |
| audio_extn_sound_trigger_update_device_status(snd_device, |
| ST_EVENT_SND_DEVICE_FREE); |
| audio_extn_listen_update_device_status(snd_device, |
| LISTEN_EVENT_SND_DEVICE_FREE); |
| goto err; |
| } |
| audio_extn_dev_arbi_acquire(snd_device); |
| audio_route_apply_and_update_path(adev->audio_route, device_name); |
| |
| if (SND_DEVICE_OUT_HEADPHONES == snd_device && |
| !adev->native_playback_enabled && |
| audio_is_true_native_stream_active(adev)) { |
| ALOGD("%s: %d: napb: enabling native mode in hardware", |
| __func__, __LINE__); |
| audio_route_apply_and_update_path(adev->audio_route, |
| "true-native-mode"); |
| adev->native_playback_enabled = true; |
| } |
| if (((snd_device == SND_DEVICE_IN_HANDSET_6MIC) || |
| (snd_device == SND_DEVICE_IN_HANDSET_QMIC)) && |
| (audio_extn_ffv_get_stream() == adev_get_active_input(adev))) { |
| ALOGD("%s: init ec ref loopback", __func__); |
| audio_extn_ffv_init_ec_ref_loopback(adev, snd_device); |
| } |
| } |
| return 0; |
| err: |
| adev->snd_dev_ref_cnt[snd_device]--; |
| return -EINVAL;; |
| } |
| |
| int disable_snd_device(struct audio_device *adev, |
| snd_device_t snd_device) |
| { |
| int i, num_devices = 0; |
| snd_device_t new_snd_devices[SND_DEVICE_OUT_END]; |
| char device_name[DEVICE_NAME_MAX_SIZE] = {0}; |
| |
| if (snd_device < SND_DEVICE_MIN || |
| snd_device >= SND_DEVICE_MAX) { |
| ALOGE("%s: Invalid sound device %d", __func__, snd_device); |
| return -EINVAL; |
| } |
| |
| if (platform_get_snd_device_name_extn(adev->platform, snd_device, device_name) < 0) { |
| ALOGE("%s: Invalid sound device returned", __func__); |
| return -EINVAL; |
| } |
| |
| if (adev->snd_dev_ref_cnt[snd_device] <= 0) { |
| ALOGE("%s: device ref cnt is already 0", __func__); |
| return -EINVAL; |
| } |
| |
| adev->snd_dev_ref_cnt[snd_device]--; |
| |
| |
| if (adev->snd_dev_ref_cnt[snd_device] == 0) { |
| ALOGD("%s: snd_device(%d: %s)", __func__, snd_device, device_name); |
| |
| audio_extn_dsm_feedback_enable(adev, snd_device, false); |
| |
| if (platform_can_enable_spkr_prot_on_device(snd_device) && |
| audio_extn_spkr_prot_is_enabled()) { |
| audio_extn_spkr_prot_stop_processing(snd_device); |
| |
| // when speaker device is disabled, reset swap. |
| // will be renabled on usecase start |
| platform_set_swap_channels(adev, false); |
| } else if (platform_split_snd_device(adev->platform, |
| snd_device, |
| &num_devices, |
| new_snd_devices) == 0) { |
| for (i = 0; i < num_devices; i++) { |
| disable_snd_device(adev, new_snd_devices[i]); |
| } |
| platform_set_speaker_gain_in_combo(adev, snd_device, false); |
| } else { |
| audio_route_reset_and_update_path(adev->audio_route, device_name); |
| } |
| |
| if (snd_device == SND_DEVICE_OUT_BT_A2DP) { |
| audio_extn_a2dp_stop_playback(); |
| adev->a2dp_started = false; |
| } else if (snd_device == SND_DEVICE_IN_BT_A2DP) |
| audio_extn_a2dp_stop_capture(); |
| else if ((snd_device == SND_DEVICE_OUT_HDMI) || |
| (snd_device == SND_DEVICE_OUT_DISPLAY_PORT)) |
| adev->is_channel_status_set = false; |
| else if ((snd_device == SND_DEVICE_OUT_HEADPHONES) && |
| adev->native_playback_enabled) { |
| ALOGD("%s: %d: napb: disabling native mode in hardware", |
| __func__, __LINE__); |
| audio_route_reset_and_update_path(adev->audio_route, |
| "true-native-mode"); |
| adev->native_playback_enabled = false; |
| } else if ((snd_device == SND_DEVICE_OUT_HEADPHONES) && |
| adev->asrc_mode_enabled) { |
| ALOGD("%s: %d: disabling asrc mode in hardware", __func__, __LINE__); |
| disable_asrc_mode(adev); |
| audio_route_apply_and_update_path(adev->audio_route, "hph-lowpower-mode"); |
| } else if (((snd_device == SND_DEVICE_IN_HANDSET_6MIC) || |
| (snd_device == SND_DEVICE_IN_HANDSET_QMIC)) && |
| (audio_extn_ffv_get_stream() == adev_get_active_input(adev))) { |
| ALOGD("%s: deinit ec ref loopback", __func__); |
| audio_extn_ffv_deinit_ec_ref_loopback(adev, snd_device); |
| } |
| |
| audio_extn_utils_release_snd_device(snd_device); |
| } else { |
| if (platform_split_snd_device(adev->platform, |
| snd_device, |
| &num_devices, |
| new_snd_devices) == 0) { |
| for (i = 0; i < num_devices; i++) { |
| adev->snd_dev_ref_cnt[new_snd_devices[i]]--; |
| } |
| } |
| } |
| |
| return 0; |
| } |
| |
| /* |
| legend: |
| uc - existing usecase |
| new_uc - new usecase |
| d1, d11, d2 - SND_DEVICE enums |
| a1, a2 - corresponding ANDROID device enums |
| B1, B2 - backend strings |
| |
| case 1 |
| uc->dev d1 (a1) B1 |
| new_uc->dev d1 (a1), d2 (a2) B1, B2 |
| |
| resolution: disable and enable uc->dev on d1 |
| |
| case 2 |
| uc->dev d1 (a1) B1 |
| new_uc->dev d11 (a1) B1 |
| |
| resolution: need to switch uc since d1 and d11 are related |
| (e.g. speaker and voice-speaker) |
| use ANDROID_DEVICE_OUT enums to match devices since SND_DEVICE enums may vary |
| |
| case 3 |
| uc->dev d1 (a1) B1 |
| new_uc->dev d2 (a2) B2 |
| |
| resolution: no need to switch uc |
| |
| case 4 |
| uc->dev d1 (a1) B1 |
| new_uc->dev d2 (a2) B1 |
| |
| resolution: disable enable uc-dev on d2 since backends match |
| we cannot enable two streams on two different devices if they |
| share the same backend. e.g. if offload is on speaker device using |
| QUAD_MI2S backend and a low-latency stream is started on voice-handset |
| using the same backend, offload must also be switched to voice-handset. |
| |
| case 5 |
| uc->dev d1 (a1) B1 |
| new_uc->dev d1 (a1), d2 (a2) B1 |
| |
| resolution: disable enable uc-dev on d2 since backends match |
| we cannot enable two streams on two different devices if they |
| share the same backend. |
| |
| case 6 |
| uc->dev d1 (a1) B1 |
| new_uc->dev d2 (a1) B2 |
| |
| resolution: no need to switch |
| |
| case 7 |
| uc->dev d1 (a1), d2 (a2) B1, B2 |
| new_uc->dev d1 (a1) B1 |
| |
| resolution: no need to switch |
| |
| case 8 |
| uc->dev d1 (a1) B1 |
| new_uc->dev d11 (a1), d2 (a2) B1, B2 |
| resolution: compared to case 1, for this case, d1 and d11 are related |
| then need to do the same as case 2 to siwtch to new uc |
| |
| case 9 |
| uc->dev d1 (a1), d2(a2) B1 B2 |
| new_uc->dev d1 (a1), d22 (a2) B1, B2 |
| resolution: disable enable uc-dev on d2 since backends match |
| we cannot enable two streams on two different devices if they |
| share the same backend. This is special case for combo use case |
| with a2dp and sco devices which uses same backend. |
| e.g. speaker-a2dp and speaker-btsco |
| */ |
| static snd_device_t derive_playback_snd_device(void * platform, |
| struct audio_usecase *uc, |
| struct audio_usecase *new_uc, |
| snd_device_t new_snd_device) |
| { |
| struct listnode a1, a2; |
| |
| snd_device_t d1 = uc->out_snd_device; |
| snd_device_t d2 = new_snd_device; |
| |
| list_init(&a1); |
| list_init(&a2); |
| |
| switch (uc->type) { |
| case TRANSCODE_LOOPBACK_RX : |
| assign_devices(&a1, &uc->stream.inout->out_config.device_list); |
| assign_devices(&a2, &new_uc->stream.inout->out_config.device_list); |
| break; |
| default : |
| assign_devices(&a1, &uc->stream.out->device_list); |
| assign_devices(&a2, &new_uc->stream.out->device_list); |
| break; |
| } |
| |
| // Treat as a special case when a1 and a2 are not disjoint |
| if (!compare_devices(&a1, &a2) && |
| compare_devices_for_any_match(&a1 ,&a2)) { |
| snd_device_t d3[2]; |
| int num_devices = 0; |
| int ret = platform_split_snd_device(platform, |
| list_length(&a1) > 1 ? d1 : d2, |
| &num_devices, |
| d3); |
| if (ret < 0) { |
| if (ret != -ENOSYS) { |
| ALOGW("%s failed to split snd_device %d", |
| __func__, |
| list_length(&a1) > 1 ? d1 : d2); |
| } |
| goto end; |
| } |
| |
| if (platform_check_backends_match(d3[0], d3[1])) { |
| return d2; // case 5 |
| } else { |
| if ((list_length(&a1) > 1) && (list_length(&a2) > 1) && |
| platform_check_backends_match(d1, d2)) |
| return d2; //case 9 |
| if (list_length(&a1) > 1) |
| return d1; //case 7 |
| // check if d1 is related to any of d3's |
| if (d1 == d3[0] || d1 == d3[1]) |
| return d1; // case 1 |
| else |
| return d3[1]; // case 8 |
| } |
| } else { |
| if (platform_check_backends_match(d1, d2)) { |
| return d2; // case 2, 4 |
| } else { |
| return d1; // case 6, 3 |
| } |
| } |
| |
| end: |
| return d2; // return whatever was calculated before. |
| } |
| |
| static void check_usecases_codec_backend(struct audio_device *adev, |
| struct audio_usecase *uc_info, |
| snd_device_t snd_device) |
| { |
| struct listnode *node; |
| struct audio_usecase *usecase; |
| bool switch_device[AUDIO_USECASE_MAX]; |
| snd_device_t uc_derive_snd_device; |
| snd_device_t derive_snd_device[AUDIO_USECASE_MAX]; |
| snd_device_t split_snd_devices[SND_DEVICE_OUT_END]; |
| int i, num_uc_to_switch = 0, num_devices = 0; |
| int status = 0; |
| bool force_restart_session = false; |
| /* |
| * This function is to make sure that all the usecases that are active on |
| * the hardware codec backend are always routed to any one device that is |
| * handled by the hardware codec. |
| * For example, if low-latency and deep-buffer usecases are currently active |
| * on speaker and out_set_parameters(headset) is received on low-latency |
| * output, then we have to make sure deep-buffer is also switched to headset, |
| * because of the limitation that both the devices cannot be enabled |
| * at the same time as they share the same backend. |
| */ |
| /* |
| * This call is to check if we need to force routing for a particular stream |
| * If there is a backend configuration change for the device when a |
| * new stream starts, then ADM needs to be closed and re-opened with the new |
| * configuraion. This call check if we need to re-route all the streams |
| * associated with the backend. Touch tone + 24 bit + native playback. |
| */ |
| bool force_routing = platform_check_and_set_codec_backend_cfg(adev, uc_info, |
| snd_device); |
| /* For a2dp device reconfigure all active sessions |
| * with new AFE encoder format based on a2dp state |
| */ |
| if ((SND_DEVICE_OUT_BT_A2DP == snd_device || |
| SND_DEVICE_OUT_SPEAKER_AND_BT_A2DP == snd_device || |
| SND_DEVICE_OUT_SPEAKER_SAFE_AND_BT_A2DP == snd_device) && |
| audio_extn_a2dp_is_force_device_switch()) { |
| force_routing = true; |
| force_restart_session = true; |
| } |
| |
| /* |
| * Island cfg and power mode config needs to set before AFE port start. |
| * Set force routing in case of voice device was enable before. |
| */ |
| if (uc_info->type == VOICE_CALL && |
| voice_extn_is_voice_power_mode_supported() && |
| is_supported_conc_usecase_for_power_mode_call(adev) && |
| platform_check_and_update_island_power_status(adev->platform, |
| uc_info, |
| snd_device)) { |
| force_routing = true; |
| ALOGD("%s:becf: force routing %d for power mode supported device", |
| __func__, force_routing); |
| } |
| ALOGD("%s:becf: force routing %d", __func__, force_routing); |
| |
| /* Disable all the usecases on the shared backend other than the |
| * specified usecase. |
| */ |
| for (i = 0; i < AUDIO_USECASE_MAX; i++) |
| switch_device[i] = false; |
| |
| list_for_each(node, &adev->usecase_list) { |
| usecase = node_to_item(node, struct audio_usecase, list); |
| |
| ALOGD("%s:becf: (%d) check_usecases curr device: %s, usecase device:%s " |
| "backends match %d",__func__, i, |
| platform_get_snd_device_name(snd_device), |
| platform_get_snd_device_name(usecase->out_snd_device), |
| platform_check_backends_match(snd_device, usecase->out_snd_device)); |
| if ((usecase->type != PCM_CAPTURE) && (usecase != uc_info) && |
| (usecase->type != PCM_PASSTHROUGH)) { |
| uc_derive_snd_device = derive_playback_snd_device(adev->platform, |
| usecase, uc_info, snd_device); |
| if (((uc_derive_snd_device != usecase->out_snd_device) || force_routing) && |
| (is_codec_backend_out_device_type(&usecase->device_list) || |
| compare_device_type(&usecase->device_list, AUDIO_DEVICE_OUT_AUX_DIGITAL) || |
| compare_device_type(&usecase->device_list, AUDIO_DEVICE_OUT_USB_DEVICE) || |
| compare_device_type(&usecase->device_list, AUDIO_DEVICE_OUT_USB_HEADSET) || |
| is_a2dp_out_device_type(&usecase->device_list) || |
| is_sco_out_device_type(&usecase->device_list)) && |
| ((force_restart_session) || |
| (platform_check_backends_match(snd_device, usecase->out_snd_device)))) { |
| ALOGD("%s:becf: check_usecases (%s) is active on (%s) - disabling ..", |
| __func__, use_case_table[usecase->id], |
| platform_get_snd_device_name(usecase->out_snd_device)); |
| disable_audio_route(adev, usecase); |
| switch_device[usecase->id] = true; |
| /* Enable existing usecase on derived playback device */ |
| derive_snd_device[usecase->id] = uc_derive_snd_device; |
| num_uc_to_switch++; |
| } |
| } |
| } |
| |
| ALOGD("%s:becf: check_usecases num.of Usecases to switch %d", __func__, |
| num_uc_to_switch); |
| |
| if (num_uc_to_switch) { |
| /* All streams have been de-routed. Disable the device */ |
| |
| /* Make sure the previous devices to be disabled first and then enable the |
| selected devices */ |
| list_for_each(node, &adev->usecase_list) { |
| usecase = node_to_item(node, struct audio_usecase, list); |
| if (switch_device[usecase->id]) { |
| /* Check if output sound device to be switched can be split and if any |
| of the split devices match with derived sound device */ |
| if (platform_split_snd_device(adev->platform, usecase->out_snd_device, |
| &num_devices, split_snd_devices) == 0) { |
| adev->snd_dev_ref_cnt[usecase->out_snd_device]--; |
| for (i = 0; i < num_devices; i++) { |
| /* Disable devices that do not match with derived sound device */ |
| if (split_snd_devices[i] != derive_snd_device[usecase->id]) |
| disable_snd_device(adev, split_snd_devices[i]); |
| } |
| } else { |
| disable_snd_device(adev, usecase->out_snd_device); |
| } |
| } |
| } |
| |
| list_for_each(node, &adev->usecase_list) { |
| usecase = node_to_item(node, struct audio_usecase, list); |
| if (switch_device[usecase->id]) { |
| if (platform_split_snd_device(adev->platform, usecase->out_snd_device, |
| &num_devices, split_snd_devices) == 0) { |
| /* Enable derived sound device only if it does not match with |
| one of the split sound devices. This is because the matching |
| sound device was not disabled */ |
| bool should_enable = true; |
| for (i = 0; i < num_devices; i++) { |
| if (derive_snd_device[usecase->id] == split_snd_devices[i]) { |
| should_enable = false; |
| break; |
| } |
| } |
| if (should_enable) |
| enable_snd_device(adev, derive_snd_device[usecase->id]); |
| } else { |
| enable_snd_device(adev, derive_snd_device[usecase->id]); |
| } |
| } |
| } |
| |
| /* Re-route all the usecases on the shared backend other than the |
| specified usecase to new snd devices */ |
| list_for_each(node, &adev->usecase_list) { |
| usecase = node_to_item(node, struct audio_usecase, list); |
| /* Update the out_snd_device only before enabling the audio route */ |
| if (switch_device[usecase->id]) { |
| usecase->out_snd_device = derive_snd_device[usecase->id]; |
| ALOGD("%s:becf: enabling usecase (%s) on (%s)", __func__, |
| use_case_table[usecase->id], |
| platform_get_snd_device_name(usecase->out_snd_device)); |
| /* Update voc calibration before enabling Voice/VoIP route */ |
| if (usecase->type == VOICE_CALL || usecase->type == VOIP_CALL) |
| status = platform_switch_voice_call_device_post(adev->platform, |
| usecase->out_snd_device, |
| platform_get_input_snd_device( |
| adev->platform, NULL, |
| &uc_info->device_list, |
| usecase->type)); |
| enable_audio_route(adev, usecase); |
| if (usecase->stream.out && usecase->id == USECASE_AUDIO_PLAYBACK_VOIP) { |
| out_set_voip_volume(&usecase->stream.out->stream, |
| usecase->stream.out->volume_l, |
| usecase->stream.out->volume_r); |
| } |
| } |
| } |
| } |
| } |
| |
| static void check_usecases_capture_codec_backend(struct audio_device *adev, |
| struct audio_usecase *uc_info, |
| snd_device_t snd_device) |
| { |
| struct listnode *node; |
| struct audio_usecase *usecase; |
| bool switch_device[AUDIO_USECASE_MAX]; |
| int i, num_uc_to_switch = 0; |
| int backend_check_cond = is_codec_backend_out_device_type(&uc_info->device_list); |
| int status = 0; |
| |
| bool force_routing = platform_check_and_set_capture_codec_backend_cfg(adev, uc_info, |
| snd_device); |
| ALOGD("%s:becf: force routing %d", __func__, force_routing); |
| |
| /* |
| * Make sure out devices is checked against out codec backend device and |
| * also in devices against in codec backend. Checking out device against in |
| * codec backend or vice versa causes issues. |
| */ |
| if (uc_info->type == PCM_CAPTURE) |
| backend_check_cond = is_codec_backend_in_device_type(&uc_info->device_list); |
| |
| /* |
| * Island cfg and power mode config needs to set before AFE port start. |
| * Set force routing in case of voice device was enable before. |
| */ |
| |
| if (uc_info->type == VOICE_CALL && |
| voice_extn_is_voice_power_mode_supported() && |
| is_supported_conc_usecase_for_power_mode_call(adev) && |
| platform_check_and_update_island_power_status(adev->platform, |
| uc_info, |
| snd_device)) { |
| force_routing = true; |
| ALOGD("%s:becf: force routing %d for power mode supported device", |
| __func__, force_routing); |
| } |
| |
| /* |
| * This function is to make sure that all the active capture usecases |
| * are always routed to the same input sound device. |
| * For example, if audio-record and voice-call usecases are currently |
| * active on speaker(rx) and speaker-mic (tx) and out_set_parameters(earpiece) |
| * is received for voice call then we have to make sure that audio-record |
| * usecase is also switched to earpiece i.e. voice-dmic-ef, |
| * because of the limitation that two devices cannot be enabled |
| * at the same time if they share the same backend. |
| */ |
| for (i = 0; i < AUDIO_USECASE_MAX; i++) |
| switch_device[i] = false; |
| |
| list_for_each(node, &adev->usecase_list) { |
| usecase = node_to_item(node, struct audio_usecase, list); |
| /* |
| * TODO: Enhance below condition to handle BT sco/USB multi recording |
| */ |
| |
| bool capture_uc_needs_routing = usecase->type != PCM_PLAYBACK && (usecase != uc_info && |
| (usecase->in_snd_device != snd_device || force_routing)); |
| bool call_proxy_snd_device = platform_is_call_proxy_snd_device(snd_device) || |
| platform_is_call_proxy_snd_device(usecase->in_snd_device); |
| if (capture_uc_needs_routing && !call_proxy_snd_device && |
| ((backend_check_cond && |
| (is_codec_backend_in_device_type(&usecase->device_list) || |
| (usecase->type == VOIP_CALL))) || |
| ((uc_info->type == VOICE_CALL && |
| is_single_device_type_equal(&usecase->device_list, |
| AUDIO_DEVICE_IN_VOICE_CALL)) || |
| platform_check_backends_match(snd_device,\ |
| usecase->in_snd_device))) && |
| (usecase->id != USECASE_AUDIO_SPKR_CALIB_TX)) { |
| ALOGD("%s: Usecase (%s) is active on (%s) - disabling ..", |
| __func__, use_case_table[usecase->id], |
| platform_get_snd_device_name(usecase->in_snd_device)); |
| disable_audio_route(adev, usecase); |
| switch_device[usecase->id] = true; |
| num_uc_to_switch++; |
| } |
| } |
| |
| if (num_uc_to_switch) { |
| /* All streams have been de-routed. Disable the device */ |
| |
| /* Make sure the previous devices to be disabled first and then enable the |
| selected devices */ |
| list_for_each(node, &adev->usecase_list) { |
| usecase = node_to_item(node, struct audio_usecase, list); |
| if (switch_device[usecase->id]) { |
| disable_snd_device(adev, usecase->in_snd_device); |
| } |
| } |
| |
| list_for_each(node, &adev->usecase_list) { |
| usecase = node_to_item(node, struct audio_usecase, list); |
| if (switch_device[usecase->id]) { |
| enable_snd_device(adev, snd_device); |
| } |
| } |
| |
| /* Re-route all the usecases on the shared backend other than the |
| specified usecase to new snd devices */ |
| list_for_each(node, &adev->usecase_list) { |
| usecase = node_to_item(node, struct audio_usecase, list); |
| /* Update the in_snd_device only before enabling the audio route */ |
| if (switch_device[usecase->id] ) { |
| usecase->in_snd_device = snd_device; |
| /* Update voc calibration before enabling Voice/VoIP route */ |
| if (usecase->type == VOICE_CALL || usecase->type == VOIP_CALL) { |
| snd_device_t voip_snd_device; |
| voip_snd_device = platform_get_output_snd_device(adev->platform, |
| usecase->stream.out, |
| usecase->type); |
| status = platform_switch_voice_call_device_post(adev->platform, |
| voip_snd_device, |
| usecase->in_snd_device); |
| } |
| enable_audio_route(adev, usecase); |
| } |
| } |
| } |
| } |
| |
| static void reset_hdmi_sink_caps(struct stream_out *out) { |
| int i = 0; |
| |
| for (i = 0; i<= MAX_SUPPORTED_CHANNEL_MASKS; i++) { |
| out->supported_channel_masks[i] = 0; |
| } |
| for (i = 0; i<= MAX_SUPPORTED_FORMATS; i++) { |
| out->supported_formats[i] = 0; |
| } |
| for (i = 0; i<= MAX_SUPPORTED_SAMPLE_RATES; i++) { |
| out->supported_sample_rates[i] = 0; |
| } |
| } |
| |
| /* must be called with hw device mutex locked */ |
| static int read_hdmi_sink_caps(struct stream_out *out) |
| { |
| int ret = 0, i = 0, j = 0; |
| int channels = platform_edid_get_max_channels_v2(out->dev->platform, |
| out->extconn.cs.controller, |
| out->extconn.cs.stream); |
| |
| reset_hdmi_sink_caps(out); |
| |
| /* Cache ext disp type */ |
| ret = platform_get_ext_disp_type_v2(adev->platform, |
| out->extconn.cs.controller, |
| out->extconn.cs.stream); |
| if(ret < 0) { |
| ALOGE("%s: Failed to query disp type, ret:%d", __func__, ret); |
| return -EINVAL; |
| } |
| |
| switch (channels) { |
| case 8: |
| ALOGV("%s: HDMI supports 7.1 channels", __func__); |
| out->supported_channel_masks[i++] = AUDIO_CHANNEL_OUT_7POINT1; |
| out->supported_channel_masks[i++] = AUDIO_CHANNEL_OUT_6POINT1; |
| case 6: |
| ALOGV("%s: HDMI supports 5.1 channels", __func__); |
| out->supported_channel_masks[i++] = AUDIO_CHANNEL_OUT_5POINT1; |
| out->supported_channel_masks[i++] = AUDIO_CHANNEL_OUT_PENTA; |
| out->supported_channel_masks[i++] = AUDIO_CHANNEL_OUT_QUAD; |
| out->supported_channel_masks[i++] = AUDIO_CHANNEL_OUT_SURROUND; |
| out->supported_channel_masks[i++] = AUDIO_CHANNEL_OUT_2POINT1; |
| break; |
| default: |
| ALOGE("invalid/nonstandard channal count[%d]",channels); |
| ret = -ENOSYS; |
| break; |
| } |
| |
| // check channel format caps |
| i = 0; |
| if (platform_is_edid_supported_format_v2(out->dev->platform, AUDIO_FORMAT_AC3, |
| out->extconn.cs.controller, |
| out->extconn.cs.stream)) { |
| ALOGV(":%s HDMI supports AC3/EAC3 formats", __func__); |
| out->supported_formats[i++] = AUDIO_FORMAT_AC3; |
| //Adding EAC3/EAC3_JOC formats if AC3 is supported by the sink. |
| //EAC3/EAC3_JOC will be converted to AC3 for decoding if needed |
| out->supported_formats[i++] = AUDIO_FORMAT_E_AC3; |
| out->supported_formats[i++] = AUDIO_FORMAT_E_AC3_JOC; |
| } |
| |
| if (platform_is_edid_supported_format_v2(out->dev->platform, AUDIO_FORMAT_DOLBY_TRUEHD, |
| out->extconn.cs.controller, |
| out->extconn.cs.stream)) { |
| ALOGV(":%s HDMI supports TRUE HD format", __func__); |
| out->supported_formats[i++] = AUDIO_FORMAT_DOLBY_TRUEHD; |
| } |
| |
| if (platform_is_edid_supported_format_v2(out->dev->platform, AUDIO_FORMAT_DTS, |
| out->extconn.cs.controller, |
| out->extconn.cs.stream)) { |
| ALOGV(":%s HDMI supports DTS format", __func__); |
| out->supported_formats[i++] = AUDIO_FORMAT_DTS; |
| } |
| |
| if (platform_is_edid_supported_format_v2(out->dev->platform, AUDIO_FORMAT_DTS_HD, |
| out->extconn.cs.controller, |
| out->extconn.cs.stream)) { |
| ALOGV(":%s HDMI supports DTS HD format", __func__); |
| out->supported_formats[i++] = AUDIO_FORMAT_DTS_HD; |
| } |
| |
| if (platform_is_edid_supported_format_v2(out->dev->platform, AUDIO_FORMAT_IEC61937, |
| out->extconn.cs.controller, |
| out->extconn.cs.stream)) { |
| ALOGV(":%s HDMI supports IEC61937 format", __func__); |
| out->supported_formats[i++] = AUDIO_FORMAT_IEC61937; |
| } |
| |
| |
| // check sample rate caps |
| i = 0; |
| for (j = 0; j < MAX_SUPPORTED_SAMPLE_RATES; j++) { |
| if (platform_is_edid_supported_sample_rate_v2(out->dev->platform, out_hdmi_sample_rates[j], |
| out->extconn.cs.controller, |
| out->extconn.cs.stream)) { |
| ALOGV(":%s HDMI supports sample rate:%d", __func__, out_hdmi_sample_rates[j]); |
| out->supported_sample_rates[i++] = out_hdmi_sample_rates[j]; |
| } |
| } |
| |
| return ret; |
| } |
| |
| static inline ssize_t read_usb_sup_sample_rates(bool is_playback __unused, |
| uint32_t *supported_sample_rates __unused, |
| uint32_t max_rates __unused) |
| { |
| ssize_t count = audio_extn_usb_get_sup_sample_rates(is_playback, |
| supported_sample_rates, |
| max_rates); |
| ssize_t i = 0; |
| |
| for (i=0; i<count; i++) { |
| ALOGV("%s %s %d", __func__, is_playback ? "P" : "C", |
| supported_sample_rates[i]); |
| } |
| return count; |
| } |
| |
| static inline int read_usb_sup_channel_masks(bool is_playback, |
| audio_channel_mask_t *supported_channel_masks, |
| uint32_t max_masks) |
| { |
| int channels = audio_extn_usb_get_max_channels(is_playback); |
| int channel_count; |
| uint32_t num_masks = 0; |
| if (channels > MAX_HIFI_CHANNEL_COUNT) |
| channels = MAX_HIFI_CHANNEL_COUNT; |
| |
| if (is_playback) { |
| // start from 2 channels as framework currently doesn't support mono. |
| if (channels >= FCC_2) { |
| supported_channel_masks[num_masks++] = audio_channel_out_mask_from_count(FCC_2); |
| } |
| for (channel_count = FCC_2; |
| channel_count <= channels && num_masks < max_masks; |
| ++channel_count) { |
| supported_channel_masks[num_masks++] = |
| audio_channel_mask_for_index_assignment_from_count(channel_count); |
| } |
| } else { |
| // For capture we report all supported channel masks from 1 channel up. |
| channel_count = MIN_CHANNEL_COUNT; |
| // audio_channel_in_mask_from_count() does the right conversion to either positional or |
| // indexed mask |
| for ( ; channel_count <= channels && num_masks < max_masks; channel_count++) { |
| audio_channel_mask_t mask = AUDIO_CHANNEL_NONE; |
| if (channel_count <= FCC_2) { |
| mask = audio_channel_in_mask_from_count(channel_count); |
| supported_channel_masks[num_masks++] = mask; |
| } |
| const audio_channel_mask_t index_mask = |
| audio_channel_mask_for_index_assignment_from_count(channel_count); |
| if (mask != index_mask && num_masks < max_masks) { // ensure index mask added. |
| supported_channel_masks[num_masks++] = index_mask; |
| } |
| } |
| } |
| |
| for (size_t i = 0; i < num_masks; ++i) { |
| ALOGV("%s: %s supported ch %d supported_channel_masks[%zu] %08x num_masks %d", __func__, |
| is_playback ? "P" : "C", channels, i, supported_channel_masks[i], num_masks); |
| } |
| return num_masks; |
| } |
| |
| static inline int read_usb_sup_formats(bool is_playback __unused, |
| audio_format_t *supported_formats, |
| uint32_t max_formats __unused) |
| { |
| int bitwidth = audio_extn_usb_get_max_bit_width(is_playback); |
| switch (bitwidth) { |
| case 24: |
| // XXX : usb.c returns 24 for s24 and s24_le? |
| supported_formats[0] = AUDIO_FORMAT_PCM_24_BIT_PACKED; |
| break; |
| case 32: |
| supported_formats[0] = AUDIO_FORMAT_PCM_32_BIT; |
| break; |
| case 16: |
| default : |
| supported_formats[0] = AUDIO_FORMAT_PCM_16_BIT; |
| break; |
| } |
| ALOGV("%s: %s supported format %d", __func__, |
| is_playback ? "P" : "C", bitwidth); |
| return 1; |
| } |
| |
| static inline int read_usb_sup_params_and_compare(bool is_playback, |
| audio_format_t *format, |
| audio_format_t *supported_formats, |
| uint32_t max_formats, |
| audio_channel_mask_t *mask, |
| audio_channel_mask_t *supported_channel_masks, |
| uint32_t max_masks, |
| uint32_t *rate, |
| uint32_t *supported_sample_rates, |
| uint32_t max_rates) { |
| int ret = 0; |
| int num_formats; |
| int num_masks; |
| int num_rates; |
| int i; |
| |
| num_formats = read_usb_sup_formats(is_playback, supported_formats, |
| max_formats); |
| num_masks = read_usb_sup_channel_masks(is_playback, supported_channel_masks, |
| max_masks); |
| |
| num_rates = read_usb_sup_sample_rates(is_playback, |
| supported_sample_rates, max_rates); |
| |
| #define LUT(table, len, what, dflt) \ |
| for (i=0; i<len && (table[i] != what); i++); \ |
| if (i==len) { ret |= (what == dflt ? 0 : -1); what=table[0]; } |
| |
| LUT(supported_formats, num_formats, *format, AUDIO_FORMAT_DEFAULT); |
| LUT(supported_channel_masks, num_masks, *mask, AUDIO_CHANNEL_NONE); |
| LUT(supported_sample_rates, num_rates, *rate, 0); |
| |
| #undef LUT |
| return ret < 0 ? -EINVAL : 0; // HACK TBD |
| } |
| |
| audio_usecase_t get_usecase_id_from_usecase_type(const struct audio_device *adev, |
| usecase_type_t type) |
| { |
| struct audio_usecase *usecase; |
| struct listnode *node; |
| |
| list_for_each(node, &adev->usecase_list) { |
| usecase = node_to_item(node, struct audio_usecase, list); |
| if (usecase->type == type) { |
| ALOGV("%s: usecase id %d", __func__, usecase->id); |
| return usecase->id; |
| } |
| } |
| return USECASE_INVALID; |
| } |
| |
| struct audio_usecase *get_usecase_from_list(const struct audio_device *adev, |
| audio_usecase_t uc_id) |
| { |
| struct audio_usecase *usecase; |
| struct listnode *node; |
| |
| list_for_each(node, &adev->usecase_list) { |
| usecase = node_to_item(node, struct audio_usecase, list); |
| if (usecase->id == uc_id) |
| return usecase; |
| } |
| return NULL; |
| } |
| |
| /* |
| * is a true native playback active |
| */ |
| bool audio_is_true_native_stream_active(struct audio_device *adev) |
| { |
| bool active = false; |
| int i = 0; |
| struct listnode *node; |
| |
| if (NATIVE_AUDIO_MODE_TRUE_44_1 != platform_get_native_support()) { |
| ALOGV("%s:napb: not in true mode or non hdphones device", |
| __func__); |
| active = false; |
| goto exit; |
| } |
| |
| list_for_each(node, &adev->usecase_list) { |
| struct audio_usecase *uc; |
| uc = node_to_item(node, struct audio_usecase, list); |
| struct stream_out *curr_out = |
| (struct stream_out*) uc->stream.out; |
| |
| if (curr_out && PCM_PLAYBACK == uc->type) { |
| ALOGD("%s:napb: (%d) (%s)id (%d) sr %d bw " |
| "(%d) device %s", __func__, i++, use_case_table[uc->id], |
| uc->id, curr_out->sample_rate, |
| curr_out->bit_width, |
| platform_get_snd_device_name(uc->out_snd_device)); |
| |
| if (is_offload_usecase(uc->id) && |
| (curr_out->sample_rate == OUTPUT_SAMPLING_RATE_44100)) { |
| active = true; |
| ALOGD("%s:napb:native stream detected", __func__); |
| } |
| } |
| } |
| exit: |
| return active; |
| } |
| |
| uint32_t adev_get_dsp_bit_width_enforce_mode() |
| { |
| if (adev == NULL) { |
| ALOGE("%s: adev is null. Disable DSP bit width enforce mode.\n", __func__); |
| return 0; |
| } |
| return adev->dsp_bit_width_enforce_mode; |
| } |
| |
| static uint32_t adev_init_dsp_bit_width_enforce_mode(struct mixer *mixer) |
| { |
| char value[PROPERTY_VALUE_MAX]; |
| int trial; |
| uint32_t dsp_bit_width_enforce_mode = 0; |
| |
| if (!mixer) { |
| ALOGE("%s: adev mixer is null. cannot update DSP bitwidth.\n", |
| __func__); |
| return 0; |
| } |
| |
| if (property_get("persist.vendor.audio_hal.dsp_bit_width_enforce_mode", |
| value, NULL) > 0) { |
| trial = atoi(value); |
| switch (trial) { |
| case 16: |
| dsp_bit_width_enforce_mode = 16; |
| break; |
| case 24: |
| dsp_bit_width_enforce_mode = 24; |
| break; |
| case 32: |
| dsp_bit_width_enforce_mode = 32; |
| break; |
| default: |
| dsp_bit_width_enforce_mode = 0; |
| ALOGD("%s Dynamic DSP bitwidth config is disabled.", __func__); |
| break; |
| } |
| } |
| |
| return dsp_bit_width_enforce_mode; |
| } |
| |
| static void audio_enable_asm_bit_width_enforce_mode(struct mixer *mixer, |
| uint32_t enforce_mode, |
| bool enable) |
| { |
| struct mixer_ctl *ctl = NULL; |
| const char *mixer_ctl_name = "ASM Bit Width"; |
| uint32_t asm_bit_width_mode = 0; |
| |
| if (enforce_mode == 0) { |
| ALOGD("%s: DSP bitwidth feature is disabled.", __func__); |
| return; |
| } |
| |
| ctl = mixer_get_ctl_by_name(mixer, mixer_ctl_name); |
| if (!ctl) { |
| ALOGE("%s: Could not get ctl for mixer cmd - %s", |
| __func__, mixer_ctl_name); |
| return; |
| } |
| |
| if (enable) |
| asm_bit_width_mode = enforce_mode; |
| else |
| asm_bit_width_mode = 0; |
| |
| ALOGV("%s DSP bit width feature status is %d width=%d", |
| __func__, enable, asm_bit_width_mode); |
| if (mixer_ctl_set_value(ctl, 0, asm_bit_width_mode) < 0) |
| ALOGE("%s: Could not set ASM biwidth %d", __func__, |
| asm_bit_width_mode); |
| |
| return; |
| } |
| |
| /* |
| * if native DSD playback active |
| */ |
| bool audio_is_dsd_native_stream_active(struct audio_device *adev) |
| { |
| bool active = false; |
| struct listnode *node = NULL; |
| struct audio_usecase *uc = NULL; |
| struct stream_out *curr_out = NULL; |
| |
| list_for_each(node, &adev->usecase_list) { |
| uc = node_to_item(node, struct audio_usecase, list); |
| curr_out = (struct stream_out*) uc->stream.out; |
| |
| if (curr_out && PCM_PLAYBACK == uc->type && |
| (DSD_NATIVE_BACKEND == platform_get_backend_index(uc->out_snd_device))) { |
| active = true; |
| ALOGV("%s:DSD playback is active", __func__); |
| break; |
| } |
| } |
| return active; |
| } |
| |
| static bool force_device_switch(struct audio_usecase *usecase) |
| { |
| bool ret = false; |
| bool is_it_true_mode = false; |
| |
| if (usecase->type == PCM_CAPTURE || |
| usecase->type == TRANSCODE_LOOPBACK_RX || |
| usecase->type == TRANSCODE_LOOPBACK_TX) { |
| return false; |
| } |
| |
| if(usecase->stream.out == NULL) { |
| ALOGE("%s: stream.out is NULL", __func__); |
| return false; |
| } |
| |
| if (is_offload_usecase(usecase->id) && |
| (usecase->stream.out->sample_rate == OUTPUT_SAMPLING_RATE_44100) && |
| (compare_device_type(&usecase->stream.out->device_list, AUDIO_DEVICE_OUT_WIRED_HEADSET) || |
| compare_device_type(&usecase->stream.out->device_list, AUDIO_DEVICE_OUT_WIRED_HEADPHONE))) { |
| is_it_true_mode = (NATIVE_AUDIO_MODE_TRUE_44_1 == platform_get_native_support()? true : false); |
| if ((is_it_true_mode && !adev->native_playback_enabled) || |
| (!is_it_true_mode && adev->native_playback_enabled)){ |
| ret = true; |
| ALOGD("napb: time to toggle native mode"); |
| } |
| } |
| |
| // Force all a2dp output devices to reconfigure for proper AFE encode format |
| //Also handle a case where in earlier a2dp start failed as A2DP stream was |
| //in suspended state, hence try to trigger a retry when we again get a routing request. |
| if(is_a2dp_out_device_type(&usecase->stream.out->device_list) && |
| audio_extn_a2dp_is_force_device_switch()) { |
| ALOGD("Force a2dp device switch to update new encoder config"); |
| ret = true; |
| } |
| |
| if (usecase->stream.out->stream_config_changed) { |
| ALOGD("Force stream_config_changed to update iec61937 transmission config"); |
| return true; |
| } |
| return ret; |
| } |
| |
| static void stream_app_type_cfg_init(struct stream_app_type_cfg *cfg) |
| { |
| cfg->gain[0] = cfg->gain[1] = APP_TYPE_GAIN_DEFAULT; |
| } |
| |
| bool is_btsco_device(snd_device_t out_snd_device, snd_device_t in_snd_device) |
| { |
| bool ret=false; |
| if ((out_snd_device == SND_DEVICE_OUT_BT_SCO || |
| out_snd_device == SND_DEVICE_OUT_BT_SCO_WB || |
| out_snd_device == SND_DEVICE_OUT_BT_SCO_SWB) || |
| in_snd_device == SND_DEVICE_IN_BT_SCO_MIC_WB_NREC || |
| in_snd_device == SND_DEVICE_IN_BT_SCO_MIC_WB || |
| in_snd_device == SND_DEVICE_IN_BT_SCO_MIC_SWB || |
| in_snd_device == SND_DEVICE_IN_BT_SCO_MIC_NREC || |
| in_snd_device == SND_DEVICE_IN_BT_SCO_MIC || |
| in_snd_device == SND_DEVICE_IN_BT_SCO_MIC_SWB_NREC) |
| ret = true; |
| |
| return ret; |
| } |
| |
| bool is_a2dp_device(snd_device_t out_snd_device) |
| { |
| bool ret=false; |
| if (out_snd_device == SND_DEVICE_OUT_BT_A2DP) |
| ret = true; |
| |
| return ret; |
| } |
| |
| bool is_bt_soc_on(struct audio_device *adev) |
| { |
| struct mixer_ctl *ctl; |
| char *mixer_ctl_name = "BT SOC status"; |
| ctl = mixer_get_ctl_by_name(adev->mixer, mixer_ctl_name); |
| bool bt_soc_status = true; |
| if (!ctl) { |
| ALOGE("%s: Could not get ctl for mixer cmd - %s", |
| __func__, mixer_ctl_name); |
| /*This is to ensure we dont break targets which dont have the kernel change*/ |
| return true; |
| } |
| bt_soc_status = mixer_ctl_get_value(ctl, 0); |
| ALOGD("BT SOC status: %d",bt_soc_status); |
| return bt_soc_status; |
| } |
| |
| static int configure_btsco_sample_rate(snd_device_t snd_device) |
| { |
| struct mixer_ctl *ctl = NULL; |
| struct mixer_ctl *ctl_sr_rx = NULL, *ctl_sr_tx = NULL, *ctl_sr = NULL; |
| char *rate_str = NULL; |
| bool is_rx_dev = true; |
| |
| if (is_btsco_device(snd_device, snd_device)) { |
| ctl_sr_tx = mixer_get_ctl_by_name(adev->mixer, "BT SampleRate TX"); |
| ctl_sr_rx = mixer_get_ctl_by_name(adev->mixer, "BT SampleRate RX"); |
| if (!ctl_sr_tx || !ctl_sr_rx) { |
| ctl_sr = mixer_get_ctl_by_name(adev->mixer, "BT SampleRate"); |
| if (!ctl_sr) |
| return -ENOSYS; |
| } |
| |
| switch (snd_device) { |
| case SND_DEVICE_OUT_BT_SCO: |
| rate_str = "KHZ_8"; |
| break; |
| case SND_DEVICE_IN_BT_SCO_MIC_NREC: |
| case SND_DEVICE_IN_BT_SCO_MIC: |
| rate_str = "KHZ_8"; |
| is_rx_dev = false; |
| break; |
| case SND_DEVICE_OUT_BT_SCO_WB: |
| rate_str = "KHZ_16"; |
| break; |
| case SND_DEVICE_IN_BT_SCO_MIC_WB_NREC: |
| case SND_DEVICE_IN_BT_SCO_MIC_WB: |
| rate_str = "KHZ_16"; |
| is_rx_dev = false; |
| break; |
| default: |
| return 0; |
| } |
| |
| ctl = (ctl_sr == NULL) ? (is_rx_dev ? ctl_sr_rx : ctl_sr_tx) : ctl_sr; |
| if (mixer_ctl_set_enum_by_string(ctl, rate_str) != 0) |
| return -ENOSYS; |
| } |
| return 0; |
| } |
| |
| int out_standby_l(struct audio_stream *stream); |
| |
| struct stream_in *adev_get_active_input(const struct audio_device *adev) |
| { |
| struct listnode *node; |
| struct stream_in *last_active_in = NULL; |
| |
| /* Get last added active input. |
| * TODO: We may use a priority mechanism to pick highest priority active source */ |
| list_for_each(node, &adev->usecase_list) |
| { |
| struct audio_usecase *usecase = node_to_item(node, struct audio_usecase, list); |
| if (usecase->type == PCM_CAPTURE && usecase->stream.in != NULL) |
| last_active_in = usecase->stream.in; |
| } |
| |
| return last_active_in; |
| } |
| |
| struct stream_in *get_voice_communication_input(const struct audio_device *adev) |
| { |
| struct listnode *node; |
| |
| /* First check active inputs with voice communication source and then |
| * any input if audio mode is in communication */ |
| list_for_each(node, &adev->usecase_list) |
| { |
| struct audio_usecase *usecase = node_to_item(node, struct audio_usecase, list); |
| if (usecase->type == PCM_CAPTURE && usecase->stream.in != NULL && |
| usecase->stream.in->source == AUDIO_SOURCE_VOICE_COMMUNICATION) |
| return usecase->stream.in; |
| } |
| if (adev->mode == AUDIO_MODE_IN_COMMUNICATION) |
| return adev_get_active_input(adev); |
| |
| return NULL; |
| } |
| |
| /* |
| * Aligned with policy.h |
| */ |
| static inline int source_priority(int inputSource) |
| { |
| switch (inputSource) { |
| case AUDIO_SOURCE_VOICE_COMMUNICATION: |
| return 9; |
| case AUDIO_SOURCE_CAMCORDER: |
| return 8; |
| case AUDIO_SOURCE_VOICE_PERFORMANCE: |
| return 7; |
| case AUDIO_SOURCE_UNPROCESSED: |
| return 6; |
| case AUDIO_SOURCE_MIC: |
| return 5; |
| case AUDIO_SOURCE_ECHO_REFERENCE: |
| return 4; |
| case AUDIO_SOURCE_FM_TUNER: |
| return 3; |
| case AUDIO_SOURCE_VOICE_RECOGNITION: |
| return 2; |
| case AUDIO_SOURCE_HOTWORD: |
| return 1; |
| default: |
| break; |
| } |
| return 0; |
| } |
| |
| static struct stream_in *get_priority_input(struct audio_device *adev) |
| { |
| struct listnode *node; |
| struct audio_usecase *usecase; |
| int last_priority = 0, priority; |
| struct stream_in *priority_in = NULL; |
| struct stream_in *in; |
| |
| list_for_each(node, &adev->usecase_list) { |
| usecase = node_to_item(node, struct audio_usecase, list); |
| if (usecase->type == PCM_CAPTURE) { |
| in = usecase->stream.in; |
| if (!in) |
| continue; |
| priority = source_priority(in->source); |
| |
| if (priority > last_priority) { |
| last_priority = priority; |
| priority_in = in; |
| } |
| } |
| } |
| return priority_in; |
| } |
| |
| int select_devices(struct audio_device *adev, audio_usecase_t uc_id) |
| { |
| snd_device_t out_snd_device = SND_DEVICE_NONE; |
| snd_device_t in_snd_device = SND_DEVICE_NONE; |
| struct audio_usecase *usecase = NULL; |
| struct audio_usecase *vc_usecase = NULL; |
| struct audio_usecase *voip_usecase = NULL; |
| struct audio_usecase *hfp_usecase = NULL; |
| struct stream_out stream_out; |
| audio_usecase_t hfp_ucid; |
| int status = 0; |
| |
| ALOGD("%s for use case (%s)", __func__, use_case_table[uc_id]); |
| |
| usecase = get_usecase_from_list(adev, uc_id); |
| if (usecase == NULL) { |
| ALOGE("%s: Could not find the usecase(%d)", __func__, uc_id); |
| return -EINVAL; |
| } |
| |
| if ((usecase->type == VOICE_CALL) || |
| (usecase->type == VOIP_CALL) || |
| (usecase->type == PCM_HFP_CALL)|| |
| (usecase->type == ICC_CALL) || |
| (usecase->type == SYNTH_LOOPBACK)) { |
| if(usecase->stream.out == NULL) { |
| ALOGE("%s: stream.out is NULL", __func__); |
| return -EINVAL; |
| } |
| if (compare_device_type(&usecase->device_list, AUDIO_DEVICE_OUT_BUS)) { |
| out_snd_device = audio_extn_auto_hal_get_output_snd_device(adev, |
| uc_id); |
| in_snd_device = audio_extn_auto_hal_get_input_snd_device(adev, |
| uc_id); |
| } else { |
| out_snd_device = platform_get_output_snd_device(adev->platform, |
| usecase->stream.out, usecase->type); |
| in_snd_device = platform_get_input_snd_device(adev->platform, |
| NULL, |
| &usecase->stream.out->device_list, |
| usecase->type); |
| } |
| assign_devices(&usecase->device_list, &usecase->stream.out->device_list); |
| } else if (usecase->type == TRANSCODE_LOOPBACK_RX) { |
| if (usecase->stream.inout == NULL) { |
| ALOGE("%s: stream.inout is NULL", __func__); |
| return -EINVAL; |
| } |
| assign_devices(&stream_out.device_list, &usecase->stream.inout->out_config.device_list); |
| stream_out.sample_rate = usecase->stream.inout->out_config.sample_rate; |
| stream_out.format = usecase->stream.inout->out_config.format; |
| stream_out.channel_mask = usecase->stream.inout->out_config.channel_mask; |
| out_snd_device = platform_get_output_snd_device(adev->platform, &stream_out, usecase->type); |
| assign_devices(&usecase->device_list, |
| &usecase->stream.inout->out_config.device_list); |
| } else if (usecase->type == TRANSCODE_LOOPBACK_TX ) { |
| if (usecase->stream.inout == NULL) { |
| ALOGE("%s: stream.inout is NULL", __func__); |
| return -EINVAL; |
| } |
| struct listnode out_devices; |
| list_init(&out_devices); |
| in_snd_device = platform_get_input_snd_device(adev->platform, NULL, |
| &out_devices, usecase->type); |
| assign_devices(&usecase->device_list, |
| &usecase->stream.inout->in_config.device_list); |
| } else { |
| /* |
| * If the voice call is active, use the sound devices of voice call usecase |
| * so that it would not result any device switch. All the usecases will |
| * be switched to new device when select_devices() is called for voice call |
| * usecase. This is to avoid switching devices for voice call when |
| * check_usecases_codec_backend() is called below. |
| * choose voice call device only if the use case device is |
| * also using the codec backend |
| */ |
| if (voice_is_in_call(adev) && adev->mode != AUDIO_MODE_NORMAL) { |
| vc_usecase = get_usecase_from_list(adev, |
| get_usecase_id_from_usecase_type(adev, VOICE_CALL)); |
| if ((vc_usecase) && ((is_codec_backend_out_device_type(&vc_usecase->device_list) && |
| is_codec_backend_out_device_type(&usecase->device_list)) || |
| (is_codec_backend_out_device_type(&vc_usecase->device_list) && |
| is_codec_backend_in_device_type(&usecase->device_list)) || |
| is_single_device_type_equal(&vc_usecase->device_list, |
| AUDIO_DEVICE_OUT_HEARING_AID) || |
| is_single_device_type_equal(&usecase->device_list, |
| AUDIO_DEVICE_IN_VOICE_CALL) || |
| (is_single_device_type_equal(&usecase->device_list, |
| AUDIO_DEVICE_IN_USB_HEADSET) && |
| is_single_device_type_equal(&vc_usecase->device_list, |
| AUDIO_DEVICE_OUT_USB_HEADSET))|| |
| (is_single_device_type_equal(&usecase->device_list, |
| AUDIO_DEVICE_IN_BLUETOOTH_SCO_HEADSET) && |
| is_codec_backend_out_device_type(&vc_usecase->device_list)))) { |
| in_snd_device = vc_usecase->in_snd_device; |
| out_snd_device = vc_usecase->out_snd_device; |
| } |
| } else if (voice_extn_compress_voip_is_active(adev)) { |
| bool out_snd_device_backend_match = true; |
| voip_usecase = get_usecase_from_list(adev, USECASE_COMPRESS_VOIP_CALL); |
| if ((voip_usecase != NULL) && |
| (usecase->type == PCM_PLAYBACK) && |
| (usecase->stream.out != NULL)) { |
| out_snd_device_backend_match = platform_check_backends_match( |
| voip_usecase->out_snd_device, |
| platform_get_output_snd_device( |
| adev->platform, |
| usecase->stream.out, usecase->type)); |
| } |
| if ((voip_usecase) && (is_codec_backend_out_device_type(&voip_usecase->device_list) && |
| (is_codec_backend_out_device_type(&usecase->device_list) || |
| is_codec_backend_in_device_type(&usecase->device_list)) && |
| out_snd_device_backend_match && |
| (voip_usecase->stream.out != adev->primary_output))) { |
| in_snd_device = voip_usecase->in_snd_device; |
| out_snd_device = voip_usecase->out_snd_device; |
| } |
| } else if (audio_extn_hfp_is_active(adev)) { |
| hfp_ucid = audio_extn_hfp_get_usecase(); |
| hfp_usecase = get_usecase_from_list(adev, hfp_ucid); |
| if ((hfp_usecase) && is_codec_backend_out_device_type(&hfp_usecase->device_list)) { |
| in_snd_device = hfp_usecase->in_snd_device; |
| out_snd_device = hfp_usecase->out_snd_device; |
| } |
| } |
| if (usecase->type == PCM_PLAYBACK) { |
| if (usecase->stream.out == NULL) { |
| ALOGE("%s: stream.out is NULL", __func__); |
| return -EINVAL; |
| } |
| assign_devices(&usecase->device_list, &usecase->stream.out->device_list); |
| in_snd_device = SND_DEVICE_NONE; |
| if (out_snd_device == SND_DEVICE_NONE) { |
| struct stream_out *voip_out = adev->primary_output; |
| struct stream_in *voip_in = get_voice_communication_input(adev); |
| if (compare_device_type(&usecase->device_list, AUDIO_DEVICE_OUT_BUS)) |
| out_snd_device = audio_extn_auto_hal_get_output_snd_device(adev, uc_id); |
| else |
| out_snd_device = platform_get_output_snd_device(adev->platform, |
| usecase->stream.out, |
| usecase->type); |
| voip_usecase = get_usecase_from_list(adev, USECASE_AUDIO_PLAYBACK_VOIP); |
| |
| if (voip_usecase) |
| voip_out = voip_usecase->stream.out; |
| |
| if (usecase->stream.out == voip_out && voip_in != NULL) |
| select_devices(adev, voip_in->usecase); |
| } |
| } else if (usecase->type == PCM_CAPTURE) { |
| if (usecase->stream.in == NULL) { |
| ALOGE("%s: stream.in is NULL", __func__); |
| return -EINVAL; |
| } |
| assign_devices(&usecase->device_list, &usecase->stream.in->device_list); |
| out_snd_device = SND_DEVICE_NONE; |
| if (in_snd_device == SND_DEVICE_NONE) { |
| struct listnode out_devices; |
| struct stream_in *voip_in = get_voice_communication_input(adev); |
| struct stream_in *priority_in = NULL; |
| |
| list_init(&out_devices); |
| if (voip_in != NULL) { |
| struct audio_usecase *voip_usecase = get_usecase_from_list(adev, |
| USECASE_AUDIO_PLAYBACK_VOIP); |
| |
| usecase->stream.in->enable_ec_port = false; |
| |
| bool is_ha_usecase = adev->ha_proxy_enable ? |
| usecase->id == USECASE_AUDIO_RECORD_AFE_PROXY2 : |
| usecase->id == USECASE_AUDIO_RECORD_AFE_PROXY; |
| if (is_ha_usecase) { |
| reassign_device_list(&out_devices, AUDIO_DEVICE_OUT_TELEPHONY_TX, ""); |
| } else if (voip_usecase) { |
| assign_devices(&out_devices, &voip_usecase->stream.out->device_list); |
| } else if (adev->primary_output && |
| !adev->primary_output->standby) { |
| assign_devices(&out_devices, &adev->primary_output->device_list); |
| } else { |
| /* forcing speaker o/p device to get matching i/p pair |
| in case o/p is not routed from same primary HAL */ |
| reassign_device_list(&out_devices, AUDIO_DEVICE_OUT_SPEAKER, ""); |
| } |
| priority_in = voip_in; |
| } else { |
| /* get the input with the highest priority source*/ |
| priority_in = get_priority_input(adev); |
| |
| if (!priority_in || |
| audio_extn_auto_hal_overwrite_priority_for_auto(usecase->stream.in)) |
| priority_in = usecase->stream.in; |
| } |
| if (compare_device_type(&usecase->device_list, AUDIO_DEVICE_IN_BUS)){ |
| in_snd_device = audio_extn_auto_hal_get_snd_device_for_car_audio_stream(priority_in->car_audio_stream); |
| } |
| else |
| in_snd_device = platform_get_input_snd_device(adev->platform, |
| priority_in, |
| &out_devices, |
| usecase->type); |
| } |
| } |
| } |
| |
| if (out_snd_device == usecase->out_snd_device && |
| in_snd_device == usecase->in_snd_device) { |
| |
| if (!force_device_switch(usecase)) |
| return 0; |
| } |
| |
| if (!compare_device_type(&usecase->device_list, AUDIO_DEVICE_OUT_BUS) && |
| ((is_btsco_device(out_snd_device,in_snd_device) && !adev->bt_sco_on) || |
| (is_a2dp_device(out_snd_device) && !audio_extn_a2dp_source_is_ready()))) { |
| ALOGD("SCO/A2DP is selected but they are not connected/ready hence dont route"); |
| return 0; |
| } |
| |
| if (out_snd_device != SND_DEVICE_NONE && |
| out_snd_device != adev->last_logged_snd_device[uc_id][0]) { |
| ALOGD("%s: changing use case %s output device from(%d: %s, acdb %d) to (%d: %s, acdb %d)", |
| __func__, |
| use_case_table[uc_id], |
| adev->last_logged_snd_device[uc_id][0], |
| platform_get_snd_device_name(adev->last_logged_snd_device[uc_id][0]), |
| adev->last_logged_snd_device[uc_id][0] != SND_DEVICE_NONE ? |
| platform_get_snd_device_acdb_id(adev->last_logged_snd_device[uc_id][0]) : |
| -1, |
| out_snd_device, |
| platform_get_snd_device_name(out_snd_device), |
| platform_get_snd_device_acdb_id(out_snd_device)); |
| adev->last_logged_snd_device[uc_id][0] = out_snd_device; |
| } |
| if (in_snd_device != SND_DEVICE_NONE && |
| in_snd_device != adev->last_logged_snd_device[uc_id][1]) { |
| ALOGD("%s: changing use case %s input device from(%d: %s, acdb %d) to (%d: %s, acdb %d)", |
| __func__, |
| use_case_table[uc_id], |
| adev->last_logged_snd_device[uc_id][1], |
| platform_get_snd_device_name(adev->last_logged_snd_device[uc_id][1]), |
| adev->last_logged_snd_device[uc_id][1] != SND_DEVICE_NONE ? |
| platform_get_snd_device_acdb_id(adev->last_logged_snd_device[uc_id][1]) : |
| -1, |
| in_snd_device, |
| platform_get_snd_device_name(in_snd_device), |
| platform_get_snd_device_acdb_id(in_snd_device)); |
| adev->last_logged_snd_device[uc_id][1] = in_snd_device; |
| } |
| |
| |
| /* |
| * Limitation: While in call, to do a device switch we need to disable |
| * and enable both RX and TX devices though one of them is same as current |
| * device. |
| */ |
| if ((usecase->type == VOICE_CALL) && |
| (usecase->in_snd_device != SND_DEVICE_NONE) && |
| (usecase->out_snd_device != SND_DEVICE_NONE)) { |
| status = platform_switch_voice_call_device_pre(adev->platform); |
| } |
| |
| if (((usecase->type == VOICE_CALL) || |
| (usecase->type == VOIP_CALL)) && |
| (usecase->out_snd_device != SND_DEVICE_NONE)) { |
| /* Disable sidetone only if voice/voip call already exists */ |
| if (voice_is_call_state_active_in_call(adev) || |
| voice_extn_compress_voip_is_started(adev)) |
| voice_set_sidetone(adev, usecase->out_snd_device, false); |
| |
| /* Disable aanc only if voice call exists */ |
| if (voice_is_call_state_active_in_call(adev)) |
| voice_check_and_update_aanc_path(adev, usecase->out_snd_device, false); |
| } |
| |
| if ((out_snd_device == SND_DEVICE_OUT_SPEAKER_AND_BT_A2DP || |
| out_snd_device == SND_DEVICE_OUT_SPEAKER_SAFE_AND_BT_A2DP) && |
| (!audio_extn_a2dp_source_is_ready())) { |
| ALOGW("%s: A2DP profile is not ready, routing to speaker only", __func__); |
| if (out_snd_device == SND_DEVICE_OUT_SPEAKER_SAFE_AND_BT_A2DP) |
| out_snd_device = SND_DEVICE_OUT_SPEAKER_SAFE; |
| else |
| out_snd_device = SND_DEVICE_OUT_SPEAKER; |
| } |
| |
| /* Disable current sound devices */ |
| if (usecase->out_snd_device != SND_DEVICE_NONE) { |
| disable_audio_route(adev, usecase); |
| disable_snd_device(adev, usecase->out_snd_device); |
| } |
| |
| if (usecase->in_snd_device != SND_DEVICE_NONE) { |
| disable_audio_route(adev, usecase); |
| disable_snd_device(adev, usecase->in_snd_device); |
| } |
| |
| /* Applicable only on the targets that has external modem. |
| * New device information should be sent to modem before enabling |
| * the devices to reduce in-call device switch time. |
| */ |
| if ((usecase->type == VOICE_CALL) && |
| (usecase->in_snd_device != SND_DEVICE_NONE) && |
| (usecase->out_snd_device != SND_DEVICE_NONE)) { |
| status = platform_switch_voice_call_enable_device_config(adev->platform, |
| out_snd_device, |
| in_snd_device); |
| } |
| |
| /* Enable new sound devices */ |
| if (out_snd_device != SND_DEVICE_NONE) { |
| check_usecases_codec_backend(adev, usecase, out_snd_device); |
| check_and_configure_headphone(adev, usecase, out_snd_device); |
| if (platform_check_codec_asrc_support(adev->platform)) |
| check_and_set_asrc_mode(adev, usecase, out_snd_device); |
| enable_snd_device(adev, out_snd_device); |
| /* Enable haptics device for haptic usecase */ |
| if (usecase->id == USECASE_AUDIO_PLAYBACK_WITH_HAPTICS) |
| enable_snd_device(adev, SND_DEVICE_OUT_HAPTICS); |
| } |
| |
| if (in_snd_device != SND_DEVICE_NONE) { |
| check_usecases_capture_codec_backend(adev, usecase, in_snd_device); |
| enable_snd_device(adev, in_snd_device); |
| } |
| |
| if (usecase->type == VOICE_CALL || usecase->type == VOIP_CALL) |
| status = platform_switch_voice_call_device_post(adev->platform, |
| out_snd_device, |
| in_snd_device); |
| |
| usecase->in_snd_device = in_snd_device; |
| usecase->out_snd_device = out_snd_device; |
| |
| audio_extn_utils_update_stream_app_type_cfg_for_usecase(adev, |
| usecase); |
| if (usecase->type == PCM_PLAYBACK) { |
| if ((24 == usecase->stream.out->bit_width) && |
| compare_device_type(&usecase->stream.out->device_list, AUDIO_DEVICE_OUT_SPEAKER)) { |
| usecase->stream.out->app_type_cfg.sample_rate = DEFAULT_OUTPUT_SAMPLING_RATE; |
| } else if ((out_snd_device == SND_DEVICE_OUT_HDMI || |
| out_snd_device == SND_DEVICE_OUT_USB_HEADSET || |
| out_snd_device == SND_DEVICE_OUT_DISPLAY_PORT) && |
| (usecase->stream.out->sample_rate >= OUTPUT_SAMPLING_RATE_44100)) { |
| /* |
| * To best utlize DSP, check if the stream sample rate is supported/multiple of |
| * configured device sample rate, if not update the COPP rate to be equal to the |
| * device sample rate, else open COPP at stream sample rate |
| */ |
| platform_check_and_update_copp_sample_rate(adev->platform, out_snd_device, |
| usecase->stream.out->sample_rate, |
| &usecase->stream.out->app_type_cfg.sample_rate); |
| } else if (((out_snd_device != SND_DEVICE_OUT_HEADPHONES_44_1 && |
| out_snd_device != SND_DEVICE_OUT_HEADPHONES && |
| out_snd_device != SND_DEVICE_OUT_HEADPHONES_HIFI_FILTER && |
| !audio_is_true_native_stream_active(adev)) && |
| usecase->stream.out->sample_rate == OUTPUT_SAMPLING_RATE_44100) || |
| (usecase->stream.out->sample_rate < OUTPUT_SAMPLING_RATE_44100)) { |
| usecase->stream.out->app_type_cfg.sample_rate = DEFAULT_OUTPUT_SAMPLING_RATE; |
| } |
| } |
| enable_audio_route(adev, usecase); |
| |
| if (uc_id == USECASE_AUDIO_PLAYBACK_VOIP) { |
| struct stream_in *voip_in = get_voice_communication_input(adev); |
| struct audio_usecase *voip_in_usecase = NULL; |
| voip_in_usecase = get_usecase_from_list(adev, USECASE_AUDIO_RECORD_VOIP); |
| if (voip_in != NULL && |
| voip_in_usecase != NULL && |
| !(out_snd_device == AUDIO_DEVICE_OUT_SPEAKER || |
| out_snd_device == AUDIO_DEVICE_OUT_SPEAKER_SAFE) && |
| (voip_in_usecase->in_snd_device == |
| platform_get_input_snd_device(adev->platform, voip_in, |
| &usecase->stream.out->device_list,usecase->type))) { |
| /* |
| * if VOIP TX is enabled before VOIP RX, needs to re-route the TX path |
| * for enabling echo-reference-voip with correct port |
| */ |
| ALOGD("%s: VOIP TX is enabled before VOIP RX,needs to re-route the TX path",__func__); |
| disable_audio_route(adev, voip_in_usecase); |
| disable_snd_device(adev, voip_in_usecase->in_snd_device); |
| enable_snd_device(adev, voip_in_usecase->in_snd_device); |
| enable_audio_route(adev, voip_in_usecase); |
| } |
| } |
| |
| |
| audio_extn_qdsp_set_device(usecase); |
| |
| /* If input stream is already running then effect needs to be |
| applied on the new input device that's being enabled here. */ |
| if (in_snd_device != SND_DEVICE_NONE) |
| check_and_enable_effect(adev); |
| |
| if (usecase->type == VOICE_CALL || usecase->type == VOIP_CALL) { |
| /* Enable aanc only if voice call exists */ |
| if (voice_is_call_state_active_in_call(adev)) |
| voice_check_and_update_aanc_path(adev, out_snd_device, true); |
| |
| /* Enable sidetone only if other voice/voip call already exists */ |
| if (voice_is_call_state_active_in_call(adev) || |
| voice_extn_compress_voip_is_started(adev)) |
| voice_set_sidetone(adev, out_snd_device, true); |
| } |
| |
| /* Applicable only on the targets that has external modem. |
| * Enable device command should be sent to modem only after |
| * enabling voice call mixer controls |
| */ |
| if (usecase->type == VOICE_CALL) |
| status = platform_switch_voice_call_usecase_route_post(adev->platform, |
| out_snd_device, |
| in_snd_device); |
| |
| if (is_btsco_device(out_snd_device, in_snd_device) || is_a2dp_device(out_snd_device)) { |
| struct stream_in *in = adev_get_active_input(adev); |
| if (usecase->type == VOIP_CALL) { |
| if (in != NULL && !in->standby) { |
| if (is_bt_soc_on(adev) == false){ |
| ALOGD("BT SCO MIC disconnected while in connection"); |
| if (in->pcm != NULL) |
| pcm_stop(in->pcm); |
| } |
| } |
| if ((usecase->stream.out != NULL) && (usecase->stream.out != adev->primary_output) |
| && usecase->stream.out->started) { |
| if (is_bt_soc_on(adev) == false) { |
| ALOGD("BT SCO/A2DP disconnected while in connection"); |
| out_standby_l(&usecase->stream.out->stream.common); |
| } |
| } |
| } else if ((usecase->stream.out != NULL) && |
| !(usecase->stream.out->flags & AUDIO_OUTPUT_FLAG_COMPRESS_OFFLOAD) && |
| (usecase->type != TRANSCODE_LOOPBACK_TX) && |
| (usecase->type != TRANSCODE_LOOPBACK_RX) && |
| (usecase->type != PCM_CAPTURE) && |
| usecase->stream.out->started) { |
| if (is_bt_soc_on(adev) == false) { |
| ALOGD("BT SCO/A2dp disconnected while in connection"); |
| out_standby_l(&usecase->stream.out->stream.common); |
| } |
| } |
| } |
| |
| if (usecase->type != PCM_CAPTURE && usecase == voip_usecase) { |
| struct stream_out *voip_out = voip_usecase->stream.out; |
| audio_extn_utils_send_app_type_gain(adev, |
| voip_out->app_type_cfg.app_type, |
| &voip_out->app_type_cfg.gain[0]); |
| } |
| |
| ALOGD("%s: done",__func__); |
| |
| return status; |
| } |
| |
| static int stop_input_stream(struct stream_in *in) |
| { |
| int ret = 0; |
| struct audio_usecase *uc_info; |
| |
| if (in == NULL) { |
| ALOGE("%s: stream_in ptr is NULL", __func__); |
| return -EINVAL; |
| } |
| |
| struct audio_device *adev = in->dev; |
| struct stream_in *priority_in = NULL; |
| |
| ALOGV("%s: enter: usecase(%d: %s)", __func__, |
| in->usecase, use_case_table[in->usecase]); |
| uc_info = get_usecase_from_list(adev, in->usecase); |
| if (uc_info == NULL) { |
| ALOGE("%s: Could not find the usecase (%d) in the list", |
| __func__, in->usecase); |
| return -EINVAL; |
| } |
| |
| priority_in = get_priority_input(adev); |
| |
| if (audio_extn_ext_hw_plugin_usecase_stop(adev->ext_hw_plugin, uc_info)) |
| ALOGE("%s: failed to stop ext hw plugin", __func__); |
| |
| /* Close in-call recording streams */ |
| voice_check_and_stop_incall_rec_usecase(adev, in); |
| |
| /* 1. Disable stream specific mixer controls */ |
| disable_audio_route(adev, uc_info); |
| |
| /* 2. Disable the tx device */ |
| disable_snd_device(adev, uc_info->in_snd_device); |
| |
| if (is_loopback_input_device(get_device_types(&in->device_list))) |
| audio_extn_keep_alive_stop(KEEP_ALIVE_OUT_PRIMARY); |
| |
| list_remove(&uc_info->list); |
| free(uc_info); |
| |
| if (priority_in == in) { |
| priority_in = get_priority_input(adev); |
| if (priority_in) { |
| if (is_usb_in_device_type(&priority_in->device_list)) { |
| if (audio_extn_usb_connected(NULL)) |
| select_devices(adev, priority_in->usecase); |
| } else { |
| select_devices(adev, priority_in->usecase); |
| } |
| } |
| } |
| |
| enable_gcov(); |
| ALOGV("%s: exit: status(%d)", __func__, ret); |
| return ret; |
| } |
| |
| int start_input_stream(struct stream_in *in) |
| { |
| /* 1. Enable output device and stream routing controls */ |
| int ret = 0; |
| struct audio_usecase *uc_info; |
| |
| if (in == NULL) { |
| ALOGE("%s: stream_in ptr is NULL", __func__); |
| return -EINVAL; |
| } |
| |
| struct audio_device *adev = in->dev; |
| struct pcm_config config = in->config; |
| int usecase = platform_update_usecase_from_source(in->source,in->usecase); |
| |
| if (get_usecase_from_list(adev, usecase) == NULL) |
| in->usecase = usecase; |
| ALOGD("%s: enter: stream(%p)usecase(%d: %s)", |
| __func__, &in->stream, in->usecase, use_case_table[in->usecase]); |
| |
| if (CARD_STATUS_OFFLINE == in->card_status|| |
| CARD_STATUS_OFFLINE == adev->card_status) { |
| ALOGW("in->card_status or adev->card_status offline, try again"); |
| ret = -EIO; |
| goto error_config; |
| } |
| |
| if (is_sco_in_device_type(&in->device_list)) { |
| if (!adev->bt_sco_on || audio_extn_a2dp_source_is_ready()) { |
| ALOGE("%s: SCO profile is not ready, return error", __func__); |
| ret = -EIO; |
| goto error_config; |
| } |
| } |
| |
| /* Check if source matches incall recording usecase criteria */ |
| ret = voice_check_and_set_incall_rec_usecase(adev, in); |
| if (ret) |
| goto error_config; |
| else |
| ALOGV("%s: usecase(%d)", __func__, in->usecase); |
| |
| if (audio_extn_cin_attached_usecase(in)) |
| audio_extn_cin_acquire_usecase(in); |
| |
| if (get_usecase_from_list(adev, in->usecase) != NULL) { |
| ALOGE("%s: use case assigned already in use, stream(%p)usecase(%d: %s)", |
| __func__, &in->stream, in->usecase, use_case_table[in->usecase]); |
| ret = -EINVAL; |
| goto error_config; |
| } |
| |
| in->pcm_device_id = platform_get_pcm_device_id(in->usecase, PCM_CAPTURE); |
| if (in->pcm_device_id < 0) { |
| ALOGE("%s: Could not find PCM device id for the usecase(%d)", |
| __func__, in->usecase); |
| ret = -EINVAL; |
| goto error_config; |
| } |
| |
| uc_info = (struct audio_usecase *)calloc(1, sizeof(struct audio_usecase)); |
| |
| if (!uc_info) { |
| ret = -ENOMEM; |
| goto error_config; |
| } |
| |
| uc_info->id = in->usecase; |
| uc_info->type = PCM_CAPTURE; |
| uc_info->stream.in = in; |
| list_init(&uc_info->device_list); |
| assign_devices(&uc_info->device_list, &in->device_list); |
| uc_info->in_snd_device = SND_DEVICE_NONE; |
| uc_info->out_snd_device = SND_DEVICE_NONE; |
| |
| list_add_tail(&adev->usecase_list, &uc_info->list); |
| audio_streaming_hint_start(); |
| audio_extn_perf_lock_acquire(&adev->perf_lock_handle, 0, |
| adev->perf_lock_opts, |
| adev->perf_lock_opts_size); |
| select_devices(adev, in->usecase); |
| |
| if (audio_extn_ext_hw_plugin_usecase_start(adev->ext_hw_plugin, uc_info)) |
| ALOGE("%s: failed to start ext hw plugin", __func__); |
| |
| android_atomic_acquire_cas(true, false, &(in->capture_stopped)); |
| |
| if (audio_extn_cin_attached_usecase(in)) { |
| ret = audio_extn_cin_open_input_stream(in); |
| if (ret) |
| goto error_open; |
| else |
| goto done_open; |
| } |
| |
| if (in->usecase == USECASE_AUDIO_RECORD_MMAP) { |
| if (in->pcm == NULL || !pcm_is_ready(in->pcm)) { |
| ALOGE("%s: pcm stream not ready", __func__); |
| goto error_open; |
| } |
| ret = pcm_start(in->pcm); |
| if (ret < 0) { |
| ALOGE("%s: MMAP pcm_start failed ret %d", __func__, ret); |
| goto error_open; |
| } |
| } else { |
| unsigned int flags = PCM_IN | PCM_MONOTONIC; |
| unsigned int pcm_open_retry_count = 0; |
| |
| if ((in->usecase == USECASE_AUDIO_RECORD_AFE_PROXY) || |
| (in->usecase == USECASE_AUDIO_RECORD_AFE_PROXY2)) { |
| flags |= PCM_MMAP | PCM_NOIRQ; |
| pcm_open_retry_count = PROXY_OPEN_RETRY_COUNT; |
| } else if (in->realtime) { |
| flags |= PCM_MMAP | PCM_NOIRQ; |
| } |
| |
| if (audio_extn_ffv_get_stream() == in) { |
| ALOGD("%s: ffv stream, update pcm config", __func__); |
| audio_extn_ffv_update_pcm_config(&config); |
| } |
| ALOGV("%s: Opening PCM device card_id(%d) device_id(%d), channels %d", |
| __func__, adev->snd_card, in->pcm_device_id, in->config.channels); |
| |
| while (1) { |
| ATRACE_BEGIN("pcm_in_open"); |
| in->pcm = pcm_open(adev->snd_card, in->pcm_device_id, |
| flags, &config); |
| ATRACE_END(); |
| if (errno == ENETRESET && !pcm_is_ready(in->pcm)) { |
| ALOGE("%s: pcm_open failed errno:%d\n", __func__, errno); |
| adev->card_status = CARD_STATUS_OFFLINE; |
| in->card_status = CARD_STATUS_OFFLINE; |
| ret = -EIO; |
| goto error_open; |
| } |
| |
| if (in->pcm == NULL || !pcm_is_ready(in->pcm)) { |
| ALOGE("%s: %s", __func__, pcm_get_error(in->pcm)); |
| if (in->pcm != NULL) { |
| pcm_close(in->pcm); |
| in->pcm = NULL; |
| } |
| if (pcm_open_retry_count == 0) { |
| ret = -EIO; |
| goto error_open; |
| } |
| pcm_open_retry_count--; |
| usleep(PROXY_OPEN_WAIT_TIME * 1000); |
| continue; |
| } |
| break; |
| } |
| |
| ALOGV("%s: pcm_prepare", __func__); |
| ATRACE_BEGIN("pcm_in_prepare"); |
| ret = pcm_prepare(in->pcm); |
| ATRACE_END(); |
| if (ret < 0) { |
| ALOGE("%s: pcm_prepare returned %d", __func__, ret); |
| pcm_close(in->pcm); |
| in->pcm = NULL; |
| goto error_open; |
| } |
| register_in_stream(in); |
| if (in->realtime) { |
| ATRACE_BEGIN("pcm_in_start"); |
| ret = pcm_start(in->pcm); |
| ATRACE_END(); |
| if (ret < 0) { |
| ALOGE("%s: RT pcm_start failed ret %d", __func__, ret); |
| pcm_close(in->pcm); |
| in->pcm = NULL; |
| goto error_open; |
| } |
| } |
| } |
| |
| check_and_enable_effect(adev); |
| audio_extn_audiozoom_set_microphone_direction(in, in->zoom); |
| audio_extn_audiozoom_set_microphone_field_dimension(in, in->direction); |
| |
| if (is_loopback_input_device(get_device_types(&in->device_list))) |
| audio_extn_keep_alive_start(KEEP_ALIVE_OUT_PRIMARY); |
| |
| done_open: |
| audio_streaming_hint_end(); |
| audio_extn_perf_lock_release(&adev->perf_lock_handle); |
| ALOGD("%s: exit", __func__); |
| enable_gcov(); |
| return ret; |
| |
| error_open: |
| audio_streaming_hint_end(); |
| audio_extn_perf_lock_release(&adev->perf_lock_handle); |
| stop_input_stream(in); |
| |
| error_config: |
| if (audio_extn_cin_attached_usecase(in)) |
| audio_extn_cin_close_input_stream(in); |
| /* |
| * sleep 50ms to allow sufficient time for kernel |
| * drivers to recover incases like SSR. |
| */ |
| usleep(50000); |
| ALOGD("%s: exit: status(%d)", __func__, ret); |
| enable_gcov(); |
| return ret; |
| } |
| |
| void lock_input_stream(struct stream_in *in) |
| { |
| pthread_mutex_lock(&in->pre_lock); |
| pthread_mutex_lock(&in->lock); |
| pthread_mutex_unlock(&in->pre_lock); |
| } |
| |
| void lock_output_stream(struct stream_out *out) |
| { |
| pthread_mutex_lock(&out->pre_lock); |
| pthread_mutex_lock(&out->lock); |
| pthread_mutex_unlock(&out->pre_lock); |
| } |
| |
| /* must be called with out->lock locked */ |
| static int send_offload_cmd_l(struct stream_out* out, int command) |
| { |
| struct offload_cmd *cmd = (struct offload_cmd *)calloc(1, sizeof(struct offload_cmd)); |
| |
| if (!cmd) { |
| ALOGE("failed to allocate mem for command 0x%x", command); |
| return -ENOMEM; |
| } |
| |
| ALOGVV("%s %d", __func__, command); |
| |
| cmd->cmd = command; |
| list_add_tail(&out->offload_cmd_list, &cmd->node); |
| pthread_cond_signal(&out->offload_cond); |
| return 0; |
| } |
| |
| /* must be called with out->lock */ |
| static void stop_compressed_output_l(struct stream_out *out) |
| { |
| pthread_mutex_lock(&out->latch_lock); |
| out->offload_state = OFFLOAD_STATE_IDLE; |
| pthread_mutex_unlock(&out->latch_lock); |
| |
| out->playback_started = 0; |
| out->send_new_metadata = 1; |
| if (out->compr != NULL) { |
| compress_stop(out->compr); |
| while (out->offload_thread_blocked) { |
| pthread_cond_wait(&out->cond, &out->lock); |
| } |
| } |
| } |
| |
| bool is_interactive_usecase(audio_usecase_t uc_id) |
| { |
| unsigned int i; |
| for (i = 0; i < sizeof(interactive_usecases)/sizeof(interactive_usecases[0]); i++) { |
| if (uc_id == interactive_usecases[i]) |
| return true; |
| } |
| return false; |
| } |
| |
| static audio_usecase_t get_interactive_usecase(struct audio_device *adev) |
| { |
| audio_usecase_t ret_uc = USECASE_INVALID; |
| unsigned int intract_uc_index; |
| unsigned int num_usecase = sizeof(interactive_usecases)/sizeof(interactive_usecases[0]); |
| |
| ALOGV("%s: num_usecase: %d", __func__, num_usecase); |
| for (intract_uc_index = 0; intract_uc_index < num_usecase; intract_uc_index++) { |
| if (!(adev->interactive_usecase_state & (0x1 << intract_uc_index))) { |
| adev->interactive_usecase_state |= 0x1 << intract_uc_index; |
| ret_uc = interactive_usecases[intract_uc_index]; |
| break; |
| } |
| } |
| |
| ALOGV("%s: Interactive usecase is %d", __func__, ret_uc); |
| return ret_uc; |
| } |
| |
| static void free_interactive_usecase(struct audio_device *adev, |
| audio_usecase_t uc_id) |
| { |
| unsigned int interact_uc_index; |
| unsigned int num_usecase = sizeof(interactive_usecases)/sizeof(interactive_usecases[0]); |
| |
| for (interact_uc_index = 0; interact_uc_index < num_usecase; interact_uc_index++) { |
| if (interactive_usecases[interact_uc_index] == uc_id) { |
| adev->interactive_usecase_state &= ~(0x1 << interact_uc_index); |
| break; |
| } |
| } |
| ALOGV("%s: free Interactive usecase %d", __func__, uc_id); |
| } |
| |
| bool is_offload_usecase(audio_usecase_t uc_id) |
| { |
| unsigned int i; |
| for (i = 0; i < sizeof(offload_usecases)/sizeof(offload_usecases[0]); i++) { |
| if (uc_id == offload_usecases[i]) |
| return true; |
| } |
| return false; |
| } |
| |
| static audio_usecase_t get_offload_usecase(struct audio_device *adev, bool is_compress) |
| { |
| audio_usecase_t ret_uc = USECASE_INVALID; |
| unsigned int offload_uc_index; |
| unsigned int num_usecase = sizeof(offload_usecases)/sizeof(offload_usecases[0]); |
| if (!adev->multi_offload_enable) { |
| if (!is_compress) |
| ret_uc = USECASE_AUDIO_PLAYBACK_OFFLOAD2; |
| else |
| ret_uc = USECASE_AUDIO_PLAYBACK_OFFLOAD; |
| |
| pthread_mutex_lock(&adev->lock); |
| if (get_usecase_from_list(adev, ret_uc) != NULL) |
| ret_uc = USECASE_INVALID; |
| pthread_mutex_unlock(&adev->lock); |
| |
| return ret_uc; |
| } |
| |
| ALOGV("%s: num_usecase: %d", __func__, num_usecase); |
| for (offload_uc_index = 0; offload_uc_index < num_usecase; offload_uc_index++) { |
| if (!(adev->offload_usecases_state & (0x1 << offload_uc_index))) { |
| adev->offload_usecases_state |= 0x1 << offload_uc_index; |
| ret_uc = offload_usecases[offload_uc_index]; |
| break; |
| } |
| } |
| |
| ALOGV("%s: offload usecase is %d", __func__, ret_uc); |
| return ret_uc; |
| } |
| |
| static void free_offload_usecase(struct audio_device *adev, |
| audio_usecase_t uc_id) |
| { |
| unsigned int offload_uc_index; |
| unsigned int num_usecase = sizeof(offload_usecases)/sizeof(offload_usecases[0]); |
| |
| if (!adev->multi_offload_enable) |
| return; |
| |
| for (offload_uc_index = 0; offload_uc_index < num_usecase; offload_uc_index++) { |
| if (offload_usecases[offload_uc_index] == uc_id) { |
| adev->offload_usecases_state &= ~(0x1 << offload_uc_index); |
| break; |
| } |
| } |
| ALOGV("%s: free offload usecase %d", __func__, uc_id); |
| } |
| |
| static void *offload_thread_loop(void *context) |
| { |
| struct stream_out *out = (struct stream_out *) context; |
| struct listnode *item; |
| int ret = 0; |
| |
| setpriority(PRIO_PROCESS, 0, ANDROID_PRIORITY_AUDIO); |
| set_sched_policy(0, SP_FOREGROUND); |
| prctl(PR_SET_NAME, (unsigned long)"Offload Callback", 0, 0, 0); |
| |
| ALOGV("%s", __func__); |
| lock_output_stream(out); |
| out->offload_state = OFFLOAD_STATE_IDLE; |
| out->playback_started = 0; |
| for (;;) { |
| struct offload_cmd *cmd = NULL; |
| stream_callback_event_t event; |
| bool send_callback = false; |
| |
| ALOGVV("%s offload_cmd_list %d out->offload_state %d", |
| __func__, list_empty(&out->offload_cmd_list), |
| out->offload_state); |
| if (list_empty(&out->offload_cmd_list)) { |
| ALOGV("%s SLEEPING", __func__); |
| pthread_cond_wait(&out->offload_cond, &out->lock); |
| ALOGV("%s RUNNING", __func__); |
| continue; |
| } |
| |
| item = list_head(&out->offload_cmd_list); |
| cmd = node_to_item(item, struct offload_cmd, node); |
| list_remove(item); |
| |
| ALOGVV("%s STATE %d CMD %d out->compr %p", |
| __func__, out->offload_state, cmd->cmd, out->compr); |
| |
| if (cmd->cmd == OFFLOAD_CMD_EXIT) { |
| free(cmd); |
| break; |
| } |
| |
| // allow OFFLOAD_CMD_ERROR reporting during standby |
| // this is needed to handle failures during compress_open |
| // Note however that on a pause timeout, the stream is closed |
| // and no offload usecase will be active. Therefore this |
| // special case is needed for compress_open failures alone |
| if (cmd->cmd != OFFLOAD_CMD_ERROR && |
| out->compr == NULL) { |
| ALOGE("%s: Compress handle is NULL", __func__); |
| free(cmd); |
| pthread_cond_signal(&out->cond); |
| continue; |
| } |
| out->offload_thread_blocked = true; |
| pthread_mutex_unlock(&out->lock); |
| send_callback = false; |
| switch(cmd->cmd) { |
| case OFFLOAD_CMD_WAIT_FOR_BUFFER: |
| ALOGD("copl(%p):calling compress_wait", out); |
| compress_wait(out->compr, -1); |
| ALOGD("copl(%p):out of compress_wait", out); |
| send_callback = true; |
| event = STREAM_CBK_EVENT_WRITE_READY; |
| break; |
| case OFFLOAD_CMD_PARTIAL_DRAIN: |
| ret = compress_next_track(out->compr); |
| if(ret == 0) { |
| ALOGD("copl(%p):calling compress_partial_drain", out); |
| ret = compress_partial_drain(out->compr); |
| ALOGD("copl(%p):out of compress_partial_drain", out); |
| if (ret < 0) |
| ret = -errno; |
| } |
| else if (ret == -ETIMEDOUT) |
| ret = compress_drain(out->compr); |
| else |
| ALOGE("%s: Next track returned error %d",__func__, ret); |
| if (-ENETRESET != ret && !(-EINTR == ret && |
| CARD_STATUS_OFFLINE == out->card_status)) { |
| send_callback = true; |
| pthread_mutex_lock(&out->lock); |
| out->send_new_metadata = 1; |
| out->send_next_track_params = true; |
| pthread_mutex_unlock(&out->lock); |
| event = STREAM_CBK_EVENT_DRAIN_READY; |
| ALOGV("copl(%p):send drain callback, ret %d", out, ret); |
| } else |
| ALOGI("%s: Block drain ready event during SSR", __func__); |
| break; |
| case OFFLOAD_CMD_DRAIN: |
| ALOGD("copl(%p):calling compress_drain", out); |
| ret = compress_drain(out->compr); |
| ALOGD("copl(%p):out of compress_drain", out); |
| // EINTR check avoids drain interruption due to SSR |
| if (-ENETRESET != ret && !(-EINTR == ret && |
| CARD_STATUS_OFFLINE == out->card_status)) { |
| send_callback = true; |
| event = STREAM_CBK_EVENT_DRAIN_READY; |
| } else |
| ALOGI("%s: Block drain ready event during SSR", __func__); |
| break; |
| case OFFLOAD_CMD_ERROR: |
| ALOGD("copl(%p): sending error callback to AF", out); |
| send_callback = true; |
| event = STREAM_CBK_EVENT_ERROR; |
| break; |
| default: |
| ALOGE("%s unknown command received: %d", __func__, cmd->cmd); |
| break; |
| } |
| lock_output_stream(out); |
| out->offload_thread_blocked = false; |
| pthread_cond_signal(&out->cond); |
| if (send_callback && out->client_callback) { |
| ALOGVV("%s: sending client_callback event %d", __func__, event); |
| out->client_callback(event, NULL, out->client_cookie); |
| } |
| free(cmd); |
| } |
| |
| pthread_cond_signal(&out->cond); |
| while (!list_empty(&out->offload_cmd_list)) { |
| item = list_head(&out->offload_cmd_list); |
| list_remove(item); |
| free(node_to_item(item, struct offload_cmd, node)); |
| } |
| pthread_mutex_unlock(&out->lock); |
| |
| return NULL; |
| } |
| |
| static int create_offload_callback_thread(struct stream_out *out) |
| { |
| pthread_cond_init(&out->offload_cond, (const pthread_condattr_t *) NULL); |
| list_init(&out->offload_cmd_list); |
| pthread_create(&out->offload_thread, (const pthread_attr_t *) NULL, |
| offload_thread_loop, out); |
| return 0; |
| } |
| |
| static int destroy_offload_callback_thread(struct stream_out *out) |
| { |
| lock_output_stream(out); |
| stop_compressed_output_l(out); |
| send_offload_cmd_l(out, OFFLOAD_CMD_EXIT); |
| |
| pthread_mutex_unlock(&out->lock); |
| pthread_join(out->offload_thread, (void **) NULL); |
| pthread_cond_destroy(&out->offload_cond); |
| |
| return 0; |
| } |
| |
| static int stop_output_stream(struct stream_out *out) |
| { |
| int ret = 0; |
| struct audio_usecase *uc_info; |
| struct audio_device *adev = out->dev; |
| bool has_voip_usecase = |
| get_usecase_from_list(adev, USECASE_AUDIO_PLAYBACK_VOIP) != NULL; |
| |
| ALOGV("%s: enter: usecase(%d: %s)", __func__, |
| out->usecase, use_case_table[out->usecase]); |
| uc_info = get_usecase_from_list(adev, out->usecase); |
| if (uc_info == NULL) { |
| ALOGE("%s: Could not find the usecase (%d) in the list", |
| __func__, out->usecase); |
| return -EINVAL; |
| } |
| |
| out->a2dp_muted = false; |
| |
| if (audio_extn_ext_hw_plugin_usecase_stop(adev->ext_hw_plugin, uc_info)) |
| ALOGE("%s: failed to stop ext hw plugin", __func__); |
| |
| if (is_offload_usecase(out->usecase) && |
| !(audio_extn_passthru_is_passthrough_stream(out))) { |
| if (adev->visualizer_stop_output != NULL) |
| adev->visualizer_stop_output(out->handle, out->pcm_device_id); |
| |
| audio_extn_dts_remove_state_notifier_node(out->usecase); |
| |
| if (adev->offload_effects_stop_output != NULL) |
| adev->offload_effects_stop_output(out->handle, out->pcm_device_id); |
| } else if (out->usecase == USECASE_AUDIO_PLAYBACK_ULL || |
| out->usecase == USECASE_AUDIO_PLAYBACK_MMAP) { |
| audio_low_latency_hint_end(); |
| } |
| |
| if (out->usecase == USECASE_INCALL_MUSIC_UPLINK || |
| out->usecase == USECASE_INCALL_MUSIC_UPLINK2) { |
| voice_set_device_mute_flag(adev, false); |
| } |
| |
| /* 1. Get and set stream specific mixer controls */ |
| disable_audio_route(adev, uc_info); |
| |
| /* 2. Disable the rx device */ |
| disable_snd_device(adev, uc_info->out_snd_device); |
| if (out->usecase == USECASE_AUDIO_PLAYBACK_WITH_HAPTICS) |
| disable_snd_device(adev, SND_DEVICE_OUT_HAPTICS); |
| |
| audio_extn_extspk_update(adev->extspk); |
| |
| if (is_offload_usecase(out->usecase)) { |
| audio_enable_asm_bit_width_enforce_mode(adev->mixer, |
| adev->dsp_bit_width_enforce_mode, |
| false); |
| } |
| if (is_usb_out_device_type(&out->device_list)) { |
| ret = audio_extn_usb_check_and_set_svc_int(uc_info, |
| false); |
| |
| if (ret != 0) |
| check_usecases_codec_backend(adev, uc_info, uc_info->out_snd_device); |
| /* default service interval was successfully updated, |
| reopen USB backend with new service interval */ |
| ret = 0; |
| } |
| |
| list_remove(&uc_info->list); |
| out->started = 0; |
| if (is_offload_usecase(out->usecase) && |
| (audio_extn_passthru_is_passthrough_stream(out))) { |
| ALOGV("Disable passthrough , reset mixer to pcm"); |
| /* NO_PASSTHROUGH */ |
| #ifdef AUDIO_GKI_ENABLED |
| /* out->compr_config.codec->reserved[0] is for compr_passthr */ |
| out->compr_config.codec->reserved[0] = 0; |
| #else |
| out->compr_config.codec->compr_passthr = 0; |
| #endif |
| audio_extn_passthru_on_stop(out); |
| audio_extn_dolby_set_dap_bypass(adev, DAP_STATE_ON); |
| } |
| |
| /* Must be called after removing the usecase from list */ |
| if (compare_device_type(&out->device_list, AUDIO_DEVICE_OUT_AUX_DIGITAL)) |
| audio_extn_keep_alive_start(KEEP_ALIVE_OUT_HDMI); |
| |
| if (out->ip_hdlr_handle) { |
| ret = audio_extn_ip_hdlr_intf_close(out->ip_hdlr_handle, true, out); |
| if (ret < 0) |
| ALOGE("%s: audio_extn_ip_hdlr_intf_close failed %d",__func__, ret); |
| } |
| |
| /* trigger voip input to reroute when voip output changes to hearing aid */ |
| if (has_voip_usecase || |
| compare_device_type(&out->device_list, AUDIO_DEVICE_OUT_SPEAKER_SAFE)) { |
| struct listnode *node; |
| struct audio_usecase *usecase; |
| list_for_each(node, &adev->usecase_list) { |
| usecase = node_to_item(node, struct audio_usecase, list); |
| if (usecase->type == PCM_PLAYBACK || usecase == uc_info || |
| (usecase->type == PCM_CAPTURE && |
| usecase->id != USECASE_AUDIO_RECORD_VOIP && |
| usecase->id != USECASE_AUDIO_RECORD_VOIP_LOW_LATENCY)) |
| continue; |
| |
| ALOGD("%s: select_devices at usecase(%d: %s) after removing the usecase(%d: %s)", |
| __func__, usecase->id, use_case_table[usecase->id], |
| out->usecase, use_case_table[out->usecase]); |
| select_devices(adev, usecase->id); |
| } |
| } |
| |
| free(uc_info); |
| ALOGV("%s: exit: status(%d)", __func__, ret); |
| return ret; |
| } |
| |
| struct pcm* pcm_open_prepare_helper(unsigned int snd_card, unsigned int pcm_device_id, |
| unsigned int flags, unsigned int pcm_open_retry_count, |
| struct pcm_config *config) |
| { |
| struct pcm* pcm = NULL; |
| |
| while (1) { |
| pcm = pcm_open(snd_card, pcm_device_id, flags, config); |
| if (pcm == NULL || !pcm_is_ready(pcm)) { |
| ALOGE("%s: %s", __func__, pcm_get_error(pcm)); |
| if (pcm != NULL) { |
| pcm_close(pcm); |
| pcm = NULL; |
| } |
| if (pcm_open_retry_count == 0) |
| return NULL; |
| |
| pcm_open_retry_count--; |
| usleep(PROXY_OPEN_WAIT_TIME * 1000); |
| continue; |
| } |
| break; |
| } |
| |
| if (pcm_is_ready(pcm)) { |
| int ret = pcm_prepare(pcm); |
| if (ret < 0) { |
| ALOGE("%s: pcm_prepare returned %d", __func__, ret); |
| pcm_close(pcm); |
| pcm = NULL; |
| } |
| } |
| |
| return pcm; |
| } |
| |
| int start_output_stream(struct stream_out *out) |
| { |
| int ret = 0; |
| struct audio_usecase *uc_info; |
| struct audio_device *adev = out->dev; |
| char mixer_ctl_name[128]; |
| struct mixer_ctl *ctl = NULL; |
| char* perf_mode[] = {"ULL", "ULL_PP", "LL"}; |
| bool a2dp_combo = false; |
| bool is_haptic_usecase = (out->usecase == USECASE_AUDIO_PLAYBACK_WITH_HAPTICS) ? true: false; |
| |
| ATRACE_BEGIN("start_output_stream"); |
| if ((out->usecase < 0) || (out->usecase >= AUDIO_USECASE_MAX)) { |
| ret = -EINVAL; |
| goto error_config; |
| } |
| |
| ALOGD("%s: enter: stream(%p)usecase(%d: %s) devices(%#x) is_haptic_usecase(%d)", |
| __func__, &out->stream, out->usecase, use_case_table[out->usecase], |
| get_device_types(&out->device_list), is_haptic_usecase); |
| |
| bool is_speaker_active = compare_device_type(&out->device_list, |
| AUDIO_DEVICE_OUT_SPEAKER); |
| bool is_speaker_safe_active = compare_device_type(&out->device_list, |
| AUDIO_DEVICE_OUT_SPEAKER_SAFE); |
| |
| if (CARD_STATUS_OFFLINE == out->card_status || |
| CARD_STATUS_OFFLINE == adev->card_status) { |
| ALOGW("out->card_status or adev->card_status offline, try again"); |
| ret = -EIO; |
| goto error_fatal; |
| } |
| |
| //Update incall music usecase to reflect correct voice session |
| if (out->flags & AUDIO_OUTPUT_FLAG_INCALL_MUSIC) { |
| ret = voice_extn_check_and_set_incall_music_usecase(adev, out); |
| if (ret != 0) { |
| ALOGE("%s: Incall music delivery usecase cannot be set error:%d", |
| __func__, ret); |
| goto error_config; |
| } |
| } |
| |
| if (is_a2dp_out_device_type(&out->device_list)) { |
| if (!audio_extn_a2dp_source_is_ready()) { |
| if (is_speaker_active || is_speaker_safe_active) { |
| a2dp_combo = true; |
| } else { |
| if (!is_offload_usecase(out->usecase)) { |
| ALOGE("%s: A2DP profile is not ready, return error", __func__); |
| ret = -EAGAIN; |
| goto error_config; |
| } |
| } |
| } |
| } |
| if (is_sco_out_device_type(&out->device_list)) { |
| if (!adev->bt_sco_on) { |
| if (is_speaker_active) { |
| //combo usecase just by pass a2dp |
| ALOGW("%s: SCO is not connected, route it to speaker", __func__); |
| reassign_device_list(&out->device_list, AUDIO_DEVICE_OUT_SPEAKER, ""); |
| } else { |
| ALOGE("%s: SCO profile is not ready, return error", __func__); |
| ret = -EAGAIN; |
| goto error_config; |
| } |
| } |
| } |
| |
| out->pcm_device_id = platform_get_pcm_device_id(out->usecase, PCM_PLAYBACK); |
| if (out->pcm_device_id < 0) { |
| ALOGE("%s: Invalid PCM device id(%d) for the usecase(%d)", |
| __func__, out->pcm_device_id, out->usecase); |
| ret = -EINVAL; |
| goto error_open; |
| } |
| |
| if (is_haptic_usecase) { |
| adev->haptic_pcm_device_id = platform_get_pcm_device_id( |
| USECASE_AUDIO_PLAYBACK_HAPTICS, PCM_PLAYBACK); |
| if (adev->haptic_pcm_device_id < 0) { |
| ALOGE("%s: Invalid Haptics pcm device id(%d) for the usecase(%d)", |
| __func__, adev->haptic_pcm_device_id, out->usecase); |
| ret = -EINVAL; |
| goto error_config; |
| } |
| } |
| |
| uc_info = (struct audio_usecase *)calloc(1, sizeof(struct audio_usecase)); |
| |
| if (!uc_info) { |
| ret = -ENOMEM; |
| goto error_config; |
| } |
| |
| uc_info->id = out->usecase; |
| uc_info->type = PCM_PLAYBACK; |
| uc_info->stream.out = out; |
| list_init(&uc_info->device_list); |
| assign_devices(&uc_info->device_list, &out->device_list); |
| uc_info->in_snd_device = SND_DEVICE_NONE; |
| uc_info->out_snd_device = SND_DEVICE_NONE; |
| |
| /* This must be called before adding this usecase to the list */ |
| if (is_usb_out_device_type(&out->device_list)) { |
| audio_extn_usb_check_and_set_svc_int(uc_info, true); |
| /* USB backend is not reopened immediately. |
| This is eventually done as part of select_devices */ |
| } |
| |
| list_add_tail(&adev->usecase_list, &uc_info->list); |
| |
| audio_streaming_hint_start(); |
| audio_extn_perf_lock_acquire(&adev->perf_lock_handle, 0, |
| adev->perf_lock_opts, |
| adev->perf_lock_opts_size); |
| |
| if (compare_device_type(&out->device_list, AUDIO_DEVICE_OUT_AUX_DIGITAL)) { |
| audio_extn_keep_alive_stop(KEEP_ALIVE_OUT_HDMI); |
| if (audio_extn_passthru_is_enabled() && |
| audio_extn_passthru_is_passthrough_stream(out)) { |
| audio_extn_passthru_on_start(out); |
| } |
| } |
| |
| if (is_a2dp_out_device_type(&out->device_list) && |
| (!audio_extn_a2dp_source_is_ready())) { |
| if (!a2dp_combo) { |
| check_a2dp_restore_l(adev, out, false); |
| } else { |
| struct listnode dev; |
| list_init(&dev); |
| assign_devices(&dev, &out->device_list); |
| if (compare_device_type(&dev, AUDIO_DEVICE_OUT_SPEAKER_SAFE)) |
| reassign_device_list(&out->device_list, |
| AUDIO_DEVICE_OUT_SPEAKER_SAFE, ""); |
| else |
| reassign_device_list(&out->device_list, |
| AUDIO_DEVICE_OUT_SPEAKER, ""); |
| select_devices(adev, out->usecase); |
| assign_devices(&out->device_list, &dev); |
| } |
| } else { |
| select_devices(adev, out->usecase); |
| if (is_a2dp_out_device_type(&out->device_list) && |
| !adev->a2dp_started) { |
| if (is_speaker_active || is_speaker_safe_active) { |
| struct listnode dev; |
| list_init(&dev); |
| assign_devices(&dev, &out->device_list); |
| if (compare_device_type(&dev, AUDIO_DEVICE_OUT_SPEAKER_SAFE)) |
| reassign_device_list(&out->device_list, |
| AUDIO_DEVICE_OUT_SPEAKER_SAFE, ""); |
| else |
| reassign_device_list(&out->device_list, |
| AUDIO_DEVICE_OUT_SPEAKER, ""); |
| select_devices(adev, out->usecase); |
| assign_devices(&out->device_list, &dev); |
| } else { |
| ret = -EINVAL; |
| goto error_open; |
| } |
| } |
| } |
| |
| if (out->usecase == USECASE_INCALL_MUSIC_UPLINK || |
| out->usecase == USECASE_INCALL_MUSIC_UPLINK2) { |
| voice_set_device_mute_flag(adev, true); |
| } |
| |
| if (audio_extn_ext_hw_plugin_usecase_start(adev->ext_hw_plugin, uc_info)) |
| ALOGE("%s: failed to start ext hw plugin", __func__); |
| |
| ALOGV("%s: Opening PCM device card_id(%d) device_id(%d) format(%#x)", |
| __func__, adev->snd_card, out->pcm_device_id, out->config.format); |
| |
| if (out->usecase == USECASE_AUDIO_PLAYBACK_MMAP) { |
| ALOGD("%s: Starting MMAP stream", __func__); |
| if (out->pcm == NULL || !pcm_is_ready(out->pcm)) { |
| ALOGE("%s: pcm stream not ready", __func__); |
| goto error_open; |
| } |
| ret = pcm_start(out->pcm); |
| if (ret < 0) { |
| ALOGE("%s: MMAP pcm_start failed ret %d", __func__, ret); |
| goto error_open; |
| } |
| out_set_mmap_volume(&out->stream, out->volume_l, out->volume_r); |
| } else if (!is_offload_usecase(out->usecase)) { |
| unsigned int flags = PCM_OUT; |
| unsigned int pcm_open_retry_count = 0; |
| if (out->usecase == USECASE_AUDIO_PLAYBACK_AFE_PROXY) { |
| flags |= PCM_MMAP | PCM_NOIRQ; |
| pcm_open_retry_count = PROXY_OPEN_RETRY_COUNT; |
| } else if (out->realtime) { |
| flags |= PCM_MMAP | PCM_NOIRQ | PCM_MONOTONIC; |
| } else |
| flags |= PCM_MONOTONIC; |
| |
| if ((adev->vr_audio_mode_enabled) && |
| (out->flags & AUDIO_OUTPUT_FLAG_RAW)) { |
| snprintf(mixer_ctl_name, sizeof(mixer_ctl_name), |
| "PCM_Dev %d Topology", out->pcm_device_id); |
| ctl = mixer_get_ctl_by_name(adev->mixer, mixer_ctl_name); |
| if (!ctl) { |
| ALOGI("%s: Could not get ctl for mixer cmd might be ULL - %s", |
| __func__, mixer_ctl_name); |
| } else { |
| //if success use ULLPP |
| ALOGI("%s: mixer ctrl %s succeeded setting up ULL for %d", |
| __func__, mixer_ctl_name, out->pcm_device_id); |
| //There is a still a possibility that some sessions |
| // that request for FAST|RAW when 3D audio is active |
| //can go through ULLPP. Ideally we expects apps to |
| //listen to audio focus and stop concurrent playback |
| //Also, we will look for mode flag (voice_in_communication) |
| //before enabling the realtime flag. |
| mixer_ctl_set_enum_by_string(ctl, perf_mode[1]); |
| } |
| } |
| |
| platform_set_stream_channel_map(adev->platform, out->channel_mask, |
| out->pcm_device_id, -1, &out->channel_map_param.channel_map[0]); |
| |
| out->pcm = pcm_open_prepare_helper(adev->snd_card, out->pcm_device_id, |
| flags, pcm_open_retry_count, |
| &(out->config)); |
| if (out->pcm == NULL) { |
| ret = -EIO; |
| goto error_open; |
| } |
| |
| if (is_haptic_usecase) { |
| adev->haptic_pcm = pcm_open_prepare_helper(adev->snd_card, |
| adev->haptic_pcm_device_id, |
| flags, pcm_open_retry_count, |
| &(adev->haptics_config)); |
| // failure to open haptics pcm shouldnt stop audio, |
| // so do not close audio pcm in case of error |
| |
| if (property_get_bool("vendor.audio.enable_haptic_audio_sync", false)) { |
| ALOGD("%s: enable haptic audio synchronization", __func__); |
| platform_set_qtime(adev->platform, out->pcm_device_id, adev->haptic_pcm_device_id); |
| } |
| } |
| |
| // apply volume for voip playback after path is set up |
| if (out->usecase == USECASE_AUDIO_PLAYBACK_VOIP) |
| out_set_voip_volume(&out->stream, out->volume_l, out->volume_r); |
| else if ((out->usecase == USECASE_AUDIO_PLAYBACK_LOW_LATENCY || out->usecase == USECASE_AUDIO_PLAYBACK_DEEP_BUFFER || |
| out->usecase == USECASE_AUDIO_PLAYBACK_ULL) && (out->apply_volume)) { |
| out_set_pcm_volume(&out->stream, out->volume_l, out->volume_r); |
| out->apply_volume = false; |
| } else if (audio_extn_auto_hal_is_bus_device_usecase(out->usecase)) { |
| out_set_pcm_volume(&out->stream, out->volume_l, out->volume_r); |
| } |
| } else { |
| /* |
| * set custom channel map if: |
| * 1. neither mono nor stereo clips i.e. channels > 2 OR |
| * 2. custom channel map has been set by client |
| * else default channel map of FC/FR/FL can always be set to DSP |
| */ |
| if (popcount(out->channel_mask) > 2 || out->channel_map_param.channel_map[0]) |
| platform_set_stream_channel_map(adev->platform, out->channel_mask, |
| out->pcm_device_id, -1, &out->channel_map_param.channel_map[0]); |
| audio_enable_asm_bit_width_enforce_mode(adev->mixer, |
| adev->dsp_bit_width_enforce_mode, |
| true); |
| out->pcm = NULL; |
| ATRACE_BEGIN("compress_open"); |
| out->compr = compress_open(adev->snd_card, |
| out->pcm_device_id, |
| COMPRESS_IN, &out->compr_config); |
| ATRACE_END(); |
| if (errno == ENETRESET && !is_compress_ready(out->compr)) { |
| ALOGE("%s: compress_open failed errno:%d\n", __func__, errno); |
| adev->card_status = CARD_STATUS_OFFLINE; |
| out->card_status = CARD_STATUS_OFFLINE; |
| ret = -EIO; |
| goto error_open; |
| } |
| |
| if (out->compr && !is_compress_ready(out->compr)) { |
| ALOGE("%s: failed /w error %s", __func__, compress_get_error(out->compr)); |
| compress_close(out->compr); |
| out->compr = NULL; |
| ret = -EIO; |
| goto error_open; |
| } |
| /* compress_open sends params of the track, so reset the flag here */ |
| out->is_compr_metadata_avail = false; |
| |
| if (out->client_callback) |
| compress_nonblock(out->compr, out->non_blocking); |
| |
| /* Since small bufs uses blocking writes, a write will be blocked |
| for the default max poll time (20s) in the event of an SSR. |
| Reduce the poll time to observe and deal with SSR faster. |
| */ |
| if (!out->non_blocking) { |
| compress_set_max_poll_wait(out->compr, 1000); |
| } |
| |
| audio_extn_utils_compress_set_render_mode(out); |
| audio_extn_utils_compress_set_clk_rec_mode(uc_info); |
| |
| audio_extn_dts_create_state_notifier_node(out->usecase); |
| audio_extn_dts_notify_playback_state(out->usecase, 0, out->sample_rate, |
| popcount(out->channel_mask), |
| out->playback_started); |
| |
| #ifdef DS1_DOLBY_DDP_ENABLED |
| if (audio_extn_utils_is_dolby_format(out->format)) |
| audio_extn_dolby_send_ddp_endp_params(adev); |
| #endif |
| if (!(audio_extn_passthru_is_passthrough_stream(out)) && |
| (out->sample_rate != 176400 && out->sample_rate <= 192000)) { |
| if (adev->visualizer_start_output != NULL) |
| adev->visualizer_start_output(out->handle, out->pcm_device_id); |
| if (adev->offload_effects_start_output != NULL) |
| adev->offload_effects_start_output(out->handle, out->pcm_device_id, adev->mixer); |
| audio_extn_check_and_set_dts_hpx_state(adev); |
| } |
| |
| if (compare_device_type(&out->device_list, AUDIO_DEVICE_OUT_BUS)) { |
| /* Update cached volume from media to offload/direct stream */ |
| struct listnode *node = NULL; |
| list_for_each(node, &adev->active_outputs_list) { |
| streams_output_ctxt_t *out_ctxt = node_to_item(node, |
| streams_output_ctxt_t, |
| list); |
| if (out_ctxt->output->usecase == USECASE_AUDIO_PLAYBACK_MEDIA) { |
| out->volume_l = out_ctxt->output->volume_l; |
| out->volume_r = out_ctxt->output->volume_r; |
| } |
| } |
| out_set_compr_volume(&out->stream, |
| out->volume_l, out->volume_r); |
| } |
| } |
| |
| if (ret == 0) { |
| register_out_stream(out); |
| if (out->realtime) { |
| if (out->pcm == NULL || !pcm_is_ready(out->pcm)) { |
| ALOGE("%s: pcm stream not ready", __func__); |
| goto error_open; |
| } |
| ATRACE_BEGIN("pcm_start"); |
| ret = pcm_start(out->pcm); |
| ATRACE_END(); |
| if (ret < 0) |
| goto error_open; |
| } |
| } |
| audio_streaming_hint_end(); |
| audio_extn_perf_lock_release(&adev->perf_lock_handle); |
| ALOGD("%s: exit", __func__); |
| |
| if (out->usecase == USECASE_AUDIO_PLAYBACK_ULL || |
| out->usecase == USECASE_AUDIO_PLAYBACK_MMAP) { |
| audio_low_latency_hint_start(); |
| } |
| |
| if (out->ip_hdlr_handle) { |
| ret = audio_extn_ip_hdlr_intf_open(out->ip_hdlr_handle, true, out, out->usecase); |
| if (ret < 0) |
| ALOGE("%s: audio_extn_ip_hdlr_intf_open failed %d",__func__, ret); |
| } |
| |
| // consider a scenario where on pause lower layers are tear down. |
| // so on resume, swap mixer control need to be sent only when |
| // backend is active, hence rather than sending from enable device |
| // sending it from start of streamtream |
| |
| platform_set_swap_channels(adev, true); |
| |
| ATRACE_END(); |
| enable_gcov(); |
| return ret; |
| error_open: |
| if (adev->haptic_pcm) { |
| pcm_close(adev->haptic_pcm); |
| adev->haptic_pcm = NULL; |
| } |
| audio_streaming_hint_end(); |
| audio_extn_perf_lock_release(&adev->perf_lock_handle); |
| stop_output_stream(out); |
| error_fatal: |
| /* |
| * sleep 50ms to allow sufficient time for kernel |
| * drivers to recover incases like SSR. |
| */ |
| usleep(50000); |
| error_config: |
| ATRACE_END(); |
| enable_gcov(); |
| return ret; |
| } |
| |
| static int check_input_parameters(uint32_t sample_rate, |
| audio_format_t format, |
| int channel_count, |
| bool is_usb_hifi) |
| { |
| int ret = 0; |
| |
| if (((format != AUDIO_FORMAT_PCM_16_BIT) && (format != AUDIO_FORMAT_PCM_8_24_BIT) && |
| (format != AUDIO_FORMAT_PCM_24_BIT_PACKED) && (format != AUDIO_FORMAT_PCM_32_BIT) && |
| (format != AUDIO_FORMAT_PCM_FLOAT)) && |
| !voice_extn_compress_voip_is_format_supported(format) && |
| !audio_extn_compr_cap_format_supported(format) && |
| !audio_extn_cin_format_supported(format)) |
| ret = -EINVAL; |
| |
| int max_channel_count = is_usb_hifi ? MAX_HIFI_CHANNEL_COUNT : MAX_CHANNEL_COUNT; |
| if ((channel_count < MIN_CHANNEL_COUNT) || (channel_count > max_channel_count)) { |
| ALOGE("%s: unsupported channel count (%d) passed Min / Max (%d / %d)", __func__, |
| channel_count, MIN_CHANNEL_COUNT, max_channel_count); |
| return -EINVAL; |
| } |
| |
| switch (channel_count) { |
| case 1: |
| case 2: |
| case 3: |
| case 4: |
| case 6: |
| case 8: |
| case 10: |
| case 12: |
| case 14: |
| break; |
| default: |
| ret = -EINVAL; |
| } |
| |
| switch (sample_rate) { |
| case 8000: |
| case 11025: |
| case 12000: |
| case 16000: |
| case 22050: |
| case 24000: |
| case 32000: |
| case 44100: |
| case 48000: |
| case 88200: |
| case 96000: |
| case 176400: |
| case 192000: |
| break; |
| default: |
| ret = -EINVAL; |
| } |
| |
| return ret; |
| } |
| |
| |
| /** Add a value in a list if not already present. |
| * @return true if value was successfully inserted or already present, |
| * false if the list is full and does not contain the value. |
| */ |
| static bool register_uint(uint32_t value, uint32_t* list, size_t list_length) { |
| for (size_t i = 0; i < list_length; i++) { |
| if (list[i] == value) return true; // value is already present |
| if (list[i] == 0) { // no values in this slot |
| list[i] = value; |
| return true; // value inserted |
| } |
| } |
| return false; // could not insert value |
| } |
| |
| /** Add channel_mask in supported_channel_masks if not already present. |
| * @return true if channel_mask was successfully inserted or already present, |
| * false if supported_channel_masks is full and does not contain channel_mask. |
| */ |
| static void register_channel_mask(audio_channel_mask_t channel_mask, |
| audio_channel_mask_t supported_channel_masks[static MAX_SUPPORTED_CHANNEL_MASKS]) { |
| ALOGE_IF(!register_uint(channel_mask, supported_channel_masks, MAX_SUPPORTED_CHANNEL_MASKS), |
| "%s: stream can not declare supporting its channel_mask %x", __func__, channel_mask); |
| } |
| |
| /** Add format in supported_formats if not already present. |
| * @return true if format was successfully inserted or already present, |
| * false if supported_formats is full and does not contain format. |
| */ |
| static void register_format(audio_format_t format, |
| audio_format_t supported_formats[static MAX_SUPPORTED_FORMATS]) { |
| ALOGE_IF(!register_uint(format, supported_formats, MAX_SUPPORTED_FORMATS), |
| "%s: stream can not declare supporting its format %x", __func__, format); |
| } |
| /** Add sample_rate in supported_sample_rates if not already present. |
| * @return true if sample_rate was successfully inserted or already present, |
| * false if supported_sample_rates is full and does not contain sample_rate. |
| */ |
| static void register_sample_rate(uint32_t sample_rate, |
| uint32_t supported_sample_rates[static MAX_SUPPORTED_SAMPLE_RATES]) { |
| ALOGE_IF(!register_uint(sample_rate, supported_sample_rates, MAX_SUPPORTED_SAMPLE_RATES), |
| "%s: stream can not declare supporting its sample rate %x", __func__, sample_rate); |
| } |
| |
| static inline uint32_t lcm(uint32_t num1, uint32_t num2) |
| { |
| uint32_t high = num1, low = num2, temp = 0; |
| |
| if (!num1 || !num2) |
| return 0; |
| |
| if (num1 < num2) { |
| high = num2; |
| low = num1; |
| } |
| |
| while (low != 0) { |
| temp = low; |
| low = high % low; |
| high = temp; |
| } |
| return (num1 * num2)/high; |
| } |
| |
| static inline uint32_t nearest_multiple(uint32_t num, uint32_t multiplier) |
| { |
| uint32_t remainder = 0; |
| |
| if (!multiplier) |
| return num; |
| |
| remainder = num % multiplier; |
| if (remainder) |
| num += (multiplier - remainder); |
| |
| return num; |
| } |
| |
| static size_t get_stream_buffer_size(size_t duration_ms, |
| uint32_t sample_rate, |
| audio_format_t format, |
| int channel_count, |
| bool is_low_latency) |
| { |
| size_t size = 0; |
| uint32_t bytes_per_period_sample = 0; |
| |
| size = (sample_rate * duration_ms) / 1000; |
| if (is_low_latency) |
| size = configured_low_latency_capture_period_size; |
| |
| bytes_per_period_sample = audio_bytes_per_sample(format) * channel_count; |
| size *= audio_bytes_per_sample(format) * channel_count; |
| |
| /* make sure the size is multiple of 32 bytes and additionally multiple of |
| * the frame_size (required for 24bit samples and non-power-of-2 channel counts) |
| * At 48 kHz mono 16-bit PCM: |
| * 5.000 ms = 240 frames = 15*16*1*2 = 480, a whole multiple of 32 (15) |
| * 3.333 ms = 160 frames = 10*16*1*2 = 320, a whole multiple of 32 (10) |
| * Also, make sure the size is multiple of bytes per period sample |
| */ |
| size = nearest_multiple(size, lcm(32, bytes_per_period_sample)); |
| |
| return size; |
| } |
| |
| static size_t get_input_buffer_size(uint32_t sample_rate, |
| audio_format_t format, |
| int channel_count, |
| bool is_low_latency) |
| { |
| /* Don't know if USB HIFI in this context so use true to be conservative */ |
| if (check_input_parameters(sample_rate, format, channel_count, |
| true /*is_usb_hifi */) != 0) |
| return 0; |
| |
| return get_stream_buffer_size(AUDIO_CAPTURE_PERIOD_DURATION_MSEC, |
| sample_rate, |
| format, |
| channel_count, |
| is_low_latency); |
| } |
| |
| size_t get_output_period_size(uint32_t sample_rate, |
| audio_format_t format, |
| int channel_count, |
| int duration /*in millisecs*/) |
| { |
| size_t size = 0; |
| uint32_t bytes_per_sample = audio_bytes_per_sample(format); |
| |
| if ((duration == 0) || (sample_rate == 0) || |
| (bytes_per_sample == 0) || (channel_count == 0)) { |
| ALOGW("Invalid config duration %d sr %d bps %d ch %d", duration, sample_rate, |
| bytes_per_sample, channel_count); |
| return -EINVAL; |
| } |
| |
| size = (sample_rate * |
| duration * |
| bytes_per_sample * |
| channel_count) / 1000; |
| /* |
| * To have same PCM samples for all channels, the buffer size requires to |
| * be multiple of (number of channels * bytes per sample) |
| * For writes to succeed, the buffer must be written at address which is multiple of 32 |
| */ |
| size = ALIGN(size, (bytes_per_sample * channel_count * 32)); |
| |
| return (size/(channel_count * bytes_per_sample)); |
| } |
| |
| static uint64_t get_actual_pcm_frames_rendered(struct stream_out *out, struct timespec *timestamp) |
| { |
| uint64_t actual_frames_rendered = 0; |
| uint64_t written_frames = 0; |
| uint64_t kernel_frames = 0; |
| uint64_t dsp_frames = 0; |
| uint64_t signed_frames = 0; |
| size_t kernel_buffer_size = 0; |
| |
| /* This adjustment accounts for buffering after app processor. |
| * It is based on estimated DSP latency per use case, rather than exact. |
| */ |
| dsp_frames = platform_render_latency(out) * |
| out->sample_rate / 1000000LL; |
| |
| pthread_mutex_lock(&out->position_query_lock); |
| written_frames = out->written / |
| (audio_bytes_per_sample(out->hal_ip_format) * popcount(out->channel_mask)); |
| |
| /* not querying actual state of buffering in kernel as it would involve an ioctl call |
| * which then needs protection, this causes delay in TS query for pcm_offload usecase |
| * hence only estimate. |
| */ |
| kernel_buffer_size = out->compr_config.fragment_size * out->compr_config.fragments; |
| kernel_frames = kernel_buffer_size / |
| (audio_bytes_per_sample(out->hal_op_format) * popcount(out->channel_mask)); |
| |
| if (written_frames >= (kernel_frames + dsp_frames)) |
| signed_frames = written_frames - kernel_frames - dsp_frames; |
| |
| if (signed_frames > 0) { |
| actual_frames_rendered = signed_frames; |
| if (timestamp != NULL ) |
| *timestamp = out->writeAt; |
| } else if (timestamp != NULL) { |
| clock_gettime(CLOCK_MONOTONIC, timestamp); |
| } |
| pthread_mutex_unlock(&out->position_query_lock); |
| |
| ALOGVV("%s signed frames %lld written frames %lld kernel frames %lld dsp frames %lld", |
| __func__, signed_frames, written_frames, kernel_frames, dsp_frames); |
| |
| return actual_frames_rendered; |
| } |
| |
| static uint32_t out_get_sample_rate(const struct audio_stream *stream) |
| { |
| struct stream_out *out = (struct stream_out *)stream; |
| |
| return out->sample_rate; |
| } |
| |
| static int out_set_sample_rate(struct audio_stream *stream __unused, |
| uint32_t rate __unused) |
| { |
| return -ENOSYS; |
| } |
| |
| static size_t out_get_buffer_size(const struct audio_stream *stream) |
| { |
| struct stream_out *out = (struct stream_out *)stream; |
| |
| if (is_interactive_usecase(out->usecase)) { |
| return out->config.period_size * out->config.period_count; |
| } else if (out->flags & AUDIO_OUTPUT_FLAG_COMPRESS_OFFLOAD) { |
| if (out->flags & AUDIO_OUTPUT_FLAG_TIMESTAMP) |
| return out->compr_config.fragment_size - sizeof(struct snd_codec_metadata); |
| else |
| return out->compr_config.fragment_size; |
| } else if(out->usecase == USECASE_COMPRESS_VOIP_CALL) |
| return voice_extn_compress_voip_out_get_buffer_size(out); |
| else if (is_offload_usecase(out->usecase) && |
| out->flags == AUDIO_OUTPUT_FLAG_DIRECT) |
| return out->hal_fragment_size; |
| |
| return out->config.period_size * out->af_period_multiplier * |
| audio_stream_out_frame_size((const struct audio_stream_out *)stream); |
| } |
| |
| static uint32_t out_get_channels(const struct audio_stream *stream) |
| { |
| struct stream_out *out = (struct stream_out *)stream; |
| |
| return out->channel_mask; |
| } |
| |
| static audio_format_t out_get_format(const struct audio_stream *stream) |
| { |
| struct stream_out *out = (struct stream_out *)stream; |
| |
| return out->format; |
| } |
| |
| static int out_set_format(struct audio_stream *stream __unused, |
| audio_format_t format __unused) |
| { |
| return -ENOSYS; |
| } |
| |
| static int out_standby(struct audio_stream *stream) |
| { |
| struct stream_out *out = (struct stream_out *)stream; |
| struct audio_device *adev = out->dev; |
| bool do_stop = true; |
| |
| ALOGD("%s: enter: stream (%p) usecase(%d: %s)", __func__, |
| stream, out->usecase, use_case_table[out->usecase]); |
| |
| lock_output_stream(out); |
| if (!out->standby) { |
| if (adev->adm_deregister_stream) |
| adev->adm_deregister_stream(adev->adm_data, out->handle); |
| |
| if (is_offload_usecase(out->usecase)) { |
| stop_compressed_output_l(out); |
| } |
| |
| pthread_mutex_lock(&adev->lock); |
| out->standby = true; |
| if (out->usecase == USECASE_COMPRESS_VOIP_CALL) { |
| voice_extn_compress_voip_close_output_stream(stream); |
| out->started = 0; |
| pthread_mutex_unlock(&adev->lock); |
| pthread_mutex_unlock(&out->lock); |
| ALOGD("VOIP output entered standby"); |
| return 0; |
| } else if (!is_offload_usecase(out->usecase)) { |
| if (out->pcm) { |
| pcm_close(out->pcm); |
| out->pcm = NULL; |
| } |
| if (out->usecase == USECASE_AUDIO_PLAYBACK_WITH_HAPTICS) { |
| if (adev->haptic_pcm) { |
| pcm_close(adev->haptic_pcm); |
| adev->haptic_pcm = NULL; |
| } |
| |
| if (adev->haptic_buffer != NULL) { |
| free(adev->haptic_buffer); |
| adev->haptic_buffer = NULL; |
| adev->haptic_buffer_size = 0; |
| } |
| adev->haptic_pcm_device_id = 0; |
| } |
| |
| if (out->usecase == USECASE_AUDIO_PLAYBACK_MMAP) { |
| do_stop = out->playback_started; |
| out->playback_started = false; |
| |
| if (out->mmap_shared_memory_fd >= 0) { |
| ALOGV("%s: closing mmap_shared_memory_fd = %d", |
| __func__, out->mmap_shared_memory_fd); |
| close(out->mmap_shared_memory_fd); |
| out->mmap_shared_memory_fd = -1; |
| } |
| } |
| } else { |
| ALOGD("copl(%p):standby", out); |
| out->send_next_track_params = false; |
| out->is_compr_metadata_avail = false; |
| out->gapless_mdata.encoder_delay = 0; |
| out->gapless_mdata.encoder_padding = 0; |
| if (out->compr != NULL) { |
| compress_close(out->compr); |
| out->compr = NULL; |
| } |
| } |
| if (do_stop) { |
| stop_output_stream(out); |
| } |
| // if fm is active route on selected device in UI |
| audio_extn_fm_route_on_selected_device(adev, &out->device_list); |
| pthread_mutex_unlock(&adev->lock); |
| } |
| pthread_mutex_unlock(&out->lock); |
| ALOGV("%s: exit", __func__); |
| return 0; |
| } |
| |
| static int out_on_error(struct audio_stream *stream) |
| { |
| struct stream_out *out = (struct stream_out *)stream; |
| int status = 0; |
| |
| lock_output_stream(out); |
| // always send CMD_ERROR for offload streams, this |
| // is needed e.g. when SSR happens within compress_open |
| // since the stream is active, offload_callback_thread is also active. |
| if (out->flags & AUDIO_OUTPUT_FLAG_COMPRESS_OFFLOAD) { |
| stop_compressed_output_l(out); |
| } |
| pthread_mutex_unlock(&out->lock); |
| |
| status = out_standby(&out->stream.common); |
| |
| lock_output_stream(out); |
| if (out->flags & AUDIO_OUTPUT_FLAG_COMPRESS_OFFLOAD) { |
| send_offload_cmd_l(out, OFFLOAD_CMD_ERROR); |
| } |
| |
| if (is_offload_usecase(out->usecase) && out->card_status == CARD_STATUS_OFFLINE) { |
| ALOGD("Setting previous card status if offline"); |
| out->prev_card_status_offline = true; |
| } |
| |
| pthread_mutex_unlock(&out->lock); |
| |
| return status; |
| } |
| |
| /* |
| * standby implementation without locks, assumes that the callee already |
| * has taken adev and out lock. |
| */ |
| int out_standby_l(struct audio_stream *stream) |
| { |
| struct stream_out *out = (struct stream_out *)stream; |
| struct audio_device *adev = out->dev; |
| |
| ALOGD("%s: enter: stream (%p) usecase(%d: %s)", __func__, |
| stream, out->usecase, use_case_table[out->usecase]); |
| |
| if (!out->standby) { |
| ATRACE_BEGIN("out_standby_l"); |
| if (adev->adm_deregister_stream) |
| adev->adm_deregister_stream(adev->adm_data, out->handle); |
| |
| if (is_offload_usecase(out->usecase)) { |
| stop_compressed_output_l(out); |
| } |
| |
| out->standby = true; |
| if (out->usecase == USECASE_COMPRESS_VOIP_CALL) { |
| voice_extn_compress_voip_close_output_stream(stream); |
| out->started = 0; |
| ALOGD("VOIP output entered standby"); |
| ATRACE_END(); |
| return 0; |
| } else if (!is_offload_usecase(out->usecase)) { |
| if (out->pcm) { |
| pcm_close(out->pcm); |
| out->pcm = NULL; |
| } |
| if (out->usecase == USECASE_AUDIO_PLAYBACK_WITH_HAPTICS) { |
| if (adev->haptic_pcm) { |
| pcm_close(adev->haptic_pcm); |
| adev->haptic_pcm = NULL; |
| } |
| |
| if (adev->haptic_buffer != NULL) { |
| free(adev->haptic_buffer); |
| adev->haptic_buffer = NULL; |
| adev->haptic_buffer_size = 0; |
| } |
| adev->haptic_pcm_device_id = 0; |
| } |
| } else { |
| ALOGD("copl(%p):standby", out); |
| out->send_next_track_params = false; |
| out->is_compr_metadata_avail = false; |
| out->gapless_mdata.encoder_delay = 0; |
| out->gapless_mdata.encoder_padding = 0; |
| if (out->compr != NULL) { |
| compress_close(out->compr); |
| out->compr = NULL; |
| } |
| } |
| stop_output_stream(out); |
| ATRACE_END(); |
| } |
| ALOGV("%s: exit", __func__); |
| return 0; |
| } |
| |
| static int out_dump(const struct audio_stream *stream, int fd) |
| { |
| struct stream_out *out = (struct stream_out *)stream; |
| |
| // We try to get the lock for consistency, |
| // but it isn't necessary for these variables. |
| // If we're not in standby, we may be blocked on a write. |
| const bool locked = (pthread_mutex_trylock(&out->lock) == 0); |
| dprintf(fd, " Standby: %s\n", out->standby ? "yes" : "no"); |
| dprintf(fd, " Frames written: %lld\n", (long long)out->written); |
| #ifndef LINUX_ENABLED |
| char buffer[256]; // for statistics formatting |
| if (!is_offload_usecase(out->usecase)) { |
| simple_stats_to_string(&out->fifo_underruns, buffer, sizeof(buffer)); |
| dprintf(fd, " Fifo frame underruns: %s\n", buffer); |
| } |
| |
| if (out->start_latency_ms.n > 0) { |
| simple_stats_to_string(&out->start_latency_ms, buffer, sizeof(buffer)); |
| dprintf(fd, " Start latency ms: %s\n", buffer); |
| } |
| #endif |
| if (locked) { |
| pthread_mutex_unlock(&out->lock); |
| } |
| |
| #ifndef LINUX_ENABLED |
| // dump error info |
| (void)error_log_dump( |
| out->error_log, fd, " " /* prefix */, 0 /* lines */, 0 /* limit_ns */); |
| #endif |
| return 0; |
| } |
| |
| static int parse_compress_metadata(struct stream_out *out, struct str_parms *parms) |
| { |
| int ret = 0; |
| char value[32]; |
| |
| if (!out || !parms) { |
| ALOGE("%s: return invalid ",__func__); |
| return -EINVAL; |
| } |
| |
| ret = audio_extn_parse_compress_metadata(out, parms); |
| |
| ret = str_parms_get_str(parms, AUDIO_OFFLOAD_CODEC_DELAY_SAMPLES, value, sizeof(value)); |
| if (ret >= 0) { |
| out->gapless_mdata.encoder_delay = atoi(value); //whats a good limit check? |
| } |
| ret = str_parms_get_str(parms, AUDIO_OFFLOAD_CODEC_PADDING_SAMPLES, value, sizeof(value)); |
| if (ret >= 0) { |
| out->gapless_mdata.encoder_padding = atoi(value); |
| } |
| |
| ALOGV("%s new encoder delay %u and padding %u", __func__, |
| out->gapless_mdata.encoder_delay, out->gapless_mdata.encoder_padding); |
| |
| return 0; |
| } |
| |
| static bool output_drives_call(struct audio_device *adev, struct stream_out *out) |
| { |
| return out == adev->primary_output || out == adev->voice_tx_output; |
| } |
| |
| // note: this call is safe only if the stream_cb is |
| // removed first in close_output_stream (as is done now). |
| static void out_snd_mon_cb(void * stream, struct str_parms * parms) |
| { |
| if (!stream || !parms) |
| return; |
| |
| struct stream_out *out = (struct stream_out *)stream; |
| struct audio_device *adev = out->dev; |
| |
| card_status_t status; |
| int card; |
| if (parse_snd_card_status(parms, &card, &status) < 0) |
| return; |
| |
| pthread_mutex_lock(&adev->lock); |
| bool valid_cb = (card == adev->snd_card); |
| pthread_mutex_unlock(&adev->lock); |
| |
| if (!valid_cb) |
| return; |
| |
| lock_output_stream(out); |
| if (out->card_status != status) |
| out->card_status = status; |
| pthread_mutex_unlock(&out->lock); |
| |
| ALOGI("out_snd_mon_cb for card %d usecase %s, status %s", card, |
| use_case_table[out->usecase], |
| status == CARD_STATUS_OFFLINE ? "offline" : "online"); |
| |
| if (status == CARD_STATUS_OFFLINE) { |
| out_on_error(stream); |
| if (voice_is_call_state_active(adev) && |
| out == adev->primary_output) { |
| ALOGD("%s: SSR/PDR occurred, end all calls\n", __func__); |
| pthread_mutex_lock(&adev->lock); |
| voice_stop_call(adev); |
| adev->mode = AUDIO_MODE_NORMAL; |
| pthread_mutex_unlock(&adev->lock); |
| } |
| } |
| return; |
| } |
| |
| int route_output_stream(struct stream_out *out, |
| struct listnode *devices) |
| { |
| struct audio_device *adev = out->dev; |
| int ret = 0; |
| struct listnode new_devices; |
| bool bypass_a2dp = false; |
| bool reconfig = false; |
| unsigned long service_interval = 0; |
| |
| ALOGD("%s: enter: usecase(%d: %s) devices %x", |
| __func__, out->usecase, use_case_table[out->usecase], get_device_types(devices)); |
| |
| list_init(&new_devices); |
| assign_devices(&new_devices, devices); |
| |
| lock_output_stream(out); |
| pthread_mutex_lock(&adev->lock); |
| |
| /* |
| * When HDMI cable is unplugged the music playback is paused and |
| * the policy manager sends routing=0. But the audioflinger continues |
| * to write data until standby time (3sec). As the HDMI core is |
| * turned off, the write gets blocked. |
| * Avoid this by routing audio to speaker until standby. |
| */ |
| if (is_single_device_type_equal(&out->device_list, |
| AUDIO_DEVICE_OUT_AUX_DIGITAL) && |
| list_empty(&new_devices) && |
| !audio_extn_passthru_is_passthrough_stream(out) && |
| (platform_get_edid_info(adev->platform) != 0) /* HDMI disconnected */) { |
| reassign_device_list(&new_devices, AUDIO_DEVICE_OUT_SPEAKER, ""); |
| } |
| /* |
| * When A2DP is disconnected the |
| * music playback is paused and the policy manager sends routing=0 |
| * But the audioflinger continues to write data until standby time |
| * (3sec). As BT is turned off, the write gets blocked. |
| * Avoid this by routing audio to speaker until standby. |
| */ |
| if (is_a2dp_out_device_type(&out->device_list) && |
| list_empty(&new_devices) && |
| !audio_extn_a2dp_source_is_ready() && |
| !adev->bt_sco_on) { |
| reassign_device_list(&new_devices, AUDIO_DEVICE_OUT_SPEAKER, ""); |
| } |
| /* |
| * When USB headset is disconnected the music playback paused |
| * and the policy manager send routing=0. But if the USB is connected |
| * back before the standby time, AFE is not closed and opened |
| * when USB is connected back. So routing to speker will guarantee |
| * AFE reconfiguration and AFE will be opend once USB is connected again |
| */ |
| if (is_usb_out_device_type(&out->device_list) && |
| list_empty(&new_devices) && |
| !audio_extn_usb_connected(NULL)) { |
| if (adev->mode == AUDIO_MODE_IN_CALL || adev->mode == AUDIO_MODE_IN_COMMUNICATION) |
| reassign_device_list(&new_devices, AUDIO_DEVICE_OUT_EARPIECE, ""); |
| else |
| reassign_device_list(&new_devices, AUDIO_DEVICE_OUT_SPEAKER, ""); |
| } |
| /* To avoid a2dp to sco overlapping / BT device improper state |
| * check with BT lib about a2dp streaming support before routing |
| */ |
| if (is_a2dp_out_device_type(&new_devices)) { |
| if (!audio_extn_a2dp_source_is_ready()) { |
| if (compare_device_type(&new_devices, AUDIO_DEVICE_OUT_SPEAKER) || |
| compare_device_type(&new_devices, AUDIO_DEVICE_OUT_SPEAKER_SAFE)) { |
| //combo usecase just by pass a2dp |
| ALOGW("%s: A2DP profile is not ready,routing to speaker only", __func__); |
| bypass_a2dp = true; |
| } else { |
| ALOGE("%s: A2DP profile is not ready,ignoring routing request", __func__); |
| /* update device to a2dp and don't route as BT returned error |
| * However it is still possible a2dp routing called because |
| * of current active device disconnection (like wired headset) |
| */ |
| assign_devices(&out->device_list, &new_devices); |
| pthread_mutex_unlock(&adev->lock); |
| pthread_mutex_unlock(&out->lock); |
| goto error; |
| } |
| } |
| } |
| |
| // Workaround: If routing to an non existing usb device, fail gracefully |
| // The routing request will otherwise block during 10 second |
| if (is_usb_out_device_type(&new_devices)) { |
| struct str_parms *parms = |
| str_parms_create_str(get_usb_device_address(&new_devices)); |
| if (!parms) |
| goto error; |
| if (!audio_extn_usb_connected(NULL)) { |
| ALOGW("%s: ignoring rerouting to non existing USB card", __func__); |
| pthread_mutex_unlock(&adev->lock); |
| pthread_mutex_unlock(&out->lock); |
| str_parms_destroy(parms); |
| ret = -ENOSYS; |
| goto error; |
| } |
| str_parms_destroy(parms); |
| } |
| |
| // Workaround: If routing to an non existing hdmi device, fail gracefully |
| if (compare_device_type(&new_devices, AUDIO_DEVICE_OUT_AUX_DIGITAL) && |
| (platform_get_edid_info_v2(adev->platform, |
| out->extconn.cs.controller, |
| out->extconn.cs.stream) != 0)) { |
| ALOGW("out_set_parameters() ignoring rerouting to non existing HDMI/DP"); |
| pthread_mutex_unlock(&adev->lock); |
| pthread_mutex_unlock(&out->lock); |
| ret = -ENOSYS; |
| goto error; |
| } |
| |
| /* |
| * select_devices() call below switches all the usecases on the same |
| * backend to the new device. Refer to check_usecases_codec_backend() in |
| * the select_devices(). But how do we undo this? |
| * |
| * For example, music playback is active on headset (deep-buffer usecase) |
| * and if we go to ringtones and select a ringtone, low-latency usecase |
| * will be started on headset+speaker. As we can't enable headset+speaker |
| * and headset devices at the same time, select_devices() switches the music |
| * playback to headset+speaker while starting low-lateny usecase for ringtone. |
| * So when the ringtone playback is completed, how do we undo the same? |
| * |
| * We are relying on the out_set_parameters() call on deep-buffer output, |
| * once the ringtone playback is ended. |
| * NOTE: We should not check if the current devices are same as new devices. |
| * Because select_devices() must be called to switch back the music |
| * playback to headset. |
| */ |
| if (!list_empty(&new_devices)) { |
| bool same_dev = compare_devices(&out->device_list, &new_devices); |
| assign_devices(&out->device_list, &new_devices); |
| |
| if (output_drives_call(adev, out)) { |
| if (!voice_is_call_state_active(adev)) { |
| if (adev->mode == AUDIO_MODE_IN_CALL) { |
| adev->current_call_output = out; |
| ret = voice_start_call(adev); |
| } |
| } else { |
| platform_is_volume_boost_supported_device(adev->platform, &new_devices); |
| adev->current_call_output = out; |
| voice_update_devices_for_all_voice_usecases(adev); |
| } |
| } |
| |
| if (is_usb_out_device_type(&out->device_list)) { |
| service_interval = audio_extn_usb_find_service_interval(false, true /*playback*/); |
| audio_extn_usb_set_service_interval(true /*playback*/, |
| service_interval, |
| &reconfig); |
| ALOGD("%s, svc_int(%ld),reconfig(%d)",__func__,service_interval, reconfig); |
| } |
| |
| if (!out->standby) { |
| if (!same_dev) { |
| ALOGV("update routing change"); |
| audio_extn_perf_lock_acquire(&adev->perf_lock_handle, 0, |
| adev->perf_lock_opts, |
| adev->perf_lock_opts_size); |
| if (adev->adm_on_routing_change) |
| adev->adm_on_routing_change(adev->adm_data, |
| out->handle); |
| } |
| if (!bypass_a2dp) { |
| select_devices(adev, out->usecase); |
| } else { |
| if (compare_device_type(&new_devices, AUDIO_DEVICE_OUT_SPEAKER_SAFE)) |
| reassign_device_list(&out->device_list, AUDIO_DEVICE_OUT_SPEAKER_SAFE, ""); |
| else |
| reassign_device_list(&out->device_list, AUDIO_DEVICE_OUT_SPEAKER, ""); |
| select_devices(adev, out->usecase); |
| assign_devices(&out->device_list, &new_devices); |
| } |
| |
| if (!same_dev) { |
| // on device switch force swap, lower functions will make sure |
| // to check if swap is allowed or not. |
| platform_set_swap_channels(adev, true); |
| audio_extn_perf_lock_release(&adev->perf_lock_handle); |
| } |
| pthread_mutex_lock(&out->latch_lock); |
| if (!is_a2dp_out_device_type(&out->device_list) || audio_extn_a2dp_source_is_ready()) { |
| if (out->a2dp_muted) { |
| out->a2dp_muted = false; |
| if (is_offload_usecase(out->usecase)) |
| out_set_compr_volume(&out->stream, out->volume_l, out->volume_r); |
| else if (out->usecase != USECASE_AUDIO_PLAYBACK_VOIP) |
| out_set_pcm_volume(&out->stream, out->volume_l, out->volume_r); |
| } |
| } |
| if (out->usecase == USECASE_AUDIO_PLAYBACK_VOIP && !out->a2dp_muted) |
| out_set_voip_volume(&out->stream, out->volume_l, out->volume_r); |
| pthread_mutex_unlock(&out->latch_lock); |
| } |
| } |
| |
| pthread_mutex_unlock(&adev->lock); |
| pthread_mutex_unlock(&out->lock); |
| |
| /*handles device and call state changes*/ |
| audio_extn_extspk_update(adev->extspk); |
| |
| error: |
| ALOGV("%s: exit: code(%d)", __func__, ret); |
| return ret; |
| } |
| |
| static int out_set_parameters(struct audio_stream *stream, const char *kvpairs) |
| { |
| struct stream_out *out = (struct stream_out *)stream; |
| struct audio_device *adev = out->dev; |
| struct str_parms *parms; |
| char value[32]; |
| int ret = 0, err; |
| int ext_controller = -1; |
| int ext_stream = -1; |
| |
| ALOGD("%s: enter: usecase(%d: %s) kvpairs: %s", |
| __func__, out->usecase, use_case_table[out->usecase], kvpairs); |
| parms = str_parms_create_str(kvpairs); |
| if (!parms) |
| goto error; |
| |
| err = platform_get_controller_stream_from_params(parms, &ext_controller, |
| &ext_stream); |
| if (err == 0) { |
| out->extconn.cs.controller = ext_controller; |
| out->extconn.cs.stream = ext_stream; |
| ALOGD("%s: usecase(%s) new controller/stream (%d/%d)", __func__, |
| use_case_table[out->usecase], out->extconn.cs.controller, |
| out->extconn.cs.stream); |
| } |
| |
| if (out == adev->primary_output) { |
| pthread_mutex_lock(&adev->lock); |
| audio_extn_set_parameters(adev, parms); |
| pthread_mutex_unlock(&adev->lock); |
| } |
| if (is_offload_usecase(out->usecase)) { |
| lock_output_stream(out); |
| parse_compress_metadata(out, parms); |
| |
| audio_extn_dts_create_state_notifier_node(out->usecase); |
| audio_extn_dts_notify_playback_state(out->usecase, 0, out->sample_rate, |
| popcount(out->channel_mask), |
| out->playback_started); |
| |
| pthread_mutex_unlock(&out->lock); |
| } |
| |
| err = str_parms_get_str(parms, AUDIO_PARAMETER_DUAL_MONO, value, |
| sizeof(value)); |
| if (err >= 0) { |
| if (!strncmp("true", value, sizeof("true")) || atoi(value)) |
| audio_extn_send_dual_mono_mixing_coefficients(out); |
| } |
| |
| err = str_parms_get_str(parms, AUDIO_PARAMETER_STREAM_PROFILE, value, sizeof(value)); |
| if (err >= 0) { |
| strlcpy(out->profile, value, sizeof(out->profile)); |
| ALOGV("updating stream profile with value '%s'", out->profile); |
| lock_output_stream(out); |
| audio_extn_utils_update_stream_output_app_type_cfg(adev->platform, |
| &adev->streams_output_cfg_list, |
| &out->device_list, out->flags, |
| out->hal_op_format, |
| out->sample_rate, out->bit_width, |
| out->channel_mask, out->profile, |
| &out->app_type_cfg); |
| pthread_mutex_unlock(&out->lock); |
| } |
| |
| //suspend, resume handling block |
| //remove QOS only if vendor.audio.hal.dynamic.qos.config.supported is set to true |
| // and vendor.audio.hal.output.suspend.supported is set to true |
| if (out->hal_output_suspend_supported && out->dynamic_pm_qos_config_supported) { |
| //check suspend parameter only for low latency and if the property |
| //is enabled |
| if (str_parms_get_str(parms, "suspend_playback", value, sizeof(value)) >= 0) { |
| ALOGI("%s: got suspend_playback %s", __func__, value); |
| lock_output_stream(out); |
| if (!strncmp(value, "false", 5)) { |
| //suspend_playback=false is supposed to set QOS value back to 75% |
| //the mixer control sent with value Enable will achieve that |
| ret = audio_route_apply_and_update_path(adev->audio_route, out->pm_qos_mixer_path); |
| } else if (!strncmp (value, "true", 4)) { |
| //suspend_playback=true is supposed to remove QOS value |
| //resetting the mixer control will set the default value |
| //for the mixer control which is Disable and this removes the QOS vote |
| ret = audio_route_reset_and_update_path(adev->audio_route, out->pm_qos_mixer_path); |
| } else { |
| ALOGE("%s: Wrong value sent for suspend_playback, expected true/false," |
| " got %s", __func__, value); |
| ret = -1; |
| } |
| |
| if (ret != 0) { |
| ALOGE("%s: %s mixer ctl failed with %d, ignore suspend/resume setparams", |
| __func__, out->pm_qos_mixer_path, ret); |
| } |
| |
| pthread_mutex_unlock(&out->lock); |
| } |
| } |
| |
| //end suspend, resume handling block |
| str_parms_destroy(parms); |
| error: |
| ALOGV("%s: exit: code(%d)", __func__, ret); |
| return ret; |
| } |
| |
| static int in_set_microphone_direction(const struct audio_stream_in *stream, |
| audio_microphone_direction_t dir) { |
| struct stream_in *in = (struct stream_in *)stream; |
| |
| ALOGVV("%s: standby %d source %d dir %d", __func__, in->standby, in->source, dir); |
| |
| in->direction = dir; |
| |
| if (in->standby) |
| return 0; |
| |
| return audio_extn_audiozoom_set_microphone_direction(in, dir); |
| } |
| |
| static int in_set_microphone_field_dimension(const struct audio_stream_in *stream, float zoom) { |
| struct stream_in *in = (struct stream_in *)stream; |
| |
| ALOGVV("%s: standby %d source %d zoom %f", __func__, in->standby, in->source, zoom); |
| |
| if (zoom > 1.0 || zoom < -1.0) |
| return -EINVAL; |
| |
| in->zoom = zoom; |
| |
| if (in->standby) |
| return 0; |
| |
| return audio_extn_audiozoom_set_microphone_field_dimension(in, zoom); |
| } |
| |
| |
| static bool stream_get_parameter_channels(struct str_parms *query, |
| struct str_parms *reply, |
| audio_channel_mask_t *supported_channel_masks) { |
| int ret = -1; |
| char value[512]; |
| bool first = true; |
| size_t i, j; |
| |
| if (str_parms_has_key(query, AUDIO_PARAMETER_STREAM_SUP_CHANNELS)) { |
| ret = 0; |
| value[0] = '\0'; |
| i = 0; |
| while (supported_channel_masks[i] != 0) { |
| for (j = 0; j < ARRAY_SIZE(channels_name_to_enum_table); j++) { |
| if (channels_name_to_enum_table[j].value == supported_channel_masks[i]) { |
| if (!first) |
| strlcat(value, "|", sizeof(value)); |
| |
| strlcat(value, channels_name_to_enum_table[j].name, sizeof(value)); |
| first = false; |
| break; |
| } |
| } |
| i++; |
| } |
| str_parms_add_str(reply, AUDIO_PARAMETER_STREAM_SUP_CHANNELS, value); |
| } |
| return ret == 0; |
| } |
| |
| static bool stream_get_parameter_formats(struct str_parms *query, |
| struct str_parms *reply, |
| audio_format_t *supported_formats) { |
| int ret = -1; |
| char value[256]; |
| size_t i, j; |
| bool first = true; |
| |
| if (str_parms_has_key(query, AUDIO_PARAMETER_STREAM_SUP_FORMATS)) { |
| ret = 0; |
| value[0] = '\0'; |
| i = 0; |
| while (supported_formats[i] != 0) { |
| for (j = 0; j < ARRAY_SIZE(formats_name_to_enum_table); j++) { |
| if (formats_name_to_enum_table[j].value == supported_formats[i]) { |
| if (!first) { |
| strlcat(value, "|", sizeof(value)); |
| } |
| strlcat(value, formats_name_to_enum_table[j].name, sizeof(value)); |
| first = false; |
| break; |
| } |
| } |
| i++; |
| } |
| str_parms_add_str(reply, AUDIO_PARAMETER_STREAM_SUP_FORMATS, value); |
| } |
| return ret == 0; |
| } |
| |
| static bool stream_get_parameter_rates(struct str_parms *query, |
| struct str_parms *reply, |
| uint32_t *supported_sample_rates) { |
| |
| int i; |
| char value[256]; |
| int ret = -1; |
| if (str_parms_has_key(query, AUDIO_PARAMETER_STREAM_SUP_SAMPLING_RATES)) { |
| ret = 0; |
| value[0] = '\0'; |
| i=0; |
| int cursor = 0; |
| while (supported_sample_rates[i]) { |
| int avail = sizeof(value) - cursor; |
| ret = snprintf(value + cursor, avail, "%s%d", |
| cursor > 0 ? "|" : "", |
| supported_sample_rates[i]); |
| if (ret < 0 || ret >= avail) { |
| // if cursor is at the last element of the array |
| // overwrite with \0 is duplicate work as |
| // snprintf already put a \0 in place. |
| // else |
| // we had space to write the '|' at value[cursor] |
| // (which will be overwritten) or no space to fill |
| // the first element (=> cursor == 0) |
| value[cursor] = '\0'; |
| break; |
| } |
| cursor += ret; |
| ++i; |
| } |
| str_parms_add_str(reply, AUDIO_PARAMETER_STREAM_SUP_SAMPLING_RATES, |
| value); |
| } |
| return ret >= 0; |
| } |
| |
| static char* out_get_parameters(const struct audio_stream *stream, const char *keys) |
| { |
| struct stream_out *out = (struct stream_out *)stream; |
| struct str_parms *query = str_parms_create_str(keys); |
| char *str = (char*) NULL; |
| char value[256]; |
| struct str_parms *reply = str_parms_create(); |
| size_t i, j; |
| int ret; |
| bool first = true; |
| |
| if (!query || !reply) { |
| if (reply) { |
| str_parms_destroy(reply); |
| } |
| if (query) { |
| str_parms_destroy(query); |
| } |
| ALOGE("out_get_parameters: failed to allocate mem for query or reply"); |
| return NULL; |
| } |
| |
| ALOGV("%s: %s enter: keys - %s", __func__, use_case_table[out->usecase], keys); |
| ret = str_parms_get_str(query, AUDIO_PARAMETER_STREAM_SUP_CHANNELS, value, sizeof(value)); |
| if (ret >= 0) { |
| value[0] = '\0'; |
| i = 0; |
| while (out->supported_channel_masks[i] != 0) { |
| for (j = 0; j < ARRAY_SIZE(channels_name_to_enum_table); j++) { |
| if (channels_name_to_enum_table[j].value == out->supported_channel_masks[i]) { |
| if (!first) { |
| strlcat(value, "|", sizeof(value)); |
| } |
| strlcat(value, channels_name_to_enum_table[j].name, sizeof(value)); |
| first = false; |
| break; |
| } |
| } |
| i++; |
| } |
| str_parms_add_str(reply, AUDIO_PARAMETER_STREAM_SUP_CHANNELS, value); |
| str = str_parms_to_str(reply); |
| } else { |
| voice_extn_out_get_parameters(out, query, reply); |
| str = str_parms_to_str(reply); |
| } |
| |
| |
| ret = str_parms_get_str(query, "is_direct_pcm_track", value, sizeof(value)); |
| if (ret >= 0) { |
| value[0] = '\0'; |
| if (out->flags & AUDIO_OUTPUT_FLAG_DIRECT && |
| !(out->flags & AUDIO_OUTPUT_FLAG_COMPRESS_OFFLOAD)) { |
| ALOGV("in direct_pcm"); |
| strlcat(value, "true", sizeof(value)); |
| } else { |
| ALOGV("not in direct_pcm"); |
| strlcat(value, "false", sizeof(value)); |
| } |
| str_parms_add_str(reply, "is_direct_pcm_track", value); |
| if (str) |
| free(str); |
| str = str_parms_to_str(reply); |
| } |
| |
| ret = str_parms_get_str(query, AUDIO_PARAMETER_STREAM_SUP_FORMATS, value, sizeof(value)); |
| if (ret >= 0) { |
| value[0] = '\0'; |
| i = 0; |
| first = true; |
| while (out->supported_formats[i] != 0) { |
| for (j = 0; j < ARRAY_SIZE(formats_name_to_enum_table); j++) { |
| if (formats_name_to_enum_table[j].value == out->supported_formats[i]) { |
| if (!first) { |
| strlcat(value, "|", sizeof(value)); |
| } |
| strlcat(value, formats_name_to_enum_table[j].name, sizeof(value)); |
| first = false; |
| break; |
| } |
| } |
| i++; |
| } |
| str_parms_add_str(reply, AUDIO_PARAMETER_STREAM_SUP_FORMATS, value); |
| if (str) |
| free(str); |
| str = str_parms_to_str(reply); |
| } |
| |
| ret = str_parms_get_str(query, AUDIO_PARAMETER_STREAM_SUP_SAMPLING_RATES, value, sizeof(value)); |
| if (ret >= 0) { |
| value[0] = '\0'; |
| i = 0; |
| first = true; |
| while (out->supported_sample_rates[i] != 0) { |
| for (j = 0; j < ARRAY_SIZE(out_sample_rates_name_to_enum_table); j++) { |
| if (out_sample_rates_name_to_enum_table[j].value == out->supported_sample_rates[i]) { |
| if (!first) { |
| strlcat(value, "|", sizeof(value)); |
| } |
| strlcat(value, out_sample_rates_name_to_enum_table[j].name, sizeof(value)); |
| first = false; |
| break; |
| } |
| } |
| i++; |
| } |
| str_parms_add_str(reply, AUDIO_PARAMETER_STREAM_SUP_SAMPLING_RATES, value); |
| if (str) |
| free(str); |
| str = str_parms_to_str(reply); |
| } |
| |
| if (str_parms_get_str(query, "supports_hw_suspend", value, sizeof(value)) >= 0) { |
| //only low latency track supports suspend_resume |
| str_parms_add_int(reply, "supports_hw_suspend", |
| (out->hal_output_suspend_supported)); |
| if (str) |
| free(str); |
| str = str_parms_to_str(reply); |
| } |
| |
| |
| str_parms_destroy(query); |
| str_parms_destroy(reply); |
| ALOGV("%s: exit: returns - %s", __func__, str); |
| return str; |
| } |
| |
| static uint32_t out_get_latency(const struct audio_stream_out *stream) |
| { |
| uint32_t period_ms; |
| struct stream_out *out = (struct stream_out *)stream; |
| uint32_t latency = 0; |
| |
| if (is_offload_usecase(out->usecase)) { |
| lock_output_stream(out); |
| latency = audio_extn_utils_compress_get_dsp_latency(out); |
| pthread_mutex_unlock(&out->lock); |
| } else if ((out->realtime) || |
| (out->usecase == USECASE_AUDIO_PLAYBACK_MMAP)) { |
| // since the buffer won't be filled up faster than realtime, |
| // return a smaller number |
| if (out->config.rate) |
| period_ms = (out->af_period_multiplier * out->config.period_size * |
| 1000) / (out->config.rate); |
| else |
| period_ms = 0; |
| latency = period_ms + platform_render_latency(out) / 1000; |
| } else { |
| latency = (out->config.period_count * out->config.period_size * 1000) / |
| (out->config.rate); |
| if (out->usecase == USECASE_AUDIO_PLAYBACK_DEEP_BUFFER) |
| latency += platform_render_latency(out)/1000; |
| } |
| |
| if (!out->standby && is_a2dp_out_device_type(&out->device_list)) |
| latency += audio_extn_a2dp_get_encoder_latency(); |
| |
| ALOGV("%s: Latency %d", __func__, latency); |
| return latency; |
| } |
| |
| static float AmpToDb(float amplification) |
| { |
| float db = DSD_VOLUME_MIN_DB; |
| if (amplification > 0) { |
| db = 20 * log10(amplification); |
| if(db < DSD_VOLUME_MIN_DB) |
| return DSD_VOLUME_MIN_DB; |
| } |
| return db; |
| } |
| |
| static int out_set_mmap_volume(struct audio_stream_out *stream, float left, |
| float right) |
| { |
| struct stream_out *out = (struct stream_out *)stream; |
| long volume = 0; |
| char mixer_ctl_name[128] = ""; |
| struct audio_device *adev = out->dev; |
| struct mixer_ctl *ctl = NULL; |
| int pcm_device_id = platform_get_pcm_device_id(out->usecase, |
| PCM_PLAYBACK); |
| |
| snprintf(mixer_ctl_name, sizeof(mixer_ctl_name), |
| "Playback %d Volume", pcm_device_id); |
| ctl = mixer_get_ctl_by_name(adev->mixer, mixer_ctl_name); |
| if (!ctl) { |
| ALOGE("%s: Could not get ctl for mixer cmd - %s", |
| __func__, mixer_ctl_name); |
| return -EINVAL; |
| } |
| if (left != right) |
| ALOGW("%s: Left and right channel volume mismatch:%f,%f", |
| __func__, left, right); |
| volume = (long)(left * (MMAP_PLAYBACK_VOLUME_MAX*1.0)); |
| if (mixer_ctl_set_value(ctl, 0, volume) < 0){ |
| ALOGE("%s:ctl for mixer cmd - %s, volume %ld returned error", |
| __func__, mixer_ctl_name, volume); |
| return -EINVAL; |
| } |
| return 0; |
| } |
| |
| static int out_set_compr_volume(struct audio_stream_out *stream, float left, |
| float right) |
| { |
| struct stream_out *out = (struct stream_out *)stream; |
| long volume[2]; |
| char mixer_ctl_name[128]; |
| struct audio_device *adev = out->dev; |
| struct mixer_ctl *ctl; |
| int pcm_device_id = platform_get_pcm_device_id(out->usecase, |
| PCM_PLAYBACK); |
| |
| snprintf(mixer_ctl_name, sizeof(mixer_ctl_name), |
| "Compress Playback %d Volume", pcm_device_id); |
| ctl = mixer_get_ctl_by_name(adev->mixer, mixer_ctl_name); |
| if (!ctl) { |
| ALOGE("%s: Could not get ctl for mixer cmd - %s", |
| __func__, mixer_ctl_name); |
| return -EINVAL; |
| } |
| ALOGE("%s:ctl for mixer cmd - %s, left %f, right %f", |
| __func__, mixer_ctl_name, left, right); |
| volume[0] = (int)(left * COMPRESS_PLAYBACK_VOLUME_MAX); |
| volume[1] = (int)(right * COMPRESS_PLAYBACK_VOLUME_MAX); |
| mixer_ctl_set_array(ctl, volume, sizeof(volume)/sizeof(volume[0])); |
| |
| return 0; |
| } |
| |
| static int out_set_voip_volume(struct audio_stream_out *stream, float left, |
| float right) |
| { |
| struct stream_out *out = (struct stream_out *)stream; |
| char mixer_ctl_name[] = "App Type Gain"; |
| struct audio_device *adev = out->dev; |
| struct mixer_ctl *ctl; |
| long set_values[4]; |
| |
| if (!is_valid_volume(left, right)) { |
| ALOGE("%s: Invalid stream volume for left=%f, right=%f", |
| __func__, left, right); |
| return -EINVAL; |
| } |
| |
| ctl = mixer_get_ctl_by_name(adev->mixer, mixer_ctl_name); |
| if (!ctl) { |
| ALOGE("%s: Could not get ctl for mixer cmd - %s", |
| __func__, mixer_ctl_name); |
| return -EINVAL; |
| } |
| |
| set_values[0] = 0; //0: Rx Session 1:Tx Session |
| set_values[1] = out->app_type_cfg.app_type; |
| set_values[2] = (long)(left * VOIP_PLAYBACK_VOLUME_MAX); |
| set_values[3] = (long)(right * VOIP_PLAYBACK_VOLUME_MAX); |
| |
| mixer_ctl_set_array(ctl, set_values, ARRAY_SIZE(set_values)); |
| return 0; |
| } |
| |
| static int out_set_pcm_volume(struct audio_stream_out *stream, float left, |
| float right) |
| { |
| struct stream_out *out = (struct stream_out *)stream; |
| /* Volume control for pcm playback */ |
| if (left != right) { |
| return -EINVAL; |
| } else { |
| char mixer_ctl_name[128]; |
| struct audio_device *adev = out->dev; |
| struct mixer_ctl *ctl; |
| int pcm_device_id = platform_get_pcm_device_id(out->usecase, PCM_PLAYBACK); |
| snprintf(mixer_ctl_name, sizeof(mixer_ctl_name), "Playback %d Volume", pcm_device_id); |
| ctl = mixer_get_ctl_by_name(adev->mixer, mixer_ctl_name); |
| if (!ctl) { |
| ALOGE("%s : Could not get ctl for mixer cmd - %s", __func__, mixer_ctl_name); |
| return -EINVAL; |
| } |
| |
| int volume = (int) (left * PCM_PLAYBACK_VOLUME_MAX); |
| int ret = mixer_ctl_set_value(ctl, 0, volume); |
| if (ret < 0) { |
| ALOGE("%s: Could not set ctl, error:%d ", __func__, ret); |
| return -EINVAL; |
| } |
| |
| ALOGV("%s : Pcm set volume value %d left %f", __func__, volume, left); |
| |
| return 0; |
| } |
| } |
| |
| static int out_set_volume(struct audio_stream_out *stream, float left, |
| float right) |
| { |
| struct stream_out *out = (struct stream_out *)stream; |
| int volume[2]; |
| int ret = 0; |
| |
| ALOGD("%s: called with left_vol=%f, right_vol=%f", __func__, left, right); |
| if (out->usecase == USECASE_AUDIO_PLAYBACK_MULTI_CH) { |
| /* only take left channel into account: the API is for stereo anyway */ |
| out->muted = (left == 0.0f); |
| return 0; |
| } else if (is_offload_usecase(out->usecase)) { |
| if (audio_extn_passthru_is_passthrough_stream(out)) { |
| /* |
| * Set mute or umute on HDMI passthrough stream. |
| * Only take left channel into account. |
| * Mute is 0 and unmute 1 |
| */ |
| audio_extn_passthru_set_volume(out, (left == 0.0f)); |
| } else if (out->format == AUDIO_FORMAT_DSD){ |
| char mixer_ctl_name[128] = "DSD Volume"; |
| struct audio_device *adev = out->dev; |
| struct mixer_ctl *ctl = mixer_get_ctl_by_name(adev->mixer, mixer_ctl_name); |
| |
| if (!ctl) { |
| ALOGE("%s: Could not get ctl for mixer cmd - %s", |
| __func__, mixer_ctl_name); |
| return -EINVAL; |
| } |
| volume[0] = (long)(AmpToDb(left)); |
| volume[1] = (long)(AmpToDb(right)); |
| mixer_ctl_set_array(ctl, volume, sizeof(volume)/sizeof(volume[0])); |
| return 0; |
| } else if (compare_device_type(&out->device_list, AUDIO_DEVICE_OUT_BUS) && |
| (out->car_audio_stream == CAR_AUDIO_STREAM_MEDIA)) { |
| ALOGD("%s: Overriding offload set volume for media bus stream", __func__); |
| struct listnode *node = NULL; |
| list_for_each(node, &adev->active_outputs_list) { |
| streams_output_ctxt_t *out_ctxt = node_to_item(node, |
| streams_output_ctxt_t, |
| list); |
| if (out_ctxt->output->usecase == USECASE_AUDIO_PLAYBACK_MEDIA) { |
| out->volume_l = out_ctxt->output->volume_l; |
| out->volume_r = out_ctxt->output->volume_r; |
| } |
| } |
| pthread_mutex_lock(&out->latch_lock); |
| if (!out->a2dp_muted) { |
| ret = out_set_compr_volume(&out->stream, out->volume_l, out->volume_r); |
| } |
| pthread_mutex_unlock(&out->latch_lock); |
| return ret; |
| } else { |
| pthread_mutex_lock(&out->latch_lock); |
| ALOGV("%s: compress mute %d", __func__, out->a2dp_muted); |
| if (!out->a2dp_muted) |
| ret = out_set_compr_volume(stream, left, right); |
| out->volume_l = left; |
| out->volume_r = right; |
| pthread_mutex_unlock(&out->latch_lock); |
| return ret; |
| } |
| } else if (out->usecase == USECASE_AUDIO_PLAYBACK_VOIP) { |
| out->app_type_cfg.gain[0] = (int)(left * VOIP_PLAYBACK_VOLUME_MAX); |
| out->app_type_cfg.gain[1] = (int)(right * VOIP_PLAYBACK_VOLUME_MAX); |
| pthread_mutex_lock(&out->latch_lock); |
| if (!out->standby) { |
| audio_extn_utils_send_app_type_gain(out->dev, |
| out->app_type_cfg.app_type, |
| &out->app_type_cfg.gain[0]); |
| if (!out->a2dp_muted) |
| ret = out_set_voip_volume(stream, left, right); |
| } |
| out->volume_l = left; |
| out->volume_r = right; |
| pthread_mutex_unlock(&out->latch_lock); |
| return ret; |
| } else if (out->usecase == USECASE_AUDIO_PLAYBACK_MMAP) { |
| ALOGV("%s: MMAP set volume called", __func__); |
| if (!out->standby) |
| ret = out_set_mmap_volume(stream, left, right); |
| out->volume_l = left; |
| out->volume_r = right; |
| return ret; |
| } else if (out->usecase == USECASE_AUDIO_PLAYBACK_LOW_LATENCY || |
| out->usecase == USECASE_AUDIO_PLAYBACK_DEEP_BUFFER || |
| out->usecase == USECASE_AUDIO_PLAYBACK_ULL) { |
| pthread_mutex_lock(&out->latch_lock); |
| /* Volume control for pcm playback */ |
| if (!out->standby && !out->a2dp_muted) |
| ret = out_set_pcm_volume(stream, left, right); |
| else |
| out->apply_volume = true; |
| |
| out->volume_l = left; |
| out->volume_r = right; |
| pthread_mutex_unlock(&out->latch_lock); |
| return ret; |
| } else if (audio_extn_auto_hal_is_bus_device_usecase(out->usecase)) { |
| ALOGV("%s: bus device set volume called", __func__); |
| pthread_mutex_lock(&out->latch_lock); |
| if (!out->standby && !out->a2dp_muted) |
| ret = out_set_pcm_volume(stream, left, right); |
| out->volume_l = left; |
| out->volume_r = right; |
| pthread_mutex_unlock(&out->latch_lock); |
| return ret; |
| } |
| |
| return -ENOSYS; |
| } |
| |
| static void update_frames_written(struct stream_out *out, size_t bytes) |
| { |
| size_t bpf = 0; |
| |
| if (is_offload_usecase(out->usecase) && !out->non_blocking && |
| !(out->flags & AUDIO_OUTPUT_FLAG_COMPRESS_OFFLOAD)) |
| bpf = 1; |
| else if (!is_offload_usecase(out->usecase)) |
| bpf = audio_bytes_per_sample(out->format) * |
| audio_channel_count_from_out_mask(out->channel_mask); |
| |
| pthread_mutex_lock(&out->position_query_lock); |
| if (bpf != 0) { |
| out->written += bytes / bpf; |
| clock_gettime(CLOCK_MONOTONIC, &out->writeAt); |
| } |
| pthread_mutex_unlock(&out->position_query_lock); |
| } |
| |
| int split_and_write_audio_haptic_data(struct stream_out *out, |
| const void *buffer, size_t bytes_to_write) |
| { |
| struct audio_device *adev = out->dev; |
| |
| int ret = 0; |
| size_t channel_count = audio_channel_count_from_out_mask(out->channel_mask); |
| size_t bytes_per_sample = audio_bytes_per_sample(out->format); |
| size_t frame_size = channel_count * bytes_per_sample; |
| size_t frame_count = bytes_to_write / frame_size; |
| |
| bool force_haptic_path = |
| property_get_bool("vendor.audio.test_haptic", false); |
| |
| // extract Haptics data from Audio buffer |
| bool alloc_haptic_buffer = false; |
| int haptic_channel_count = adev->haptics_config.channels; |
| size_t haptic_frame_size = bytes_per_sample * haptic_channel_count; |
| size_t audio_frame_size = frame_size - haptic_frame_size; |
| size_t total_haptic_buffer_size = frame_count * haptic_frame_size; |
| |
| if (adev->haptic_buffer == NULL) { |
| alloc_haptic_buffer = true; |
| } else if (adev->haptic_buffer_size < total_haptic_buffer_size) { |
| free(adev->haptic_buffer); |
| adev->haptic_buffer_size = 0; |
| alloc_haptic_buffer = true; |
| } |
| |
| if (alloc_haptic_buffer) { |
| adev->haptic_buffer = (uint8_t *)calloc(1, total_haptic_buffer_size); |
| if(adev->haptic_buffer == NULL) { |
| ALOGE("%s: failed to allocate mem for dev->haptic_buffer", __func__); |
| return -ENOMEM; |
| } |
| adev->haptic_buffer_size = total_haptic_buffer_size; |
| } |
| |
| size_t src_index = 0, aud_index = 0, hap_index = 0; |
| uint8_t *audio_buffer = (uint8_t *)buffer; |
| uint8_t *haptic_buffer = adev->haptic_buffer; |
| |
| // This is required for testing only. This works for stereo data only. |
| // One channel is fed to audio stream and other to haptic stream for testing. |
| if (force_haptic_path) |
| audio_frame_size = haptic_frame_size = bytes_per_sample; |
| |
| for (size_t i = 0; i < frame_count; i++) { |
| memcpy(audio_buffer + aud_index, audio_buffer + src_index, |
| audio_frame_size); |
| aud_index += audio_frame_size; |
| src_index += audio_frame_size; |
| |
| if (adev->haptic_pcm) |
| memcpy(haptic_buffer + hap_index, audio_buffer + src_index, |
| haptic_frame_size); |
| hap_index += haptic_frame_size; |
| src_index += haptic_frame_size; |
| |
| // This is required for testing only. |
| // Discard haptic channel data. |
| if (force_haptic_path) |
| src_index += haptic_frame_size; |
| } |
| |
| // write to audio pipeline |
| ret = pcm_write(out->pcm, (void *)audio_buffer, |
| frame_count * audio_frame_size); |
| |
| // write to haptics pipeline |
| if (adev->haptic_pcm) |
| ret = pcm_write(adev->haptic_pcm, (void *)adev->haptic_buffer, |
| frame_count * haptic_frame_size); |
| |
| return ret; |
| } |
| |
| #ifdef NO_AUDIO_OUT |
| static ssize_t out_write_for_no_output(struct audio_stream_out *stream, |
| const void *buffer __unused, size_t bytes) |
| { |
| struct stream_out *out = (struct stream_out *)stream; |
| |
| /* No Output device supported other than BT for playback. |
| * Sleep for the amount of buffer duration |
| */ |
| lock_output_stream(out); |
| usleep(bytes * 1000000 / audio_stream_out_frame_size( |
| (const struct audio_stream_out *)&out->stream) / |
| out_get_sample_rate(&out->stream.common)); |
| pthread_mutex_unlock(&out->lock); |
| return bytes; |
| } |
| #endif |
| |
| static ssize_t out_write(struct audio_stream_out *stream, const void *buffer, |
| size_t bytes) |
| { |
| struct stream_out *out = (struct stream_out *)stream; |
| struct audio_device *adev = out->dev; |
| ssize_t ret = 0; |
| int channels = 0; |
| const size_t frame_size = audio_stream_out_frame_size(stream); |
| const size_t frames = (frame_size != 0) ? bytes / frame_size : bytes; |
| struct audio_usecase *usecase = NULL; |
| uint32_t compr_passthr = 0; |
| |
| ATRACE_BEGIN("out_write"); |
| lock_output_stream(out); |
| |
| if (CARD_STATUS_OFFLINE == out->card_status) { |
| |
| if (out->flags & AUDIO_OUTPUT_FLAG_COMPRESS_OFFLOAD) { |
| /*during SSR for compress usecase we should return error to flinger*/ |
| ALOGD(" copl %s: sound card is not active/SSR state", __func__); |
| pthread_mutex_unlock(&out->lock); |
| ATRACE_END(); |
| return -ENETRESET; |
| } else { |
| ALOGD(" %s: sound card is not active/SSR state", __func__); |
| ret= -EIO; |
| goto exit; |
| } |
| } |
| |
| if (audio_extn_passthru_should_drop_data(out)) { |
| ALOGV(" %s : Drop data as compress passthrough session is going on", __func__); |
| ret = -EIO; |
| goto exit; |
| } |
| |
| if (out->usecase == USECASE_AUDIO_PLAYBACK_MMAP) { |
| ret = -EINVAL; |
| goto exit; |
| } |
| |
| if (compare_device_type(&out->device_list, AUDIO_DEVICE_OUT_AUX_DIGITAL) && |
| !out->is_iec61937_info_available) { |
| |
| if (!audio_extn_passthru_is_passthrough_stream(out)) { |
| out->is_iec61937_info_available = true; |
| } else if (audio_extn_passthru_is_enabled()) { |
| audio_extn_passthru_update_stream_configuration(adev, out, buffer, bytes); |
| out->is_iec61937_info_available = true; |
| |
| if((out->format == AUDIO_FORMAT_DTS) || |
| (out->format == AUDIO_FORMAT_DTS_HD)) { |
| ret = audio_extn_passthru_update_dts_stream_configuration(out, |
| buffer, bytes); |
| if (ret) { |
| if (ret != -ENOSYS) { |
| out->is_iec61937_info_available = false; |
| ALOGD("iec61937 transmission info not yet updated retry"); |
| } |
| } else if (!out->standby) { |
| /* if stream has started and after that there is |
| * stream config change (iec transmission config) |
| * then trigger select_device to update backend configuration. |
| */ |
| out->stream_config_changed = true; |
| pthread_mutex_lock(&adev->lock); |
| select_devices(adev, out->usecase); |
| if (!audio_extn_passthru_is_supported_backend_edid_cfg(adev, out)) { |
| pthread_mutex_unlock(&adev->lock); |
| ret = -EINVAL; |
| goto exit; |
| } |
| pthread_mutex_unlock(&adev->lock); |
| out->stream_config_changed = false; |
| out->is_iec61937_info_available = true; |
| } |
| } |
| |
| #ifdef AUDIO_GKI_ENABLED |
| /* out->compr_config.codec->reserved[0] is for compr_passthr */ |
| compr_passthr = out->compr_config.codec->reserved[0]; |
| #else |
| compr_passthr = out->compr_config.codec->compr_passthr; |
| #endif |
| |
| if ((channels < (int)audio_channel_count_from_out_mask(out->channel_mask)) && |
| (compr_passthr == PASSTHROUGH) && |
| (out->is_iec61937_info_available == true)) { |
| ALOGE("%s: ERROR: Unsupported channel config in passthrough mode", __func__); |
| ret = -EINVAL; |
| goto exit; |
| } |
| } |
| } |
| |
| if (is_a2dp_out_device_type(&out->device_list) && |
| (audio_extn_a2dp_source_is_suspended())) { |
| if (!(compare_device_type(&out->device_list, AUDIO_DEVICE_OUT_SPEAKER) || |
| compare_device_type(&out->device_list, AUDIO_DEVICE_OUT_SPEAKER_SAFE))) { |
| if (!is_offload_usecase(out->usecase)) { |
| ret = -EIO; |
| goto exit; |
| } |
| } |
| } |
| |
| if (is_usb_out_device_type(&out->device_list) && |
| !audio_extn_usb_connected(NULL)) { |
| ret = -EIO; |
| goto exit; |
| } |
| |
| if (out->standby) { |
| out->standby = false; |
| const int64_t startNs = systemTime(SYSTEM_TIME_MONOTONIC); |
| |
| pthread_mutex_lock(&adev->lock); |
| if (out->usecase == USECASE_COMPRESS_VOIP_CALL) |
| ret = voice_extn_compress_voip_start_output_stream(out); |
| else |
| ret = start_output_stream(out); |
| /* ToDo: If use case is compress offload should return 0 */ |
| if (ret != 0) { |
| out->standby = true; |
| pthread_mutex_unlock(&adev->lock); |
| goto exit; |
| } |
| out->started = 1; |
| out->last_fifo_valid = false; // we're coming out of standby, last_fifo isn't valid. |
| |
| if ((last_known_cal_step != -1) && (adev->platform != NULL)) { |
| ALOGD("%s: retry previous failed cal level set", __func__); |
| platform_send_gain_dep_cal(adev->platform, last_known_cal_step); |
| last_known_cal_step = -1; |
| } |
| pthread_mutex_unlock(&adev->lock); |
| |
| if ((out->is_iec61937_info_available == true) && |
| (audio_extn_passthru_is_passthrough_stream(out))&& |
| (!audio_extn_passthru_is_supported_backend_edid_cfg(adev, out))) { |
| ret = -EINVAL; |
| goto exit; |
| } |
| if (out->set_dual_mono) |
| audio_extn_send_dual_mono_mixing_coefficients(out); |
| |
| #ifndef LINUX_ENABLED |
| // log startup time in ms. |
| simple_stats_log( |
| &out->start_latency_ms, (systemTime(SYSTEM_TIME_MONOTONIC) - startNs) * 1e-6); |
| #endif |
| } |
| |
| if (adev->is_channel_status_set == false && |
| compare_device_type(&out->device_list, |
| AUDIO_DEVICE_OUT_AUX_DIGITAL)) { |
| audio_utils_set_hdmi_channel_status(out, (void *)buffer, bytes); |
| adev->is_channel_status_set = true; |
| } |
| |
| if ((adev->use_old_pspd_mix_ctrl == true) && |
| (out->pspd_coeff_sent == false)) { |
| /* |
| * Need to resend pspd coefficients after stream started for |
| * older kernel version as it does not save the coefficients |
| * and also stream has to be started for coeff to apply. |
| */ |
| usecase = get_usecase_from_list(adev, out->usecase); |
| if (usecase != NULL) { |
| audio_extn_set_custom_mtmx_params_v2(adev, usecase, true); |
| out->pspd_coeff_sent = true; |
| } |
| } |
| |
| if (is_offload_usecase(out->usecase)) { |
| ALOGVV("copl(%p): writing buffer (%zu bytes) to compress device", out, bytes); |
| if (out->send_new_metadata) { |
| ALOGD("copl(%p):send new gapless metadata", out); |
| compress_set_gapless_metadata(out->compr, &out->gapless_mdata); |
| out->send_new_metadata = 0; |
| if (out->send_next_track_params && out->is_compr_metadata_avail) { |
| ALOGD("copl(%p):send next track params in gapless", out); |
| compress_set_next_track_param(out->compr, &(out->compr_config.codec->options)); |
| out->send_next_track_params = false; |
| out->is_compr_metadata_avail = false; |
| } |
| } |
| if (!(out->flags & AUDIO_OUTPUT_FLAG_COMPRESS_OFFLOAD) && |
| (out->convert_buffer) != NULL) { |
| |
| if ((bytes > out->hal_fragment_size)) { |
| ALOGW("Error written bytes %zu > %d (fragment_size)", |
| bytes, out->hal_fragment_size); |
| pthread_mutex_unlock(&out->lock); |
| ATRACE_END(); |
| return -EINVAL; |
| } else { |
| audio_format_t dst_format = out->hal_op_format; |
| audio_format_t src_format = out->hal_ip_format; |
| |
| /* prevent division-by-zero */ |
| uint32_t bitwidth_src = format_to_bitwidth_table[src_format]; |
| uint32_t bitwidth_dst = format_to_bitwidth_table[dst_format]; |
| if ((bitwidth_src == 0) || (bitwidth_dst == 0)) { |
| ALOGE("%s: Error bitwidth == 0", __func__); |
| pthread_mutex_unlock(&out->lock); |
| ATRACE_END(); |
| return -EINVAL; |
| } |
| |
| uint32_t frames = bytes / format_to_bitwidth_table[src_format]; |
| uint32_t bytes_to_write = frames * format_to_bitwidth_table[dst_format]; |
| |
| memcpy_by_audio_format(out->convert_buffer, |
| dst_format, |
| buffer, |
| src_format, |
| frames); |
| |
| ret = compress_write(out->compr, out->convert_buffer, |
| bytes_to_write); |
| |
| /*Convert written bytes in audio flinger format*/ |
| if (ret > 0) |
| ret = ((ret * format_to_bitwidth_table[out->format]) / |
| format_to_bitwidth_table[dst_format]); |
| } |
| } else |
| ret = compress_write(out->compr, buffer, bytes); |
| |
| if ((ret < 0 || ret == (ssize_t)bytes) && !out->non_blocking) |
| update_frames_written(out, bytes); |
| |
| if (ret < 0) |
| ret = -errno; |
| ALOGVV("%s: writing buffer (%zu bytes) to compress device returned %d", __func__, bytes, (int)ret); |
| /*msg to cb thread only if non blocking write is enabled*/ |
| if (ret >= 0 && ret < (ssize_t)bytes && out->non_blocking) { |
| ALOGD("No space available in compress driver, post msg to cb thread"); |
| send_offload_cmd_l(out, OFFLOAD_CMD_WAIT_FOR_BUFFER); |
| } else if (-ENETRESET == ret) { |
| ALOGE("copl %s: received sound card offline state on compress write", __func__); |
| out->card_status = CARD_STATUS_OFFLINE; |
| pthread_mutex_unlock(&out->lock); |
| out_on_error(&out->stream.common); |
| ATRACE_END(); |
| return ret; |
| } |
| |
| /* Call compr start only when non-zero bytes of data is there to be rendered */ |
| if (!out->playback_started && ret > 0) { |
| int status = compress_start(out->compr); |
| if (status < 0) { |
| ret = status; |
| ALOGE("%s: compr start failed with err %d", __func__, errno); |
| goto exit; |
| } |
| audio_extn_dts_eagle_fade(adev, true, out); |
| out->playback_started = 1; |
| pthread_mutex_lock(&out->latch_lock); |
| out->offload_state = OFFLOAD_STATE_PLAYING; |
| pthread_mutex_unlock(&out->latch_lock); |
| |
| audio_extn_dts_notify_playback_state(out->usecase, 0, out->sample_rate, |
| popcount(out->channel_mask), |
| out->playback_started); |
| } |
| pthread_mutex_unlock(&out->lock); |
| ATRACE_END(); |
| return ret; |
| } else { |
| if (out->pcm) { |
| size_t bytes_to_write = bytes; |
| if (out->muted) |
| memset((void *)buffer, 0, bytes); |
| ALOGV("%s: frames=%zu, frame_size=%zu, bytes_to_write=%zu", |
| __func__, frames, frame_size, bytes_to_write); |
| |
| if (out->usecase == USECASE_INCALL_MUSIC_UPLINK || |
| out->usecase == USECASE_INCALL_MUSIC_UPLINK2 || |
| (out->usecase == USECASE_AUDIO_PLAYBACK_VOIP && |
| !audio_extn_utils_is_vendor_enhanced_fwk())) { |
| size_t channel_count = audio_channel_count_from_out_mask(out->channel_mask); |
| int16_t *src = (int16_t *)buffer; |
| int16_t *dst = (int16_t *)buffer; |
| |
| LOG_ALWAYS_FATAL_IF(channel_count > 2 || |
| out->format != AUDIO_FORMAT_PCM_16_BIT, |
| "out_write called for %s use case with wrong properties", |
| use_case_table[out->usecase]); |
| |
| /* |
| * FIXME: this can be removed once audio flinger mixer supports |
| * mono output |
| */ |
| |
| /* |
| * Code below goes over each frame in the buffer and adds both |
| * L and R samples and then divides by 2 to convert to mono |
| */ |
| if (channel_count == 2) { |
| for (size_t i = 0; i < frames ; i++, dst++, src += 2) { |
| *dst = (int16_t)(((int32_t)src[0] + (int32_t)src[1]) >> 1); |
| } |
| bytes_to_write /= 2; |
| } |
| } |
| |
| // Note: since out_get_presentation_position() is called alternating with out_write() |
| // by AudioFlinger, we can check underruns using the prior timestamp read. |
| // (Alternately we could check if the buffer is empty using pcm_get_htimestamp(). |
| if (out->last_fifo_valid) { |
| // compute drain to see if there is an underrun. |
| const int64_t current_ns = systemTime(SYSTEM_TIME_MONOTONIC); // sys call |
| int64_t time_diff_ns = current_ns - out->last_fifo_time_ns; |
| int64_t frames_by_time = |
| ((time_diff_ns > 0) && (time_diff_ns < (INT64_MAX / out->config.rate))) ? |
| (time_diff_ns * out->config.rate / NANOS_PER_SECOND) : 0; |
| const int64_t underrun = frames_by_time - out->last_fifo_frames_remaining; |
| |
| if (underrun > 0) { |
| #ifndef LINUX_ENABLED |
| simple_stats_log(&out->fifo_underruns, underrun); |
| #endif |
| |
| ALOGW("%s: underrun(%lld) " |
| "frames_by_time(%lld) > out->last_fifo_frames_remaining(%lld)", |
| __func__, |
| (long long)out->fifo_underruns.n, |
| (long long)frames_by_time, |
| (long long)out->last_fifo_frames_remaining); |
| } |
| out->last_fifo_valid = false; // we're writing below, mark fifo info as stale. |
| } |
| |
| ALOGVV("%s: writing buffer (%zu bytes) to pcm device", __func__, bytes); |
| |
| long ns = 0; |
| |
| if (out->config.rate) |
| ns = pcm_bytes_to_frames(out->pcm, bytes)*1000000000LL/ |
| out->config.rate; |
| |
| request_out_focus(out, ns); |
| bool use_mmap = is_mmap_usecase(out->usecase) || out->realtime; |
| |
| if (use_mmap) |
| ret = pcm_mmap_write(out->pcm, (void *)buffer, bytes_to_write); |
| else if (out->hal_op_format != out->hal_ip_format && |
| out->convert_buffer != NULL) { |
| |
| memcpy_by_audio_format(out->convert_buffer, |
| out->hal_op_format, |
| buffer, |
| out->hal_ip_format, |
| out->config.period_size * out->config.channels); |
| |
| ret = pcm_write(out->pcm, out->convert_buffer, |
| (out->config.period_size * |
| out->config.channels * |
| format_to_bitwidth_table[out->hal_op_format])); |
| } else { |
| /* |
| * To avoid underrun in DSP when the application is not pumping |
| * data at required rate, check for the no. of bytes and ignore |
| * pcm_write if it is less than actual buffer size. |
| * It is a work around to a change in compress VOIP driver. |
| */ |
| if ((out->flags & AUDIO_OUTPUT_FLAG_VOIP_RX) && |
| bytes < (out->config.period_size * out->config.channels * |
| audio_bytes_per_sample(out->format))) { |
| size_t voip_buf_size = |
| out->config.period_size * out->config.channels * |
| audio_bytes_per_sample(out->format); |
| ALOGE("%s:VOIP underrun: bytes received %zu, required:%zu\n", |
| __func__, bytes, voip_buf_size); |
| usleep(((uint64_t)voip_buf_size - bytes) * |
| 1000000 / audio_stream_out_frame_size(stream) / |
| out_get_sample_rate(&out->stream.common)); |
| ret = 0; |
| } else { |
| if (out->usecase == USECASE_AUDIO_PLAYBACK_WITH_HAPTICS) |
| ret = split_and_write_audio_haptic_data(out, buffer, bytes); |
| else |
| ret = pcm_write(out->pcm, (void *)buffer, bytes_to_write); |
| } |
| } |
| |
| release_out_focus(out); |
| |
| if (ret < 0) |
| ret = -errno; |
| else if (ret > 0) |
| ret = -EINVAL; |
| } |
| } |
| |
| exit: |
| update_frames_written(out, bytes); |
| if (-ENETRESET == ret) { |
| out->card_status = CARD_STATUS_OFFLINE; |
| } |
| pthread_mutex_unlock(&out->lock); |
| |
| if (ret != 0) { |
| if (out->pcm) |
| ALOGE("%s: error %d, %s", __func__, (int)ret, pcm_get_error(out->pcm)); |
| if (out->usecase == USECASE_COMPRESS_VOIP_CALL) { |
| pthread_mutex_lock(&adev->lock); |
| voice_extn_compress_voip_close_output_stream(&out->stream.common); |
| out->started = 0; |
| pthread_mutex_unlock(&adev->lock); |
| out->standby = true; |
| } |
| out_on_error(&out->stream.common); |
| if (!(out->flags & AUDIO_OUTPUT_FLAG_COMPRESS_OFFLOAD)) { |
| /* prevent division-by-zero */ |
| uint32_t stream_size = audio_stream_out_frame_size(stream); |
| uint32_t srate = out_get_sample_rate(&out->stream.common); |
| |
| if ((stream_size == 0) || (srate == 0)) { |
| ALOGE("%s: stream_size= %d, srate = %d", __func__, stream_size, srate); |
| ATRACE_END(); |
| return -EINVAL; |
| } |
| usleep((uint64_t)bytes * 1000000 / stream_size / srate); |
| } |
| if (audio_extn_passthru_is_passthrough_stream(out)) { |
| //ALOGE("%s: write error, ret = %zd", __func__, ret); |
| ATRACE_END(); |
| return ret; |
| } |
| } |
| ATRACE_END(); |
| return bytes; |
| } |
| |
| static int out_get_render_position(const struct audio_stream_out *stream, |
| uint32_t *dsp_frames) |
| { |
| struct stream_out *out = (struct stream_out *)stream; |
| |
| if (dsp_frames == NULL) |
| return -EINVAL; |
| |
| *dsp_frames = 0; |
| if (is_offload_usecase(out->usecase)) { |
| ssize_t ret = 0; |
| |
| /* Below piece of code is not guarded against any lock beacuse audioFliner serializes |
| * this operation and adev_close_output_stream(where out gets reset). |
| */ |
| if (!out->non_blocking && !(out->flags & AUDIO_OUTPUT_FLAG_COMPRESS_OFFLOAD)) { |
| *dsp_frames = get_actual_pcm_frames_rendered(out, NULL); |
| ALOGVV("dsp_frames %d sampleRate %d",(int)*dsp_frames,out->sample_rate); |
| adjust_frames_for_device_delay(out, dsp_frames); |
| return 0; |
| } |
| |
| lock_output_stream(out); |
| if (out->compr != NULL && out->non_blocking) { |
| ret = compress_get_tstamp(out->compr, (unsigned long *)dsp_frames, |
| &out->sample_rate); |
| if (ret < 0) |
| ret = -errno; |
| ALOGVV("%s rendered frames %d sample_rate %d", |
| __func__, *dsp_frames, out->sample_rate); |
| } |
| if (-ENETRESET == ret) { |
| ALOGE(" ERROR: sound card not active Unable to get time stamp from compress driver"); |
| out->card_status = CARD_STATUS_OFFLINE; |
| ret = -EINVAL; |
| } else if(ret < 0) { |
| ALOGE(" ERROR: Unable to get time stamp from compress driver"); |
| ret = -EINVAL; |
| } else if (out->card_status == CARD_STATUS_OFFLINE) { |
| /* |
| * Handle corner case where compress session is closed during SSR |
| * and timestamp is queried |
| */ |
| ALOGE(" ERROR: sound card not active, return error"); |
| ret = -EINVAL; |
| } else if (out->prev_card_status_offline) { |
| ALOGE("ERROR: previously sound card was offline,return error"); |
| ret = -EINVAL; |
| } else { |
| ret = 0; |
| adjust_frames_for_device_delay(out, dsp_frames); |
| } |
| pthread_mutex_unlock(&out->lock); |
| return ret; |
| } else if (audio_is_linear_pcm(out->format)) { |
| *dsp_frames = out->written; |
| adjust_frames_for_device_delay(out, dsp_frames); |
| return 0; |
| } else |
| return -EINVAL; |
| } |
| |
| static int out_add_audio_effect(const struct audio_stream *stream __unused, |
| effect_handle_t effect __unused) |
| { |
| return 0; |
| } |
| |
| static int out_remove_audio_effect(const struct audio_stream *stream __unused, |
| effect_handle_t effect __unused) |
| { |
| return 0; |
| } |
| |
| static int out_get_next_write_timestamp(const struct audio_stream_out *stream __unused, |
| int64_t *timestamp __unused) |
| { |
| return -ENOSYS; |
| } |
| |
| static int out_get_presentation_position(const struct audio_stream_out *stream, |
| uint64_t *frames, struct timespec *timestamp) |
| { |
| struct stream_out *out = (struct stream_out *)stream; |
| int ret = -ENODATA; |
| unsigned long dsp_frames; |
| |
| /* below piece of code is not guarded against any lock because audioFliner serializes |
| * this operation and adev_close_output_stream( where out gets reset). |
| */ |
| if (is_offload_usecase(out->usecase) && !out->non_blocking && |
| !(out->flags & AUDIO_OUTPUT_FLAG_COMPRESS_OFFLOAD)) { |
| *frames = get_actual_pcm_frames_rendered(out, timestamp); |
| ALOGVV("frames %lld playedat %lld",(long long int)*frames, |
| timestamp->tv_sec * 1000000LL + timestamp->tv_nsec / 1000); |
| return 0; |
| } |
| |
| lock_output_stream(out); |
| |
| if (is_offload_usecase(out->usecase) && out->compr != NULL && out->non_blocking) { |
| ret = compress_get_tstamp(out->compr, &dsp_frames, |
| &out->sample_rate); |
| // Adjustment accounts for A2dp encoder latency with offload usecases |
| // Note: Encoder latency is returned in ms. |
| if (is_a2dp_out_device_type(&out->device_list)) { |
| unsigned long offset = |
| (audio_extn_a2dp_get_encoder_latency() * out->sample_rate / 1000); |
| dsp_frames = (dsp_frames > offset) ? (dsp_frames - offset) : 0; |
| } |
| ALOGVV("%s rendered frames %ld sample_rate %d", |
| __func__, dsp_frames, out->sample_rate); |
| *frames = dsp_frames; |
| if (ret < 0) |
| ret = -errno; |
| if (-ENETRESET == ret) { |
| ALOGE(" ERROR: sound card not active Unable to get time stamp from compress driver"); |
| out->card_status = CARD_STATUS_OFFLINE; |
| ret = -EINVAL; |
| } else |
| ret = 0; |
| /* this is the best we can do */ |
| clock_gettime(CLOCK_MONOTONIC, timestamp); |
| } else { |
| if (out->pcm) { |
| unsigned int avail; |
| if (pcm_get_htimestamp(out->pcm, &avail, timestamp) == 0) { |
| uint64_t signed_frames = 0; |
| uint64_t frames_temp = 0; |
| |
| if (out->kernel_buffer_size > avail) { |
| frames_temp = out->last_fifo_frames_remaining = out->kernel_buffer_size - avail; |
| } else { |
| ALOGW("%s: avail:%u > kernel_buffer_size:%zu clamping!", |
| __func__, avail, out->kernel_buffer_size); |
| avail = out->kernel_buffer_size; |
| frames_temp = out->last_fifo_frames_remaining = 0; |
| } |
| out->last_fifo_valid = true; |
| out->last_fifo_time_ns = audio_utils_ns_from_timespec(timestamp); |
| |
| if (out->written >= frames_temp) |
| signed_frames = out->written - frames_temp; |
| |
| ALOGVV("%s: frames:%lld avail:%u kernel_buffer_size:%zu", |
| __func__, (long long)signed_frames, avail, out->kernel_buffer_size); |
| |
| // This adjustment accounts for buffering after app processor. |
| // It is based on estimated DSP latency per use case, rather than exact. |
| frames_temp = platform_render_latency(out) * |
| out->sample_rate / 1000000LL; |
| if (signed_frames >= frames_temp) |
| signed_frames -= frames_temp; |
| |
| // Adjustment accounts for A2dp encoder latency with non offload usecases |
| // Note: Encoder latency is returned in ms, while platform_render_latency in us. |
| if (is_a2dp_out_device_type(&out->device_list)) { |
| frames_temp = audio_extn_a2dp_get_encoder_latency() * out->sample_rate / 1000; |
| if (signed_frames >= frames_temp) |
| signed_frames -= frames_temp; |
| } |
| |
| // It would be unusual for this value to be negative, but check just in case ... |
| *frames = signed_frames; |
| ret = 0; |
| } |
| } else if (out->card_status == CARD_STATUS_OFFLINE || |
| // audioflinger still needs position updates when A2DP is suspended |
| (is_a2dp_out_device_type(&out->device_list) && audio_extn_a2dp_source_is_suspended())) { |
| *frames = out->written; |
| clock_gettime(CLOCK_MONOTONIC, timestamp); |
| if (is_offload_usecase(out->usecase)) |
| ret = -EINVAL; |
| else |
| ret = 0; |
| } |
| } |
| pthread_mutex_unlock(&out->lock); |
| return ret; |
| } |
| |
| static int out_set_callback(struct audio_stream_out *stream, |
| stream_callback_t callback, void *cookie) |
| { |
| struct stream_out *out = (struct stream_out *)stream; |
| int ret; |
| |
| ALOGV("%s", __func__); |
| lock_output_stream(out); |
| out->client_callback = callback; |
| out->client_cookie = cookie; |
| if (out->adsp_hdlr_stream_handle) { |
| ret = audio_extn_adsp_hdlr_stream_set_callback( |
| out->adsp_hdlr_stream_handle, |
| callback, |
| cookie); |
| if (ret) |
| ALOGW("%s:adsp hdlr callback registration failed %d", |
| __func__, ret); |
| } |
| pthread_mutex_unlock(&out->lock); |
| return 0; |
| } |
| |
| static int out_pause(struct audio_stream_out* stream) |
| { |
| struct stream_out *out = (struct stream_out *)stream; |
| int status = -ENOSYS; |
| ALOGV("%s", __func__); |
| if (is_offload_usecase(out->usecase)) { |
| ALOGD("copl(%p):pause compress driver", out); |
| status = -ENODATA; |
| lock_output_stream(out); |
| pthread_mutex_lock(&out->latch_lock); |
| if (out->compr != NULL && out->offload_state == OFFLOAD_STATE_PLAYING) { |
| if (out->card_status != CARD_STATUS_OFFLINE) |
| status = compress_pause(out->compr); |
| |
| out->offload_state = OFFLOAD_STATE_PAUSED; |
| |
| if (audio_extn_passthru_is_active()) { |
| ALOGV("offload use case, pause passthru"); |
| audio_extn_passthru_on_pause(out); |
| } |
| |
| audio_extn_dts_eagle_fade(adev, false, out); |
| audio_extn_dts_notify_playback_state(out->usecase, 0, |
| out->sample_rate, popcount(out->channel_mask), |
| 0); |
| } |
| pthread_mutex_unlock(&out->latch_lock); |
| pthread_mutex_unlock(&out->lock); |
| } |
| return status; |
| } |
| |
| static int out_resume(struct audio_stream_out* stream) |
| { |
| struct stream_out *out = (struct stream_out *)stream; |
| int status = -ENOSYS; |
| ALOGV("%s", __func__); |
| if (is_offload_usecase(out->usecase)) { |
| ALOGD("copl(%p):resume compress driver", out); |
| status = -ENODATA; |
| lock_output_stream(out); |
| pthread_mutex_lock(&out->latch_lock); |
| if (out->compr != NULL && out->offload_state == OFFLOAD_STATE_PAUSED) { |
| if (out->card_status != CARD_STATUS_OFFLINE) { |
| status = compress_resume(out->compr); |
| } |
| if (!status) { |
| out->offload_state = OFFLOAD_STATE_PLAYING; |
| } |
| audio_extn_dts_eagle_fade(adev, true, out); |
| audio_extn_dts_notify_playback_state(out->usecase, 0, out->sample_rate, |
| popcount(out->channel_mask), 1); |
| } |
| pthread_mutex_unlock(&out->latch_lock); |
| pthread_mutex_unlock(&out->lock); |
| } |
| return status; |
| } |
| |
| static int out_drain(struct audio_stream_out* stream, audio_drain_type_t type ) |
| { |
| struct stream_out *out = (struct stream_out *)stream; |
| int status = -ENOSYS; |
| ALOGV("%s", __func__); |
| if (is_offload_usecase(out->usecase)) { |
| lock_output_stream(out); |
| if (type == AUDIO_DRAIN_EARLY_NOTIFY) |
| status = send_offload_cmd_l(out, OFFLOAD_CMD_PARTIAL_DRAIN); |
| else |
| status = send_offload_cmd_l(out, OFFLOAD_CMD_DRAIN); |
| pthread_mutex_unlock(&out->lock); |
| } |
| return status; |
| } |
| |
| static int out_flush(struct audio_stream_out* stream) |
| { |
| struct stream_out *out = (struct stream_out *)stream; |
| ALOGV("%s", __func__); |
| if (is_offload_usecase(out->usecase)) { |
| ALOGD("copl(%p):calling compress flush", out); |
| lock_output_stream(out); |
| pthread_mutex_lock(&out->latch_lock); |
| if (out->offload_state == OFFLOAD_STATE_PAUSED) { |
| pthread_mutex_unlock(&out->latch_lock); |
| stop_compressed_output_l(out); |
| } else { |
| ALOGW("%s called in invalid state %d", __func__, out->offload_state); |
| pthread_mutex_unlock(&out->latch_lock); |
| } |
| out->written = 0; |
| pthread_mutex_unlock(&out->lock); |
| ALOGD("copl(%p):out of compress flush", out); |
| return 0; |
| } |
| return -ENOSYS; |
| } |
| |
| static int out_stop(const struct audio_stream_out* stream) |
| { |
| struct stream_out *out = (struct stream_out *)stream; |
| struct audio_device *adev = out->dev; |
| int ret = -ENOSYS; |
| |
| ALOGV("%s", __func__); |
| pthread_mutex_lock(&adev->lock); |
| if (out->usecase == USECASE_AUDIO_PLAYBACK_MMAP && !out->standby && |
| out->playback_started && out->pcm != NULL) { |
| pcm_stop(out->pcm); |
| ret = stop_output_stream(out); |
| out->playback_started = false; |
| } |
| pthread_mutex_unlock(&adev->lock); |
| return ret; |
| } |
| |
| static int out_start(const struct audio_stream_out* stream) |
| { |
| struct stream_out *out = (struct stream_out *)stream; |
| struct audio_device *adev = out->dev; |
| int ret = -ENOSYS; |
| |
| ALOGV("%s", __func__); |
| pthread_mutex_lock(&adev->lock); |
| if (out->usecase == USECASE_AUDIO_PLAYBACK_MMAP && !out->standby && |
| !out->playback_started && out->pcm != NULL) { |
| ret = start_output_stream(out); |
| if (ret == 0) { |
| out->playback_started = true; |
| } |
| } |
| pthread_mutex_unlock(&adev->lock); |
| return ret; |
| } |
| |
| /* |
| * Modify config->period_count based on min_size_frames |
| */ |
| static void adjust_mmap_period_count(struct pcm_config *config, int32_t min_size_frames) |
| { |
| int periodCountRequested = (min_size_frames + config->period_size - 1) |
| / config->period_size; |
| int periodCount = MMAP_PERIOD_COUNT_MIN; |
| |
| ALOGV("%s original config.period_size = %d config.period_count = %d", |
| __func__, config->period_size, config->period_count); |
| |
| while (periodCount < periodCountRequested && (periodCount * 2) < MMAP_PERIOD_COUNT_MAX) { |
| periodCount *= 2; |
| } |
| config->period_count = periodCount; |
| |
| ALOGV("%s requested config.period_count = %d", __func__, config->period_count); |
| } |
| |
| // Read offset for the positional timestamp from a persistent vendor property. |
| // This is to workaround apparent inaccuracies in the timing information that |
| // is used by the AAudio timing model. The inaccuracies can cause glitches. |
| static int64_t get_mmap_out_time_offset() { |
| const int32_t kDefaultOffsetMicros = 0; |
| int32_t mmap_time_offset_micros = property_get_int32( |
| "persist.vendor.audio.out_mmap_delay_micros", kDefaultOffsetMicros); |
| ALOGI("mmap_time_offset_micros = %d for output", mmap_time_offset_micros); |
| return mmap_time_offset_micros * (int64_t)1000; |
| } |
| |
| static int out_create_mmap_buffer(const struct audio_stream_out *stream, |
| int32_t min_size_frames, |
| struct audio_mmap_buffer_info *info) |
| { |
| struct stream_out *out = (struct stream_out *)stream; |
| struct audio_device *adev = out->dev; |
| int ret = 0; |
| unsigned int offset1 = 0; |
| unsigned int frames1 = 0; |
| const char *step = ""; |
| uint32_t mmap_size; |
| uint32_t buffer_size; |
| |
| ALOGD("%s", __func__); |
| lock_output_stream(out); |
| pthread_mutex_lock(&adev->lock); |
| |
| if (CARD_STATUS_OFFLINE == out->card_status || |
| CARD_STATUS_OFFLINE == adev->card_status) { |
| ALOGW("out->card_status or adev->card_status offline, try again"); |
| ret = -EIO; |
| goto exit; |
| } |
| if (info == NULL || !(min_size_frames > 0 && min_size_frames < INT32_MAX)) { |
| ALOGE("%s: info = %p, min_size_frames = %d", __func__, info, min_size_frames); |
| ret = -EINVAL; |
| goto exit; |
| } |
| if (out->usecase != USECASE_AUDIO_PLAYBACK_MMAP || !out->standby) { |
| ALOGE("%s: usecase = %d, standby = %d", __func__, out->usecase, out->standby); |
| ret = -ENOSYS; |
| goto exit; |
| } |
| out->pcm_device_id = platform_get_pcm_device_id(out->usecase, PCM_PLAYBACK); |
| if (out->pcm_device_id < 0) { |
| ALOGE("%s: Invalid PCM device id(%d) for the usecase(%d)", |
| __func__, out->pcm_device_id, out->usecase); |
| ret = -EINVAL; |
| goto exit; |
| } |
| |
| adjust_mmap_period_count(&out->config, min_size_frames); |
| |
| ALOGD("%s: Opening PCM device card_id(%d) device_id(%d), channels %d", |
| __func__, adev->snd_card, out->pcm_device_id, out->config.channels); |
| out->pcm = pcm_open(adev->snd_card, out->pcm_device_id, |
| (PCM_OUT | PCM_MMAP | PCM_NOIRQ | PCM_MONOTONIC), &out->config); |
| if (errno == ENETRESET && !pcm_is_ready(out->pcm)) { |
| ALOGE("%s: pcm_open failed errno:%d\n", __func__, errno); |
| out->card_status = CARD_STATUS_OFFLINE; |
| adev->card_status = CARD_STATUS_OFFLINE; |
| ret = -EIO; |
| goto exit; |
| } |
| |
| if (out->pcm == NULL || !pcm_is_ready(out->pcm)) { |
| step = "open"; |
| ret = -ENODEV; |
| goto exit; |
| } |
| ret = pcm_mmap_begin(out->pcm, &info->shared_memory_address, &offset1, &frames1); |
| if (ret < 0) { |
| step = "begin"; |
| goto exit; |
| } |
| |
| info->flags = 0; |
| info->buffer_size_frames = pcm_get_buffer_size(out->pcm); |
| buffer_size = pcm_frames_to_bytes(out->pcm, info->buffer_size_frames); |
| info->burst_size_frames = out->config.period_size; |
| ret = platform_get_mmap_data_fd(adev->platform, |
| out->pcm_device_id, 0 /*playback*/, |
| &info->shared_memory_fd, |
| &mmap_size); |
| if (ret < 0) { |
| // Fall back to non exclusive mode |
| info->shared_memory_fd = pcm_get_poll_fd(out->pcm); |
| } else { |
| out->mmap_shared_memory_fd = info->shared_memory_fd; // for closing later |
| ALOGV("%s: opened mmap_shared_memory_fd = %d", __func__, out->mmap_shared_memory_fd); |
| |
| if (mmap_size < buffer_size) { |
| step = "mmap"; |
| goto exit; |
| } |
| info->flags |= AUDIO_MMAP_APPLICATION_SHAREABLE; |
| } |
| memset(info->shared_memory_address, 0, pcm_frames_to_bytes(out->pcm, |
| info->buffer_size_frames)); |
| |
| ret = pcm_mmap_commit(out->pcm, 0, MMAP_PERIOD_SIZE); |
| if (ret < 0) { |
| step = "commit"; |
| goto exit; |
| } |
| |
| out->mmap_time_offset_nanos = get_mmap_out_time_offset(); |
| |
| out->standby = false; |
| ret = 0; |
| |
| ALOGD("%s: got mmap buffer address %p info->buffer_size_frames %d", |
| __func__, info->shared_memory_address, info->buffer_size_frames); |
| |
| exit: |
| if (ret != 0) { |
| if (out->pcm == NULL) { |
| ALOGE("%s: %s - %d", __func__, step, ret); |
| } else { |
| ALOGE("%s: %s %s", __func__, step, pcm_get_error(out->pcm)); |
| pcm_close(out->pcm); |
| out->pcm = NULL; |
| } |
| } |
| pthread_mutex_unlock(&adev->lock); |
| pthread_mutex_unlock(&out->lock); |
| return ret; |
| } |
| |
| static int out_get_mmap_position(const struct audio_stream_out *stream, |
| struct audio_mmap_position *position) |
| { |
| struct stream_out *out = (struct stream_out *)stream; |
| ALOGVV("%s", __func__); |
| if (position == NULL) { |
| return -EINVAL; |
| } |
| if (out->usecase != USECASE_AUDIO_PLAYBACK_MMAP) { |
| ALOGE("%s: called on %s", __func__, use_case_table[out->usecase]); |
| return -ENOSYS; |
| } |
| if (out->pcm == NULL) { |
| return -ENOSYS; |
| } |
| |
| struct timespec ts = { 0, 0 }; |
| int ret = pcm_mmap_get_hw_ptr(out->pcm, (unsigned int *)&position->position_frames, &ts); |
| if (ret < 0) { |
| ALOGE("%s: %s", __func__, pcm_get_error(out->pcm)); |
| return ret; |
| } |
| position->time_nanoseconds = ts.tv_sec*1000000000LL + ts.tv_nsec |
| + out->mmap_time_offset_nanos; |
| return 0; |
| } |
| |
| |
| /** audio_stream_in implementation **/ |
| static uint32_t in_get_sample_rate(const struct audio_stream *stream) |
| { |
| struct stream_in *in = (struct stream_in *)stream; |
| |
| return in->config.rate; |
| } |
| |
| static int in_set_sample_rate(struct audio_stream *stream __unused, |
| uint32_t rate __unused) |
| { |
| return -ENOSYS; |
| } |
| |
| static size_t in_get_buffer_size(const struct audio_stream *stream) |
| { |
| struct stream_in *in = (struct stream_in *)stream; |
| |
| if(in->usecase == USECASE_COMPRESS_VOIP_CALL) |
| return voice_extn_compress_voip_in_get_buffer_size(in); |
| else if(audio_extn_compr_cap_usecase_supported(in->usecase)) |
| return audio_extn_compr_cap_get_buffer_size(in->config.format); |
| else if(audio_extn_cin_attached_usecase(in)) |
| return audio_extn_cin_get_buffer_size(in); |
| |
| return in->config.period_size * in->af_period_multiplier * |
| audio_stream_in_frame_size((const struct audio_stream_in *)stream); |
| } |
| |
| static uint32_t in_get_channels(const struct audio_stream *stream) |
| { |
| struct stream_in *in = (struct stream_in *)stream; |
| |
| return in->channel_mask; |
| } |
| |
| static audio_format_t in_get_format(const struct audio_stream *stream) |
| { |
| struct stream_in *in = (struct stream_in *)stream; |
| |
| return in->format; |
| } |
| |
| static int in_set_format(struct audio_stream *stream __unused, |
| audio_format_t format __unused) |
| { |
| return -ENOSYS; |
| } |
| |
| static int in_standby(struct audio_stream *stream) |
| { |
| struct stream_in *in = (struct stream_in *)stream; |
| struct audio_device *adev = in->dev; |
| int status = 0; |
| ALOGD("%s: enter: stream (%p) usecase(%d: %s)", __func__, |
| stream, in->usecase, use_case_table[in->usecase]); |
| bool do_stop = true; |
| |
| lock_input_stream(in); |
| if (!in->standby && in->is_st_session) { |
| ALOGD("%s: sound trigger pcm stop lab", __func__); |
| audio_extn_sound_trigger_stop_lab(in); |
| if (adev->num_va_sessions > 0) |
| adev->num_va_sessions--; |
| in->standby = 1; |
| } |
| |
| if (!in->standby) { |
| if (adev->adm_deregister_stream) |
| adev->adm_deregister_stream(adev->adm_data, in->capture_handle); |
| |
| pthread_mutex_lock(&adev->lock); |
| in->standby = true; |
| if (in->usecase == USECASE_COMPRESS_VOIP_CALL) { |
| do_stop = false; |
| voice_extn_compress_voip_close_input_stream(stream); |
| ALOGD("VOIP input entered standby"); |
| } else if (in->usecase == USECASE_AUDIO_RECORD_MMAP) { |
| do_stop = in->capture_started; |
| in->capture_started = false; |
| if (in->mmap_shared_memory_fd >= 0) { |
| ALOGV("%s: closing mmap_shared_memory_fd = %d", |
| __func__, in->mmap_shared_memory_fd); |
| close(in->mmap_shared_memory_fd); |
| in->mmap_shared_memory_fd = -1; |
| } |
| } else { |
| if (audio_extn_cin_attached_usecase(in)) |
| audio_extn_cin_close_input_stream(in); |
| } |
| |
| if (in->pcm) { |
| ATRACE_BEGIN("pcm_in_close"); |
| pcm_close(in->pcm); |
| ATRACE_END(); |
| in->pcm = NULL; |
| } |
| |
| if (do_stop) |
| status = stop_input_stream(in); |
| |
| if (in->source == AUDIO_SOURCE_VOICE_RECOGNITION) { |
| if (adev->num_va_sessions > 0) |
| adev->num_va_sessions--; |
| } |
| |
| pthread_mutex_unlock(&adev->lock); |
| } |
| pthread_mutex_unlock(&in->lock); |
| ALOGV("%s: exit: status(%d)", __func__, status); |
| return status; |
| } |
| |
| static int in_dump(const struct audio_stream *stream, |
| int fd) |
| { |
| struct stream_in *in = (struct stream_in *)stream; |
| |
| // We try to get the lock for consistency, |
| // but it isn't necessary for these variables. |
| // If we're not in standby, we may be blocked on a read. |
| const bool locked = (pthread_mutex_trylock(&in->lock) == 0); |
| dprintf(fd, " Standby: %s\n", in->standby ? "yes" : "no"); |
| dprintf(fd, " Frames read: %lld\n", (long long)in->frames_read); |
| dprintf(fd, " Frames muted: %lld\n", (long long)in->frames_muted); |
| #ifndef LINUX_ENABLED |
| char buffer[256]; // for statistics formatting |
| if (in->start_latency_ms.n > 0) { |
| simple_stats_to_string(&in->start_latency_ms, buffer, sizeof(buffer)); |
| dprintf(fd, " Start latency ms: %s\n", buffer); |
| } |
| #endif |
| if (locked) { |
| pthread_mutex_unlock(&in->lock); |
| } |
| #ifndef LINUX_ENABLED |
| // dump error info |
| (void)error_log_dump( |
| in->error_log, fd, " " /* prefix */, 0 /* lines */, 0 /* limit_ns */); |
| #endif |
| return 0; |
| } |
| |
| static void in_snd_mon_cb(void * stream, struct str_parms * parms) |
| { |
| if (!stream || !parms) |
| return; |
| |
| struct stream_in *in = (struct stream_in *)stream; |
| struct audio_device *adev = in->dev; |
| |
| card_status_t status; |
| int card; |
| if (parse_snd_card_status(parms, &card, &status) < 0) |
| return; |
| |
| pthread_mutex_lock(&adev->lock); |
| bool valid_cb = (card == adev->snd_card); |
| pthread_mutex_unlock(&adev->lock); |
| |
| if (!valid_cb) |
| return; |
| |
| lock_input_stream(in); |
| if (in->card_status != status) |
| in->card_status = status; |
| pthread_mutex_unlock(&in->lock); |
| |
| ALOGW("in_snd_mon_cb for card %d usecase %s, status %s", card, |
| use_case_table[in->usecase], |
| status == CARD_STATUS_OFFLINE ? "offline" : "online"); |
| |
| // a better solution would be to report error back to AF and let |
| // it put the stream to standby |
| if (status == CARD_STATUS_OFFLINE) |
| in_standby(&in->stream.common); |
| |
| return; |
| } |
| |
| int route_input_stream(struct stream_in *in, |
| struct listnode *devices, |
| audio_source_t source) |
| { |
| struct audio_device *adev = in->dev; |
| int ret = 0; |
| |
| lock_input_stream(in); |
| pthread_mutex_lock(&adev->lock); |
| |
| /* no audio source uses val == 0 */ |
| if ((in->source != source) && (source != AUDIO_SOURCE_DEFAULT)) { |
| in->source = source; |
| if ((in->source == AUDIO_SOURCE_VOICE_COMMUNICATION) && |
| (in->dev->mode == AUDIO_MODE_IN_COMMUNICATION) && |
| (voice_extn_compress_voip_is_format_supported(in->format)) && |
| (in->config.rate == 8000 || in->config.rate == 16000 || |
| in->config.rate == 32000 || in->config.rate == 48000 ) && |
| (audio_channel_count_from_in_mask(in->channel_mask) == 1)) { |
| ret = voice_extn_compress_voip_open_input_stream(in); |
| if (ret != 0) { |
| ALOGE("%s: Compress voip input cannot be opened, error:%d", |
| __func__, ret); |
| } |
| } |
| } |
| |
| if (!compare_devices(&in->device_list, devices) && !list_empty(devices) && |
| is_audio_in_device_type(devices)) { |
| // Workaround: If routing to an non existing usb device, fail gracefully |
| // The routing request will otherwise block during 10 second |
| struct str_parms *usb_addr = |
| str_parms_create_str(get_usb_device_address(devices)); |
| if (is_usb_in_device_type(devices) && usb_addr && |
| !audio_extn_usb_connected(NULL)) { |
| ALOGW("%s: ignoring rerouting to non existing USB", __func__); |
| ret = -ENOSYS; |
| } else { |
| /* If recording is in progress, change the tx device to new device */ |
| assign_devices(&in->device_list, devices); |
| if (!in->standby && !in->is_st_session) { |
| ALOGV("update input routing change"); |
| // inform adm before actual routing to prevent glitches. |
| if (adev->adm_on_routing_change) { |
| adev->adm_on_routing_change(adev->adm_data, |
| in->capture_handle); |
| ret = select_devices(adev, in->usecase); |
| if (in->usecase == USECASE_AUDIO_RECORD_LOW_LATENCY) |
| adev->adm_routing_changed = true; |
| } |
| } |
| } |
| if (usb_addr) |
| str_parms_destroy(usb_addr); |
| } |
| pthread_mutex_unlock(&adev->lock); |
| pthread_mutex_unlock(&in->lock); |
| |
| ALOGV("%s: exit: status(%d)", __func__, ret); |
| return ret; |
| } |
| |
| static int in_set_parameters(struct audio_stream *stream, const char *kvpairs) |
| { |
| struct stream_in *in = (struct stream_in *)stream; |
| struct audio_device *adev = in->dev; |
| struct str_parms *parms; |
| char value[32]; |
| int err = 0; |
| |
| ALOGD("%s: enter: kvpairs=%s", __func__, kvpairs); |
| parms = str_parms_create_str(kvpairs); |
| |
| if (!parms) |
| goto error; |
| lock_input_stream(in); |
| pthread_mutex_lock(&adev->lock); |
| |
| err = str_parms_get_str(parms, AUDIO_PARAMETER_STREAM_PROFILE, value, sizeof(value)); |
| if (err >= 0) { |
| strlcpy(in->profile, value, sizeof(in->profile)); |
| ALOGV("updating stream profile with value '%s'", in->profile); |
| audio_extn_utils_update_stream_input_app_type_cfg(adev->platform, |
| &adev->streams_input_cfg_list, |
| &in->device_list, in->flags, in->format, |
| in->sample_rate, in->bit_width, |
| in->profile, &in->app_type_cfg); |
| } |
| |
| pthread_mutex_unlock(&adev->lock); |
| pthread_mutex_unlock(&in->lock); |
| |
| str_parms_destroy(parms); |
| error: |
| return 0; |
| } |
| |
| static char* in_get_parameters(const struct audio_stream *stream, |
| const char *keys) |
| { |
| struct stream_in *in = (struct stream_in *)stream; |
| struct str_parms *query = str_parms_create_str(keys); |
| char *str; |
| struct str_parms *reply = str_parms_create(); |
| |
| if (!query || !reply) { |
| if (reply) { |
| str_parms_destroy(reply); |
| } |
| if (query) { |
| str_parms_destroy(query); |
| } |
| ALOGE("in_get_parameters: failed to create query or reply"); |
| return NULL; |
| } |
| |
| ALOGV("%s: enter: keys - %s %s ", __func__, use_case_table[in->usecase], keys); |
| |
| voice_extn_in_get_parameters(in, query, reply); |
| |
| stream_get_parameter_channels(query, reply, |
| &in->supported_channel_masks[0]); |
| stream_get_parameter_formats(query, reply, |
| &in->supported_formats[0]); |
| stream_get_parameter_rates(query, reply, |
| &in->supported_sample_rates[0]); |
| str = str_parms_to_str(reply); |
| str_parms_destroy(query); |
| str_parms_destroy(reply); |
| |
| ALOGV("%s: exit: returns - %s", __func__, str); |
| return str; |
| } |
| |
| static int in_set_gain(struct audio_stream_in *stream, |
| float gain) |
| { |
| struct stream_in *in = (struct stream_in *)stream; |
| char mixer_ctl_name[128]; |
| struct mixer_ctl *ctl; |
| int ctl_value; |
| |
| ALOGV("%s: gain %f", __func__, gain); |
| |
| if (stream == NULL) |
| return -EINVAL; |
| |
| /* in_set_gain() only used to silence MMAP capture for now */ |
| if (in->usecase != USECASE_AUDIO_RECORD_MMAP) |
| return -ENOSYS; |
| |
| snprintf(mixer_ctl_name, sizeof(mixer_ctl_name), "Capture %d Volume", in->pcm_device_id); |
| |
| ctl = mixer_get_ctl_by_name(in->dev->mixer, mixer_ctl_name); |
| if (!ctl) { |
| ALOGW("%s: Could not get ctl for mixer cmd - %s", |
| __func__, mixer_ctl_name); |
| return -ENOSYS; |
| } |
| |
| if (gain < RECORD_GAIN_MIN) |
| gain = RECORD_GAIN_MIN; |
| else if (gain > RECORD_GAIN_MAX) |
| gain = RECORD_GAIN_MAX; |
| ctl_value = (int)(RECORD_VOLUME_CTL_MAX * gain); |
| |
| mixer_ctl_set_value(ctl, 0, ctl_value); |
| |
| return 0; |
| } |
| |
| static ssize_t in_read(struct audio_stream_in *stream, void *buffer, |
| size_t bytes) |
| { |
| struct stream_in *in = (struct stream_in *)stream; |
| |
| if (in == NULL) { |
| ALOGE("%s: stream_in ptr is NULL", __func__); |
| return -EINVAL; |
| } |
| |
| struct audio_device *adev = in->dev; |
| int ret = -1; |
| size_t bytes_read = 0, frame_size = 0; |
| |
| lock_input_stream(in); |
| |
| if (in->is_st_session) { |
| ALOGVV(" %s: reading on st session bytes=%zu", __func__, bytes); |
| /* Read from sound trigger HAL */ |
| audio_extn_sound_trigger_read(in, buffer, bytes); |
| if (in->standby) { |
| if (adev->num_va_sessions < UINT_MAX) |
| adev->num_va_sessions++; |
| in->standby = 0; |
| } |
| pthread_mutex_unlock(&in->lock); |
| return bytes; |
| } |
| |
| if (in->usecase == USECASE_AUDIO_RECORD_MMAP) { |
| ret = -ENOSYS; |
| goto exit; |
| } |
| |
| if (in->usecase == USECASE_AUDIO_RECORD_LOW_LATENCY && |
| !in->standby && adev->adm_routing_changed) { |
| ret = -ENOSYS; |
| goto exit; |
| } |
| |
| if (in->standby) { |
| const int64_t startNs = systemTime(SYSTEM_TIME_MONOTONIC); |
| |
| pthread_mutex_lock(&adev->lock); |
| if (in->usecase == USECASE_COMPRESS_VOIP_CALL) |
| ret = voice_extn_compress_voip_start_input_stream(in); |
| else |
| ret = start_input_stream(in); |
| if (!ret && in->source == AUDIO_SOURCE_VOICE_RECOGNITION) { |
| if (adev->num_va_sessions < UINT_MAX) |
| adev->num_va_sessions++; |
| } |
| pthread_mutex_unlock(&adev->lock); |
| if (ret != 0) { |
| goto exit; |
| } |
| in->standby = 0; |
| #ifndef LINUX_ENABLED |
| // log startup time in ms. |
| simple_stats_log( |
| &in->start_latency_ms, (systemTime(SYSTEM_TIME_MONOTONIC) - startNs) * 1e-6); |
| #endif |
| } |
| |
| /* Avoid read if capture_stopped is set */ |
| if (android_atomic_acquire_load(&(in->capture_stopped)) > 0) { |
| ALOGD("%s: force stopped catpure session, ignoring read request", __func__); |
| ret = -EINVAL; |
| goto exit; |
| } |
| |
| // what's the duration requested by the client? |
| long ns = 0; |
| |
| if (in->pcm && in->config.rate) |
| ns = pcm_bytes_to_frames(in->pcm, bytes)*1000000000LL/ |
| in->config.rate; |
| |
| ret = request_in_focus(in, ns); |
| if (ret != 0) |
| goto exit; |
| bool use_mmap = is_mmap_usecase(in->usecase) || in->realtime; |
| |
| if (audio_extn_cin_attached_usecase(in)) { |
| ret = audio_extn_cin_read(in, buffer, bytes, &bytes_read); |
| } else if (in->pcm) { |
| if (audio_extn_ssr_get_stream() == in) { |
| ret = audio_extn_ssr_read(stream, buffer, bytes); |
| } else if (audio_extn_compr_cap_usecase_supported(in->usecase)) { |
| ret = audio_extn_compr_cap_read(in, buffer, bytes); |
| } else if (use_mmap) { |
| ret = pcm_mmap_read(in->pcm, buffer, bytes); |
| } else if (audio_extn_ffv_get_stream() == in) { |
| ret = audio_extn_ffv_read(stream, buffer, bytes); |
| } else { |
| ret = pcm_read(in->pcm, buffer, bytes); |
| /* data from DSP comes in 24_8 format, convert it to 8_24 */ |
| if (!ret && bytes > 0 && (in->format == AUDIO_FORMAT_PCM_8_24_BIT)) { |
| if (audio_extn_utils_convert_format_24_8_to_8_24(buffer, bytes) |
| != bytes) { |
| ret = -EINVAL; |
| goto exit; |
| } |
| } else if (ret < 0) { |
| ret = -errno; |
| } |
| } |
| /* bytes read is always set to bytes for non compress usecases */ |
| bytes_read = bytes; |
| } |
| |
| release_in_focus(in); |
| |
| /* |
| * Instead of writing zeroes here, we could trust the hardware to always |
| * provide zeroes when muted. This is also muted with voice recognition |
| * usecases so that other clients do not have access to voice recognition |
| * data. |
| */ |
| if ((ret == 0 && voice_get_mic_mute(adev) && |
| !voice_is_in_call_rec_stream(in) && |
| (in->usecase != USECASE_AUDIO_RECORD_AFE_PROXY && |
| in->usecase != USECASE_AUDIO_RECORD_AFE_PROXY2)) || |
| (adev->num_va_sessions && |
| in->source != AUDIO_SOURCE_VOICE_RECOGNITION && |
| property_get_bool("persist.vendor.audio.va_concurrency_mute_enabled", |
| false))) |
| memset(buffer, 0, bytes); |
| |
| exit: |
| frame_size = audio_stream_in_frame_size(stream); |
| if (frame_size > 0) |
| in->frames_read += bytes_read/frame_size; |
| |
| if (-ENETRESET == ret) |
| in->card_status = CARD_STATUS_OFFLINE; |
| pthread_mutex_unlock(&in->lock); |
| |
| if (ret != 0) { |
| if (in->usecase == USECASE_COMPRESS_VOIP_CALL) { |
| pthread_mutex_lock(&adev->lock); |
| voice_extn_compress_voip_close_input_stream(&in->stream.common); |
| pthread_mutex_unlock(&adev->lock); |
| in->standby = true; |
| } |
| if (!audio_extn_cin_attached_usecase(in)) { |
| bytes_read = bytes; |
| memset(buffer, 0, bytes); |
| } |
| in_standby(&in->stream.common); |
| if (in->usecase == USECASE_AUDIO_RECORD_LOW_LATENCY) |
| adev->adm_routing_changed = false; |
| ALOGV("%s: read failed status %d- sleeping for buffer duration", __func__, ret); |
| usleep((uint64_t)bytes * 1000000 / audio_stream_in_frame_size(stream) / |
| in_get_sample_rate(&in->stream.common)); |
| } |
| return bytes_read; |
| } |
| |
| static uint32_t in_get_input_frames_lost(struct audio_stream_in *stream __unused) |
| { |
| return 0; |
| } |
| |
| static int in_get_capture_position(const struct audio_stream_in *stream, |
| int64_t *frames, int64_t *time) |
| { |
| if (stream == NULL || frames == NULL || time == NULL) { |
| return -EINVAL; |
| } |
| struct stream_in *in = (struct stream_in *)stream; |
| int ret = -ENOSYS; |
| |
| lock_input_stream(in); |
| // note: ST sessions do not close the alsa pcm driver synchronously |
| // on standby. Therefore, we may return an error even though the |
| // pcm stream is still opened. |
| if (in->standby) { |
| ALOGE_IF(in->pcm != NULL && !in->is_st_session, |
| "%s stream in standby but pcm not NULL for non ST session", __func__); |
| goto exit; |
| } |
| if (in->pcm) { |
| struct timespec timestamp; |
| unsigned int avail; |
| if (pcm_get_htimestamp(in->pcm, &avail, ×tamp) == 0) { |
| *frames = in->frames_read + avail; |
| *time = timestamp.tv_sec * 1000000000LL + timestamp.tv_nsec |
| - platform_capture_latency(in) * 1000LL; |
| //Adjustment accounts for A2dp decoder latency for recording usecase |
| // Note: decoder latency is returned in ms, while platform_capture_latency in ns. |
| if (is_a2dp_in_device_type(&in->device_list)) |
| *time -= audio_extn_a2dp_get_decoder_latency() * 1000000LL; |
| ret = 0; |
| } |
| } |
| exit: |
| pthread_mutex_unlock(&in->lock); |
| return ret; |
| } |
| |
| static int in_update_effect_list(bool add, effect_handle_t effect, |
| struct listnode *head) |
| { |
| struct listnode *node; |
| struct in_effect_list *elist = NULL; |
| struct in_effect_list *target = NULL; |
| int ret = 0; |
| |
| if (!head) |
| return ret; |
| |
| list_for_each(node, head) { |
| elist = node_to_item(node, struct in_effect_list, list); |
| if (elist->handle == effect) { |
| target = elist; |
| break; |
| } |
| } |
| |
| if (add) { |
| if (target) { |
| ALOGD("effect %p already exist", effect); |
| return ret; |
| } |
| |
| target = (struct in_effect_list *) |
| calloc(1, sizeof(struct in_effect_list)); |
| |
| if (!target) { |
| ALOGE("%s:fail to allocate memory", __func__); |
| return -ENOMEM; |
| } |
| |
| target->handle = effect; |
| list_add_tail(head, &target->list); |
| } else { |
| if (target) { |
| list_remove(&target->list); |
| free(target); |
| } |
| } |
| |
| return ret; |
| } |
| |
| static int add_remove_audio_effect(const struct audio_stream *stream, |
| effect_handle_t effect, |
| bool enable) |
| { |
| struct stream_in *in = (struct stream_in *)stream; |
| int status = 0; |
| effect_descriptor_t desc; |
| |
| status = (*effect)->get_descriptor(effect, &desc); |
| ALOGV("%s: status %d in->standby %d enable:%d", __func__, status, in->standby, enable); |
| |
| if (status != 0) |
| return status; |
| |
| lock_input_stream(in); |
| pthread_mutex_lock(&in->dev->lock); |
| if ((in->source == AUDIO_SOURCE_VOICE_COMMUNICATION || |
| in->source == AUDIO_SOURCE_VOICE_RECOGNITION || |
| adev->mode == AUDIO_MODE_IN_COMMUNICATION) && |
| (memcmp(&desc.type, FX_IID_AEC, sizeof(effect_uuid_t)) == 0)) { |
| |
| in_update_effect_list(enable, effect, &in->aec_list); |
| enable = !list_empty(&in->aec_list); |
| if (enable == in->enable_aec) |
| goto exit; |
| |
| in->enable_aec = enable; |
| ALOGD("AEC enable %d", enable); |
| |
| if (in->source == AUDIO_SOURCE_VOICE_COMMUNICATION || |
| in->dev->mode == AUDIO_MODE_IN_COMMUNICATION) { |
| in->dev->enable_voicerx = enable; |
| struct audio_usecase *usecase; |
| struct listnode *node; |
| list_for_each(node, &in->dev->usecase_list) { |
| usecase = node_to_item(node, struct audio_usecase, list); |
| if (usecase->type == PCM_PLAYBACK) |
| select_devices(in->dev, usecase->id); |
| } |
| } |
| if (!in->standby) { |
| if (enable_disable_effect(in->dev, EFFECT_AEC, enable) == ENOSYS) |
| select_devices(in->dev, in->usecase); |
| } |
| |
| } |
| if (memcmp(&desc.type, FX_IID_NS, sizeof(effect_uuid_t)) == 0) { |
| |
| in_update_effect_list(enable, effect, &in->ns_list); |
| enable = !list_empty(&in->ns_list); |
| if (enable == in->enable_ns) |
| goto exit; |
| |
| in->enable_ns = enable; |
| ALOGD("NS enable %d", enable); |
| if (!in->standby) { |
| if (in->source == AUDIO_SOURCE_VOICE_COMMUNICATION || |
| in->dev->mode == AUDIO_MODE_IN_COMMUNICATION) { |
| if (enable_disable_effect(in->dev, EFFECT_NS, enable) == ENOSYS) |
| select_devices(in->dev, in->usecase); |
| } else |
| select_devices(in->dev, in->usecase); |
| } |
| } |
| exit: |
| pthread_mutex_unlock(&in->dev->lock); |
| pthread_mutex_unlock(&in->lock); |
| |
| return 0; |
| } |
| |
| static int in_add_audio_effect(const struct audio_stream *stream, |
| effect_handle_t effect) |
| { |
| ALOGV("%s: effect %p", __func__, effect); |
| return add_remove_audio_effect(stream, effect, true); |
| } |
| |
| static int in_remove_audio_effect(const struct audio_stream *stream, |
| effect_handle_t effect) |
| { |
| ALOGV("%s: effect %p", __func__, effect); |
| return add_remove_audio_effect(stream, effect, false); |
| } |
| |
| static int in_stop(const struct audio_stream_in* stream) |
| { |
| struct stream_in *in = (struct stream_in *)stream; |
| struct audio_device *adev = in->dev; |
| |
| int ret = -ENOSYS; |
| ALOGV("%s", __func__); |
| pthread_mutex_lock(&adev->lock); |
| if (in->usecase == USECASE_AUDIO_RECORD_MMAP && !in->standby && |
| in->capture_started && in->pcm != NULL) { |
| pcm_stop(in->pcm); |
| ret = stop_input_stream(in); |
| in->capture_started = false; |
| } |
| pthread_mutex_unlock(&adev->lock); |
| return ret; |
| } |
| |
| static int in_start(const struct audio_stream_in* stream) |
| { |
| struct stream_in *in = (struct stream_in *)stream; |
| struct audio_device *adev = in->dev; |
| int ret = -ENOSYS; |
| |
| ALOGV("%s in %p", __func__, in); |
| pthread_mutex_lock(&adev->lock); |
| if (in->usecase == USECASE_AUDIO_RECORD_MMAP && !in->standby && |
| !in->capture_started && in->pcm != NULL) { |
| if (!in->capture_started) { |
| ret = start_input_stream(in); |
| if (ret == 0) { |
| in->capture_started = true; |
| } |
| } |
| } |
| pthread_mutex_unlock(&adev->lock); |
| return ret; |
| } |
| |
| // Read offset for the positional timestamp from a persistent vendor property. |
| // This is to workaround apparent inaccuracies in the timing information that |
| // is used by the AAudio timing model. The inaccuracies can cause glitches. |
| static int64_t in_get_mmap_time_offset() { |
| const int32_t kDefaultOffsetMicros = 0; |
| int32_t mmap_time_offset_micros = property_get_int32( |
| "persist.vendor.audio.in_mmap_delay_micros", kDefaultOffsetMicros); |
| ALOGI("mmap_time_offset_micros = %d for input", mmap_time_offset_micros); |
| return mmap_time_offset_micros * (int64_t)1000; |
| } |
| |
| static int in_create_mmap_buffer(const struct audio_stream_in *stream, |
| int32_t min_size_frames, |
| struct audio_mmap_buffer_info *info) |
| { |
| struct stream_in *in = (struct stream_in *)stream; |
| struct audio_device *adev = in->dev; |
| int ret = 0; |
| unsigned int offset1 = 0; |
| unsigned int frames1 = 0; |
| const char *step = ""; |
| uint32_t mmap_size = 0; |
| uint32_t buffer_size = 0; |
| |
| pthread_mutex_lock(&adev->lock); |
| ALOGV("%s in %p", __func__, in); |
| |
| if (CARD_STATUS_OFFLINE == in->card_status|| |
| CARD_STATUS_OFFLINE == adev->card_status) { |
| ALOGW("in->card_status or adev->card_status offline, try again"); |
| ret = -EIO; |
| goto exit; |
| } |
| |
| if (info == NULL || !(min_size_frames > 0 && min_size_frames < INT32_MAX)) { |
| ALOGE("%s invalid argument info %p min_size_frames %d", __func__, info, min_size_frames); |
| ret = -EINVAL; |
| goto exit; |
| } |
| if (in->usecase != USECASE_AUDIO_RECORD_MMAP || !in->standby) { |
| ALOGE("%s: usecase = %d, standby = %d", __func__, in->usecase, in->standby); |
| ALOGV("%s in %p", __func__, in); |
| ret = -ENOSYS; |
| goto exit; |
| } |
| in->pcm_device_id = platform_get_pcm_device_id(in->usecase, PCM_CAPTURE); |
| if (in->pcm_device_id < 0) { |
| ALOGE("%s: Invalid PCM device id(%d) for the usecase(%d)", |
| __func__, in->pcm_device_id, in->usecase); |
| ret = -EINVAL; |
| goto exit; |
| } |
| |
| adjust_mmap_period_count(&in->config, min_size_frames); |
| |
| ALOGV("%s: Opening PCM device card_id(%d) device_id(%d), channels %d", |
| __func__, adev->snd_card, in->pcm_device_id, in->config.channels); |
| in->pcm = pcm_open(adev->snd_card, in->pcm_device_id, |
| (PCM_IN | PCM_MMAP | PCM_NOIRQ | PCM_MONOTONIC), &in->config); |
| if (errno == ENETRESET && !pcm_is_ready(in->pcm)) { |
| ALOGE("%s: pcm_open failed errno:%d\n", __func__, errno); |
| in->card_status = CARD_STATUS_OFFLINE; |
| adev->card_status = CARD_STATUS_OFFLINE; |
| ret = -EIO; |
| goto exit; |
| } |
| |
| if (in->pcm == NULL || !pcm_is_ready(in->pcm)) { |
| step = "open"; |
| ret = -ENODEV; |
| goto exit; |
| } |
| |
| ret = pcm_mmap_begin(in->pcm, &info->shared_memory_address, &offset1, &frames1); |
| if (ret < 0) { |
| step = "begin"; |
| goto exit; |
| } |
| |
| info->flags = 0; |
| info->buffer_size_frames = pcm_get_buffer_size(in->pcm); |
| buffer_size = pcm_frames_to_bytes(in->pcm, info->buffer_size_frames); |
| info->burst_size_frames = in->config.period_size; |
| ret = platform_get_mmap_data_fd(adev->platform, |
| in->pcm_device_id, 1 /*capture*/, |
| &info->shared_memory_fd, |
| &mmap_size); |
| if (ret < 0) { |
| // Fall back to non exclusive mode |
| info->shared_memory_fd = pcm_get_poll_fd(in->pcm); |
| } else { |
| in->mmap_shared_memory_fd = info->shared_memory_fd; // for closing later |
| ALOGV("%s: opened mmap_shared_memory_fd = %d", __func__, in->mmap_shared_memory_fd); |
| |
| if (mmap_size < buffer_size) { |
| step = "mmap"; |
| goto exit; |
| } |
| info->flags |= AUDIO_MMAP_APPLICATION_SHAREABLE; |
| } |
| |
| memset(info->shared_memory_address, 0, buffer_size); |
| |
| ret = pcm_mmap_commit(in->pcm, 0, MMAP_PERIOD_SIZE); |
| if (ret < 0) { |
| step = "commit"; |
| goto exit; |
| } |
| |
| in->mmap_time_offset_nanos = in_get_mmap_time_offset(); |
| |
| in->standby = false; |
| ret = 0; |
| |
| ALOGV("%s: got mmap buffer address %p info->buffer_size_frames %d", |
| __func__, info->shared_memory_address, info->buffer_size_frames); |
| |
| exit: |
| if (ret != 0) { |
| if (in->pcm == NULL) { |
| ALOGE("%s: %s - %d", __func__, step, ret); |
| } else { |
| ALOGE("%s: %s %s", __func__, step, pcm_get_error(in->pcm)); |
| pcm_close(in->pcm); |
| in->pcm = NULL; |
| } |
| } |
| pthread_mutex_unlock(&adev->lock); |
| return ret; |
| } |
| |
| static int in_get_mmap_position(const struct audio_stream_in *stream, |
| struct audio_mmap_position *position) |
| { |
| struct stream_in *in = (struct stream_in *)stream; |
| ALOGVV("%s", __func__); |
| if (position == NULL) { |
| return -EINVAL; |
| } |
| lock_input_stream(in); |
| if (in->usecase != USECASE_AUDIO_RECORD_MMAP) { |
| pthread_mutex_unlock(&in->lock); |
| return -ENOSYS; |
| } |
| if (in->pcm == NULL) { |
| pthread_mutex_unlock(&in->lock); |
| return -ENOSYS; |
| } |
| struct timespec ts = { 0, 0 }; |
| int ret = pcm_mmap_get_hw_ptr(in->pcm, (unsigned int *)&position->position_frames, &ts); |
| if (ret < 0) { |
| ALOGE("%s: %s", __func__, pcm_get_error(in->pcm)); |
| pthread_mutex_unlock(&in->lock); |
| return ret; |
| } |
| position->time_nanoseconds = ts.tv_sec*1000000000LL + ts.tv_nsec |
| + in->mmap_time_offset_nanos; |
| pthread_mutex_unlock(&in->lock); |
| return 0; |
| } |
| |
| static int in_get_active_microphones(const struct audio_stream_in *stream, |
| struct audio_microphone_characteristic_t *mic_array, |
| size_t *mic_count) { |
| struct stream_in *in = (struct stream_in *)stream; |
| struct audio_device *adev = in->dev; |
| ALOGVV("%s", __func__); |
| |
| lock_input_stream(in); |
| pthread_mutex_lock(&adev->lock); |
| int ret = platform_get_active_microphones(adev->platform, |
| audio_channel_count_from_in_mask(in->channel_mask), |
| in->usecase, mic_array, mic_count); |
| pthread_mutex_unlock(&adev->lock); |
| pthread_mutex_unlock(&in->lock); |
| |
| return ret; |
| } |
| |
| static int adev_get_microphones(const struct audio_hw_device *dev, |
| struct audio_microphone_characteristic_t *mic_array, |
| size_t *mic_count) { |
| struct audio_device *adev = (struct audio_device *)dev; |
| ALOGVV("%s", __func__); |
| |
| pthread_mutex_lock(&adev->lock); |
| int ret = platform_get_microphones(adev->platform, mic_array, mic_count); |
| pthread_mutex_unlock(&adev->lock); |
| |
| return ret; |
| } |
| |
| static void in_update_sink_metadata(struct audio_stream_in *stream, |
| const struct sink_metadata *sink_metadata) { |
| |
| if (stream == NULL |
| || sink_metadata == NULL |
| || sink_metadata->tracks == NULL) { |
| return; |
| } |
| |
| int error = 0; |
| struct stream_in *in = (struct stream_in *)stream; |
| struct audio_device *adev = in->dev; |
| struct listnode devices; |
| bool is_ha_usecase = false; |
| |
| list_init(&devices); |
| |
| if (sink_metadata->track_count != 0) |
| reassign_device_list(&devices, sink_metadata->tracks->dest_device, ""); |
| |
| lock_input_stream(in); |
| pthread_mutex_lock(&adev->lock); |
| ALOGV("%s: in->usecase: %d, device: %x", __func__, in->usecase, get_device_types(&devices)); |
| |
| is_ha_usecase = adev->ha_proxy_enable ? |
| in->usecase == USECASE_AUDIO_RECORD_AFE_PROXY2 : |
| in->usecase == USECASE_AUDIO_RECORD_AFE_PROXY; |
| if (is_ha_usecase && !list_empty(&devices) |
| && adev->voice_tx_output != NULL) { |
| /* Use the rx device from afe-proxy record to route voice call because |
| there is no routing if tx device is on primary hal and rx device |
| is on other hal during voice call. */ |
| assign_devices(&adev->voice_tx_output->device_list, &devices); |
| |
| if (!voice_is_call_state_active(adev)) { |
| if (adev->mode == AUDIO_MODE_IN_CALL) { |
| adev->current_call_output = adev->voice_tx_output; |
| error = voice_start_call(adev); |
| if (error != 0) |
| ALOGE("%s: start voice call failed %d", __func__, error); |
| } |
| } else { |
| adev->current_call_output = adev->voice_tx_output; |
| voice_update_devices_for_all_voice_usecases(adev); |
| } |
| } |
| |
| pthread_mutex_unlock(&adev->lock); |
| pthread_mutex_unlock(&in->lock); |
| } |
| |
| int adev_open_output_stream(struct audio_hw_device *dev, |
| audio_io_handle_t handle, |
| audio_devices_t devices, |
| audio_output_flags_t flags, |
| struct audio_config *config, |
| struct audio_stream_out **stream_out, |
| const char *address) |
| { |
| struct audio_device *adev = (struct audio_device *)dev; |
| struct stream_out *out; |
| int ret = 0, ip_hdlr_stream = 0, ip_hdlr_dev = 0; |
| audio_format_t format; |
| struct adsp_hdlr_stream_cfg hdlr_stream_cfg; |
| bool is_direct_passthough = false; |
| bool is_hdmi = devices & AUDIO_DEVICE_OUT_AUX_DIGITAL; |
| bool is_usb_dev = audio_is_usb_out_device(devices) && |
| (devices != AUDIO_DEVICE_OUT_USB_ACCESSORY); |
| bool direct_dev = is_hdmi || is_usb_dev; |
| bool use_db_as_primary = |
| property_get_bool("vendor.audio.feature.deepbuffer_as_primary.enable", |
| false); |
| bool force_haptic_path = |
| property_get_bool("vendor.audio.test_haptic", false); |
| bool is_voip_rx = flags & AUDIO_OUTPUT_FLAG_VOIP_RX; |
| #ifdef AUDIO_GKI_ENABLED |
| __s32 *generic_dec; |
| #endif |
| pthread_mutexattr_t latch_attr; |
| |
| if (is_usb_dev && (!audio_extn_usb_connected(NULL))) { |
| is_usb_dev = false; |
| devices = AUDIO_DEVICE_OUT_SPEAKER; |
| ALOGW("%s: ignore set device to non existing USB card, use output device(%#x)", |
| __func__, devices); |
| if (config->format == AUDIO_FORMAT_DEFAULT) |
| config->format = AUDIO_FORMAT_PCM_16_BIT; |
| if (config->sample_rate == 0) |
| config->sample_rate = DEFAULT_OUTPUT_SAMPLING_RATE; |
| if (config->channel_mask == AUDIO_CHANNEL_NONE) |
| config->channel_mask = AUDIO_CHANNEL_OUT_STEREO; |
| } |
| |
| *stream_out = NULL; |
| |
| out = (struct stream_out *)calloc(1, sizeof(struct stream_out)); |
| |
| ALOGD("%s: enter: format(%#x) sample_rate(%d) channel_mask(%#x) devices(%#x) flags(%#x)\ |
| stream_handle(%p) address(%s)", __func__, config->format, config->sample_rate, config->channel_mask, |
| devices, flags, &out->stream, address); |
| |
| |
| if (!out) { |
| return -ENOMEM; |
| } |
| |
| pthread_mutex_init(&out->lock, (const pthread_mutexattr_t *) NULL); |
| pthread_mutex_init(&out->pre_lock, (const pthread_mutexattr_t *) NULL); |
| pthread_mutexattr_init(&latch_attr); |
| pthread_mutexattr_settype(&latch_attr, PTHREAD_MUTEX_RECURSIVE); |
| pthread_mutex_init(&out->latch_lock, &latch_attr); |
| pthread_mutexattr_destroy(&latch_attr); |
| pthread_mutex_init(&out->position_query_lock, (const pthread_mutexattr_t *) NULL); |
| pthread_cond_init(&out->cond, (const pthread_condattr_t *) NULL); |
| |
| if (devices == AUDIO_DEVICE_NONE) |
| devices = AUDIO_DEVICE_OUT_SPEAKER; |
| |
| out->flags = flags; |
| list_init(&out->device_list); |
| update_device_list(&out->device_list, devices, address, true /* add devices */); |
| out->dev = adev; |
| out->hal_op_format = out->hal_ip_format = format = out->format = config->format; |
| out->sample_rate = config->sample_rate; |
| out->channel_mask = config->channel_mask; |
| if (out->channel_mask == AUDIO_CHANNEL_NONE) |
| out->supported_channel_masks[0] = AUDIO_CHANNEL_OUT_STEREO; |
| else |
| out->supported_channel_masks[0] = out->channel_mask; |
| out->handle = handle; |
| out->bit_width = CODEC_BACKEND_DEFAULT_BIT_WIDTH; |
| out->non_blocking = 0; |
| out->convert_buffer = NULL; |
| out->started = 0; |
| out->a2dp_muted = false; |
| out->hal_output_suspend_supported = 0; |
| out->dynamic_pm_qos_config_supported = 0; |
| out->set_dual_mono = false; |
| out->prev_card_status_offline = false; |
| out->pspd_coeff_sent = false; |
| out->mmap_shared_memory_fd = -1; // not open |
| |
| if ((flags & AUDIO_OUTPUT_FLAG_BD) && |
| (property_get_bool("vendor.audio.matrix.limiter.enable", false))) |
| platform_set_device_params(out, DEVICE_PARAM_LIMITER_ID, 1); |
| |
| if (direct_dev && |
| (audio_is_linear_pcm(out->format) || |
| config->format == AUDIO_FORMAT_DEFAULT) && |
| out->flags == AUDIO_OUTPUT_FLAG_NONE) { |
| audio_format_t req_format = config->format; |
| audio_channel_mask_t req_channel_mask = config->channel_mask; |
| uint32_t req_sample_rate = config->sample_rate; |
| |
| pthread_mutex_lock(&adev->lock); |
| if (is_hdmi) { |
| ALOGV("AUDIO_DEVICE_OUT_AUX_DIGITAL and DIRECT|OFFLOAD, check hdmi caps"); |
| ret = read_hdmi_sink_caps(out); |
| if (config->sample_rate == 0) |
| config->sample_rate = DEFAULT_OUTPUT_SAMPLING_RATE; |
| if (config->channel_mask == AUDIO_CHANNEL_NONE) |
| config->channel_mask = AUDIO_CHANNEL_OUT_5POINT1; |
| if (config->format == AUDIO_FORMAT_DEFAULT) |
| config->format = AUDIO_FORMAT_PCM_16_BIT; |
| } else if (is_usb_dev) { |
| ret = read_usb_sup_params_and_compare(true /*is_playback*/, |
| &config->format, |
| &out->supported_formats[0], |
| MAX_SUPPORTED_FORMATS, |
| &config->channel_mask, |
| &out->supported_channel_masks[0], |
| MAX_SUPPORTED_CHANNEL_MASKS, |
| &config->sample_rate, |
| &out->supported_sample_rates[0], |
| MAX_SUPPORTED_SAMPLE_RATES); |
| ALOGV("plugged dev USB ret %d", ret); |
| } |
| |
| pthread_mutex_unlock(&adev->lock); |
| if (ret != 0) { |
| if (ret == -ENOSYS) { |
| /* ignore and go with default */ |
| ret = 0; |
| } |
| // For MMAP NO IRQ, allow conversions in ADSP |
| else if (is_hdmi || (flags & AUDIO_OUTPUT_FLAG_MMAP_NOIRQ) == 0) |
| goto error_open; |
| else { |
| ALOGE("error reading direct dev sink caps"); |
| goto error_open; |
| } |
| |
| if (req_sample_rate != 0 && config->sample_rate != req_sample_rate) |
| config->sample_rate = req_sample_rate; |
| if (req_channel_mask != AUDIO_CHANNEL_NONE && config->channel_mask != req_channel_mask) |
| config->channel_mask = req_channel_mask; |
| if (req_format != AUDIO_FORMAT_DEFAULT && config->format != req_format) |
| config->format = req_format; |
| } |
| |
| out->sample_rate = config->sample_rate; |
| out->channel_mask = config->channel_mask; |
| out->format = config->format; |
| if (is_hdmi) { |
| out->usecase = USECASE_AUDIO_PLAYBACK_HIFI; |
| out->config = pcm_config_hdmi_multi; |
| } else if (flags & AUDIO_OUTPUT_FLAG_MMAP_NOIRQ) { |
| out->usecase = USECASE_AUDIO_PLAYBACK_MMAP; |
| out->config = pcm_config_mmap_playback; |
| out->stream.start = out_start; |
| out->stream.stop = out_stop; |
| out->stream.create_mmap_buffer = out_create_mmap_buffer; |
| out->stream.get_mmap_position = out_get_mmap_position; |
| } else { |
| out->usecase = USECASE_AUDIO_PLAYBACK_HIFI; |
| out->config = pcm_config_hifi; |
| } |
| |
| out->config.rate = out->sample_rate; |
| out->config.channels = audio_channel_count_from_out_mask(out->channel_mask); |
| if (is_hdmi) { |
| out->config.period_size = HDMI_MULTI_PERIOD_BYTES / (out->config.channels * |
| audio_bytes_per_sample(out->format)); |
| } |
| out->config.format = pcm_format_from_audio_format(out->format); |
| } |
| |
| /* validate bus device address */ |
| if (compare_device_type(&out->device_list, AUDIO_DEVICE_OUT_BUS)) { |
| /* extract car audio stream index */ |
| out->car_audio_stream = |
| audio_extn_auto_hal_get_car_audio_stream_from_address(address); |
| if (out->car_audio_stream < 0) { |
| ALOGE("%s: invalid car audio stream %x", |
| __func__, out->car_audio_stream); |
| ret = -EINVAL; |
| goto error_open; |
| } |
| ALOGV("%s: car_audio_stream %x", __func__, out->car_audio_stream); |
| } |
| |
| /* Check for VOIP usecase */ |
| if (is_voip_rx) { |
| if (!voice_extn_is_compress_voip_supported()) { |
| if (out->sample_rate == 8000 || out->sample_rate == 16000 || |
| out->sample_rate == 32000 || out->sample_rate == 48000) { |
| out->channel_mask = config->channel_mask; |
| out->usecase = USECASE_AUDIO_PLAYBACK_VOIP; |
| out->format = AUDIO_FORMAT_PCM_16_BIT; |
| out->volume_l = INVALID_OUT_VOLUME; |
| out->volume_r = INVALID_OUT_VOLUME; |
| |
| out->config = default_pcm_config_voip_copp; |
| out->config.rate = out->sample_rate; |
| uint32_t channel_count = |
| audio_channel_count_from_out_mask(out->channel_mask); |
| out->config.channels = channel_count; |
| |
| uint32_t buffer_size = get_stream_buffer_size(DEFAULT_VOIP_BUF_DURATION_MS, |
| out->sample_rate, out->format, |
| channel_count, false); |
| uint32_t frame_size = audio_bytes_per_sample(out->format) * channel_count; |
| if (frame_size != 0) |
| out->config.period_size = buffer_size / frame_size; |
| else |
| ALOGW("%s: frame size is 0 for format %#x", __func__, out->format); |
| } |
| } else { |
| if ((out->dev->mode == AUDIO_MODE_IN_COMMUNICATION || |
| voice_extn_compress_voip_is_active(out->dev)) && |
| (voice_extn_compress_voip_is_config_supported(config))) { |
| ret = voice_extn_compress_voip_open_output_stream(out); |
| if (ret != 0) { |
| ALOGE("%s: Compress voip output cannot be opened, error:%d", |
| __func__, ret); |
| goto error_open; |
| } |
| } else { |
| out->usecase = GET_USECASE_AUDIO_PLAYBACK_PRIMARY(use_db_as_primary); |
| out->config = GET_PCM_CONFIG_AUDIO_PLAYBACK_PRIMARY(use_db_as_primary); |
| } |
| } |
| } else if (audio_is_linear_pcm(out->format) && |
| out->flags == AUDIO_OUTPUT_FLAG_NONE && is_usb_dev) { |
| out->channel_mask = config->channel_mask; |
| out->sample_rate = config->sample_rate; |
| out->format = config->format; |
| out->usecase = USECASE_AUDIO_PLAYBACK_HIFI; |
| // does this change? |
| out->config = is_hdmi ? pcm_config_hdmi_multi : pcm_config_hifi; |
| out->config.rate = config->sample_rate; |
| out->config.channels = audio_channel_count_from_out_mask(out->channel_mask); |
| out->config.period_size = HDMI_MULTI_PERIOD_BYTES / (out->config.channels * |
| audio_bytes_per_sample(config->format)); |
| out->config.format = pcm_format_from_audio_format(out->format); |
| } else if ((out->flags & AUDIO_OUTPUT_FLAG_COMPRESS_OFFLOAD) || |
| (out->flags == AUDIO_OUTPUT_FLAG_DIRECT)) { |
| pthread_mutex_lock(&adev->lock); |
| bool offline = (adev->card_status == CARD_STATUS_OFFLINE); |
| pthread_mutex_unlock(&adev->lock); |
| |
| // reject offload during card offline to allow |
| // fallback to s/w paths |
| if (offline) { |
| ret = -ENODEV; |
| goto error_open; |
| } |
| |
| if (config->offload_info.version != AUDIO_INFO_INITIALIZER.version || |
| config->offload_info.size != AUDIO_INFO_INITIALIZER.size) { |
| ALOGE("%s: Unsupported Offload information", __func__); |
| ret = -EINVAL; |
| goto error_open; |
| } |
| |
| if (config->offload_info.format == 0) |
| config->offload_info.format = config->format; |
| if (config->offload_info.sample_rate == 0) |
| config->offload_info.sample_rate = config->sample_rate; |
| |
| if (!is_supported_format(config->offload_info.format) && |
| !audio_extn_passthru_is_supported_format(config->offload_info.format)) { |
| ALOGE("%s: Unsupported audio format %x " , __func__, config->offload_info.format); |
| ret = -EINVAL; |
| goto error_open; |
| } |
| |
| /* TrueHD only supported for 48k multiples (48k, 96k, 192k) */ |
| if ((config->offload_info.format == AUDIO_FORMAT_DOLBY_TRUEHD) && |
| (audio_extn_passthru_is_passthrough_stream(out)) && |
| !((config->sample_rate == 48000) || |
| (config->sample_rate == 96000) || |
| (config->sample_rate == 192000))) { |
| ALOGE("%s: Unsupported sample rate %d for audio format %x", |
| __func__, config->sample_rate, config->offload_info.format); |
| ret = -EINVAL; |
| goto error_open; |
| } |
| |
| out->compr_config.codec = (struct snd_codec *) |
| calloc(1, sizeof(struct snd_codec)); |
| |
| if (!out->compr_config.codec) { |
| ret = -ENOMEM; |
| goto error_open; |
| } |
| |
| out->stream.pause = out_pause; |
| out->stream.resume = out_resume; |
| out->stream.flush = out_flush; |
| out->stream.set_callback = out_set_callback; |
| if (out->flags & AUDIO_OUTPUT_FLAG_COMPRESS_OFFLOAD) { |
| out->stream.drain = out_drain; |
| out->usecase = get_offload_usecase(adev, true /* is_compress */); |
| ALOGV("Compress Offload usecase .. usecase selected %d", out->usecase); |
| } else { |
| out->usecase = get_offload_usecase(adev, false /* is_compress */); |
| ALOGV("non-offload DIRECT_usecase ... usecase selected %d ", out->usecase); |
| } |
| |
| if (out->flags & AUDIO_OUTPUT_FLAG_FAST) { |
| ALOGD("%s: Setting latency mode to true", __func__); |
| #ifdef AUDIO_GKI_ENABLED |
| /* out->compr_config.codec->reserved[1] is for flags */ |
| out->compr_config.codec->reserved[1] |= audio_extn_utils_get_perf_mode_flag(); |
| #else |
| out->compr_config.codec->flags |= audio_extn_utils_get_perf_mode_flag(); |
| #endif |
| } |
| |
| if (out->usecase == USECASE_INVALID) { |
| if (compare_device_type(&out->device_list, AUDIO_DEVICE_OUT_AUX_DIGITAL) && |
| config->format == 0 && config->sample_rate == 0 && |
| config->channel_mask == 0) { |
| ALOGI("%s dummy open to query sink capability",__func__); |
| out->usecase = USECASE_AUDIO_PLAYBACK_OFFLOAD; |
| } else { |
| ALOGE("%s, Max allowed OFFLOAD usecase reached ... ", __func__); |
| ret = -EEXIST; |
| goto error_open; |
| } |
| } |
| |
| if (config->offload_info.channel_mask) |
| out->channel_mask = config->offload_info.channel_mask; |
| else if (config->channel_mask) { |
| out->channel_mask = config->channel_mask; |
| config->offload_info.channel_mask = config->channel_mask; |
| } else { |
| ALOGE("out->channel_mask not set for OFFLOAD/DIRECT usecase"); |
| ret = -EINVAL; |
| goto error_open; |
| } |
| |
| format = out->format = config->offload_info.format; |
| out->sample_rate = config->offload_info.sample_rate; |
| |
| out->bit_width = CODEC_BACKEND_DEFAULT_BIT_WIDTH; |
| |
| out->compr_config.codec->id = get_snd_codec_id(config->offload_info.format); |
| if (audio_extn_utils_is_dolby_format(config->offload_info.format)) { |
| audio_extn_dolby_send_ddp_endp_params(adev); |
| audio_extn_dolby_set_dmid(adev); |
| } |
| |
| out->compr_config.codec->sample_rate = |
| config->offload_info.sample_rate; |
| out->compr_config.codec->bit_rate = |
| config->offload_info.bit_rate; |
| out->compr_config.codec->ch_in = |
| audio_channel_count_from_out_mask(out->channel_mask); |
| out->compr_config.codec->ch_out = out->compr_config.codec->ch_in; |
| /* Update bit width only for non passthrough usecases. |
| * For passthrough usecases, the output will always be opened @16 bit |
| */ |
| if (!audio_extn_passthru_is_passthrough_stream(out)) |
| out->bit_width = AUDIO_OUTPUT_BIT_WIDTH; |
| |
| if (out->flags & AUDIO_OUTPUT_FLAG_TIMESTAMP) |
| #ifdef AUDIO_GKI_ENABLED |
| /* out->compr_config.codec->reserved[1] is for flags */ |
| out->compr_config.codec->reserved[1] |= COMPRESSED_TIMESTAMP_FLAG; |
| ALOGVV("%s : out->compr_config.codec->flags -> (%#x) ", __func__, out->compr_config.codec->reserved[1]); |
| #else |
| out->compr_config.codec->flags |= COMPRESSED_TIMESTAMP_FLAG; |
| ALOGVV("%s : out->compr_config.codec->flags -> (%#x) ", __func__, out->compr_config.codec->flags); |
| #endif |
| |
| /*TODO: Do we need to change it for passthrough */ |
| out->compr_config.codec->format = SND_AUDIOSTREAMFORMAT_RAW; |
| |
| if ((config->offload_info.format & AUDIO_FORMAT_MAIN_MASK) == AUDIO_FORMAT_AAC) |
| out->compr_config.codec->format = SND_AUDIOSTREAMFORMAT_RAW; |
| else if ((config->offload_info.format & AUDIO_FORMAT_MAIN_MASK) == AUDIO_FORMAT_AAC_ADTS) |
| out->compr_config.codec->format = SND_AUDIOSTREAMFORMAT_MP4ADTS; |
| else if ((config->offload_info.format & AUDIO_FORMAT_MAIN_MASK) == AUDIO_FORMAT_AAC_LATM) |
| out->compr_config.codec->format = SND_AUDIOSTREAMFORMAT_MP4LATM; |
| |
| if ((config->offload_info.format & AUDIO_FORMAT_MAIN_MASK) == |
| AUDIO_FORMAT_PCM) { |
| |
| /*Based on platform support, configure appropriate alsa format for corresponding |
| *hal input format. |
| */ |
| out->compr_config.codec->format = hal_format_to_alsa( |
| config->offload_info.format); |
| |
| out->hal_op_format = alsa_format_to_hal( |
| out->compr_config.codec->format); |
| out->hal_ip_format = out->format; |
| |
| /*for direct non-compress playback populate bit_width based on selected alsa format as |
| *hal input format and alsa format might differ based on platform support. |
| */ |
| out->bit_width = audio_bytes_per_sample( |
| out->hal_op_format) << 3; |
| |
| out->compr_config.fragments = DIRECT_PCM_NUM_FRAGMENTS; |
| |
| if (property_get_bool("vendor.audio.offload.buffer.duration.enabled", false)) { |
| if ((config->offload_info.duration_us >= MIN_OFFLOAD_BUFFER_DURATION_MS * 1000) && |
| (config->offload_info.duration_us <= MAX_OFFLOAD_BUFFER_DURATION_MS * 1000)) |
| out->info.duration_us = (int64_t)config->offload_info.duration_us; |
| } |
| |
| /* Check if alsa session is configured with the same format as HAL input format, |
| * if not then derive correct fragment size needed to accomodate the |
| * conversion of HAL input format to alsa format. |
| */ |
| audio_extn_utils_update_direct_pcm_fragment_size(out); |
| |
| /*if hal input and output fragment size is different this indicates HAL input format is |
| *not same as the alsa format |
| */ |
| if (out->hal_fragment_size != out->compr_config.fragment_size) { |
| /*Allocate a buffer to convert input data to the alsa configured format. |
| *size of convert buffer is equal to the size required to hold one fragment size |
| *worth of pcm data, this is because flinger does not write more than fragment_size |
| */ |
| out->convert_buffer = calloc(1,out->compr_config.fragment_size); |
| if (out->convert_buffer == NULL){ |
| ALOGE("Allocation failed for convert buffer for size %d", out->compr_config.fragment_size); |
| ret = -ENOMEM; |
| goto error_open; |
| } |
| } |
| } else if (audio_extn_passthru_is_passthrough_stream(out)) { |
| out->compr_config.fragment_size = |
| audio_extn_passthru_get_buffer_size(&config->offload_info); |
| out->compr_config.fragments = COMPRESS_OFFLOAD_NUM_FRAGMENTS; |
| } else { |
| out->compr_config.fragment_size = |
| platform_get_compress_offload_buffer_size(&config->offload_info); |
| out->compr_config.fragments = COMPRESS_OFFLOAD_NUM_FRAGMENTS; |
| } |
| |
| if (out->flags & AUDIO_OUTPUT_FLAG_TIMESTAMP) { |
| out->compr_config.fragment_size += sizeof(struct snd_codec_metadata); |
| } |
| if (config->offload_info.format == AUDIO_FORMAT_FLAC) { |
| #ifdef AUDIO_GKI_ENABLED |
| generic_dec = |
| &(out->compr_config.codec->options.generic.reserved[1]); |
| ((struct snd_generic_dec_flac *)generic_dec)->sample_size = |
| AUDIO_OUTPUT_BIT_WIDTH; |
| #else |
| out->compr_config.codec->options.flac_dec.sample_size = AUDIO_OUTPUT_BIT_WIDTH; |
| #endif |
| } |
| |
| if (config->offload_info.format == AUDIO_FORMAT_APTX) { |
| audio_extn_send_aptx_dec_bt_addr_to_dsp(out); |
| } |
| |
| if (flags & AUDIO_OUTPUT_FLAG_NON_BLOCKING) |
| out->non_blocking = 1; |
| |
| if ((flags & AUDIO_OUTPUT_FLAG_TIMESTAMP) && |
| (flags & AUDIO_OUTPUT_FLAG_HW_AV_SYNC)) { |
| out->render_mode = RENDER_MODE_AUDIO_STC_MASTER; |
| } else if(flags & AUDIO_OUTPUT_FLAG_TIMESTAMP) { |
| out->render_mode = RENDER_MODE_AUDIO_MASTER; |
| } else { |
| out->render_mode = RENDER_MODE_AUDIO_NO_TIMESTAMP; |
| } |
| |
| memset(&out->channel_map_param, 0, |
| sizeof(struct audio_out_channel_map_param)); |
| |
| out->send_new_metadata = 1; |
| out->send_next_track_params = false; |
| out->is_compr_metadata_avail = false; |
| out->offload_state = OFFLOAD_STATE_IDLE; |
| out->playback_started = 0; |
| out->writeAt.tv_sec = 0; |
| out->writeAt.tv_nsec = 0; |
| |
| audio_extn_dts_create_state_notifier_node(out->usecase); |
| |
| ALOGV("%s: offloaded output offload_info version %04x bit rate %d", |
| __func__, config->offload_info.version, |
| config->offload_info.bit_rate); |
| |
| /* Check if DSD audio format is supported in codec |
| * and there is no active native DSD use case |
| */ |
| |
| if ((config->format == AUDIO_FORMAT_DSD) && |
| (!platform_check_codec_dsd_support(adev->platform) || |
| audio_is_dsd_native_stream_active(adev))) { |
| ret = -EINVAL; |
| goto error_open; |
| } |
| |
| /* Disable gapless if any of the following is true |
| * passthrough playback |
| * AV playback |
| * non compressed Direct playback |
| */ |
| if (audio_extn_passthru_is_passthrough_stream(out) || |
| (config->format == AUDIO_FORMAT_DSD) || |
| (config->format == AUDIO_FORMAT_IEC61937) || |
| config->offload_info.has_video || |
| !(out->flags & AUDIO_OUTPUT_FLAG_COMPRESS_OFFLOAD)) { |
| check_and_set_gapless_mode(adev, false); |
| } else |
| check_and_set_gapless_mode(adev, true); |
| |
| if (audio_extn_passthru_is_passthrough_stream(out)) { |
| out->flags |= AUDIO_OUTPUT_FLAG_COMPRESS_PASSTHROUGH; |
| } |
| if (config->format == AUDIO_FORMAT_DSD) { |
| out->flags |= AUDIO_OUTPUT_FLAG_COMPRESS_PASSTHROUGH; |
| #ifdef AUDIO_GKI_ENABLED |
| /* out->compr_config.codec->reserved[0] is for compr_passthr */ |
| out->compr_config.codec->reserved[0] = PASSTHROUGH_DSD; |
| #else |
| out->compr_config.codec->compr_passthr = PASSTHROUGH_DSD; |
| #endif |
| } |
| |
| create_offload_callback_thread(out); |
| |
| } else if (out->flags & AUDIO_OUTPUT_FLAG_INCALL_MUSIC) { |
| switch (config->sample_rate) { |
| case 0: |
| out->sample_rate = DEFAULT_OUTPUT_SAMPLING_RATE; |
| break; |
| case 8000: |
| case 16000: |
| case 48000: |
| out->sample_rate = config->sample_rate; |
| break; |
| default: |
| ALOGE("%s: Unsupported sampling rate %d for Incall Music", __func__, |
| config->sample_rate); |
| config->sample_rate = DEFAULT_OUTPUT_SAMPLING_RATE; |
| ret = -EINVAL; |
| goto error_open; |
| } |
| //FIXME: add support for MONO stream configuration when audioflinger mixer supports it |
| switch (config->channel_mask) { |
| case AUDIO_CHANNEL_NONE: |
| case AUDIO_CHANNEL_OUT_STEREO: |
| out->channel_mask = AUDIO_CHANNEL_OUT_STEREO; |
| break; |
| default: |
| ALOGE("%s: Unsupported channel mask %#x for Incall Music", __func__, |
| config->channel_mask); |
| config->channel_mask = AUDIO_CHANNEL_OUT_STEREO; |
| ret = -EINVAL; |
| goto error_open; |
| } |
| switch (config->format) { |
| case AUDIO_FORMAT_DEFAULT: |
| case AUDIO_FORMAT_PCM_16_BIT: |
| out->format = AUDIO_FORMAT_PCM_16_BIT; |
| break; |
| default: |
| ALOGE("%s: Unsupported format %#x for Incall Music", __func__, |
| config->format); |
| config->format = AUDIO_FORMAT_PCM_16_BIT; |
| ret = -EINVAL; |
| goto error_open; |
| } |
| |
| ret = voice_extn_check_and_set_incall_music_usecase(adev, out); |
| if (ret != 0) { |
| ALOGE("%s: Incall music delivery usecase cannot be set error:%d", |
| __func__, ret); |
| goto error_open; |
| } |
| } else if (is_single_device_type_equal(&out->device_list, |
| AUDIO_DEVICE_OUT_TELEPHONY_TX)) { |
| switch (config->sample_rate) { |
| case 0: |
| out->sample_rate = AFE_PROXY_SAMPLING_RATE; |
| break; |
| case 8000: |
| case 16000: |
| case 48000: |
| out->sample_rate = config->sample_rate; |
| break; |
| default: |
| ALOGE("%s: Unsupported sampling rate %d for Telephony TX", __func__, |
| config->sample_rate); |
| config->sample_rate = AFE_PROXY_SAMPLING_RATE; |
| ret = -EINVAL; |
| break; |
| } |
| //FIXME: add support for MONO stream configuration when audioflinger mixer supports it |
| switch (config->channel_mask) { |
| case AUDIO_CHANNEL_NONE: |
| out->channel_mask = AUDIO_CHANNEL_OUT_STEREO; |
| break; |
| case AUDIO_CHANNEL_OUT_STEREO: |
| out->channel_mask = config->channel_mask; |
| break; |
| default: |
| ALOGE("%s: Unsupported channel mask %#x for Telephony TX", __func__, |
| config->channel_mask); |
| config->channel_mask = AUDIO_CHANNEL_OUT_STEREO; |
| ret = -EINVAL; |
| break; |
| } |
| switch (config->format) { |
| case AUDIO_FORMAT_DEFAULT: |
| out->format = AUDIO_FORMAT_PCM_16_BIT; |
| break; |
| case AUDIO_FORMAT_PCM_16_BIT: |
| out->format = config->format; |
| break; |
| default: |
| ALOGE("%s: Unsupported format %#x for Telephony TX", __func__, |
| config->format); |
| config->format = AUDIO_FORMAT_PCM_16_BIT; |
| ret = -EINVAL; |
| break; |
| } |
| if (ret != 0) |
| goto error_open; |
| |
| out->usecase = USECASE_AUDIO_PLAYBACK_AFE_PROXY; |
| out->config = pcm_config_afe_proxy_playback; |
| out->config.rate = out->sample_rate; |
| out->config.channels = |
| audio_channel_count_from_out_mask(out->channel_mask); |
| out->config.format = pcm_format_from_audio_format(out->format); |
| adev->voice_tx_output = out; |
| } else { |
| unsigned int channels = 0; |
| /*Update config params to default if not set by the caller*/ |
| if (config->sample_rate == 0) |
| config->sample_rate = DEFAULT_OUTPUT_SAMPLING_RATE; |
| if (config->channel_mask == AUDIO_CHANNEL_NONE) |
| config->channel_mask = AUDIO_CHANNEL_OUT_STEREO; |
| if (config->format == AUDIO_FORMAT_DEFAULT) |
| config->format = AUDIO_FORMAT_PCM_16_BIT; |
| |
| channels = audio_channel_count_from_out_mask(out->channel_mask); |
| |
| if (out->flags & AUDIO_OUTPUT_FLAG_INTERACTIVE) { |
| out->usecase = get_interactive_usecase(adev); |
| out->config = pcm_config_low_latency; |
| } else if (out->flags & AUDIO_OUTPUT_FLAG_RAW) { |
| out->usecase = USECASE_AUDIO_PLAYBACK_ULL; |
| out->realtime = may_use_noirq_mode(adev, USECASE_AUDIO_PLAYBACK_ULL, |
| out->flags); |
| out->config = out->realtime ? pcm_config_rt : pcm_config_low_latency; |
| } else if (out->flags & AUDIO_OUTPUT_FLAG_MMAP_NOIRQ) { |
| out->usecase = USECASE_AUDIO_PLAYBACK_MMAP; |
| out->config = pcm_config_mmap_playback; |
| out->stream.start = out_start; |
| out->stream.stop = out_stop; |
| out->stream.create_mmap_buffer = out_create_mmap_buffer; |
| out->stream.get_mmap_position = out_get_mmap_position; |
| } else if (out->flags & AUDIO_OUTPUT_FLAG_FAST) { |
| out->usecase = USECASE_AUDIO_PLAYBACK_LOW_LATENCY; |
| out->hal_output_suspend_supported = |
| property_get_bool("vendor.audio.hal.output.suspend.supported", false); |
| out->dynamic_pm_qos_config_supported = |
| property_get_bool("vendor.audio.hal.dynamic.qos.config.supported", false); |
| if (!out->dynamic_pm_qos_config_supported) { |
| ALOGI("%s: dynamic qos voting not enabled for platform", __func__); |
| } else { |
| ALOGI("%s: dynamic qos voting enabled for platform", __func__); |
| //the mixer path will be a string similar to "low-latency-playback resume" |
| strlcpy(out->pm_qos_mixer_path, use_case_table[out->usecase], MAX_MIXER_PATH_LEN); |
| strlcat(out->pm_qos_mixer_path, |
| " resume", MAX_MIXER_PATH_LEN); |
| ALOGI("%s: created %s pm_qos_mixer_path" , __func__, |
| out->pm_qos_mixer_path); |
| } |
| out->config = pcm_config_low_latency; |
| } else if (out->flags & AUDIO_OUTPUT_FLAG_DEEP_BUFFER) { |
| out->usecase = USECASE_AUDIO_PLAYBACK_DEEP_BUFFER; |
| out->config = pcm_config_deep_buffer; |
| out->config.period_size = get_output_period_size(config->sample_rate, out->format, |
| channels, DEEP_BUFFER_OUTPUT_PERIOD_DURATION); |
| if (out->config.period_size <= 0) { |
| ALOGE("Invalid configuration period size is not valid"); |
| ret = -EINVAL; |
| goto error_open; |
| } |
| } else if (flags & AUDIO_OUTPUT_FLAG_TTS) { |
| out->usecase = USECASE_AUDIO_PLAYBACK_TTS; |
| out->config = pcm_config_deep_buffer; |
| } else if (config->channel_mask & AUDIO_CHANNEL_HAPTIC_ALL) { |
| out->usecase = USECASE_AUDIO_PLAYBACK_WITH_HAPTICS; |
| out->config = pcm_config_haptics_audio; |
| if (force_haptic_path) |
| adev->haptics_config = pcm_config_haptics_audio; |
| else |
| adev->haptics_config = pcm_config_haptics; |
| |
| channels = |
| audio_channel_count_from_out_mask(out->channel_mask & ~AUDIO_CHANNEL_HAPTIC_ALL); |
| |
| if (force_haptic_path) { |
| out->config.channels = 1; |
| adev->haptics_config.channels = 1; |
| } else |
| adev->haptics_config.channels = audio_channel_count_from_out_mask(out->channel_mask & AUDIO_CHANNEL_HAPTIC_ALL); |
| } else if (compare_device_type(&out->device_list, AUDIO_DEVICE_OUT_BUS)) { |
| ret = audio_extn_auto_hal_open_output_stream(out); |
| if (ret) { |
| ALOGE("%s: Failed to open output stream for bus device", __func__); |
| ret = -EINVAL; |
| goto error_open; |
| } |
| } else { |
| /* primary path is the default path selected if no other outputs are available/suitable */ |
| out->usecase = GET_USECASE_AUDIO_PLAYBACK_PRIMARY(use_db_as_primary); |
| out->config = GET_PCM_CONFIG_AUDIO_PLAYBACK_PRIMARY(use_db_as_primary); |
| } |
| out->hal_ip_format = format = out->format; |
| out->config.format = hal_format_to_pcm(out->hal_ip_format); |
| out->hal_op_format = pcm_format_to_hal(out->config.format); |
| out->bit_width = format_to_bitwidth_table[out->hal_op_format] << 3; |
| out->config.rate = config->sample_rate; |
| out->sample_rate = out->config.rate; |
| out->config.channels = channels; |
| if (out->hal_ip_format != out->hal_op_format) { |
| uint32_t buffer_size = out->config.period_size * |
| format_to_bitwidth_table[out->hal_op_format] * |
| out->config.channels; |
| out->convert_buffer = calloc(1, buffer_size); |
| if (out->convert_buffer == NULL){ |
| ALOGE("Allocation failed for convert buffer for size %d", |
| out->compr_config.fragment_size); |
| ret = -ENOMEM; |
| goto error_open; |
| } |
| ALOGD("Convert buffer allocated of size %d", buffer_size); |
| } |
| } |
| |
| ALOGV("%s devices:%d, format:%x, out->sample_rate:%d,out->bit_width:%d out->format:%d out->flags:%x, flags: %x usecase %d", |
| __func__, devices, format, out->sample_rate, out->bit_width, out->format, out->flags, flags, out->usecase); |
| |
| /* TODO remove this hardcoding and check why width is zero*/ |
| if (out->bit_width == 0) |
| out->bit_width = 16; |
| audio_extn_utils_update_stream_output_app_type_cfg(adev->platform, |
| &adev->streams_output_cfg_list, |
| &out->device_list, out->flags, |
| out->hal_op_format, out->sample_rate, |
| out->bit_width, out->channel_mask, out->profile, |
| &out->app_type_cfg); |
| if ((out->usecase == (audio_usecase_t)(GET_USECASE_AUDIO_PLAYBACK_PRIMARY(use_db_as_primary))) || |
| (flags & AUDIO_OUTPUT_FLAG_PRIMARY)) { |
| /* Ensure the default output is not selected twice */ |
| if(adev->primary_output == NULL) |
| adev->primary_output = out; |
| else { |
| ALOGE("%s: Primary output is already opened", __func__); |
| ret = -EEXIST; |
| goto error_open; |
| } |
| } |
| |
| /* Check if this usecase is already existing */ |
| pthread_mutex_lock(&adev->lock); |
| if ((get_usecase_from_list(adev, out->usecase) != NULL) && |
| (out->usecase != USECASE_COMPRESS_VOIP_CALL)) { |
| ALOGE("%s: Usecase (%d) is already present", __func__, out->usecase); |
| pthread_mutex_unlock(&adev->lock); |
| ret = -EEXIST; |
| goto error_open; |
| } |
| |
| pthread_mutex_unlock(&adev->lock); |
| |
| out->stream.common.get_sample_rate = out_get_sample_rate; |
| out->stream.common.set_sample_rate = out_set_sample_rate; |
| out->stream.common.get_buffer_size = out_get_buffer_size; |
| out->stream.common.get_channels = out_get_channels; |
| out->stream.common.get_format = out_get_format; |
| out->stream.common.set_format = out_set_format; |
| out->stream.common.standby = out_standby; |
| out->stream.common.dump = out_dump; |
| out->stream.common.set_parameters = out_set_parameters; |
| out->stream.common.get_parameters = out_get_parameters; |
| out->stream.common.add_audio_effect = out_add_audio_effect; |
| out->stream.common.remove_audio_effect = out_remove_audio_effect; |
| out->stream.get_latency = out_get_latency; |
| out->stream.set_volume = out_set_volume; |
| #ifdef NO_AUDIO_OUT |
| out->stream.write = out_write_for_no_output; |
| #else |
| out->stream.write = out_write; |
| #endif |
| out->stream.get_render_position = out_get_render_position; |
| out->stream.get_next_write_timestamp = out_get_next_write_timestamp; |
| out->stream.get_presentation_position = out_get_presentation_position; |
| |
| if (out->realtime) |
| out->af_period_multiplier = af_period_multiplier; |
| else |
| out->af_period_multiplier = 1; |
| |
| out->kernel_buffer_size = out->config.period_size * out->config.period_count; |
| |
| out->standby = 1; |
| out->volume_l = PLAYBACK_GAIN_MAX; |
| out->volume_r = PLAYBACK_GAIN_MAX; |
| /* out->muted = false; by calloc() */ |
| /* out->written = 0; by calloc() */ |
| |
| config->format = out->stream.common.get_format(&out->stream.common); |
| config->channel_mask = out->stream.common.get_channels(&out->stream.common); |
| config->sample_rate = out->stream.common.get_sample_rate(&out->stream.common); |
| register_format(out->format, out->supported_formats); |
| register_channel_mask(out->channel_mask, out->supported_channel_masks); |
| register_sample_rate(out->sample_rate, out->supported_sample_rates); |
| |
| #ifndef LINUX_ENABLED |
| out->error_log = error_log_create( |
| ERROR_LOG_ENTRIES, |
| 1000000000 /* aggregate consecutive identical errors within one second in ns */); |
| #endif |
| /* |
| By locking output stream before registering, we allow the callback |
| to update stream's state only after stream's initial state is set to |
| adev state. |
| */ |
| lock_output_stream(out); |
| audio_extn_snd_mon_register_listener(out, out_snd_mon_cb); |
| pthread_mutex_lock(&adev->lock); |
| out->card_status = adev->card_status; |
| pthread_mutex_unlock(&adev->lock); |
| pthread_mutex_unlock(&out->lock); |
| |
| stream_app_type_cfg_init(&out->app_type_cfg); |
| |
| *stream_out = &out->stream; |
| ALOGD("%s: Stream (%p) picks up usecase (%s)", __func__, &out->stream, |
| use_case_table[out->usecase]); |
| |
| if (out->flags & AUDIO_OUTPUT_FLAG_COMPRESS_OFFLOAD) |
| audio_extn_dts_notify_playback_state(out->usecase, 0, out->sample_rate, |
| popcount(out->channel_mask), out->playback_started); |
| /* setup a channel for client <--> adsp communication for stream events */ |
| is_direct_passthough = audio_extn_passthru_is_direct_passthrough(out); |
| if ((out->flags & AUDIO_OUTPUT_FLAG_COMPRESS_OFFLOAD) || |
| (out->flags & AUDIO_OUTPUT_FLAG_DIRECT_PCM) || |
| audio_extn_ip_hdlr_intf_supported_for_copp(adev->platform) || |
| (audio_extn_ip_hdlr_intf_supported(config->format, is_direct_passthough, false))) { |
| hdlr_stream_cfg.pcm_device_id = platform_get_pcm_device_id( |
| out->usecase, PCM_PLAYBACK); |
| hdlr_stream_cfg.flags = out->flags; |
| hdlr_stream_cfg.type = PCM_PLAYBACK; |
| ret = audio_extn_adsp_hdlr_stream_open(&out->adsp_hdlr_stream_handle, |
| &hdlr_stream_cfg); |
| if (ret) { |
| ALOGE("%s: adsp_hdlr_stream_open failed %d",__func__, ret); |
| out->adsp_hdlr_stream_handle = NULL; |
| } |
| } |
| ip_hdlr_stream = audio_extn_ip_hdlr_intf_supported(config->format, |
| is_direct_passthough, false); |
| ip_hdlr_dev = audio_extn_ip_hdlr_intf_supported_for_copp(adev->platform); |
| if (ip_hdlr_stream || ip_hdlr_dev ) { |
| ret = audio_extn_ip_hdlr_intf_init(&out->ip_hdlr_handle, NULL, NULL, adev, out->usecase); |
| if (ret < 0) { |
| ALOGE("%s: audio_extn_ip_hdlr_intf_init failed %d",__func__, ret); |
| out->ip_hdlr_handle = NULL; |
| } |
| } |
| |
| ret = io_streams_map_insert(adev, &out->stream.common, |
| out->handle, AUDIO_PATCH_HANDLE_NONE); |
| if (ret != 0) |
| goto error_open; |
| |
| out->out_ctxt.output = out; |
| |
| pthread_mutex_lock(&adev->lock); |
| list_add_tail(&adev->active_outputs_list, &out->out_ctxt.list); |
| pthread_mutex_unlock(&adev->lock); |
| |
| ALOGV("%s: exit", __func__); |
| return 0; |
| |
| error_open: |
| if (out->convert_buffer) |
| free(out->convert_buffer); |
| free(out); |
| *stream_out = NULL; |
| ALOGD("%s: exit: ret %d", __func__, ret); |
| return ret; |
| } |
| |
| void adev_close_output_stream(struct audio_hw_device *dev __unused, |
| struct audio_stream_out *stream) |
| { |
| struct stream_out *out = (struct stream_out *)stream; |
| struct audio_device *adev = out->dev; |
| int ret = 0; |
| |
| ALOGD("%s: enter:stream_handle(%s)",__func__, use_case_table[out->usecase]); |
| |
| io_streams_map_remove(adev, out->handle); |
| |
| // remove out_ctxt early to prevent the stream |
| // being opened in a race condition |
| pthread_mutex_lock(&adev->lock); |
| list_remove(&out->out_ctxt.list); |
| pthread_mutex_unlock(&adev->lock); |
| |
| // must deregister from sndmonitor first to prevent races |
| // between the callback and close_stream |
| audio_extn_snd_mon_unregister_listener(out); |
| |
| /* close adsp hdrl session before standby */ |
| if (out->adsp_hdlr_stream_handle) { |
| ret = audio_extn_adsp_hdlr_stream_close(out->adsp_hdlr_stream_handle); |
| if (ret) |
| ALOGE("%s: adsp_hdlr_stream_close failed %d",__func__, ret); |
| out->adsp_hdlr_stream_handle = NULL; |
| } |
| |
| if (out->ip_hdlr_handle) { |
| audio_extn_ip_hdlr_intf_deinit(out->ip_hdlr_handle); |
| out->ip_hdlr_handle = NULL; |
| } |
| |
| if (out->usecase == USECASE_COMPRESS_VOIP_CALL) { |
| pthread_mutex_lock(&adev->lock); |
| ret = voice_extn_compress_voip_close_output_stream(&stream->common); |
| out->started = 0; |
| pthread_mutex_unlock(&adev->lock); |
| if(ret != 0) |
| ALOGE("%s: Compress voip output cannot be closed, error:%d", |
| __func__, ret); |
| } else |
| out_standby(&stream->common); |
| |
| if (is_offload_usecase(out->usecase)) { |
| audio_extn_dts_remove_state_notifier_node(out->usecase); |
| destroy_offload_callback_thread(out); |
| free_offload_usecase(adev, out->usecase); |
| if (out->compr_config.codec != NULL) |
| free(out->compr_config.codec); |
| } |
| |
| out->a2dp_muted = false; |
| |
| if (is_interactive_usecase(out->usecase)) |
| free_interactive_usecase(adev, out->usecase); |
| |
| if (out->convert_buffer != NULL) { |
| free(out->convert_buffer); |
| out->convert_buffer = NULL; |
| } |
| |
| if (adev->voice_tx_output == out) |
| adev->voice_tx_output = NULL; |
| |
| #ifndef LINUX_ENABLED |
| error_log_destroy(out->error_log); |
| out->error_log = NULL; |
| #endif |
| if (adev->primary_output == out) |
| adev->primary_output = NULL; |
| |
| pthread_cond_destroy(&out->cond); |
| pthread_mutex_destroy(&out->lock); |
| pthread_mutex_destroy(&out->pre_lock); |
| pthread_mutex_destroy(&out->latch_lock); |
| pthread_mutex_destroy(&out->position_query_lock); |
| |
| pthread_mutex_lock(&adev->lock); |
| free(stream); |
| pthread_mutex_unlock(&adev->lock); |
| ALOGV("%s: exit", __func__); |
| } |
| |
| static int adev_set_parameters(struct audio_hw_device *dev, const char *kvpairs) |
| { |
| struct audio_device *adev = (struct audio_device *)dev; |
| struct str_parms *parms; |
| char value[32]; |
| int val; |
| int ret; |
| int status = 0; |
| bool a2dp_reconfig = false; |
| struct listnode *node; |
| int controller = -1, stream = -1; |
| |
| ALOGD("%s: enter: %s", __func__, kvpairs); |
| parms = str_parms_create_str(kvpairs); |
| |
| if (!parms) |
| goto error; |
| |
| /* notify adev and input/output streams on the snd card status */ |
| adev_snd_mon_cb((void *)adev, parms); |
| |
| ret = str_parms_get_str(parms, "SND_CARD_STATUS", value, sizeof(value)); |
| if (ret >= 0) { |
| list_for_each(node, &adev->active_outputs_list) { |
| streams_output_ctxt_t *out_ctxt = node_to_item(node, |
| streams_output_ctxt_t, |
| list); |
| out_snd_mon_cb((void *)out_ctxt->output, parms); |
| } |
| |
| list_for_each(node, &adev->active_inputs_list) { |
| streams_input_ctxt_t *in_ctxt = node_to_item(node, |
| streams_input_ctxt_t, |
| list); |
| in_snd_mon_cb((void *)in_ctxt->input, parms); |
| } |
| } |
| |
| pthread_mutex_lock(&adev->lock); |
| ret = str_parms_get_str(parms, "BT_SCO", value, sizeof(value)); |
| if (ret >= 0) { |
| /* When set to false, HAL should disable EC and NS */ |
| if (strcmp(value, AUDIO_PARAMETER_VALUE_ON) == 0){ |
| adev->bt_sco_on = true; |
| /* |
| * When ever BT_SCO=ON arrives, make sure to route |
| * all use cases to SCO device, otherwise due to delay in |
| * BT_SCO=ON and lack of synchronization with create audio patch |
| * request for SCO device, some times use case not routed properly to |
| * SCO device |
| */ |
| struct audio_usecase *usecase; |
| struct listnode *node; |
| list_for_each(node, &adev->usecase_list) { |
| usecase = node_to_item(node, struct audio_usecase, list); |
| if (usecase->stream.in && (usecase->type == PCM_CAPTURE) && |
| (!is_btsco_device(SND_DEVICE_NONE, usecase->in_snd_device)) && (is_sco_in_device_type(&usecase->stream.in->device_list))) { |
| ALOGD("BT_SCO ON, switch all in use case to it"); |
| select_devices(adev, usecase->id); |
| } |
| if (usecase->stream.out && (usecase->type == PCM_PLAYBACK || |
| usecase->type == VOICE_CALL) && |
| (!is_btsco_device(usecase->out_snd_device, SND_DEVICE_NONE)) && (is_sco_out_device_type(&usecase->stream.out->device_list))) { |
| ALOGD("BT_SCO ON, switch all out use case to it"); |
| select_devices(adev, usecase->id); |
| } |
| } |
| } |
| else { |
| adev->bt_sco_on = false; |
| audio_extn_sco_reset_configuration(); |
| } |
| } |
| |
| ret = str_parms_get_str(parms, "A2dpSuspended", value, sizeof(value)); |
| if (ret >= 0) { |
| if (!strncmp(value, "false", 5) && |
| audio_extn_a2dp_source_is_suspended()) { |
| struct audio_usecase *usecase; |
| struct listnode *node; |
| list_for_each(node, &adev->usecase_list) { |
| usecase = node_to_item(node, struct audio_usecase, list); |
| if (usecase->stream.in && (usecase->type == PCM_CAPTURE) && |
| is_sco_in_device_type(&usecase->stream.in->device_list)) { |
| ALOGD("a2dp resumed, switch bt sco mic to handset mic"); |
| reassign_device_list(&usecase->stream.in->device_list, |
| AUDIO_DEVICE_IN_BUILTIN_MIC, ""); |
| select_devices(adev, usecase->id); |
| } |
| } |
| } |
| } |
| |
| status = voice_set_parameters(adev, parms); |
| if (status != 0) |
| goto done; |
| |
| status = platform_set_parameters(adev->platform, parms); |
| if (status != 0) |
| goto done; |
| |
| ret = str_parms_get_str(parms, AUDIO_PARAMETER_KEY_BT_NREC, value, sizeof(value)); |
| if (ret >= 0) { |
| /* When set to false, HAL should disable EC and NS */ |
| if (strcmp(value, AUDIO_PARAMETER_VALUE_ON) == 0) |
| adev->bluetooth_nrec = true; |
| else |
| adev->bluetooth_nrec = false; |
| } |
| |
| ret = str_parms_get_str(parms, "screen_state", value, sizeof(value)); |
| if (ret >= 0) { |
| if (strcmp(value, AUDIO_PARAMETER_VALUE_ON) == 0) |
| adev->screen_off = false; |
| else |
| adev->screen_off = true; |
| audio_extn_sound_trigger_update_screen_status(adev->screen_off); |
| } |
| |
| ret = str_parms_get_int(parms, "rotation", &val); |
| if (ret >= 0) { |
| bool reverse_speakers = false; |
| int camera_rotation = CAMERA_ROTATION_LANDSCAPE; |
| switch (val) { |
| // FIXME: note that the code below assumes that the speakers are in the correct placement |
| // relative to the user when the device is rotated 90deg from its default rotation. This |
| // assumption is device-specific, not platform-specific like this code. |
| case 270: |
| reverse_speakers = true; |
| camera_rotation = CAMERA_ROTATION_INVERT_LANDSCAPE; |
| break; |
| case 0: |
| case 180: |
| camera_rotation = CAMERA_ROTATION_PORTRAIT; |
| break; |
| case 90: |
| camera_rotation = CAMERA_ROTATION_LANDSCAPE; |
| break; |
| default: |
| ALOGE("%s: unexpected rotation of %d", __func__, val); |
| status = -EINVAL; |
| } |
| if (status == 0) { |
| // check and set swap |
| // - check if orientation changed and speaker active |
| // - set rotation and cache the rotation value |
| adev->camera_orientation = |
| (adev->camera_orientation & ~CAMERA_ROTATION_MASK) | camera_rotation; |
| if (!audio_extn_is_maxx_audio_enabled()) |
| platform_check_and_set_swap_lr_channels(adev, reverse_speakers); |
| } |
| } |
| |
| ret = str_parms_get_str(parms, AUDIO_PARAMETER_KEY_BT_SCO_WB, value, sizeof(value)); |
| if (ret >= 0) { |
| if (strcmp(value, AUDIO_PARAMETER_VALUE_ON) == 0) |
| adev->bt_wb_speech_enabled = true; |
| else |
| adev->bt_wb_speech_enabled = false; |
| } |
| |
| ret = str_parms_get_str(parms, "bt_swb", value, sizeof(value)); |
| if (ret >= 0) { |
| val = atoi(value); |
| adev->swb_speech_mode = val; |
| } |
| |
| ret = str_parms_get_str(parms, AUDIO_PARAMETER_DEVICE_CONNECT, value, sizeof(value)); |
| if (ret >= 0) { |
| val = atoi(value); |
| audio_devices_t device = (audio_devices_t) val; |
| if (audio_is_output_device(val) && |
| (val & AUDIO_DEVICE_OUT_AUX_DIGITAL)) { |
| ALOGV("cache new ext disp type and edid"); |
| platform_get_controller_stream_from_params(parms, &controller, &stream); |
| platform_set_ext_display_device_v2(adev->platform, controller, stream); |
| ret = platform_get_ext_disp_type_v2(adev->platform, controller, stream); |
| if (ret < 0) { |
| ALOGE("%s: Failed to query disp type, ret:%d", __func__, ret); |
| } else { |
| platform_cache_edid_v2(adev->platform, controller, stream); |
| } |
| } else if (audio_is_usb_out_device(device) || audio_is_usb_in_device(device)) { |
| /* |
| * Do not allow AFE proxy port usage by WFD source when USB headset is connected. |
| * Per AudioPolicyManager, USB device is higher priority than WFD. |
| * For Voice call over USB headset, voice call audio is routed to AFE proxy ports. |
| * If WFD use case occupies AFE proxy, it may result unintended behavior while |
| * starting voice call on USB |
| */ |
| ret = str_parms_get_str(parms, "card", value, sizeof(value)); |
| if (ret >= 0) |
| audio_extn_usb_add_device(device, atoi(value)); |
| |
| if (!audio_extn_usb_is_tunnel_supported()) { |
| ALOGV("detected USB connect .. disable proxy"); |
| adev->allow_afe_proxy_usage = false; |
| } |
| } |
| } |
| |
| ret = str_parms_get_str(parms, AUDIO_PARAMETER_DEVICE_DISCONNECT, value, sizeof(value)); |
| if (ret >= 0) { |
| val = atoi(value); |
| audio_devices_t device = (audio_devices_t) val; |
| /* |
| * The HDMI / Displayport disconnect handling has been moved to |
| * audio extension to ensure that its parameters are not |
| * invalidated prior to updating sysfs of the disconnect event |
| * Invalidate will be handled by audio_extn_ext_disp_set_parameters() |
| */ |
| if (audio_is_usb_out_device(device) || audio_is_usb_in_device(device)) { |
| ret = str_parms_get_str(parms, "card", value, sizeof(value)); |
| if (ret >= 0) |
| audio_extn_usb_remove_device(device, atoi(value)); |
| |
| if (!audio_extn_usb_is_tunnel_supported()) { |
| ALOGV("detected USB disconnect .. enable proxy"); |
| adev->allow_afe_proxy_usage = true; |
| } |
| } |
| } |
| |
| audio_extn_qdsp_set_parameters(adev, parms); |
| |
| status = audio_extn_a2dp_set_parameters(parms, &a2dp_reconfig); |
| if (status >= 0 && a2dp_reconfig) { |
| struct audio_usecase *usecase; |
| struct listnode *node; |
| list_for_each(node, &adev->usecase_list) { |
| usecase = node_to_item(node, struct audio_usecase, list); |
| if ((usecase->stream.out == NULL) || (usecase->type != PCM_PLAYBACK)) |
| continue; |
| |
| if (is_a2dp_out_device_type(&usecase->device_list)) { |
| ALOGD("reconfigure a2dp... forcing device switch"); |
| audio_extn_a2dp_set_handoff_mode(true); |
| ALOGD("Switching to speaker and muting the stream before select_devices"); |
| check_a2dp_restore_l(adev, usecase->stream.out, false); |
| //force device switch to re configure encoder |
| select_devices(adev, usecase->id); |
| ALOGD("Unmuting the stream after select_devices"); |
| check_a2dp_restore_l(adev, usecase->stream.out, true); |
| audio_extn_a2dp_set_handoff_mode(false); |
| break; |
| } else if (is_offload_usecase(usecase->stream.out->usecase)) { |
| pthread_mutex_lock(&usecase->stream.out->latch_lock); |
| if (usecase->stream.out->a2dp_muted) { |
| pthread_mutex_unlock(&usecase->stream.out->latch_lock); |
| reassign_device_list(&usecase->stream.out->device_list, |
| AUDIO_DEVICE_OUT_BLUETOOTH_A2DP, ""); |
| check_a2dp_restore_l(adev, usecase->stream.out, true); |
| break; |
| } |
| pthread_mutex_unlock(&usecase->stream.out->latch_lock); |
| } |
| } |
| } |
| |
| //handle vr audio setparam |
| ret = str_parms_get_str(parms, AUDIO_PARAMETER_KEY_VR_AUDIO_MODE, |
| value, sizeof(value)); |
| if (ret >= 0) { |
| ALOGI("Setting vr mode to be %s", value); |
| if (!strncmp(value, "true", 4)) { |
| adev->vr_audio_mode_enabled = true; |
| ALOGI("Setting vr mode to true"); |
| } else if (!strncmp(value, "false", 5)) { |
| adev->vr_audio_mode_enabled = false; |
| ALOGI("Setting vr mode to false"); |
| } else { |
| ALOGI("wrong vr mode set"); |
| } |
| } |
| |
| //FIXME: to be replaced by proper video capture properties API |
| ret = str_parms_get_str(parms, AUDIO_PARAMETER_KEY_CAMERA_FACING, value, sizeof(value)); |
| if (ret >= 0) { |
| int camera_facing = CAMERA_FACING_BACK; |
| if (strcmp(value, AUDIO_PARAMETER_VALUE_FRONT) == 0) |
| camera_facing = CAMERA_FACING_FRONT; |
| else if (strcmp(value, AUDIO_PARAMETER_VALUE_BACK) == 0) |
| camera_facing = CAMERA_FACING_BACK; |
| else { |
| ALOGW("%s: invalid camera facing value: %s", __func__, value); |
| goto done; |
| } |
| adev->camera_orientation = |
| (adev->camera_orientation & ~CAMERA_FACING_MASK) | camera_facing; |
| struct audio_usecase *usecase; |
| struct listnode *node; |
| list_for_each(node, &adev->usecase_list) { |
| usecase = node_to_item(node, struct audio_usecase, list); |
| struct stream_in *in = usecase->stream.in; |
| if (usecase->type == PCM_CAPTURE && in != NULL && |
| in->source == AUDIO_SOURCE_CAMCORDER && !in->standby) { |
| select_devices(adev, in->usecase); |
| } |
| } |
| } |
| |
| audio_extn_set_parameters(adev, parms); |
| done: |
| str_parms_destroy(parms); |
| pthread_mutex_unlock(&adev->lock); |
| error: |
| ALOGV("%s: exit with code(%d)", __func__, status); |
| return status; |
| } |
| |
| static char* adev_get_parameters(const struct audio_hw_device *dev, |
| const char *keys) |
| { |
| ALOGD("%s:%s", __func__, keys); |
| |
| struct audio_device *adev = (struct audio_device *)dev; |
| struct str_parms *reply = str_parms_create(); |
| struct str_parms *query = str_parms_create_str(keys); |
| char *str; |
| char value[256] = {0}; |
| int ret = 0; |
| |
| if (!query || !reply) { |
| if (reply) { |
| str_parms_destroy(reply); |
| } |
| if (query) { |
| str_parms_destroy(query); |
| } |
| ALOGE("adev_get_parameters: failed to create query or reply"); |
| return NULL; |
| } |
| |
| //handle vr audio getparam |
| |
| ret = str_parms_get_str(query, |
| AUDIO_PARAMETER_KEY_VR_AUDIO_MODE, |
| value, sizeof(value)); |
| |
| if (ret >= 0) { |
| bool vr_audio_enabled = false; |
| pthread_mutex_lock(&adev->lock); |
| vr_audio_enabled = adev->vr_audio_mode_enabled; |
| pthread_mutex_unlock(&adev->lock); |
| |
| ALOGV("getting vr mode to %d", vr_audio_enabled); |
| |
| if (vr_audio_enabled) { |
| str_parms_add_str(reply, AUDIO_PARAMETER_KEY_VR_AUDIO_MODE, |
| "true"); |
| goto exit; |
| } else { |
| str_parms_add_str(reply, AUDIO_PARAMETER_KEY_VR_AUDIO_MODE, |
| "false"); |
| goto exit; |
| } |
| } |
| |
| pthread_mutex_lock(&adev->lock); |
| audio_extn_get_parameters(adev, query, reply); |
| voice_get_parameters(adev, query, reply); |
| audio_extn_a2dp_get_parameters(query, reply); |
| platform_get_parameters(adev->platform, query, reply); |
| audio_extn_ma_get_parameters(adev, query, reply); |
| pthread_mutex_unlock(&adev->lock); |
| |
| exit: |
| str = str_parms_to_str(reply); |
| str_parms_destroy(query); |
| str_parms_destroy(reply); |
| |
| ALOGV("%s: exit: returns - %s", __func__, str); |
| return str; |
| } |
| |
| static int adev_init_check(const struct audio_hw_device *dev __unused) |
| { |
| return 0; |
| } |
| |
| static int adev_set_voice_volume(struct audio_hw_device *dev, float volume) |
| { |
| int ret; |
| struct audio_device *adev = (struct audio_device *)dev; |
| |
| audio_extn_extspk_set_voice_vol(adev->extspk, volume); |
| |
| pthread_mutex_lock(&adev->lock); |
| /* cache volume */ |
| ret = voice_set_volume(adev, volume); |
| pthread_mutex_unlock(&adev->lock); |
| return ret; |
| } |
| |
| static int adev_set_master_volume(struct audio_hw_device *dev __unused, |
| float volume __unused) |
| { |
| return -ENOSYS; |
| } |
| |
| static int adev_get_master_volume(struct audio_hw_device *dev __unused, |
| float *volume __unused) |
| { |
| return -ENOSYS; |
| } |
| |
| static int adev_set_master_mute(struct audio_hw_device *dev __unused, |
| bool muted __unused) |
| { |
| return -ENOSYS; |
| } |
| |
| static int adev_get_master_mute(struct audio_hw_device *dev __unused, |
| bool *muted __unused) |
| { |
| return -ENOSYS; |
| } |
| |
| static int adev_set_mode(struct audio_hw_device *dev, audio_mode_t mode) |
| { |
| struct audio_device *adev = (struct audio_device *)dev; |
| struct listnode *node; |
| struct audio_usecase *usecase = NULL; |
| int ret = 0; |
| |
| pthread_mutex_lock(&adev->lock); |
| if (adev->mode != mode) { |
| ALOGD("%s: mode %d , prev_mode %d \n", __func__, mode , adev->mode); |
| adev->prev_mode = adev->mode; /* prev_mode is kept to handle voip concurrency*/ |
| adev->mode = mode; |
| if (mode == AUDIO_MODE_CALL_SCREEN) { |
| adev->current_call_output = adev->primary_output; |
| voice_start_call(adev); |
| } else if (voice_is_in_call_or_call_screen(adev) && |
| (mode == AUDIO_MODE_NORMAL || |
| (mode == AUDIO_MODE_IN_COMMUNICATION && !voice_is_call_state_active(adev)))) { |
| list_for_each(node, &adev->usecase_list) { |
| usecase = node_to_item(node, struct audio_usecase, list); |
| if (usecase->type == VOICE_CALL) |
| break; |
| } |
| if (usecase && |
| audio_is_usb_out_device(usecase->out_snd_device & AUDIO_DEVICE_OUT_ALL_USB)) { |
| ret = audio_extn_usb_check_and_set_svc_int(usecase, |
| true); |
| if (ret != 0) { |
| /* default service interval was successfully updated, |
| reopen USB backend with new service interval */ |
| check_usecases_codec_backend(adev, |
| usecase, |
| usecase->out_snd_device); |
| } |
| } |
| |
| voice_stop_call(adev); |
| platform_set_gsm_mode(adev->platform, false); |
| adev->current_call_output = NULL; |
| // restore device for other active usecases after stop call |
| list_for_each(node, &adev->usecase_list) { |
| usecase = node_to_item(node, struct audio_usecase, list); |
| select_devices(adev, usecase->id); |
| } |
| } |
| } |
| pthread_mutex_unlock(&adev->lock); |
| return 0; |
| } |
| |
| static int adev_set_mic_mute(struct audio_hw_device *dev, bool state) |
| { |
| int ret; |
| struct audio_device *adev = (struct audio_device *)dev; |
| |
| pthread_mutex_lock(&adev->lock); |
| ALOGD("%s state %d\n", __func__, state); |
| ret = voice_set_mic_mute((struct audio_device *)dev, state); |
| |
| if (adev->ext_hw_plugin) |
| ret = audio_extn_ext_hw_plugin_set_mic_mute(adev->ext_hw_plugin, state); |
| |
| adev->mic_muted = state; |
| pthread_mutex_unlock(&adev->lock); |
| |
| return ret; |
| } |
| |
| static int adev_get_mic_mute(const struct audio_hw_device *dev, bool *state) |
| { |
| *state = voice_get_mic_mute((struct audio_device *)dev); |
| return 0; |
| } |
| |
| static size_t adev_get_input_buffer_size(const struct audio_hw_device *dev __unused, |
| const struct audio_config *config) |
| { |
| int channel_count = audio_channel_count_from_in_mask(config->channel_mask); |
| |
| /* Don't know if USB HIFI in this context so use true to be conservative */ |
| if (check_input_parameters(config->sample_rate, config->format, channel_count, |
| true /*is_usb_hifi */) != 0) |
| return 0; |
| |
| return get_input_buffer_size(config->sample_rate, config->format, channel_count, |
| false /* is_low_latency: since we don't know, be conservative */); |
| } |
| |
| static bool adev_input_allow_hifi_record(struct audio_device *adev, |
| audio_devices_t devices, |
| audio_input_flags_t flags, |
| audio_source_t source) { |
| const bool allowed = true; |
| |
| if (!audio_is_usb_in_device(devices)) |
| return !allowed; |
| |
| switch (flags) { |
| case AUDIO_INPUT_FLAG_NONE: |
| break; |
| case AUDIO_INPUT_FLAG_FAST: // disallow hifi record for FAST as |
| // it affects RTD numbers over USB |
| default: |
| return !allowed; |
| } |
| |
| switch (source) { |
| case AUDIO_SOURCE_DEFAULT: |
| case AUDIO_SOURCE_MIC: |
| case AUDIO_SOURCE_UNPROCESSED: |
| break; |
| default: |
| return !allowed; |
| } |
| |
| switch (adev->mode) { |
| case 0: |
| break; |
| default: |
| return !allowed; |
| } |
| |
| return allowed; |
| } |
| |
| static int adev_update_voice_comm_input_stream(struct stream_in *in, |
| struct audio_config *config) |
| { |
| bool valid_rate = (config->sample_rate == 8000 || |
| config->sample_rate == 16000 || |
| config->sample_rate == 32000 || |
| config->sample_rate == 48000); |
| bool valid_ch = audio_channel_count_from_in_mask(in->channel_mask) == 1; |
| |
| if(!voice_extn_is_compress_voip_supported()) { |
| if (valid_rate && valid_ch) { |
| in->usecase = USECASE_AUDIO_RECORD_VOIP; |
| in->config = default_pcm_config_voip_copp; |
| in->config.period_size = VOIP_IO_BUF_SIZE(in->sample_rate, |
| DEFAULT_VOIP_BUF_DURATION_MS, |
| DEFAULT_VOIP_BIT_DEPTH_BYTE)/2; |
| } else { |
| ALOGW("%s No valid input in voip, use defaults" |
| "sample rate %u, channel mask 0x%X", |
| __func__, config->sample_rate, in->channel_mask); |
| } |
| in->config.rate = config->sample_rate; |
| in->sample_rate = config->sample_rate; |
| } else { |
| //XXX needed for voice_extn_compress_voip_open_input_stream |
| in->config.rate = config->sample_rate; |
| if ((in->dev->mode == AUDIO_MODE_IN_COMMUNICATION || |
| in->source == AUDIO_SOURCE_VOICE_COMMUNICATION || |
| voice_extn_compress_voip_is_active(in->dev)) && |
| (voice_extn_compress_voip_is_format_supported(in->format)) && |
| valid_rate && valid_ch) { |
| voice_extn_compress_voip_open_input_stream(in); |
| // update rate entries to match config from AF |
| in->config.rate = config->sample_rate; |
| in->sample_rate = config->sample_rate; |
| } else { |
| ALOGW("%s compress voip not active, use defaults", __func__); |
| } |
| } |
| return 0; |
| } |
| |
| static int adev_open_input_stream(struct audio_hw_device *dev, |
| audio_io_handle_t handle, |
| audio_devices_t devices, |
| struct audio_config *config, |
| struct audio_stream_in **stream_in, |
| audio_input_flags_t flags, |
| const char *address, |
| audio_source_t source) |
| { |
| struct audio_device *adev = (struct audio_device *)dev; |
| struct stream_in *in; |
| int ret = 0, buffer_size, frame_size; |
| int channel_count = audio_channel_count_from_in_mask(config->channel_mask); |
| bool is_low_latency = false; |
| bool channel_mask_updated = false; |
| bool is_usb_dev = audio_is_usb_in_device(devices); |
| bool may_use_hifi_record = adev_input_allow_hifi_record(adev, |
| devices, |
| flags, |
| source); |
| ALOGV("%s: enter: flags %#x, is_usb_dev %d, may_use_hifi_record %d," |
| " sample_rate %u, channel_mask %#x, format %#x", |
| __func__, flags, is_usb_dev, may_use_hifi_record, |
| config->sample_rate, config->channel_mask, config->format); |
| |
| if (is_usb_dev && (!audio_extn_usb_connected(NULL))) { |
| is_usb_dev = false; |
| devices = AUDIO_DEVICE_IN_BUILTIN_MIC; |
| ALOGW("%s: ignore set device to non existing USB card, use input device(%#x)", |
| __func__, devices); |
| } |
| |
| *stream_in = NULL; |
| |
| if (!(is_usb_dev && may_use_hifi_record)) { |
| if (config->sample_rate == 0) |
| config->sample_rate = 48000; |
| if (config->channel_mask == AUDIO_CHANNEL_NONE) |
| config->channel_mask = AUDIO_CHANNEL_IN_MONO; |
| if (config->format == AUDIO_FORMAT_DEFAULT) |
| config->format = AUDIO_FORMAT_PCM_16_BIT; |
| |
| channel_count = audio_channel_count_from_in_mask(config->channel_mask); |
| |
| if (check_input_parameters(config->sample_rate, config->format, channel_count, |
| false) != 0) |
| return -EINVAL; |
| } |
| |
| in = (struct stream_in *)calloc(1, sizeof(struct stream_in)); |
| |
| if (!in) { |
| ALOGE("failed to allocate input stream"); |
| return -ENOMEM; |
| } |
| |
| ALOGD("%s: enter: sample_rate(%d) channel_mask(%#x) devices(%#x)\ |
| stream_handle(%p) io_handle(%d) source(%d) format %x",__func__, config->sample_rate, |
| config->channel_mask, devices, &in->stream, handle, source, config->format); |
| pthread_mutex_init(&in->lock, (const pthread_mutexattr_t *) NULL); |
| pthread_mutex_init(&in->pre_lock, (const pthread_mutexattr_t *) NULL); |
| |
| in->stream.common.get_sample_rate = in_get_sample_rate; |
| in->stream.common.set_sample_rate = in_set_sample_rate; |
| in->stream.common.get_buffer_size = in_get_buffer_size; |
| in->stream.common.get_channels = in_get_channels; |
| in->stream.common.get_format = in_get_format; |
| in->stream.common.set_format = in_set_format; |
| in->stream.common.standby = in_standby; |
| in->stream.common.dump = in_dump; |
| in->stream.common.set_parameters = in_set_parameters; |
| in->stream.common.get_parameters = in_get_parameters; |
| in->stream.common.add_audio_effect = in_add_audio_effect; |
| in->stream.common.remove_audio_effect = in_remove_audio_effect; |
| in->stream.set_gain = in_set_gain; |
| in->stream.read = in_read; |
| in->stream.get_input_frames_lost = in_get_input_frames_lost; |
| in->stream.get_capture_position = in_get_capture_position; |
| in->stream.get_active_microphones = in_get_active_microphones; |
| in->stream.set_microphone_direction = in_set_microphone_direction; |
| in->stream.set_microphone_field_dimension = in_set_microphone_field_dimension; |
| in->stream.update_sink_metadata = in_update_sink_metadata; |
| |
| list_init(&in->device_list); |
| update_device_list(&in->device_list, devices, address, true); |
| in->source = source; |
| in->dev = adev; |
| in->standby = 1; |
| in->capture_handle = handle; |
| in->flags = flags; |
| in->bit_width = 16; |
| in->af_period_multiplier = 1; |
| in->direction = MIC_DIRECTION_UNSPECIFIED; |
| in->zoom = 0; |
| list_init(&in->aec_list); |
| list_init(&in->ns_list); |
| in->mmap_shared_memory_fd = -1; // not open |
| |
| ALOGV("%s: source %d, config->channel_mask %#x", __func__, source, config->channel_mask); |
| if (source == AUDIO_SOURCE_VOICE_UPLINK || |
| source == AUDIO_SOURCE_VOICE_DOWNLINK) { |
| /* Force channel config requested to mono if incall |
| record is being requested for only uplink/downlink */ |
| if (config->channel_mask != AUDIO_CHANNEL_IN_MONO) { |
| config->channel_mask = AUDIO_CHANNEL_IN_MONO; |
| ret = -EINVAL; |
| goto err_open; |
| } |
| } |
| |
| if (is_usb_dev && may_use_hifi_record) { |
| /* HiFi record selects an appropriate format, channel, rate combo |
| depending on sink capabilities*/ |
| ret = read_usb_sup_params_and_compare(false /*is_playback*/, |
| &config->format, |
| &in->supported_formats[0], |
| MAX_SUPPORTED_FORMATS, |
| &config->channel_mask, |
| &in->supported_channel_masks[0], |
| MAX_SUPPORTED_CHANNEL_MASKS, |
| &config->sample_rate, |
| &in->supported_sample_rates[0], |
| MAX_SUPPORTED_SAMPLE_RATES); |
| if (ret != 0) { |
| ret = -EINVAL; |
| goto err_open; |
| } |
| channel_count = audio_channel_count_from_in_mask(config->channel_mask); |
| } else if (config->format == AUDIO_FORMAT_DEFAULT) { |
| config->format = AUDIO_FORMAT_PCM_16_BIT; |
| } else if (property_get_bool("vendor.audio.capture.pcm.32bit.enable", false) |
| && config->format == AUDIO_FORMAT_PCM_32_BIT) { |
| in->config.format = PCM_FORMAT_S32_LE; |
| in->bit_width = 32; |
| } else if ((config->format == AUDIO_FORMAT_PCM_FLOAT) || |
| (config->format == AUDIO_FORMAT_PCM_32_BIT) || |
| (config->format == AUDIO_FORMAT_PCM_24_BIT_PACKED) || |
| (config->format == AUDIO_FORMAT_PCM_8_24_BIT)) { |
| bool ret_error = false; |
| in->bit_width = 24; |
| /* 24 bit is restricted to UNPROCESSED source only,also format supported |
| from HAL is 24_packed and 8_24 |
| *> In case of UNPROCESSED source, for 24 bit, if format requested is other than |
| 24_packed return error indicating supported format is 24_packed |
| *> In case of any other source requesting 24 bit or float return error |
| indicating format supported is 16 bit only. |
| |
| on error flinger will retry with supported format passed |
| */ |
| if ((source != AUDIO_SOURCE_UNPROCESSED) && |
| (source != AUDIO_SOURCE_CAMCORDER)) { |
| config->format = AUDIO_FORMAT_PCM_16_BIT; |
| if (config->sample_rate > 48000) |
| config->sample_rate = 48000; |
| ret_error = true; |
| } else if (!(config->format == AUDIO_FORMAT_PCM_24_BIT_PACKED || |
| config->format == AUDIO_FORMAT_PCM_8_24_BIT)) { |
| config->format = AUDIO_FORMAT_PCM_24_BIT_PACKED; |
| ret_error = true; |
| } |
| |
| if (ret_error) { |
| ret = -EINVAL; |
| goto err_open; |
| } |
| } |
| |
| in->channel_mask = config->channel_mask; |
| in->format = config->format; |
| |
| in->usecase = USECASE_AUDIO_RECORD; |
| |
| /* validate bus device address */ |
| if (compare_device_type(&in->device_list, AUDIO_DEVICE_IN_BUS)) { |
| /* extract car audio stream index */ |
| in->car_audio_stream = |
| audio_extn_auto_hal_get_car_audio_stream_from_address(address); |
| if (in->car_audio_stream < 0) { |
| ALOGE("%s: invalid car audio stream %x", |
| __func__, in->car_audio_stream); |
| ret = -EINVAL; |
| goto err_open; |
| } |
| ALOGV("%s: car_audio_stream 0x%x", __func__, in->car_audio_stream); |
| ret = audio_extn_auto_hal_open_input_stream(in); |
| if (ret) { |
| ALOGE("%s: Failed to open input stream for bus device", __func__); |
| ret = -EINVAL; |
| goto err_open; |
| } |
| } |
| |
| /* reassign use case for echo reference stream on automotive platforms */ |
| if (in->source == AUDIO_SOURCE_ECHO_REFERENCE) { |
| ret = audio_extn_auto_hal_open_echo_reference_stream(in); |
| } |
| |
| if (in->source == AUDIO_SOURCE_FM_TUNER) { |
| if(!get_usecase_from_list(adev, USECASE_AUDIO_RECORD_FM_VIRTUAL)) |
| in->usecase = USECASE_AUDIO_RECORD_FM_VIRTUAL; |
| else { |
| ret = -EINVAL; |
| goto err_open; |
| } |
| } |
| |
| if (config->sample_rate == LOW_LATENCY_CAPTURE_SAMPLE_RATE && |
| (flags & AUDIO_INPUT_FLAG_TIMESTAMP) == 0 && |
| (flags & AUDIO_INPUT_FLAG_COMPRESS) == 0 && |
| (flags & AUDIO_INPUT_FLAG_FAST) != 0) { |
| is_low_latency = true; |
| #if LOW_LATENCY_CAPTURE_USE_CASE |
| if ((flags & AUDIO_INPUT_FLAG_VOIP_TX) != 0) |
| in->usecase = USECASE_AUDIO_RECORD_VOIP_LOW_LATENCY; |
| else |
| in->usecase = USECASE_AUDIO_RECORD_LOW_LATENCY; |
| #endif |
| in->realtime = may_use_noirq_mode(adev, in->usecase, in->flags); |
| if (!in->realtime) { |
| in->config = pcm_config_audio_capture; |
| frame_size = audio_stream_in_frame_size(&in->stream); |
| buffer_size = get_input_buffer_size(config->sample_rate, |
| config->format, |
| channel_count, |
| is_low_latency); |
| in->config.period_size = buffer_size / frame_size; |
| in->config.rate = config->sample_rate; |
| in->af_period_multiplier = 1; |
| } else { |
| // period size is left untouched for rt mode playback |
| in->config = pcm_config_audio_capture_rt; |
| in->af_period_multiplier = af_period_multiplier; |
| } |
| } |
| |
| if ((config->sample_rate == LOW_LATENCY_CAPTURE_SAMPLE_RATE) && |
| ((in->flags & AUDIO_INPUT_FLAG_MMAP_NOIRQ) != 0)) { |
| in->realtime = 0; |
| in->usecase = USECASE_AUDIO_RECORD_MMAP; |
| in->config = pcm_config_mmap_capture; |
| in->config.format = pcm_format_from_audio_format(config->format); |
| in->stream.start = in_start; |
| in->stream.stop = in_stop; |
| in->stream.create_mmap_buffer = in_create_mmap_buffer; |
| in->stream.get_mmap_position = in_get_mmap_position; |
| ALOGV("%s: USECASE_AUDIO_RECORD_MMAP", __func__); |
| } else if (is_usb_dev && may_use_hifi_record) { |
| in->usecase = USECASE_AUDIO_RECORD_HIFI; |
| in->config = pcm_config_audio_capture; |
| frame_size = audio_stream_in_frame_size(&in->stream); |
| buffer_size = get_input_buffer_size(config->sample_rate, |
| config->format, |
| channel_count, |
| false /*is_low_latency*/); |
| in->config.period_size = buffer_size / frame_size; |
| in->config.rate = config->sample_rate; |
| in->config.format = pcm_format_from_audio_format(config->format); |
| switch (config->format) { |
| case AUDIO_FORMAT_PCM_32_BIT: |
| in->bit_width = 32; |
| break; |
| case AUDIO_FORMAT_PCM_24_BIT_PACKED: |
| case AUDIO_FORMAT_PCM_8_24_BIT: |
| in->bit_width = 24; |
| break; |
| default: |
| in->bit_width = 16; |
| } |
| } else if (is_single_device_type_equal(&in->device_list, |
| AUDIO_DEVICE_IN_TELEPHONY_RX) || |
| is_single_device_type_equal(&in->device_list, |
| AUDIO_DEVICE_IN_PROXY)) { |
| if (config->sample_rate == 0) |
| config->sample_rate = AFE_PROXY_SAMPLING_RATE; |
| if (config->sample_rate != 48000 && config->sample_rate != 16000 && |
| config->sample_rate != 8000) { |
| config->sample_rate = AFE_PROXY_SAMPLING_RATE; |
| ret = -EINVAL; |
| goto err_open; |
| } |
| if (config->format == AUDIO_FORMAT_DEFAULT) |
| config->format = AUDIO_FORMAT_PCM_16_BIT; |
| if (config->format != AUDIO_FORMAT_PCM_16_BIT) { |
| config->format = AUDIO_FORMAT_PCM_16_BIT; |
| ret = -EINVAL; |
| goto err_open; |
| } |
| |
| in->usecase = USECASE_AUDIO_RECORD_AFE_PROXY; |
| if (adev->ha_proxy_enable && |
| is_single_device_type_equal(&in->device_list, |
| AUDIO_DEVICE_IN_TELEPHONY_RX)) |
| in->usecase = USECASE_AUDIO_RECORD_AFE_PROXY2; |
| in->config = pcm_config_afe_proxy_record; |
| in->config.rate = config->sample_rate; |
| in->af_period_multiplier = 1; |
| } else if (in->source == AUDIO_SOURCE_VOICE_COMMUNICATION && |
| (!voice_extn_is_compress_voip_supported()) && |
| in->flags & AUDIO_INPUT_FLAG_VOIP_TX && |
| (config->sample_rate == 8000 || |
| config->sample_rate == 16000 || |
| config->sample_rate == 32000 || |
| config->sample_rate == 48000) && |
| channel_count == 1) { |
| in->usecase = USECASE_AUDIO_RECORD_VOIP; |
| in->config = pcm_config_audio_capture; |
| frame_size = audio_stream_in_frame_size(&in->stream); |
| buffer_size = get_stream_buffer_size(VOIP_CAPTURE_PERIOD_DURATION_MSEC, |
| config->sample_rate, |
| config->format, |
| channel_count, false /*is_low_latency*/); |
| in->config.period_size = buffer_size / frame_size; |
| in->config.period_count = VOIP_CAPTURE_PERIOD_COUNT; |
| in->config.rate = config->sample_rate; |
| in->af_period_multiplier = 1; |
| } else if (in->realtime) { |
| in->config = pcm_config_audio_capture_rt; |
| in->config.format = pcm_format_from_audio_format(config->format); |
| in->af_period_multiplier = af_period_multiplier; |
| } else { |
| int ret_val; |
| pthread_mutex_lock(&adev->lock); |
| ret_val = audio_extn_check_and_set_multichannel_usecase(adev, |
| in, config, &channel_mask_updated); |
| pthread_mutex_unlock(&adev->lock); |
| |
| if (!ret_val) { |
| if (channel_mask_updated == true) { |
| ALOGD("%s: return error to retry with updated channel mask (%#x)", |
| __func__, config->channel_mask); |
| ret = -EINVAL; |
| goto err_open; |
| } |
| ALOGD("%s: created multi-channel session succesfully",__func__); |
| } else if (audio_extn_compr_cap_enabled() && |
| audio_extn_compr_cap_format_supported(config->format) && |
| (in->dev->mode != AUDIO_MODE_IN_COMMUNICATION)) { |
| audio_extn_compr_cap_init(in); |
| } else if (audio_extn_cin_applicable_stream(in)) { |
| ret = audio_extn_cin_configure_input_stream(in, config); |
| if (ret) |
| goto err_open; |
| } else { |
| in->config = pcm_config_audio_capture; |
| in->config.rate = config->sample_rate; |
| in->config.format = pcm_format_from_audio_format(config->format); |
| in->format = config->format; |
| frame_size = audio_stream_in_frame_size(&in->stream); |
| buffer_size = get_input_buffer_size(config->sample_rate, |
| config->format, |
| channel_count, |
| is_low_latency); |
| /* prevent division-by-zero */ |
| if (frame_size == 0) { |
| ALOGE("%s: Error frame_size==0", __func__); |
| ret = -EINVAL; |
| goto err_open; |
| } |
| |
| in->config.period_size = buffer_size / frame_size; |
| in->af_period_multiplier = 1; |
| |
| if (in->source == AUDIO_SOURCE_VOICE_COMMUNICATION) { |
| /* optionally use VOIP usecase depending on config(s) */ |
| ret = adev_update_voice_comm_input_stream(in, config); |
| } |
| |
| if (ret) { |
| ALOGE("%s AUDIO_SOURCE_VOICE_COMMUNICATION invalid args", __func__); |
| goto err_open; |
| } |
| } |
| |
| /* assign concurrent capture usecase if record has to caried out from |
| * actual hardware input source */ |
| if (audio_extn_is_concurrent_capture_enabled() && |
| !audio_is_virtual_input_source(in->source)) { |
| /* Acquire lock to avoid two concurrent use cases initialized to |
| same pcm record use case */ |
| |
| if (in->usecase == USECASE_AUDIO_RECORD) { |
| pthread_mutex_lock(&adev->lock); |
| if (!(adev->pcm_record_uc_state)) { |
| ALOGV("%s: using USECASE_AUDIO_RECORD",__func__); |
| adev->pcm_record_uc_state = 1; |
| pthread_mutex_unlock(&adev->lock); |
| } else { |
| pthread_mutex_unlock(&adev->lock); |
| /* Assign compress record use case for second record */ |
| in->usecase = USECASE_AUDIO_RECORD_COMPRESS2; |
| in->flags |= AUDIO_INPUT_FLAG_COMPRESS; |
| ALOGV("%s: overriding usecase with USECASE_AUDIO_RECORD_COMPRESS2 and appending compress flag", __func__); |
| if (audio_extn_cin_applicable_stream(in)) { |
| in->sample_rate = config->sample_rate; |
| ret = audio_extn_cin_configure_input_stream(in, config); |
| if (ret) |
| goto err_open; |
| } |
| } |
| } |
| } |
| } |
| if (audio_extn_ssr_get_stream() != in) |
| in->config.channels = channel_count; |
| |
| in->sample_rate = in->config.rate; |
| |
| audio_extn_utils_update_stream_input_app_type_cfg(adev->platform, |
| &adev->streams_input_cfg_list, |
| &in->device_list, flags, in->format, |
| in->sample_rate, in->bit_width, |
| in->profile, &in->app_type_cfg); |
| register_format(in->format, in->supported_formats); |
| register_channel_mask(in->channel_mask, in->supported_channel_masks); |
| register_sample_rate(in->sample_rate, in->supported_sample_rates); |
| |
| #ifndef LINUX_ENABLED |
| in->error_log = error_log_create( |
| ERROR_LOG_ENTRIES, |
| 1000000000 /* aggregate consecutive identical errors within one second */); |
| #endif |
| |
| /* This stream could be for sound trigger lab, |
| get sound trigger pcm if present */ |
| audio_extn_sound_trigger_check_and_get_session(in); |
| |
| lock_input_stream(in); |
| audio_extn_snd_mon_register_listener(in, in_snd_mon_cb); |
| pthread_mutex_lock(&adev->lock); |
| in->card_status = adev->card_status; |
| pthread_mutex_unlock(&adev->lock); |
| pthread_mutex_unlock(&in->lock); |
| |
| stream_app_type_cfg_init(&in->app_type_cfg); |
| |
| *stream_in = &in->stream; |
| |
| ret = io_streams_map_insert(adev, &in->stream.common, |
| handle, AUDIO_PATCH_HANDLE_NONE); |
| if (ret != 0) |
| goto err_open; |
| |
| in->in_ctxt.input = in; |
| |
| pthread_mutex_lock(&adev->lock); |
| list_add_tail(&adev->active_inputs_list, &in->in_ctxt.list); |
| pthread_mutex_unlock(&adev->lock); |
| |
| ALOGV("%s: exit", __func__); |
| return ret; |
| |
| err_open: |
| if (in->usecase == USECASE_AUDIO_RECORD) { |
| pthread_mutex_lock(&adev->lock); |
| adev->pcm_record_uc_state = 0; |
| pthread_mutex_unlock(&adev->lock); |
| } |
| free(in); |
| *stream_in = NULL; |
| return ret; |
| } |
| |
| static void adev_close_input_stream(struct audio_hw_device *dev, |
| struct audio_stream_in *stream) |
| { |
| int ret; |
| struct stream_in *in = (struct stream_in *)stream; |
| struct audio_device *adev = (struct audio_device *)dev; |
| |
| ALOGD("%s: enter:stream_handle(%p)",__func__, in); |
| |
| if (in == NULL) { |
| ALOGE("%s: audio_stream_in ptr is NULL", __func__); |
| return; |
| } |
| io_streams_map_remove(adev, in->capture_handle); |
| |
| // remove out_ctxt early to prevent the stream |
| // being opened in a race condition |
| pthread_mutex_lock(&adev->lock); |
| list_remove(&in->in_ctxt.list); |
| pthread_mutex_unlock(&adev->lock); |
| |
| /* must deregister from sndmonitor first to prevent races |
| * between the callback and close_stream |
| */ |
| audio_extn_snd_mon_unregister_listener(stream); |
| |
| /* Disable echo reference if there are no active input, hfp call |
| * and sound trigger while closing input stream |
| */ |
| if (adev_get_active_input(adev) == NULL && |
| !audio_extn_hfp_is_active(adev) && |
| !audio_extn_sound_trigger_check_ec_ref_enable()) { |
| struct listnode out_devices; |
| list_init(&out_devices); |
| platform_set_echo_reference(adev, false, &out_devices); |
| } else |
| audio_extn_sound_trigger_update_ec_ref_status(false); |
| |
| #ifndef LINUX_ENABLED |
| error_log_destroy(in->error_log); |
| in->error_log = NULL; |
| #endif |
| |
| if (in->usecase == USECASE_COMPRESS_VOIP_CALL) { |
| pthread_mutex_lock(&adev->lock); |
| ret = voice_extn_compress_voip_close_input_stream(&stream->common); |
| pthread_mutex_unlock(&adev->lock); |
| if (ret != 0) |
| ALOGE("%s: Compress voip input cannot be closed, error:%d", |
| __func__, ret); |
| } else |
| in_standby(&stream->common); |
| |
| pthread_mutex_destroy(&in->lock); |
| pthread_mutex_destroy(&in->pre_lock); |
| |
| pthread_mutex_lock(&adev->lock); |
| if (in->usecase == USECASE_AUDIO_RECORD) { |
| adev->pcm_record_uc_state = 0; |
| } |
| |
| if (in->source == AUDIO_SOURCE_VOICE_COMMUNICATION) { |
| adev->enable_voicerx = false; |
| } |
| |
| if (audio_extn_ssr_get_stream() == in) { |
| audio_extn_ssr_deinit(); |
| } |
| |
| if (audio_extn_ffv_get_stream() == in) { |
| audio_extn_ffv_stream_deinit(); |
| } |
| |
| if (audio_extn_compr_cap_enabled() && |
| audio_extn_compr_cap_format_supported(in->config.format)) |
| audio_extn_compr_cap_deinit(); |
| |
| if (audio_extn_cin_attached_usecase(in)) |
| audio_extn_cin_free_input_stream_resources(in); |
| |
| if (in->is_st_session) { |
| ALOGV("%s: sound trigger pcm stop lab", __func__); |
| audio_extn_sound_trigger_stop_lab(in); |
| } |
| free(stream); |
| pthread_mutex_unlock(&adev->lock); |
| return; |
| } |
| |
| /* verifies input and output devices and their capabilities. |
| * |
| * This verification is required when enabling extended bit-depth or |
| * sampling rates, as not all qcom products support it. |
| * |
| * Suitable for calling only on initialization such as adev_open(). |
| * It fills the audio_device use_case_table[] array. |
| * |
| * Has a side-effect that it needs to configure audio routing / devices |
| * in order to power up the devices and read the device parameters. |
| * It does not acquire any hw device lock. Should restore the devices |
| * back to "normal state" upon completion. |
| */ |
| static int adev_verify_devices(struct audio_device *adev) |
| { |
| /* enumeration is a bit difficult because one really wants to pull |
| * the use_case, device id, etc from the hidden pcm_device_table[]. |
| * In this case there are the following use cases and device ids. |
| * |
| * [USECASE_AUDIO_PLAYBACK_DEEP_BUFFER] = {0, 0}, |
| * [USECASE_AUDIO_PLAYBACK_LOW_LATENCY] = {15, 15}, |
| * [USECASE_AUDIO_PLAYBACK_HIFI] = {1, 1}, |
| * [USECASE_AUDIO_PLAYBACK_OFFLOAD] = {9, 9}, |
| * [USECASE_AUDIO_RECORD] = {0, 0}, |
| * [USECASE_AUDIO_RECORD_LOW_LATENCY] = {15, 15}, |
| * [USECASE_VOICE_CALL] = {2, 2}, |
| * |
| * USECASE_AUDIO_PLAYBACK_OFFLOAD, USECASE_AUDIO_PLAYBACK_HIFI omitted. |
| * USECASE_VOICE_CALL omitted, but possible for either input or output. |
| */ |
| |
| /* should be the usecases enabled in adev_open_input_stream() */ |
| static const int test_in_usecases[] = { |
| USECASE_AUDIO_RECORD, |
| USECASE_AUDIO_RECORD_LOW_LATENCY, /* does not appear to be used */ |
| }; |
| /* should be the usecases enabled in adev_open_output_stream()*/ |
| static const int test_out_usecases[] = { |
| USECASE_AUDIO_PLAYBACK_DEEP_BUFFER, |
| USECASE_AUDIO_PLAYBACK_LOW_LATENCY, |
| }; |
| static const usecase_type_t usecase_type_by_dir[] = { |
| PCM_PLAYBACK, |
| PCM_CAPTURE, |
| }; |
| static const unsigned flags_by_dir[] = { |
| PCM_OUT, |
| PCM_IN, |
| }; |
| |
| size_t i; |
| unsigned dir; |
| const unsigned card_id = adev->snd_card; |
| |
| for (dir = 0; dir < 2; ++dir) { |
| const usecase_type_t usecase_type = usecase_type_by_dir[dir]; |
| const unsigned flags_dir = flags_by_dir[dir]; |
| const size_t testsize = |
| dir ? ARRAY_SIZE(test_in_usecases) : ARRAY_SIZE(test_out_usecases); |
| const int *testcases = |
| dir ? test_in_usecases : test_out_usecases; |
| const audio_devices_t audio_device = |
| dir ? AUDIO_DEVICE_IN_BUILTIN_MIC : AUDIO_DEVICE_OUT_SPEAKER; |
| |
| for (i = 0; i < testsize; ++i) { |
| const audio_usecase_t audio_usecase = testcases[i]; |
| int device_id; |
| struct pcm_params **pparams; |
| struct stream_out out; |
| struct stream_in in; |
| struct audio_usecase uc_info; |
| int retval; |
| |
| pparams = &adev->use_case_table[audio_usecase]; |
| pcm_params_free(*pparams); /* can accept null input */ |
| *pparams = NULL; |
| |
| /* find the device ID for the use case (signed, for error) */ |
| device_id = platform_get_pcm_device_id(audio_usecase, usecase_type); |
| if (device_id < 0) |
| continue; |
| |
| /* prepare structures for device probing */ |
| memset(&uc_info, 0, sizeof(uc_info)); |
| uc_info.id = audio_usecase; |
| uc_info.type = usecase_type; |
| list_init(&uc_info.device_list); |
| if (dir) { |
| memset(&in, 0, sizeof(in)); |
| list_init(&in.device_list); |
| update_device_list(&in.device_list, audio_device, "", true); |
| in.source = AUDIO_SOURCE_VOICE_COMMUNICATION; |
| uc_info.stream.in = ∈ |
| } |
| memset(&out, 0, sizeof(out)); |
| list_init(&out.device_list); |
| update_device_list(&out.device_list, audio_device, "", true); |
| uc_info.stream.out = &out; |
| update_device_list(&uc_info.device_list, audio_device, "", true); |
| uc_info.in_snd_device = SND_DEVICE_NONE; |
| uc_info.out_snd_device = SND_DEVICE_NONE; |
| list_add_tail(&adev->usecase_list, &uc_info.list); |
| |
| /* select device - similar to start_(in/out)put_stream() */ |
| retval = select_devices(adev, audio_usecase); |
| if (retval >= 0) { |
| *pparams = pcm_params_get(card_id, device_id, flags_dir); |
| #if LOG_NDEBUG == 0 |
| char info[512]; /* for possible debug info */ |
| if (*pparams) { |
| ALOGV("%s: (%s) card %d device %d", __func__, |
| dir ? "input" : "output", card_id, device_id); |
| pcm_params_to_string(*pparams, info, ARRAY_SIZE(info)); |
| } else { |
| ALOGV("%s: cannot locate card %d device %d", __func__, card_id, device_id); |
| } |
| #endif |
| } |
| |
| /* deselect device - similar to stop_(in/out)put_stream() */ |
| /* 1. Get and set stream specific mixer controls */ |
| retval = disable_audio_route(adev, &uc_info); |
| /* 2. Disable the rx device */ |
| retval = disable_snd_device(adev, |
| dir ? uc_info.in_snd_device : uc_info.out_snd_device); |
| list_remove(&uc_info.list); |
| } |
| } |
| return 0; |
| } |
| |
| int update_patch(unsigned int num_sources, |
| const struct audio_port_config *sources, |
| unsigned int num_sinks, |
| const struct audio_port_config *sinks, |
| audio_patch_handle_t handle, |
| struct audio_patch_info *p_info, |
| patch_type_t patch_type, bool new_patch) |
| { |
| ALOGV("%s: enter", __func__); |
| |
| if (p_info == NULL) { |
| ALOGE("%s: Invalid patch pointer", __func__); |
| return -EINVAL; |
| } |
| |
| if (new_patch) { |
| p_info->patch = (struct audio_patch *) calloc(1, sizeof(struct audio_patch)); |
| if (p_info->patch == NULL) { |
| ALOGE("%s: Could not allocate patch", __func__); |
| return -ENOMEM; |
| } |
| } |
| |
| p_info->patch->id = handle; |
| p_info->patch->num_sources = num_sources; |
| p_info->patch->num_sinks = num_sinks; |
| |
| for (int i = 0; i < num_sources; i++) |
| p_info->patch->sources[i] = sources[i]; |
| for (int i = 0; i < num_sinks; i++) |
| p_info->patch->sinks[i] = sinks[i]; |
| |
| p_info->patch_type = patch_type; |
| return 0; |
| } |
| |
| audio_patch_handle_t generate_patch_handle() |
| { |
| static audio_patch_handle_t patch_handle = AUDIO_PATCH_HANDLE_NONE; |
| if (++patch_handle < 0) |
| patch_handle = AUDIO_PATCH_HANDLE_NONE + 1; |
| return patch_handle; |
| } |
| |
| int adev_create_audio_patch(struct audio_hw_device *dev, |
| unsigned int num_sources, |
| const struct audio_port_config *sources, |
| unsigned int num_sinks, |
| const struct audio_port_config *sinks, |
| audio_patch_handle_t *handle) |
| { |
| int ret = 0; |
| struct audio_device *adev = (struct audio_device *)dev; |
| struct audio_patch_info *p_info = NULL; |
| patch_type_t patch_type = PATCH_NONE; |
| audio_io_handle_t io_handle = AUDIO_IO_HANDLE_NONE; |
| audio_source_t input_source = AUDIO_SOURCE_DEFAULT; |
| struct audio_stream_info *s_info = NULL; |
| struct audio_stream *stream = NULL; |
| struct listnode devices; |
| audio_devices_t device_type = AUDIO_DEVICE_NONE; |
| bool new_patch = false; |
| char addr[AUDIO_DEVICE_MAX_ADDRESS_LEN]; |
| |
| ALOGD("%s: enter: num sources %d, num_sinks %d, handle %d", __func__, |
| num_sources, num_sinks, *handle); |
| |
| if (num_sources == 0 || num_sources > AUDIO_PATCH_PORTS_MAX || |
| num_sinks == 0 || num_sinks > AUDIO_PATCH_PORTS_MAX) { |
| ALOGE("%s: Invalid patch arguments", __func__); |
| ret = -EINVAL; |
| goto done; |
| } |
| |
| if (num_sources > 1) { |
| ALOGE("%s: Multiple sources are not supported", __func__); |
| ret = -EINVAL; |
| goto done; |
| } |
| |
| if (sources == NULL || sinks == NULL) { |
| ALOGE("%s: Invalid sources or sinks port config", __func__); |
| ret = -EINVAL; |
| goto done; |
| } |
| |
| ALOGV("%s: source role %d, source type %d", __func__, |
| sources[0].type, sources[0].role); |
| list_init(&devices); |
| |
| // Populate source/sink information and fetch stream info |
| switch (sources[0].type) { |
| case AUDIO_PORT_TYPE_DEVICE: // Patch for audio capture or loopback |
| device_type = sources[0].ext.device.type; |
| strlcpy(&addr[0], &sources[0].ext.device.address[0], AUDIO_DEVICE_MAX_ADDRESS_LEN); |
| update_device_list(&devices, device_type, &addr[0], true); |
| if (sinks[0].type == AUDIO_PORT_TYPE_MIX) { |
| patch_type = PATCH_CAPTURE; |
| io_handle = sinks[0].ext.mix.handle; |
| input_source = sinks[0].ext.mix.usecase.source; |
| ALOGD("%s: Capture patch from device %x to mix %d", |
| __func__, device_type, io_handle); |
| } else { |
| // Device to device patch is not implemented. |
| // This space will need changes if audio HAL |
| // handles device to device patches in the future. |
| patch_type = PATCH_DEVICE_LOOPBACK; |
| } |
| break; |
| case AUDIO_PORT_TYPE_MIX: // Patch for audio playback |
| io_handle = sources[0].ext.mix.handle; |
| for (int i = 0; i < num_sinks; i++) { |
| device_type = sinks[i].ext.device.type; |
| strlcpy(&addr[0], &sinks[i].ext.device.address[0], AUDIO_DEVICE_MAX_ADDRESS_LEN); |
| update_device_list(&devices, device_type, &addr[0], true); |
| } |
| patch_type = PATCH_PLAYBACK; |
| ALOGD("%s: Playback patch from mix handle %d to device %x", |
| __func__, io_handle, get_device_types(&devices)); |
| break; |
| case AUDIO_PORT_TYPE_SESSION: |
| case AUDIO_PORT_TYPE_NONE: |
| ALOGE("%s: Unsupported source type %d", __func__, sources[0].type); |
| ret = -EINVAL; |
| goto done; |
| } |
| |
| pthread_mutex_lock(&adev->lock); |
| |
| // Generate patch info and update patch |
| if (*handle == AUDIO_PATCH_HANDLE_NONE) { |
| *handle = generate_patch_handle(); |
| p_info = (struct audio_patch_info *) |
| calloc(1, sizeof(struct audio_patch_info)); |
| if (p_info == NULL) { |
| ALOGE("%s: Failed to allocate memory", __func__); |
| pthread_mutex_unlock(&adev->lock); |
| ret = -ENOMEM; |
| goto done; |
| } |
| new_patch = true; |
| } else { |
| p_info = fetch_patch_info_l(adev, *handle); |
| if (p_info == NULL) { |
| ALOGE("%s: Unable to fetch patch for received patch handle %d", |
| __func__, *handle); |
| pthread_mutex_unlock(&adev->lock); |
| ret = -EINVAL; |
| goto done; |
| } |
| } |
| update_patch(num_sources, sources, num_sinks, sinks, |
| *handle, p_info, patch_type, new_patch); |
| |
| // Fetch stream info of associated mix for playback or capture patches |
| if (p_info->patch_type == PATCH_PLAYBACK || |
| p_info->patch_type == PATCH_CAPTURE) { |
| s_info = hashmapGet(adev->io_streams_map, (void *) (intptr_t) io_handle); |
| if (s_info == NULL) { |
| ALOGE("%s: Failed to obtain stream info", __func__); |
| if (new_patch) |
| free(p_info); |
| pthread_mutex_unlock(&adev->lock); |
| ret = -EINVAL; |
| goto done; |
| } |
| ALOGV("%s: Fetched stream info with io_handle %d", __func__, io_handle); |
| s_info->patch_handle = *handle; |
| stream = s_info->stream; |
| } |
| pthread_mutex_unlock(&adev->lock); |
| |
| // Update routing for stream |
| if (stream != NULL) { |
| if (p_info->patch_type == PATCH_PLAYBACK) |
| ret = route_output_stream((struct stream_out *) stream, &devices); |
| else if (p_info->patch_type == PATCH_CAPTURE) |
| ret = route_input_stream((struct stream_in *) stream, &devices, input_source); |
| if (ret < 0) { |
| pthread_mutex_lock(&adev->lock); |
| s_info->patch_handle = AUDIO_PATCH_HANDLE_NONE; |
| if (new_patch) |
| free(p_info); |
| pthread_mutex_unlock(&adev->lock); |
| ALOGE("%s: Stream routing failed for io_handle %d", __func__, io_handle); |
| goto done; |
| } |
| } |
| |
| // Add new patch to patch map |
| if (!ret && new_patch) { |
| pthread_mutex_lock(&adev->lock); |
| hashmapPut(adev->patch_map, (void *) (intptr_t) *handle, (void *) p_info); |
| ALOGD("%s: Added a new patch with handle %d", __func__, *handle); |
| pthread_mutex_unlock(&adev->lock); |
| } |
| |
| done: |
| audio_extn_hw_loopback_create_audio_patch(dev, |
| num_sources, |
| sources, |
| num_sinks, |
| sinks, |
| handle); |
| audio_extn_auto_hal_create_audio_patch(dev, |
| num_sources, |
| sources, |
| num_sinks, |
| sinks, |
| handle); |
| return ret; |
| } |
| |
| int adev_release_audio_patch(struct audio_hw_device *dev, |
| audio_patch_handle_t handle) |
| { |
| struct audio_device *adev = (struct audio_device *) dev; |
| int ret = 0; |
| audio_source_t input_source = AUDIO_SOURCE_DEFAULT; |
| struct audio_stream *stream = NULL; |
| |
| if (handle == AUDIO_PATCH_HANDLE_NONE) { |
| ALOGE("%s: Invalid patch handle %d", __func__, handle); |
| ret = -EINVAL; |
| goto done; |
| } |
| |
| ALOGD("%s: Remove patch with handle %d", __func__, handle); |
| pthread_mutex_lock(&adev->lock); |
| struct audio_patch_info *p_info = fetch_patch_info_l(adev, handle); |
| if (p_info == NULL) { |
| ALOGE("%s: Patch info not found with handle %d", __func__, handle); |
| pthread_mutex_unlock(&adev->lock); |
| ret = -EINVAL; |
| goto done; |
| } |
| struct audio_patch *patch = p_info->patch; |
| if (patch == NULL) { |
| ALOGE("%s: Patch not found for handle %d", __func__, handle); |
| pthread_mutex_unlock(&adev->lock); |
| ret = -EINVAL; |
| goto done; |
| } |
| audio_io_handle_t io_handle = AUDIO_IO_HANDLE_NONE; |
| switch (patch->sources[0].type) { |
| case AUDIO_PORT_TYPE_MIX: |
| io_handle = patch->sources[0].ext.mix.handle; |
| break; |
| case AUDIO_PORT_TYPE_DEVICE: |
| if (p_info->patch_type == PATCH_CAPTURE) |
| io_handle = patch->sinks[0].ext.mix.handle; |
| break; |
| case AUDIO_PORT_TYPE_SESSION: |
| case AUDIO_PORT_TYPE_NONE: |
| pthread_mutex_unlock(&adev->lock); |
| ret = -EINVAL; |
| goto done; |
| } |
| |
| // Remove patch and reset patch handle in stream info |
| patch_type_t patch_type = p_info->patch_type; |
| patch_map_remove_l(adev, handle); |
| if (patch_type == PATCH_PLAYBACK || |
| patch_type == PATCH_CAPTURE) { |
| struct audio_stream_info *s_info = |
| hashmapGet(adev->io_streams_map, (void *) (intptr_t) io_handle); |
| if (s_info == NULL) { |
| ALOGE("%s: stream for io_handle %d is not available", __func__, io_handle); |
| pthread_mutex_unlock(&adev->lock); |
| goto done; |
| } |
| s_info->patch_handle = AUDIO_PATCH_HANDLE_NONE; |
| stream = s_info->stream; |
| } |
| pthread_mutex_unlock(&adev->lock); |
| |
| if (stream != NULL) { |
| struct listnode devices; |
| list_init(&devices); |
| if (patch_type == PATCH_PLAYBACK) |
| ret = route_output_stream((struct stream_out *) stream, &devices); |
| else if (patch_type == PATCH_CAPTURE) |
| ret = route_input_stream((struct stream_in *) stream, &devices, input_source); |
| } |
| |
| if (ret < 0) |
| ALOGW("%s: Stream routing failed for io_handle %d", __func__, io_handle); |
| |
| done: |
| audio_extn_hw_loopback_release_audio_patch(dev, handle); |
| audio_extn_auto_hal_release_audio_patch(dev, handle); |
| |
| ALOGV("%s: Successfully released patch %d", __func__, handle); |
| return ret; |
| } |
| |
| int adev_get_audio_port(struct audio_hw_device *dev, struct audio_port *config) |
| { |
| int ret = 0; |
| |
| ret = audio_extn_hw_loopback_get_audio_port(dev, config); |
| ret |= audio_extn_auto_hal_get_audio_port(dev, config); |
| return ret; |
| } |
| |
| int adev_set_audio_port_config(struct audio_hw_device *dev, |
| const struct audio_port_config *config) |
| { |
| int ret = 0; |
| |
| ret = audio_extn_hw_loopback_set_audio_port_config(dev, config); |
| ret |= audio_extn_auto_hal_set_audio_port_config(dev, config); |
| return ret; |
| } |
| |
| static int adev_dump(const audio_hw_device_t *device __unused, |
| int fd __unused) |
| { |
| return 0; |
| } |
| |
| static int adev_close(hw_device_t *device) |
| { |
| size_t i; |
| struct audio_device *adev_temp = (struct audio_device *)device; |
| |
| if (!adev_temp) |
| return 0; |
| |
| pthread_mutex_lock(&adev_init_lock); |
| |
| if ((--audio_device_ref_count) == 0) { |
| if (audio_extn_spkr_prot_is_enabled()) |
| audio_extn_spkr_prot_deinit(); |
| audio_extn_battery_properties_listener_deinit(); |
| audio_extn_snd_mon_unregister_listener(adev); |
| audio_extn_sound_trigger_deinit(adev); |
| audio_extn_listen_deinit(adev); |
| audio_extn_qdsp_deinit(); |
| audio_extn_extspk_deinit(adev->extspk); |
| audio_extn_utils_release_streams_cfg_lists( |
| &adev->streams_output_cfg_list, |
| &adev->streams_input_cfg_list); |
| if (audio_extn_qap_is_enabled()) |
| audio_extn_qap_deinit(); |
| if (audio_extn_qaf_is_enabled()) |
| audio_extn_qaf_deinit(); |
| audio_route_free(adev->audio_route); |
| audio_extn_gef_deinit(adev); |
| free(adev->snd_dev_ref_cnt); |
| platform_deinit(adev->platform); |
| for (i = 0; i < ARRAY_SIZE(adev->use_case_table); ++i) { |
| pcm_params_free(adev->use_case_table[i]); |
| } |
| if (adev->adm_deinit) |
| adev->adm_deinit(adev->adm_data); |
| qahwi_deinit(device); |
| audio_extn_adsp_hdlr_deinit(); |
| audio_extn_snd_mon_deinit(); |
| audio_extn_hw_loopback_deinit(adev); |
| audio_extn_ffv_deinit(); |
| if (adev->device_cfg_params) { |
| free(adev->device_cfg_params); |
| adev->device_cfg_params = NULL; |
| } |
| if(adev->ext_hw_plugin) |
| audio_extn_ext_hw_plugin_deinit(adev->ext_hw_plugin); |
| audio_extn_auto_hal_deinit(); |
| free_map(adev->patch_map); |
| free_map(adev->io_streams_map); |
| free(device); |
| adev = NULL; |
| } |
| pthread_mutex_unlock(&adev_init_lock); |
| enable_gcov(); |
| return 0; |
| } |
| |
| /* This returns 1 if the input parameter looks at all plausible as a low latency period size, |
| * or 0 otherwise. A return value of 1 doesn't mean the value is guaranteed to work, |
| * just that it _might_ work. |
| */ |
| static int period_size_is_plausible_for_low_latency(int period_size) |
| { |
| switch (period_size) { |
| case 160: |
| case 192: |
| case 240: |
| case 320: |
| case 480: |
| return 1; |
| default: |
| return 0; |
| } |
| } |
| |
| static void adev_snd_mon_cb(void *cookie, struct str_parms *parms) |
| { |
| bool is_snd_card_status = false; |
| bool is_ext_device_status = false; |
| char value[32]; |
| int card = -1; |
| card_status_t status; |
| |
| if (cookie != adev || !parms) |
| return; |
| |
| if (!parse_snd_card_status(parms, &card, &status)) { |
| is_snd_card_status = true; |
| } else if (0 < str_parms_get_str(parms, "ext_audio_device", value, sizeof(value))) { |
| is_ext_device_status = true; |
| } else { |
| // not a valid event |
| return; |
| } |
| |
| pthread_mutex_lock(&adev->lock); |
| if (card == adev->snd_card || is_ext_device_status) { |
| if (is_snd_card_status && adev->card_status != status) { |
| ALOGD("%s card_status %d", __func__, status); |
| adev->card_status = status; |
| platform_snd_card_update(adev->platform, status); |
| audio_extn_fm_set_parameters(adev, parms); |
| audio_extn_auto_hal_set_parameters(adev, parms); |
| if (status == CARD_STATUS_OFFLINE) |
| audio_extn_sco_reset_configuration(); |
| } else if (is_ext_device_status) { |
| platform_set_parameters(adev->platform, parms); |
| } |
| } |
| pthread_mutex_unlock(&adev->lock); |
| return; |
| } |
| |
| /* adev lock held */ |
| int check_a2dp_restore_l(struct audio_device *adev, struct stream_out *out, bool restore) |
| { |
| struct audio_usecase *uc_info; |
| struct audio_usecase *usecase; |
| struct listnode devices; |
| struct listnode *node; |
| |
| uc_info = get_usecase_from_list(adev, out->usecase); |
| if (uc_info == NULL) { |
| ALOGE("%s: Could not find the usecase (%d) in the list", |
| __func__, out->usecase); |
| return -EINVAL; |
| } |
| list_init(&devices); |
| |
| ALOGD("%s: enter: usecase(%d: %s), a2dp muted %d", __func__, |
| out->usecase, use_case_table[out->usecase], out->a2dp_muted); |
| |
| if (restore) { |
| pthread_mutex_lock(&out->latch_lock); |
| // restore A2DP device for active usecases and unmute if required |
| if (is_a2dp_out_device_type(&out->device_list)) { |
| ALOGD("%s: restoring A2dp and unmuting stream", __func__); |
| if (uc_info->out_snd_device != SND_DEVICE_OUT_BT_A2DP) |
| select_devices(adev, uc_info->id); |
| |
| if (is_offload_usecase(out->usecase)) { |
| if (uc_info->out_snd_device == SND_DEVICE_OUT_BT_A2DP) |
| out_set_compr_volume(&out->stream, out->volume_l, out->volume_r); |
| } else if (out->usecase == USECASE_AUDIO_PLAYBACK_VOIP) { |
| out_set_voip_volume(&out->stream, out->volume_l, out->volume_r); |
| } else { |
| out_set_pcm_volume(&out->stream, out->volume_l, out->volume_r); |
| } |
| out->a2dp_muted = false; |
| } |
| pthread_mutex_unlock(&out->latch_lock); |
| } else { |
| pthread_mutex_lock(&out->latch_lock); |
| // mute stream and switch to speaker if suspended |
| if (!out->a2dp_muted && !out->standby) { |
| assign_devices(&devices, &out->device_list); |
| reassign_device_list(&out->device_list, AUDIO_DEVICE_OUT_SPEAKER, ""); |
| list_for_each(node, &adev->usecase_list) { |
| usecase = node_to_item(node, struct audio_usecase, list); |
| if ((usecase->type != PCM_CAPTURE) && (usecase != uc_info) && |
| !is_a2dp_out_device_type(&usecase->stream.out->device_list) && |
| !is_sco_out_device_type(&usecase->stream.out->device_list) && |
| platform_check_backends_match(SND_DEVICE_OUT_SPEAKER, |
| usecase->out_snd_device)) { |
| assign_devices(&out->device_list, &usecase->stream.out->device_list); |
| break; |
| } |
| } |
| if ((is_a2dp_out_device_type(&devices) && list_length(&devices) == 1) || |
| (uc_info->out_snd_device == SND_DEVICE_OUT_BT_A2DP)) { |
| out->a2dp_muted = true; |
| if (is_offload_usecase(out->usecase)) { |
| if (out->offload_state == OFFLOAD_STATE_PLAYING) |
| compress_pause(out->compr); |
| out_set_compr_volume(&out->stream, (float)0, (float)0); |
| } else { |
| if (out->usecase == USECASE_AUDIO_PLAYBACK_VOIP) |
| out_set_voip_volume(&out->stream, (float)0, (float)0); |
| else |
| out_set_pcm_volume(&out->stream, (float)0, (float)0); |
| |
| /* wait for stale pcm drained before switching to speaker */ |
| uint32_t latency = |
| (out->config.period_count * out->config.period_size * 1000) / |
| (out->config.rate); |
| usleep(latency * 1000); |
| } |
| } |
| select_devices(adev, out->usecase); |
| ALOGD("%s: switched to device:%s and stream muted:%d", __func__, |
| platform_get_snd_device_name(uc_info->out_snd_device), out->a2dp_muted); |
| if (is_offload_usecase(out->usecase)) { |
| if (out->offload_state == OFFLOAD_STATE_PLAYING) |
| compress_resume(out->compr); |
| } |
| assign_devices(&out->device_list, &devices); |
| } |
| pthread_mutex_unlock(&out->latch_lock); |
| } |
| ALOGV("%s: exit", __func__); |
| return 0; |
| } |
| |
| void adev_on_battery_status_changed(bool charging) |
| { |
| pthread_mutex_lock(&adev->lock); |
| ALOGI("%s: battery status changed to %scharging", __func__, charging ? "" : "not "); |
| adev->is_charging = charging; |
| audio_extn_sound_trigger_update_battery_status(charging); |
| pthread_mutex_unlock(&adev->lock); |
| } |
| |
| static int adev_open(const hw_module_t *module, const char *name, |
| hw_device_t **device) |
| { |
| int ret; |
| char value[PROPERTY_VALUE_MAX] = {0}; |
| char mixer_ctl_name[128] = {0}; |
| struct mixer_ctl *ctl = NULL; |
| |
| ALOGD("%s: enter", __func__); |
| if (strcmp(name, AUDIO_HARDWARE_INTERFACE) != 0) return -EINVAL; |
| |
| pthread_mutex_lock(&adev_init_lock); |
| if (audio_device_ref_count != 0){ |
| *device = &adev->device.common; |
| audio_device_ref_count++; |
| ALOGD("%s: returning existing instance of adev", __func__); |
| ALOGD("%s: exit", __func__); |
| pthread_mutex_unlock(&adev_init_lock); |
| return 0; |
| } |
| |
| adev = calloc(1, sizeof(struct audio_device)); |
| |
| if (!adev) { |
| pthread_mutex_unlock(&adev_init_lock); |
| return -ENOMEM; |
| } |
| |
| pthread_mutex_init(&adev->lock, (const pthread_mutexattr_t *) NULL); |
| |
| // register audio ext hidl at the earliest |
| audio_extn_hidl_init(); |
| #ifdef DYNAMIC_LOG_ENABLED |
| register_for_dynamic_logging("hal"); |
| #endif |
| |
| /* default audio HAL major version */ |
| uint32_t maj_version = 3; |
| if(property_get("vendor.audio.hal.maj.version", value, NULL)) |
| maj_version = atoi(value); |
| |
| adev->device.common.tag = HARDWARE_DEVICE_TAG; |
| adev->device.common.version = HARDWARE_DEVICE_API_VERSION(maj_version, 0); |
| adev->device.common.module = (struct hw_module_t *)module; |
| adev->device.common.close = adev_close; |
| |
| adev->device.init_check = adev_init_check; |
| adev->device.set_voice_volume = adev_set_voice_volume; |
| adev->device.set_master_volume = adev_set_master_volume; |
| adev->device.get_master_volume = adev_get_master_volume; |
| adev->device.set_master_mute = adev_set_master_mute; |
| adev->device.get_master_mute = adev_get_master_mute; |
| adev->device.set_mode = adev_set_mode; |
| adev->device.set_mic_mute = adev_set_mic_mute; |
| adev->device.get_mic_mute = adev_get_mic_mute; |
| adev->device.set_parameters = adev_set_parameters; |
| adev->device.get_parameters = adev_get_parameters; |
| adev->device.get_input_buffer_size = adev_get_input_buffer_size; |
| adev->device.open_output_stream = adev_open_output_stream; |
| adev->device.close_output_stream = adev_close_output_stream; |
| adev->device.open_input_stream = adev_open_input_stream; |
| adev->device.close_input_stream = adev_close_input_stream; |
| adev->device.create_audio_patch = adev_create_audio_patch; |
| adev->device.release_audio_patch = adev_release_audio_patch; |
| adev->device.get_audio_port = adev_get_audio_port; |
| adev->device.set_audio_port_config = adev_set_audio_port_config; |
| adev->device.dump = adev_dump; |
| adev->device.get_microphones = adev_get_microphones; |
| |
| /* Set the default route before the PCM stream is opened */ |
| adev->mode = AUDIO_MODE_NORMAL; |
| adev->primary_output = NULL; |
| adev->out_device = AUDIO_DEVICE_NONE; |
| adev->bluetooth_nrec = true; |
| adev->acdb_settings = TTY_MODE_OFF; |
| adev->allow_afe_proxy_usage = true; |
| adev->bt_sco_on = false; |
| /* adev->cur_hdmi_channels = 0; by calloc() */ |
| adev->snd_dev_ref_cnt = calloc(SND_DEVICE_MAX, sizeof(int)); |
| /* Init audio and voice feature */ |
| audio_extn_feature_init(); |
| voice_extn_feature_init(); |
| voice_init(adev); |
| list_init(&adev->usecase_list); |
| list_init(&adev->active_inputs_list); |
| list_init(&adev->active_outputs_list); |
| list_init(&adev->audio_patch_record_list); |
| adev->io_streams_map = hashmapCreate(AUDIO_IO_PORTS_MAX, audio_extn_utils_hash_fn, |
| audio_extn_utils_hash_eq); |
| if (!adev->io_streams_map) { |
| ALOGE("%s: Could not create io streams map", __func__); |
| ret = -ENOMEM; |
| goto adev_open_err; |
| } |
| adev->patch_map = hashmapCreate(AUDIO_PATCH_PORTS_MAX, audio_extn_utils_hash_fn, |
| audio_extn_utils_hash_eq); |
| if (!adev->patch_map) { |
| ALOGE("%s: Could not create audio patch map", __func__); |
| ret = -ENOMEM; |
| goto adev_open_err; |
| } |
| adev->cur_wfd_channels = 2; |
| adev->offload_usecases_state = 0; |
| adev->pcm_record_uc_state = 0; |
| adev->is_channel_status_set = false; |
| adev->perf_lock_opts[0] = 0x101; |
| adev->perf_lock_opts[1] = 0x20E; |
| adev->perf_lock_opts_size = 2; |
| adev->dsp_bit_width_enforce_mode = 0; |
| adev->enable_hfp = false; |
| adev->use_old_pspd_mix_ctrl = false; |
| adev->adm_routing_changed = false; |
| adev->a2dp_started = false; |
| |
| audio_extn_perf_lock_init(); |
| |
| /* Loads platform specific libraries dynamically */ |
| adev->platform = platform_init(adev); |
| if (!adev->platform) { |
| ALOGE("%s: Failed to init platform data, aborting.", __func__); |
| ret = -EINVAL; |
| goto adev_open_err; |
| } |
| |
| adev->extspk = audio_extn_extspk_init(adev); |
| if (audio_extn_qap_is_enabled()) { |
| ret = audio_extn_qap_init(adev); |
| if (ret < 0) { |
| ALOGE("%s: Failed to init platform data, aborting.", __func__); |
| goto adev_open_err; |
| } |
| adev->device.open_output_stream = audio_extn_qap_open_output_stream; |
| adev->device.close_output_stream = audio_extn_qap_close_output_stream; |
| } |
| |
| if (audio_extn_qaf_is_enabled()) { |
| ret = audio_extn_qaf_init(adev); |
| if (ret < 0) { |
| ALOGE("%s: Failed to init platform data, aborting.", __func__); |
| goto adev_open_err; |
| } |
| |
| adev->device.open_output_stream = audio_extn_qaf_open_output_stream; |
| adev->device.close_output_stream = audio_extn_qaf_close_output_stream; |
| } |
| |
| audio_extn_auto_hal_init(adev); |
| adev->ext_hw_plugin = audio_extn_ext_hw_plugin_init(adev); |
| |
| if (access(VISUALIZER_LIBRARY_PATH, R_OK) == 0) { |
| adev->visualizer_lib = dlopen(VISUALIZER_LIBRARY_PATH, RTLD_NOW); |
| if (adev->visualizer_lib == NULL) { |
| ALOGE("%s: DLOPEN failed for %s", __func__, VISUALIZER_LIBRARY_PATH); |
| } else { |
| ALOGV("%s: DLOPEN successful for %s", __func__, VISUALIZER_LIBRARY_PATH); |
| adev->visualizer_start_output = |
| (int (*)(audio_io_handle_t, int))dlsym(adev->visualizer_lib, |
| "visualizer_hal_start_output"); |
| adev->visualizer_stop_output = |
| (int (*)(audio_io_handle_t, int))dlsym(adev->visualizer_lib, |
| "visualizer_hal_stop_output"); |
| } |
| } |
| audio_extn_init(adev); |
| voice_extn_init(adev); |
| audio_extn_listen_init(adev, adev->snd_card); |
| audio_extn_gef_init(adev); |
| audio_extn_hw_loopback_init(adev); |
| audio_extn_ffv_init(adev); |
| |
| if (access(OFFLOAD_EFFECTS_BUNDLE_LIBRARY_PATH, R_OK) == 0) { |
| adev->offload_effects_lib = dlopen(OFFLOAD_EFFECTS_BUNDLE_LIBRARY_PATH, RTLD_NOW); |
| if (adev->offload_effects_lib == NULL) { |
| ALOGE("%s: DLOPEN failed for %s", __func__, |
| OFFLOAD_EFFECTS_BUNDLE_LIBRARY_PATH); |
| } else { |
| ALOGV("%s: DLOPEN successful for %s", __func__, |
| OFFLOAD_EFFECTS_BUNDLE_LIBRARY_PATH); |
| adev->offload_effects_start_output = |
| (int (*)(audio_io_handle_t, int, struct mixer *))dlsym(adev->offload_effects_lib, |
| "offload_effects_bundle_hal_start_output"); |
| adev->offload_effects_stop_output = |
| (int (*)(audio_io_handle_t, int))dlsym(adev->offload_effects_lib, |
| "offload_effects_bundle_hal_stop_output"); |
| adev->offload_effects_set_hpx_state = |
| (int (*)(bool))dlsym(adev->offload_effects_lib, |
| "offload_effects_bundle_set_hpx_state"); |
| adev->offload_effects_get_parameters = |
| (void (*)(struct str_parms *, struct str_parms *)) |
| dlsym(adev->offload_effects_lib, |
| "offload_effects_bundle_get_parameters"); |
| adev->offload_effects_set_parameters = |
| (void (*)(struct str_parms *))dlsym(adev->offload_effects_lib, |
| "offload_effects_bundle_set_parameters"); |
| } |
| } |
| |
| if (access(ADM_LIBRARY_PATH, R_OK) == 0) { |
| adev->adm_lib = dlopen(ADM_LIBRARY_PATH, RTLD_NOW); |
| if (adev->adm_lib == NULL) { |
| ALOGE("%s: DLOPEN failed for %s", __func__, ADM_LIBRARY_PATH); |
| } else { |
| ALOGV("%s: DLOPEN successful for %s", __func__, ADM_LIBRARY_PATH); |
| adev->adm_init = (adm_init_t) |
| dlsym(adev->adm_lib, "adm_init"); |
| adev->adm_deinit = (adm_deinit_t) |
| dlsym(adev->adm_lib, "adm_deinit"); |
| adev->adm_register_input_stream = (adm_register_input_stream_t) |
| dlsym(adev->adm_lib, "adm_register_input_stream"); |
| adev->adm_register_output_stream = (adm_register_output_stream_t) |
| dlsym(adev->adm_lib, "adm_register_output_stream"); |
| adev->adm_deregister_stream = (adm_deregister_stream_t) |
| dlsym(adev->adm_lib, "adm_deregister_stream"); |
| adev->adm_request_focus = (adm_request_focus_t) |
| dlsym(adev->adm_lib, "adm_request_focus"); |
| adev->adm_abandon_focus = (adm_abandon_focus_t) |
| dlsym(adev->adm_lib, "adm_abandon_focus"); |
| adev->adm_set_config = (adm_set_config_t) |
| dlsym(adev->adm_lib, "adm_set_config"); |
| adev->adm_request_focus_v2 = (adm_request_focus_v2_t) |
| dlsym(adev->adm_lib, "adm_request_focus_v2"); |
| adev->adm_is_noirq_avail = (adm_is_noirq_avail_t) |
| dlsym(adev->adm_lib, "adm_is_noirq_avail"); |
| adev->adm_on_routing_change = (adm_on_routing_change_t) |
| dlsym(adev->adm_lib, "adm_on_routing_change"); |
| adev->adm_request_focus_v2_1 = (adm_request_focus_v2_1_t) |
| dlsym(adev->adm_lib, "adm_request_focus_v2_1"); |
| } |
| } |
| |
| adev->enable_voicerx = false; |
| adev->bt_wb_speech_enabled = false; |
| adev->swb_speech_mode = SPEECH_MODE_INVALID; |
| adev->fluence_nn_usecase_id = USECASE_INVALID; |
| //initialize this to false for now, |
| //this will be set to true through set param |
| adev->vr_audio_mode_enabled = false; |
| |
| audio_extn_ds2_enable(adev); |
| *device = &adev->device.common; |
| |
| if (k_enable_extended_precision) |
| adev_verify_devices(adev); |
| |
| adev->dsp_bit_width_enforce_mode = |
| adev_init_dsp_bit_width_enforce_mode(adev->mixer); |
| |
| audio_extn_utils_update_streams_cfg_lists(adev->platform, adev->mixer, |
| &adev->streams_output_cfg_list, |
| &adev->streams_input_cfg_list); |
| |
| audio_device_ref_count++; |
| |
| int trial; |
| if (property_get("vendor.audio_hal.period_size", value, NULL) > 0) { |
| trial = atoi(value); |
| if (period_size_is_plausible_for_low_latency(trial)) { |
| pcm_config_low_latency.period_size = trial; |
| pcm_config_low_latency.start_threshold = trial / 4; |
| pcm_config_low_latency.avail_min = trial / 4; |
| configured_low_latency_capture_period_size = trial; |
| } |
| } |
| if (property_get("vendor.audio_hal.in_period_size", value, NULL) > 0) { |
| trial = atoi(value); |
| if (period_size_is_plausible_for_low_latency(trial)) { |
| configured_low_latency_capture_period_size = trial; |
| } |
| } |
| |
| adev->mic_break_enabled = property_get_bool("vendor.audio.mic_break", false); |
| |
| adev->camera_orientation = CAMERA_DEFAULT; |
| |
| if (property_get("vendor.audio_hal.period_multiplier",value,NULL) > 0) { |
| af_period_multiplier = atoi(value); |
| if (af_period_multiplier < 0) |
| af_period_multiplier = 2; |
| else if (af_period_multiplier > 4) |
| af_period_multiplier = 4; |
| |
| ALOGV("new period_multiplier = %d", af_period_multiplier); |
| } |
| |
| audio_extn_qdsp_init(adev->platform); |
| |
| adev->multi_offload_enable = property_get_bool("vendor.audio.offload.multiple.enabled", false); |
| adev->ha_proxy_enable = property_get_bool("persist.vendor.audio.ha_proxy.enabled", false); |
| pthread_mutex_unlock(&adev_init_lock); |
| |
| if (adev->adm_init) |
| adev->adm_data = adev->adm_init(); |
| |
| qahwi_init(*device); |
| audio_extn_adsp_hdlr_init(adev->mixer); |
| |
| audio_extn_snd_mon_init(); |
| pthread_mutex_lock(&adev->lock); |
| audio_extn_snd_mon_register_listener(adev, adev_snd_mon_cb); |
| adev->card_status = CARD_STATUS_ONLINE; |
| audio_extn_battery_properties_listener_init(adev_on_battery_status_changed); |
| /* |
| * if the battery state callback happens before charging can be queried, |
| * it will be guarded with the adev->lock held in the cb function and so |
| * the callback value will reflect the latest state |
| */ |
| adev->is_charging = audio_extn_battery_properties_is_charging(); |
| audio_extn_sound_trigger_init(adev); /* dependent on snd_mon_init() */ |
| audio_extn_sound_trigger_update_battery_status(adev->is_charging); |
| audio_extn_audiozoom_init(); |
| pthread_mutex_unlock(&adev->lock); |
| /* Allocate memory for Device config params */ |
| adev->device_cfg_params = (struct audio_device_config_param*) |
| calloc(platform_get_max_codec_backend(), |
| sizeof(struct audio_device_config_param)); |
| if (adev->device_cfg_params == NULL) |
| ALOGE("%s: Memory allocation failed for Device config params", __func__); |
| |
| /* |
| * Check if new PSPD matrix mixer control is supported. If not |
| * supported, then set flag so that old mixer ctrl is sent while |
| * sending pspd coefficients on older kernel version. Query mixer |
| * control for default pcm id and channel value one. |
| */ |
| snprintf(mixer_ctl_name, sizeof(mixer_ctl_name), |
| "AudStr %d ChMixer Weight Ch %d", 0, 1); |
| |
| ctl = mixer_get_ctl_by_name(adev->mixer, mixer_ctl_name); |
| if (!ctl) { |
| ALOGE("%s: ERROR. Could not get ctl for mixer cmd - %s", |
| __func__, mixer_ctl_name); |
| adev->use_old_pspd_mix_ctrl = true; |
| } |
| |
| ALOGD("%s: exit", __func__); |
| return 0; |
| |
| adev_open_err: |
| free_map(adev->patch_map); |
| free_map(adev->io_streams_map); |
| free(adev->snd_dev_ref_cnt); |
| pthread_mutex_destroy(&adev->lock); |
| free(adev); |
| adev = NULL; |
| *device = NULL; |
| pthread_mutex_unlock(&adev_init_lock); |
| return ret; |
| } |
| |
| static struct hw_module_methods_t hal_module_methods = { |
| .open = adev_open, |
| }; |
| |
| struct audio_module HAL_MODULE_INFO_SYM = { |
| .common = { |
| .tag = HARDWARE_MODULE_TAG, |
| .module_api_version = AUDIO_MODULE_API_VERSION_0_1, |
| .hal_api_version = HARDWARE_HAL_API_VERSION, |
| .id = AUDIO_HARDWARE_MODULE_ID, |
| .name = "QCOM Audio HAL", |
| .author = "The Linux Foundation", |
| .methods = &hal_module_methods, |
| }, |
| }; |