Re: [PATCH] ALSA: pcm: fix wait_time calculations

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 




On 12/06/2023 13:18, Jon Hunter wrote:
Hi Oswald,

On 05/04/2023 21:12, Oswald Buddenhagen wrote:
... in wait_for_avail() and snd_pcm_drain().

t was calculated in seconds, so it would be pretty much always zero, to
be subsequently de-facto ignored due to being max(t, 10)'d. And then it
(i.e., 10) would be treated as secs, which doesn't seem right.

However, fixing it to properly calculate msecs would potentially cause
timeouts when using twice the period size for the default timeout (which
seems reasonable to me), so instead use the buffer size plus 10 percent
to be on the safe side ... but that still seems insufficient, presumably
because the hardware typically needs a moment to fire up. To compensate
for this, we up the minimal timeout to 100ms, which is still two orders
of magnitude less than the bogus minimum.

substream->wait_time was also misinterpreted as jiffies, despite being
documented as being in msecs. Only the soc/sof driver sets it - to 500,
which looks very much like msecs were intended.

Speaking of which, shouldn't snd_pcm_drain() also use substream->
wait_time?

As a drive-by, make the debug messages on timeout less confusing.

Signed-off-by: Oswald Buddenhagen <oswald.buddenhagen@xxxxxx>
---
  sound/core/pcm_lib.c    | 11 +++++------
  sound/core/pcm_native.c |  8 ++++----
  2 files changed, 9 insertions(+), 10 deletions(-)

diff --git a/sound/core/pcm_lib.c b/sound/core/pcm_lib.c
index 02fd65993e7e..af1eb136feb0 100644
--- a/sound/core/pcm_lib.c
+++ b/sound/core/pcm_lib.c
@@ -1878,15 +1878,14 @@ static int wait_for_avail(struct snd_pcm_substream *substream,
          if (substream->wait_time) {
              wait_time = substream->wait_time;
          } else {
-            wait_time = 10;
+            wait_time = 100;
              if (runtime->rate) {
-                long t = runtime->period_size * 2 /
-                     runtime->rate;
+                long t = runtime->buffer_size * 1100 / runtime->rate;
                  wait_time = max(t, wait_time);
              }
-            wait_time = msecs_to_jiffies(wait_time * 1000);
          }
+        wait_time = msecs_to_jiffies(wait_time);
      }
      for (;;) {
@@ -1934,8 +1933,8 @@ static int wait_for_avail(struct snd_pcm_substream *substream,
          }
          if (!tout) {
              pcm_dbg(substream->pcm,
-                "%s write error (DMA or IRQ trouble?)\n",
-                is_playback ? "playback" : "capture");
+                "%s timeout (DMA or IRQ trouble?)\n",
+                is_playback ? "playback write" : "capture read");
              err = -EIO;
              break;
          }
diff --git a/sound/core/pcm_native.c b/sound/core/pcm_native.c
index 331380c2438b..94185267a7b9 100644
--- a/sound/core/pcm_native.c
+++ b/sound/core/pcm_native.c
@@ -2159,12 +2159,12 @@ static int snd_pcm_drain(struct snd_pcm_substream *substream,
          if (runtime->no_period_wakeup)
              tout = MAX_SCHEDULE_TIMEOUT;
          else {
-            tout = 10;
+            tout = 100;
              if (runtime->rate) {
-                long t = runtime->period_size * 2 / runtime->rate;
+                long t = runtime->buffer_size * 1100 / runtime->rate;
                  tout = max(t, tout);
              }
-            tout = msecs_to_jiffies(tout * 1000);
+            tout = msecs_to_jiffies(tout);
          }
          tout = schedule_timeout(tout);
@@ -2187,7 +2187,7 @@ static int snd_pcm_drain(struct snd_pcm_substream *substream,
                  result = -ESTRPIPE;
              else {
                  dev_dbg(substream->pcm->card->dev,
-                    "playback drain error (DMA or IRQ trouble?)\n");
+                    "playback drain timeout (DMA or IRQ trouble?)\n");
                  snd_pcm_stop(substream, SNDRV_PCM_STATE_SETUP);
                  result = -EIO;
              }


Sorry for not catching this sooner, but I have just noticed that one of our audio tests for Tegra is failing on v6.4-rc and bisect is pointing to this commit. Reverting this on top of the current mainline fixes it.

Sameer, any thoughts on why this change would break the volume control test for Tegra194? After this patch is applied the test fails on the 2nd test for MVC1 ...

[ 1314 ms] Starting mvc_run...
[ 3616 ms] Starting mvc_check_configs...
[ 3921 ms] Starting mvc_check_sha1sum...
[ 4379 ms] Test with gain of 40 for Linear [ 4404 ms] curve type with 32 bits 2 channels using MVC1 ... PASSED!
[ 4437 ms] Starting mvc_run...
[ 6720 ms] Starting mvc_check_configs...
[ 7029 ms] Starting mvc_check_sha1sum...
[ 7463 ms] Test with gain of 40 for Linear [ 7490 ms] curve type with 32 bits 2 channels using MVC2 ... PASSED!
[ 7541 ms] Starting mvc_run...
[ 9899 ms] Starting mvc_check_configs...
[ 10255 ms] Starting mvc_check_sha1sum...
[ 10384 ms] Error: Sha1sum check FAILED!
[ 10590 ms] Error: Incorrect output gain dB against [ 10613 ms] expected gain -122dB for channel 1 [ 10701 ms] Test with gain of -119 for Linear [ 10730 ms] curve type with 32 bits 2 channels using MVC1 ... FAILED!


If I enable the debug prints, I do see the following messages ...

 tegra-audio-graph-card sound: capture read timeout (DMA or IRQ trouble?)
 tegra-audio-graph-card sound: capture read timeout (DMA or IRQ trouble?)
 tegra-audio-graph-card sound: capture read timeout (DMA or IRQ trouble?)
 tegra-audio-graph-card sound: capture read timeout (DMA or IRQ trouble?)


Jon

--
nvpublic



[Index of Archives]     [ARM Kernel]     [Linux ARM]     [Linux ARM MSM]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]

  Powered by Linux